904 resultados para Feature discretization


Relevância:

20.00% 20.00%

Publicador:

Resumo:

The governance of climate adaptation involves the collective efforts of multiple societal actors to address problems, or to reap the benefits, associated with impacts of climate change. Governing involves the creation of institutions, rules and organizations, and the selection of normative principles to guide problem solution and institution building. We argue that actors involved in governing climate change adaptation, as climate change governance regimes evolve, inevitably must engage in making choices, for instance on problem definitions, jurisdictional levels, on modes of governance and policy instruments, and on the timing of interventions. Yet little is known about how and why these choices are made in practice, and how such choices affect the outcomes of our efforts to govern adaptation. In this introduction we review the current state of evidence and the specific contribution of the articles published in this Special Feature, which are aimed at bringing greater clarity in these matters, and thereby informing both governance theory and practice. Collectively, the contributing papers suggest that the way issues are defined has important consequences for the support for governance interventions, and their effectiveness. The articles suggest that currently the emphasis in adaptation governance is on the local and regional levels, while underscoring the benefits of interventions and governance at higher jurisdictional levels in terms of visioning and scaling-up effective approaches. The articles suggest that there is a central role of government agencies in leading governance interventions to address spillover effects, to provide public goods, and to promote the long-term perspectives for planning. They highlight the issue of justice in the governance of adaptation showing how governance measures have wide distributional consequences, including the potential to amplify existing inequalities, access to resources, or generating new injustices through distribution of risks. For several of these findings, future research directions are suggested.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We present a new discretization for the Hadamard fractional derivative, that simplifies the computations. We then apply the method to solve a fractional differential equation and a fractional variational problem with dependence on the Hadamard fractional derivative.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The estimating of the relative orientation and position of a camera is one of the integral topics in the field of computer vision. The accuracy of a certain Finnish technology company’s traffic sign inventory and localization process can be improved by utilizing the aforementioned concept. The company’s localization process uses video data produced by a vehicle installed camera. The accuracy of estimated traffic sign locations depends on the relative orientation between the camera and the vehicle. This thesis proposes a computer vision based software solution which can estimate a camera’s orientation relative to the movement direction of the vehicle by utilizing video data. The task was solved by using feature-based methods and open source software. When using simulated data sets, the camera orientation estimates had an absolute error of 0.31 degrees on average. The software solution can be integrated to be a part of the traffic sign localization pipeline of the company in question.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The use of human brain electroencephalography (EEG) signals for automatic person identi cation has been investigated for a decade. It has been found that the performance of an EEG-based person identication system highly depends on what feature to be extracted from multi-channel EEG signals. Linear methods such as Power Spectral Density and Autoregressive Model have been used to extract EEG features. However these methods assumed that EEG signals are stationary. In fact, EEG signals are complex, non-linear, non-stationary, and random in nature. In addition, other factors such as brain condition or human characteristics may have impacts on the performance, however these factors have not been investigated and evaluated in previous studies. It has been found in the literature that entropy is used to measure the randomness of non-linear time series data. Entropy is also used to measure the level of chaos of braincomputer interface systems. Therefore, this thesis proposes to study the role of entropy in non-linear analysis of EEG signals to discover new features for EEG-based person identi- cation. Five dierent entropy methods including Shannon Entropy, Approximate Entropy, Sample Entropy, Spectral Entropy, and Conditional Entropy have been proposed to extract entropy features that are used to evaluate the performance of EEG-based person identication systems and the impacts of epilepsy, alcohol, age and gender characteristics on these systems. Experiments were performed on the Australian EEG and Alcoholism datasets. Experimental results have shown that, in most cases, the proposed entropy features yield very fast person identication, yet with compatible accuracy because the feature dimension is low. In real life security operation, timely response is critical. The experimental results have also shown that epilepsy, alcohol, age and gender characteristics have impacts on the EEG-based person identication systems.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Model Driven based approach for Service Evolution in Clouds will mainly focus on the reusable evolution patterns' advantage to solve evolution problems. During the process, evolution pattern will be driven by MDA models to pattern aspects. Weaving the aspects into service based process by using Aspect-Oriented extended BPEL engine at runtime will be the dynamic feature of the evolution.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The use of digital image processing techniques is prominent in medical settings for the automatic diagnosis of diseases. Glaucoma is the second leading cause of blindness in the world and it has no cure. Currently, there are treatments to prevent vision loss, but the disease must be detected in the early stages. Thus, the objective of this work is to develop an automatic detection method of Glaucoma in retinal images. The methodology used in the study were: acquisition of image database, Optic Disc segmentation, texture feature extraction in different color models and classification of images in glaucomatous or not. We obtained results of 93% accuracy

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this article we propose a new symmetric version of the interior penalty discontinuous Galerkin finite element method for the numerical approximation of the compressible Navier-Stokes equations. Here, particular emphasis is devoted to the construction of an optimal numerical method for the evaluation of certain target functionals of practical interest, such as the lift and drag coefficients of a body immersed in a viscous fluid. With this in mind, the key ingredients in the construction of the method include: (i) An adjoint consistent imposition of the boundary conditions; (ii) An adjoint consistent reformulation of the underlying target functional of practical interest; (iii) Design of appropriate interior-penalty stabilization terms. Numerical experiments presented within this article clearly indicate the optimality of the proposed method when the error is measured in terms of both the L_2-norm, as well as for certain target functionals. Computational comparisons with other discontinuous Galerkin schemes proposed in the literature, including the second scheme of Bassi & Rebay, cf. [11], the standard SIPG method outlined in [25], and an NIPG variant of the new scheme will be undertaken.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Dissertação (mestrado)—Universidade de Brasília, Instituto de Geociências, Programa de Pós-Graduação Stricto Sensu em Geociências Aplicadas, 2016.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Thanks to the advanced technologies and social networks that allow the data to be widely shared among the Internet, there is an explosion of pervasive multimedia data, generating high demands of multimedia services and applications in various areas for people to easily access and manage multimedia data. Towards such demands, multimedia big data analysis has become an emerging hot topic in both industry and academia, which ranges from basic infrastructure, management, search, and mining to security, privacy, and applications. Within the scope of this dissertation, a multimedia big data analysis framework is proposed for semantic information management and retrieval with a focus on rare event detection in videos. The proposed framework is able to explore hidden semantic feature groups in multimedia data and incorporate temporal semantics, especially for video event detection. First, a hierarchical semantic data representation is presented to alleviate the semantic gap issue, and the Hidden Coherent Feature Group (HCFG) analysis method is proposed to capture the correlation between features and separate the original feature set into semantic groups, seamlessly integrating multimedia data in multiple modalities. Next, an Importance Factor based Temporal Multiple Correspondence Analysis (i.e., IF-TMCA) approach is presented for effective event detection. Specifically, the HCFG algorithm is integrated with the Hierarchical Information Gain Analysis (HIGA) method to generate the Importance Factor (IF) for producing the initial detection results. Then, the TMCA algorithm is proposed to efficiently incorporate temporal semantics for re-ranking and improving the final performance. At last, a sampling-based ensemble learning mechanism is applied to further accommodate the imbalanced datasets. In addition to the multimedia semantic representation and class imbalance problems, lack of organization is another critical issue for multimedia big data analysis. In this framework, an affinity propagation-based summarization method is also proposed to transform the unorganized data into a better structure with clean and well-organized information. The whole framework has been thoroughly evaluated across multiple domains, such as soccer goal event detection and disaster information management.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper describes a novel algorithm for tracking the motion of the urethra from trans-perineal ultrasound. Our work is based on the structure-from-motion paradigm and therefore handles well structures with ill-defined and partially missing boundaries. The proposed approach is particularly well-suited for video sequences of low resolution and variable levels of blurriness introduced by anatomical motion of variable speed. Our tracking method identifies feature points on a frame by frame basis using the SURF detector/descriptor. Inter-frame correspondence is achieved using nearest-neighbor matching in the feature space. The motion is estimated using a non-linear bi-quadratic model, which adequately describes the deformable motion of the urethra. Experimental results are promising and show that our algorithm performs well when compared to manual tracking.