874 resultados para correlation-based feature selection
Resumo:
Hyperspectral remote sensing exploits the electromagnetic scattering patterns of the different materials at specific wavelengths [2, 3]. Hyperspectral sensors have been developed to sample the scattered portion of the electromagnetic spectrum extending from the visible region through the near-infrared and mid-infrared, in hundreds of narrow contiguous bands [4, 5]. The number and variety of potential civilian and military applications of hyperspectral remote sensing is enormous [6, 7]. Very often, the resolution cell corresponding to a single pixel in an image contains several substances (endmembers) [4]. In this situation, the scattered energy is a mixing of the endmember spectra. A challenging task underlying many hyperspectral imagery applications is then decomposing a mixed pixel into a collection of reflectance spectra, called endmember signatures, and the corresponding abundance fractions [8–10]. Depending on the mixing scales at each pixel, the observed mixture is either linear or nonlinear [11, 12]. Linear mixing model holds approximately when the mixing scale is macroscopic [13] and there is negligible interaction among distinct endmembers [3, 14]. If, however, the mixing scale is microscopic (or intimate mixtures) [15, 16] and the incident solar radiation is scattered by the scene through multiple bounces involving several endmembers [17], the linear model is no longer accurate. Linear spectral unmixing has been intensively researched in the last years [9, 10, 12, 18–21]. It considers that a mixed pixel is a linear combination of endmember signatures weighted by the correspondent abundance fractions. Under this model, and assuming that the number of substances and their reflectance spectra are known, hyperspectral unmixing is a linear problem for which many solutions have been proposed (e.g., maximum likelihood estimation [8], spectral signature matching [22], spectral angle mapper [23], subspace projection methods [24,25], and constrained least squares [26]). In most cases, the number of substances and their reflectances are not known and, then, hyperspectral unmixing falls into the class of blind source separation problems [27]. Independent component analysis (ICA) has recently been proposed as a tool to blindly unmix hyperspectral data [28–31]. ICA is based on the assumption of mutually independent sources (abundance fractions), which is not the case of hyperspectral data, since the sum of abundance fractions is constant, implying statistical dependence among them. This dependence compromises ICA applicability to hyperspectral images as shown in Refs. [21, 32]. In fact, ICA finds the endmember signatures by multiplying the spectral vectors with an unmixing matrix, which minimizes the mutual information among sources. If sources are independent, ICA provides the correct unmixing, since the minimum of the mutual information is obtained only when sources are independent. This is no longer true for dependent abundance fractions. Nevertheless, some endmembers may be approximately unmixed. These aspects are addressed in Ref. [33]. Under the linear mixing model, the observations from a scene are in a simplex whose vertices correspond to the endmembers. Several approaches [34–36] have exploited this geometric feature of hyperspectral mixtures [35]. Minimum volume transform (MVT) algorithm [36] determines the simplex of minimum volume containing the data. The method presented in Ref. [37] is also of MVT type but, by introducing the notion of bundles, it takes into account the endmember variability usually present in hyperspectral mixtures. The MVT type approaches are complex from the computational point of view. Usually, these algorithms find in the first place the convex hull defined by the observed data and then fit a minimum volume simplex to it. For example, the gift wrapping algorithm [38] computes the convex hull of n data points in a d-dimensional space with a computational complexity of O(nbd=2cþ1), where bxc is the highest integer lower or equal than x and n is the number of samples. The complexity of the method presented in Ref. [37] is even higher, since the temperature of the simulated annealing algorithm used shall follow a log( ) law [39] to assure convergence (in probability) to the desired solution. Aiming at a lower computational complexity, some algorithms such as the pixel purity index (PPI) [35] and the N-FINDR [40] still find the minimum volume simplex containing the data cloud, but they assume the presence of at least one pure pixel of each endmember in the data. This is a strong requisite that may not hold in some data sets. In any case, these algorithms find the set of most pure pixels in the data. PPI algorithm uses the minimum noise fraction (MNF) [41] as a preprocessing step to reduce dimensionality and to improve the signal-to-noise ratio (SNR). The algorithm then projects every spectral vector onto skewers (large number of random vectors) [35, 42,43]. The points corresponding to extremes, for each skewer direction, are stored. A cumulative account records the number of times each pixel (i.e., a given spectral vector) is found to be an extreme. The pixels with the highest scores are the purest ones. N-FINDR algorithm [40] is based on the fact that in p spectral dimensions, the p-volume defined by a simplex formed by the purest pixels is larger than any other volume defined by any other combination of pixels. This algorithm finds the set of pixels defining the largest volume by inflating a simplex inside the data. ORA SIS [44, 45] is a hyperspectral framework developed by the U.S. Naval Research Laboratory consisting of several algorithms organized in six modules: exemplar selector, adaptative learner, demixer, knowledge base or spectral library, and spatial postrocessor. The first step consists in flat-fielding the spectra. Next, the exemplar selection module is used to select spectral vectors that best represent the smaller convex cone containing the data. The other pixels are rejected when the spectral angle distance (SAD) is less than a given thresh old. The procedure finds the basis for a subspace of a lower dimension using a modified Gram–Schmidt orthogonalizati on. The selected vectors are then projected onto this subspace and a simplex is found by an MV T pro cess. ORA SIS is oriented to real-time target detection from uncrewed air vehicles using hyperspectral data [46]. In this chapter we develop a new algorithm to unmix linear mixtures of endmember spectra. First, the algorithm determines the number of endmembers and the signal subspace using a newly developed concept [47, 48]. Second, the algorithm extracts the most pure pixels present in the data. Unlike other methods, this algorithm is completely automatic and unsupervised. To estimate the number of endmembers and the signal subspace in hyperspectral linear mixtures, the proposed scheme begins by estimating sign al and noise correlation matrices. The latter is based on multiple regression theory. The signal subspace is then identified by selectin g the set of signal eigenvalue s that best represents the data, in the least-square sense [48,49 ], we note, however, that VCA works with projected and with unprojected data. The extraction of the end members exploits two facts: (1) the endmembers are the vertices of a simplex and (2) the affine transformation of a simplex is also a simplex. As PPI and N-FIND R algorithms, VCA also assumes the presence of pure pixels in the data. The algorithm iteratively projects data on to a direction orthogonal to the subspace spanned by the endmembers already determined. The new end member signature corresponds to the extreme of the projection. The algorithm iterates until all end members are exhausted. VCA performs much better than PPI and better than or comparable to N-FI NDR; yet it has a computational complexity between on e and two orders of magnitude lower than N-FINDR. The chapter is structure d as follows. Section 19.2 describes the fundamentals of the proposed method. Section 19.3 and Section 19.4 evaluate the proposed algorithm using simulated and real data, respectively. Section 19.5 presents some concluding remarks.
Resumo:
The objective of this work was to propose a new selection strategy for the initial stages of sugarcane improvement, based on the methodology 'simulated individual BLUP (BLUPIS)', which promotes a dynamic allocation of individuals selected in each full-sib family, using BLUP as a base for both the genotypic effects of the referred families and plot effects. The method proposed applies to single full-sib families or those obtained from unbalanced or balanced diallel crosses, half-sib families and self-pollinated families. BLUPIS indicates the number of individuals to be selected within each family, the total number of clones to be advanced, and the number of families to contribute with selected individuals. Correlation between BLUPIS and true BLUP was 0.96, by method validation. Additionally, BLUPIS allows the identification of which replication contains the best individuals of each family.
Resumo:
The signalling function of melanin-based colouration is debated. Sexual selection theory states that ornaments should be costly to produce, maintain, wear or display to signal quality honestly to potential mates or competitors. An increasing number of studies supports the hypothesis that the degree of melanism covaries with aspects of body condition (e.g. body mass or immunity), which has contributed to change the initial perception that melanin-based colour ornaments entail no costs. Indeed, the expression of many (but not all) melanin-based colour traits is weakly sensitive to the environment but strongly heritable suggesting that these colour traits are relatively cheap to produce and maintain, thus raising the question of how such colour traits could signal quality honestly. Here I review the production, maintenance and wearing/displaying costs that can generate a correlation between melanin-based colouration and body condition, and consider other evolutionary mechanisms that can also lead to covariation between colour and body condition. Because genes controlling melanic traits can affect numerous phenotypic traits, pleiotropy could also explain a linkage between body condition and colouration. Pleiotropy may result in differently coloured individuals signalling different aspects of quality that are maintained by frequency-dependent selection or local adaptation. Colouration may therefore not signal absolute quality to potential mates or competitors (e.g. dark males may not achieve a higher fitness than pale males); otherwise genetic variation would be rapidly depleted by directional selection. As a consequence, selection on heritable melanin-based colouration may not always be directional, but mate choice may be conditional to environmental conditions (i.e. context-dependent sexual selection). Despite the interest of evolutionary biologists in the adaptive value of melanin-based colouration, its actual role in sexual selection is still poorly understood.
Resumo:
The study was conducted in two different locations in South Brazil, in tillage in the 2009/2010 season on eight sunflower hybrids, aiming to determine the path correlations and coefficients between primary and secondary characters on the main variable of achene productivity. The correlations were similar between environments. The characters of the head diameter and mass of a thousand achenes had a significant influence on sunflower productivity. Based on the magnitude of the direct and indirect effects, we highlighted all primary components on the main variable, beside the good determination coefficient and low residual effect. The secondary component, the number of achenes, despite the significant direct effect on productivity, was indirectly influenced by the primary components, making it an undesirable character for selection.
Resumo:
Background: It is yet unclear if there are differences between using electronic key feature problems (KFPs) or electronic case-based multiple choice questions (cbMCQ) for the assessment of clinical decision making. Summary of Work: Fifth year medical students were exposed to clerkships which ended with a summative exam. Assessment of knowledge per exam was done by 6-9 KFPs, 9-20 cbMCQ and 9-28 MC questions. Each KFP consisted of a case vignette and three key features (KF) using “long menu” as question format. We sought students’ perceptions of the KFPs and cbMCQs in focus groups (n of students=39). Furthermore statistical data of 11 exams (n of students=377) concerning the KFPs and (cb)MCQs were compared. Summary of Results: The analysis of the focus groups resulted in four themes reflecting students’ perceptions of KFPs and their comparison with (cb)MCQ: KFPs were perceived as (i) more realistic, (ii) more difficult, (iii) more motivating for the intense study of clinical reasoning than (cb)MCQ and (iv) showed an overall good acceptance when some preconditions are taken into account. The statistical analysis revealed that there was no difference in difficulty; however KFP showed a higher discrimination and reliability (G-coefficient) even when corrected for testing times. Correlation of the different exam parts was intermediate. Conclusions: Students perceived the KFPs as more motivating for the study of clinical reasoning. Statistically KFPs showed a higher discrimination and higher reliability than cbMCQs. Take-home messages: Including KFPs with long menu questions into summative clerkship exams seems to offer positive educational effects.
Resumo:
This paper presents a strategy for solving the feature matching problem in calibrated very wide-baseline camera settings. In this kind of settings, perspective distortion, depth discontinuities and occlusion represent enormous challenges. The proposed strategy addresses them by using geometrical information, specifically by exploiting epipolar-constraints. As a result it provides a sparse number of reliable feature points for which 3D position is accurately recovered. Special features known as junctions are used for robust matching. In particular, a strategy for refinement of junction end-point matching is proposed which enhances usual junction-based approaches. This allows to compute cross-correlation between perfectly aligned plane patches in both images, thus yielding better matching results. Evaluation of experimental results proves the effectiveness of the proposed algorithm in very wide-baseline environments.
Resumo:
Due to dynamic variability, identifying the specific conditions under which non-functional requirements (NFRs) are satisfied may be only possible at runtime. Therefore, it is necessary to consider the dynamic treatment of relevant information during the requirements specifications. The associated data can be gathered by monitoring the execution of the application and its underlying environment to support reasoning about how the current application configuration is fulfilling the established requirements. This paper presents a dynamic decision-making infrastructure to support both NFRs representation and monitoring, and to reason about the degree of satisfaction of NFRs during runtime. The infrastructure is composed of: (i) an extended feature model aligned with a domain-specific language for representing NFRs to be monitored at runtime; (ii) a monitoring infrastructure to continuously assess NFRs at runtime; and (iii) a exible decision-making process to select the best available configuration based on the satisfaction degree of the NRFs. The evaluation of the approach has shown that it is able to choose application configurations that well fit user NFRs based on runtime information. The evaluation also revealed that the proposed infrastructure provided consistent indicators regarding the best application configurations that fit user NFRs. Finally, a benefit of our approach is that it allows us to quantify the level of satisfaction with respect to NFRs specification.
Resumo:
Survival or longevity is an economically important trait in beef cattle. The main inconvenience for its inclusion in selection criteria is delayed recording of phenotypic data and the high computational demand for including survival in proportional hazard models. Thus, identification of a longevity-correlated trait that could be recorded early in life would be very useful for selection purposes. We estimated the genetic relationship of survival with productive and reproductive traits in Nellore cattle, including weaning weight (WW), post-weaning growth (PWG), muscularity (MUSC), scrotal circumference at 18 months (SC18), and heifer pregnancy (HP). Survival was measured in discrete time intervals and modeled through a sequential threshold model. Five independent bivariate Bayesian analyses were performed, accounting for cow survival and the five productive and reproductive traits. Posterior mean estimates for heritability (standard deviation in parentheses) were 0.55 (0.01) for WW, 0.25 (0.01) for PWG, 0.23 (0.01) for MUSC, and 0.48 (0.01) for SC18. The posterior mean estimates (95% confidence interval in parentheses) for the genetic correlation with survival were 0.16 (0.13-0.19), 0.30 (0.25-0.34), 0.31 (0.25-0.36), 0.07 (0.02-0.12), and 0.82 (0.78-0.86) for WW, PWG, MUSC, SC18, and HP, respectively. Based on the high genetic correlation and heritability (0.54) posterior mean estimates for HP, the expected progeny difference for HP can be used to select bulls for longevity, as well as for post-weaning gain and muscle score.
Resumo:
This study presents a decision-making method for maintenance policy selection of power plants equipment. The method is based on risk analysis concepts. The method first step consists in identifying critical equipment both for power plant operational performance and availability based on risk concepts. The second step involves the proposal of a potential maintenance policy that could be applied to critical equipment in order to increase its availability. The costs associated with each potential maintenance policy must be estimated, including the maintenance costs and the cost of failure that measures the critical equipment failure consequences for the power plant operation. Once the failure probabilities and the costs of failures are estimated, a decision-making procedure is applied to select the best maintenance policy. The decision criterion is to minimize the equipment cost of failure, considering the costs and likelihood of occurrence of failure scenarios. The method is applied to the analysis of a lubrication oil system used in gas turbines journal bearings. The turbine has more than 150 MW nominal output, installed in an open cycle thermoelectric power plant. A design modification with the installation of a redundant oil pump is proposed for lubricating oil system availability improvement. (C) 2009 Elsevier Ltd. All rights reserved.
Resumo:
A 12 week kayak training programme was evaluated in children who either had or did not have the anthropometric characteristics identified as being unique to senior elite sprint kayakers. Altogether, 234 male and female school children were screened to select 10 children with and 10 children without the identified key anthropometric characteristics. Before and after training, the children completed an all-out 2 min kayak ergometer simulation test; measures of oxygen consumption, plasma lactate and total work accomplished were recorded. In addition, a 500 m time trial was performed at weeks 3 and 12. The coaches were unaware which 20 children possessed those anthropometric characteristics deemed to favour development of kayak ability. All children improved in both the 2 min ergometer simulation test and 500 m time trial. However, boys who were selected according to favourable anthropometric characteristics showed greater improvement than those without such characteristics in the 2 min ergometer test only. In summary, in a small group of children selected according to anthropometric data unique to elite adult kayakers, 12 weeks of intensive kayak training did not influence the rate of improvement of on-water sprint kayak performance.
Resumo:
Background: Condition-dependence is a ubiquitous feature of animal life histories and has important implications for both natural and sexual selection. Mate choice, for instance, is typically based on condition-dependent signals. Theory predicts that one reason why condition-dependent signals may be special is that they allow females to scan for genes that confer high parasite resistance. Such explanations require a genetic link between immunocompetence and body condition, but existing evidence is limited to phenotypic associations. It remains unknown, therefore, whether females selecting males with good body condition simply obtain a healthy mate, or if they acquire genes for their offspring that confer high immunocompetence. Results: Here we use a cross-foster experimental design to partition the phenotypic covariance in indices of body condition and immunocompetence into genetic, maternal and environmental effects in a passerine bird, the zebra finch Taeniopygia guttata. We show that there is significant positive additive genetic covariance between an index of body condition and an index of cell-mediated immune response. In this case, genetic variance in the index of immune response explained 56% of the additive genetic variance in the index of body condition. Conclusion: Our results suggest that, in the context of sexual selection, females that assess males on the basis of condition-dependent signals may gain genes that confer high immunocompetence for their offspring. More generally, a genetic correlation between indices of body condition and imuunocompetence supports the hypothesis that parasite resistance may be an important target of natural selection. Additional work is now required to test whether genetic covariance exists among other aspects of both condition and immunocompetence.
Resumo:
Object. The goal of this paper is to analyze the extension and relationships of glomus jugulare tumor with the temporal bone and the results of its surgical treatment aiming at preservation of the facial nerve. Based on the tumor extension and its relationships with the facial nerve, new criteria to be used in the selection of different surgical approaches are proposed. Methods. Between December 1997 and December 2007, 34 patients (22 female and 12 male) with glomus jugulare tumors were treated. Their mean age was 48 years. The mean follow-up was 52.5 months. Clinical findings included hearing loss in 88%, swallowing disturbance in 50%, and facial nerve palsy in 41%. Magnetic resonance imaging demonstrated a mass in the jugular foramen in all cases, a mass in the middle ear in 97%, a cervical mass in 85%, and an intradural mass in 41%. The tumor was supplied by the external carotid artery in all cases, the internal carotid artery in 44%, and the vertebral artery in 32%. Preoperative embolization was performed in 15 cases. The approach was tailored to each patient, and 4 types of approaches were designed. The infralabyrinthine retrofacial approach (Type A) was used in 32.5%; infralabyrinthine pre- and retrofacial approach without occlusion of the external acoustic meatus (Type B) in 20.5%; infralabyrinthine pre- and retrofacial approach with occlusion of the external acoustic meatus (Type C) in 41 W. and the infralabyrinthine approach with transposition of the facial nerve and removal of the middle ear structures (Type D) in 6% of the patients. Results. Radical removal was achieved in 91% of the cases and partial removal in 9%. Among 20 patients without preoperative facial nerve dysfunction, the nerve was kept in anatomical position in 19 (95%), and facial nerve function was normal during the immediate postoperative period in 17 (85%). Six patients (17.6%) had a new lower cranial nerve deficit, but recovery of swallowing function was adequate in all cases. Voice disturbance remained in all 6 cases. Cerebrospinal fluid leakage occurred in 6 patients (17.6%), with no need for reoperation in any of them. One patient died in the postoperative period due to pulmonary complications. The global recovery, based on the Karnofsky Performance Scale (KPS), was 100% in 15% of the patients, 90% in 45%, 80% in 33%, and 70% in 6%. Conclusions. Radical removal of glomus jugulare tumor can be achieved without anterior transposition of the facial nerve. The extension of dissection, however, should be tailored to each case based on tumor blood supply, preoperative symptoms, and tumor extension. The operative field provided by the retrofacial infralabyrinthine approach, or the pre- and retrofacial approaches. with or without Closure of the external acoustic meatus, allows a wide exposure of the jugular foramen area. Global functional recovery based on the KPS is acceptable in 94% of the patients. (DOI: 10.3171/2008.10.JNS08612)
Resumo:
Motivation: This paper introduces the software EMMIX-GENE that has been developed for the specific purpose of a model-based approach to the clustering of microarray expression data, in particular, of tissue samples on a very large number of genes. The latter is a nonstandard problem in parametric cluster analysis because the dimension of the feature space (the number of genes) is typically much greater than the number of tissues. A feasible approach is provided by first selecting a subset of the genes relevant for the clustering of the tissue samples by fitting mixtures of t distributions to rank the genes in order of increasing size of the likelihood ratio statistic for the test of one versus two components in the mixture model. The imposition of a threshold on the likelihood ratio statistic used in conjunction with a threshold on the size of a cluster allows the selection of a relevant set of genes. However, even this reduced set of genes will usually be too large for a normal mixture model to be fitted directly to the tissues, and so the use of mixtures of factor analyzers is exploited to reduce effectively the dimension of the feature space of genes. Results: The usefulness of the EMMIX-GENE approach for the clustering of tissue samples is demonstrated on two well-known data sets on colon and leukaemia tissues. For both data sets, relevant subsets of the genes are able to be selected that reveal interesting clusterings of the tissues that are either consistent with the external classification of the tissues or with background and biological knowledge of these sets.
Resumo:
The selection, synthesis and chromatographic evaluation of a synthetic affinity adsorbent for human recombinant factor VIIa is described. The requirement for a metal ion-dependent immunoadsorbent step in the purification of the recombinant human clotting factor, FVIIa, has been obviated by using the X-ray crystallographic structure of the complex of tissue factor (TF) and Factor VIIa and has directed our combinatorial approach to select, synthesise and evaluate a rationally-selected affinity adsorbent from a limited library of putative ligands. The selected and optimised ligand comprises a triazine scaffold bis-substituted with 3-aminobenzoic acid and has been shown to bind selectively to FVIIa in a Ca2+-dependent manner. The adsorbent purifies FVIIa to almost identical purity (>99%), yield (99%), activation/degradation profile and impurity content (∼1000 ppm) as the current immunoadsorption process, while displaying a 10-fold higher static capacity and substantially higher reusability and durability. © 2002 Elsevier Science B.V. All rights reserved.
Resumo:
The thin-layer drying behaviour of bananas in a beat pump dehumidifier dryer was examined. Four pre-treatments (blanching, chilling, freezing and combined blanching and freezing) were applied to the bananas, which were dried at 50 degreesC with an air velocity of 3.1 m s(-1) and with the relative humidity of the inlet air of 10-35%. Three drying models, the simple model, the two-term exponential model and the Page model were examined. All models were evaluated using three statistical measures, correlation coefficient, root means square error, and mean absolute percent error. Moisture diffusivity was calculated based on the diffusion equation for an infinite cylindrical shape using the slope method. The rate of drying was higher for the pre-treatments involving freezing. The sample which was blanched only did not show any improvement in drying rate. In fact, a longer drying time resulted due to water absorption during blanching. There was no change in the rate for the chilled sample compared with the control. While all models closely fitted the drying data, the simple model showed greatest deviation from the experimental results. The two-term exponential model was found to be the best model for describing the drying curves of bananas because its parameters represent better the physical characteristics of the drying process. Moisture diffusivities of bananas were in the range 4.3-13.2 x 10(-10) m(2)s(-1). (C) 2002 Published by Elsevier Science Ltd.