927 resultados para Linear discriminant analysis
Resumo:
Habitat-based statistical models relating patterns of presence and absence of species to habitat variables could be useful to resolve conservation-related problems and highlight the causes of population declines. In this paper, we apply such a modelling approach to an endemic amphibian, the Sardinian mountain newt Euproctus platycephalus, considered by IUCN a critically endangered species. Sardinian newts inhabit freshwater habitat in streams, small lakes and pools on the island of Sardinia (Italy). Reported declines of newt populations are not yet supported by quantitative data, however, they are perceived or suspected across the species' historical range. This study represents a first attempt trying to statistically relate habitat characteristics to Sardinian newt occurrence and persistence. Linear regression analysis revealed that newts are more likely to be found in sites with colder water temperature, less riparian vegetation and, marginally, absence of fish. The implications of the results for the conservation of the species are discussed, and suggestions for the short-term management of newt inhabited sites suggested. (C) 2003 Elsevier Ltd. All rights reserved.
Resumo:
Although apolipoprotein AN (apoA-V) polymorphisms have been consistently associated with fasting triglyceride (TG) levels, their impact on postprandial lipemia remains relatively unknown. In this study, we investigate the impact of two common apoA-V polymorphisms (-1131 T>C and S19W) and apoA-V haplotypes on fasting and postprandial lipid metabolism in adults in the United Kingdom (n = 259). Compared with the wild-type TT, apoA-V -1131 TC heterozygotes had 15% (P = 0.057) and 21% (P = 0.002) higher fasting TG and postprandial TG area under the curve (AUC), respectively. Significant (P = 0.038) and nearly significant (P = 0.057) gender X genotype interactions were observed for fasting TG and TG AUC, with a greater impact of genotype in males. Lower HDL-cholesterol was associated with the rare TC genotype (P = 0.047). Significant linkage disequilibrium was found between the apoA-V -1131 T>C and the apoC-III 3238 C>G variants, with univariate analysis indicating an impact of this apoC-III single nucleotide polymorphism (SNP) on TG AUC (P = 0.015). However, in linear regression analysis, a significant independent association with TG AUC (P = 0.007) was only evident for the apoA-V -1131 T>C SNP, indicating a greater relative importance of the apoA-V genotype.
Resumo:
Craloxylum formosum Dyer is consumed throughout the year as food and medicine in Thailand. It contains large amounts of chlorogenic acid and quinic acid derivatives. The antioxidative activity of the extract was studied in refined soybean oil coating on rice crackers without any seasoning. They were stored in accelerated oxidation conditions at 40 degrees C, 80% relative humidity (RH) in the dark for 18 days. The oxidative state of each sample was monitored by analyzing of the peroxide value (PV) and thiobarbituric acid reactive substances (TBARS) as well as by odor analysis by quantitative descriptive analysis (QDA). The C formosum extract was more effective than alpha-tocopherol due to metal ions present in the crackers, which resulted in alpha-tocopherol being less effective as an antioxidant. Sensory odor attributes of rice crackers were related more closely to TBARS than to PV values by linear regression analysis. The present study indicated that C. formosum extract was a promising source of a natural food antioxidant and was effective in inhibiting lipid oxidation in rice crackers.
Resumo:
This work analyzes the use of linear discriminant models, multi-layer perceptron neural networks and wavelet networks for corporate financial distress prediction. Although simple and easy to interpret, linear models require statistical assumptions that may be unrealistic. Neural networks are able to discriminate patterns that are not linearly separable, but the large number of parameters involved in a neural model often causes generalization problems. Wavelet networks are classification models that implement nonlinear discriminant surfaces as the superposition of dilated and translated versions of a single "mother wavelet" function. In this paper, an algorithm is proposed to select dilation and translation parameters that yield a wavelet network classifier with good parsimony characteristics. The models are compared in a case study involving failed and continuing British firms in the period 1997-2000. Problems associated with over-parameterized neural networks are illustrated and the Optimal Brain Damage pruning technique is employed to obtain a parsimonious neural model. The results, supported by a re-sampling study, show that both neural and wavelet networks may be a valid alternative to classical linear discriminant models.
Resumo:
Much prior research on the structure and performance of UK real estate portfolios has relied on aggregated measures for sector and region. For these groupings to have validity, the performance of individual properties within each group should be similar. This paper analyses a sample of 1,200 properties using multiple discriminant analysis and cluster analysis techniques. It is shown that conventional property type and spatial classifications do not capture the variation in return behaviour at the individual building level. The major feature is heterogeneity - but there may be distinctions between growth and income properties and between single and multi-let properties that could help refine portfolio structures.
Resumo:
The recent global economic crisis is often associated with the development and pricing of mortgage-backed securities (i.e. MBSs) and underlying products (i.e. sub-prime mortgages). This work uses a rich database of MBS issues and represents the first attempt to price commercial MBSs (i.e. CMBSs) in the European market. Our results are consistent with research carried out in the US market and we find that bond-, mortgage-, real estate-related and multinational characteristics show different degrees of significance in explaining European CMBS spreads at issuance. Multiple linear regression analysis using a databank of CMBSs issued between 1997 and 2007 indicates a strong relationship with bond-related factors, followed by real estate and mortgage market conditions. We also find that multinational factors are significant, with country of issuance, collateral location and access to more liquid markets all being important in explaining the cost of secured funding for real estate companies. As floater coupon tranches tend to be riskier and exhibit higher spreads, we also estimate a model using this sub-set of data and results hold, hence reinforcing our findings. Finally, we estimate our model for both tranches A and B and find that real estate factors become relatively more important for the riskier investment products.
Resumo:
In this paper we explore classification techniques for ill-posed problems. Two classes are linearly separable in some Hilbert space X if they can be separated by a hyperplane. We investigate stable separability, i.e. the case where we have a positive distance between two separating hyperplanes. When the data in the space Y is generated by a compact operator A applied to the system states ∈ X, we will show that in general we do not obtain stable separability in Y even if the problem in X is stably separable. In particular, we show this for the case where a nonlinear classification is generated from a non-convergent family of linear classes in X. We apply our results to the problem of quality control of fuel cells where we classify fuel cells according to their efficiency. We can potentially classify a fuel cell using either some external measured magnetic field or some internal current. However we cannot measure the current directly since we cannot access the fuel cell in operation. The first possibility is to apply discrimination techniques directly to the measured magnetic fields. The second approach first reconstructs currents and then carries out the classification on the current distributions. We show that both approaches need regularization and that the regularized classifications are not equivalent in general. Finally, we investigate a widely used linear classification algorithm Fisher's linear discriminant with respect to its ill-posedness when applied to data generated via a compact integral operator. We show that the method cannot stay stable when the number of measurement points becomes large.
Resumo:
Using NCANDS data of US child maltreatment reports for 2009, logistic regression, probit analysis, discriminant analysis and an artificial neural network are used to determine the factors which explain the decision to place a child in out-of-home care. As well as developing a new model for 2009, a previous study using 2005 data is replicated. While there are many small differences, the four estimation techniques give broadly the same results, demonstrating the robustness of the results. Similarly, apart from age and sexual abuse, the 2005 and 2009 results are roughly similar. For 2009, child characteristics (particularly child emotional problems) are more important than the nature of the abuse and the situation of the household; while caregiver characteristics are the least important. All these models have low explanatory power.
Resumo:
A continuous tropospheric and stratospheric vertically resolved ozone time series, from 1850 to 2099, has been generated to be used as forcing in global climate models that do not include interactive chemistry. A multiple linear regression analysis of SAGE I+II satellite observations and polar ozonesonde measurements is used for the stratospheric zonal mean dataset during the well-observed period from 1979 to 2009. In addition to terms describing the mean annual cycle, the regression includes terms representing equivalent effective stratospheric chlorine (EESC) and the 11-yr solar cycle variability. The EESC regression fit coefficients, together with pre-1979 EESC values, are used to extrapolate the stratospheric ozone time series backward to 1850. While a similar procedure could be used to extrapolate into the future, coupled chemistry climate model (CCM) simulations indicate that future stratospheric ozone abundances are likely to be significantly affected by climate change, and capturing such effects through a regression model approach is not feasible. Therefore, the stratospheric ozone dataset is extended into the future (merged in 2009) with multimodel mean projections from 13 CCMs that performed a simulation until 2099 under the SRES (Special Report on Emission Scenarios) A1B greenhouse gas scenario and the A1 adjusted halogen scenario in the second round of the Chemistry-Climate Model Validation (CCMVal-2) Activity. The stratospheric zonal mean ozone time series is merged with a three-dimensional tropospheric data set extracted from simulations of the past by two CCMs (CAM3.5 and GISSPUCCINI)and of the future by one CCM (CAM3.5). The future tropospheric ozone time series continues the historical CAM3.5 simulation until 2099 following the four different Representative Concentration Pathways (RCPs). Generally good agreement is found between the historical segment of the ozone database and satellite observations, although it should be noted that total column ozone is overestimated in the southern polar latitudes during spring and tropospheric column ozone is slightly underestimated. Vertical profiles of tropospheric ozone are broadly consistent with ozonesondes and in-situ measurements, with some deviations in regions of biomass burning. The tropospheric ozone radiative forcing (RF) from the 1850s to the 2000s is 0.23Wm−2, lower than previous results. The lower value is mainly due to (i) a smaller increase in biomass burning emissions; (ii) a larger influence of stratospheric ozone depletion on upper tropospheric ozone at high southern latitudes; and possibly (iii) a larger influence of clouds (which act to reduce the net forcing) compared to previous radiative forcing calculations. Over the same period, decreases in stratospheric ozone, mainly at high latitudes, produce a RF of −0.08Wm−2, which is more negative than the central Intergovernmental Panel on Climate Change (IPCC) Fourth Assessment Report (AR4) value of −0.05Wm−2, but which is within the stated range of −0.15 to +0.05Wm−2. The more negative value is explained by the fact that the regression model simulates significant ozone depletion prior to 1979, in line with the increase in EESC and as confirmed by CCMs, while the AR4 assumed no change in stratospheric RF prior to 1979. A negative RF of similar magnitude persists into the future, although its location shifts from high latitudes to the tropics. This shift is due to increases in polar stratospheric ozone, but decreases in tropical lower stratospheric ozone, related to a strengthening of the Brewer-Dobson circulation, particularly through the latter half of the 21st century. Differences in trends in tropospheric ozone among the four RCPs are mainly driven by different methane concentrations, resulting in a range of tropospheric ozone RFs between 0.4 and 0.1Wm−2 by 2100. The ozone dataset described here has been released for the Coupled Model Intercomparison Project (CMIP5) model simulations in netCDF Climate and Forecast (CF) Metadata Convention at the PCMDI website (http://cmip-pcmdi.llnl.gov/).
Resumo:
In this paper, various types of fault detection methods for fuel cells are compared. For example, those that use a model based approach or a data driven approach or a combination of the two. The potential advantages and drawbacks of each method are discussed and comparisons between methods are made. In particular, classification algorithms are investigated, which separate a data set into classes or clusters based on some prior knowledge or measure of similarity. In particular, the application of classification methods to vectors of reconstructed currents by magnetic tomography or to vectors of magnetic field measurements directly is explored. Bases are simulated using the finite integration technique (FIT) and regularization techniques are employed to overcome ill-posedness. Fisher's linear discriminant is used to illustrate these concepts. Numerical experiments show that the ill-posedness of the magnetic tomography problem is a part of the classification problem on magnetic field measurements as well. This is independent of the particular working mode of the cell but influenced by the type of faulty behavior that is studied. The numerical results demonstrate the ill-posedness by the exponential decay behavior of the singular values for three examples of fault classes.
Resumo:
Neural field models describe the coarse-grained activity of populations of interacting neurons. Because of the laminar structure of real cortical tissue they are often studied in two spatial dimensions, where they are well known to generate rich patterns of spatiotemporal activity. Such patterns have been interpreted in a variety of contexts ranging from the understanding of visual hallucinations to the generation of electroencephalographic signals. Typical patterns include localized solutions in the form of traveling spots, as well as intricate labyrinthine structures. These patterns are naturally defined by the interface between low and high states of neural activity. Here we derive the equations of motion for such interfaces and show, for a Heaviside firing rate, that the normal velocity of an interface is given in terms of a non-local Biot-Savart type interaction over the boundaries of the high activity regions. This exact, but dimensionally reduced, system of equations is solved numerically and shown to be in excellent agreement with the full nonlinear integral equation defining the neural field. We develop a linear stability analysis for the interface dynamics that allows us to understand the mechanisms of pattern formation that arise from instabilities of spots, rings, stripes and fronts. We further show how to analyze neural field models with linear adaptation currents, and determine the conditions for the dynamic instability of spots that can give rise to breathers and traveling waves.
Resumo:
Neural field models of firing rate activity typically take the form of integral equations with space-dependent axonal delays. Under natural assumptions on the synaptic connectivity we show how one can derive an equivalent partial differential equation (PDE) model that properly treats the axonal delay terms of the integral formulation. Our analysis avoids the so-called long-wavelength approximation that has previously been used to formulate PDE models for neural activity in two spatial dimensions. Direct numerical simulations of this PDE model show instabilities of the homogeneous steady state that are in full agreement with a Turing instability analysis of the original integral model. We discuss the benefits of such a local model and its usefulness in modeling electrocortical activity. In particular, we are able to treat “patchy” connections, whereby a homogeneous and isotropic system is modulated in a spatially periodic fashion. In this case the emergence of a “lattice-directed” traveling wave predicted by a linear instability analysis is confirmed by the numerical simulation of an appropriate set of coupled PDEs.
Resumo:
Wine production is largely governed by atmospheric conditions, such as air temperature and precipitation, together with soil management and viticultural/enological practices. Therefore, anthropogenic climate change is likely to have important impacts on the winemaking sector worldwide. An important winemaking region is the Portuguese Douro Valley, which is known by its world-famous Port Wine. The identification of robust relationships between atmospheric factors and wine parameters is of great relevance for the region. A multivariate linear regression analysis of a long wine production series (1932–2010) reveals that high rainfall and cool temperatures during budburst, shoot and inflorescence development (February-March) and warm temperatures during flowering and berry development (May) are generally favourable to high production. The probabilities of occurrence of three production categories (low, normal and high) are also modelled using multinomial logistic regression. Results show that both statistical models are valuable tools for predicting the production in a given year with a lead time of 3–4 months prior to harvest. These statistical models are applied to an ensemble of 16 regional climate model experiments following the SRES A1B scenario to estimate possible future changes. Wine production is projected to increase by about 10 % by the end of the 21st century, while the occurrence of high production years is expected to increase from 25 % to over 60 %. Nevertheless, further model development will be needed to include other aspects that may shape production in the future. In particular, the rising heat stress and/or changes in ripening conditions could limit the projected production increase in future decades.
Resumo:
We introduce a novel technique in which linear regression analysis is applied to clusters of tracked cyclones to statistically assess the factors controlling cyclone development. We illustrate this technique by evaluating the differences between cyclones forming in the west and east North Atlantic (herein termed west and east Atlantic cyclones). Enhanced cyclone intensity 2 days after genesis is found to be associated with deeper upper-level troughs upstream of the cyclone center at the genesis time in both west and east Atlantic cyclones. However, whilst west Atlantic cyclones are also enhanced by the presence of strong fronts, east Atlantic cyclones are not. Instead, east Atlantic cyclones exhibit an enhancement when diabatically generated midlevel potential vorticity is present (with the enhancement being of approximately equal magnitude to that associated with the potential vorticity in the upper-level trough). This is consistent with the paradigm of latent heat release in the warm conveyor belt region playing an important role in the development of east Atlantic cyclones.
Resumo:
The quantitative effects of uniform strain and background rotation on the stability of a strip of constant vorticity (a simple shear layer) are examined. The thickness of the strip decreases in time under the strain, so it is necessary to formulate the linear stability analysis for a time-dependent basic flow. The results show that even a strain rate γ (scaled with the vorticity of the strip) as small as 0.25 suppresses the conventional Rayleigh shear instability mechanism, in the sense that the r.m.s. wave steepness cannot amplify by more than a certain factor, and must eventually decay. For γ < 0.25 the amplification factor increases as γ decreases; however, it is only 3 when γ e 0.065. Numerical simulations confirm the predictions of linear theory at small steepness and predict a threshold value necessary for the formation of coherent vortices. The results help to explain the impression from numerous simulations of two-dimensional turbulence reported in the literature that filaments of vorticity infrequently roll up into vortices. The stabilization effect may be expected to extend to two- and three-dimensional quasi-geostrophic flows.