24 resultados para principle component analysis
em Universidade Federal do Rio Grande do Norte(UFRN)
Resumo:
LOPES-DOS-SANTOS, V. , CONDE-OCAZIONEZ, S. ; NICOLELIS, M. A. L. , RIBEIRO, S. T. , TORT, A. B. L. . Neuronal assembly detection and cell membership specification by principal component analysis. Plos One, v. 6, p. e20996, 2011.
Resumo:
LOPES-DOS-SANTOS, V. , CONDE-OCAZIONEZ, S. ; NICOLELIS, M. A. L. , RIBEIRO, S. T. , TORT, A. B. L. . Neuronal assembly detection and cell membership specification by principal component analysis. Plos One, v. 6, p. e20996, 2011.
Resumo:
Recent progress in the technology for single unit recordings has given the neuroscientific community theopportunity to record the spiking activity of large neuronal populations. At the same pace, statistical andmathematical tools were developed to deal with high-dimensional datasets typical of such recordings.A major line of research investigates the functional role of subsets of neurons with significant co-firingbehavior: the Hebbian cell assemblies. Here we review three linear methods for the detection of cellassemblies in large neuronal populations that rely on principal and independent component analysis.Based on their performance in spike train simulations, we propose a modified framework that incorpo-rates multiple features of these previous methods. We apply the new framework to actual single unitrecordings and show the existence of cell assemblies in the rat hippocampus, which typically oscillate attheta frequencies and couple to different phases of the underlying field rhythm
Resumo:
Untreated effluents that reach surface water affect the aquatic life and humans. This study aimed to evaluate the wastewater s toxicity (municipal, industrial and shrimp pond effluents) released in the Estuarine Complex of Jundiaí- Potengi, Natal/RN, through chronic quantitative e qualitative toxicity tests using the test organism Mysidopsis Juniae, CRUSTACEA, MYSIDACEA (Silva, 1979). For this, a new methodology for viewing chronic effects on organisms of M. juniae was used (only renewal), based on another existing methodology to another testorganism very similar to M. Juniae, the M. Bahia (daily renewal).Toxicity tests 7 days duration were used for detecting effects on the survival and fecundity in M. juniae. Lethal Concentration 50% (LC50%) was determined by the Trimmed Spearman-Karber; Inhibition Concentration 50% (IC50%) in fecundity was determined by Linear Interpolation. ANOVA (One Way) tests (p = 0.05) were used to determinate the No Observed Effect Concentration (NOEC) and Low Observed Effect Concentration (LOEC). Effluents flows were measured and the toxic load of the effluents was estimated. Multivariate analysis - Principal Component Analysis (PCA) and Correspondence Analysis (CA) - identified the physic-chemical parameters better explain the patterns of toxicity found in survival and fecundity of M. juniae. We verified the feasibility of applying the only renewal system in chronic tests with M. Juniae. Most efluentes proved toxic on the survival and fecundity of M. Juniae, except for some shrimp pond effluents. The most toxic effluent was ETE Lagoa Aerada (LC50, 6.24%; IC50, 4.82%), ETE Quintas (LC50, 5.85%), Giselda Trigueiro Hospital (LC50, 2.05%), CLAN (LC50, 2.14%) and COTEMINAS (LC50, IC50 and 38.51%, 6.94%). The greatest toxic load was originated from ETE inefficient high flow effluents, textile effluents and CLAN. The organic load was related to the toxic effects of wastewater and hospital effluents in survival of M. Juniae, as well as heavy metals, total residual chlorine and phenols. In industrial effluents was found relationship between toxicity and organic load, phenols, oils and greases and benzene. The effects on fertility were related, in turn, with chlorine and heavy metals. Toxicity tests using other organisms of different trophic levels, as well as analysis of sediment toxicity are recommended to confirm the patterns found with M. Juniae. However, the results indicate the necessity for implementation and improvement of sewage treatment systems affluent to the Potengi s estuary
Resumo:
Ecomorphology is a science based on the idea that morphological differences among species could be associated with distinct biological and environmental pressures suffered by them. These differences can be studied employing morphological and biometric indexes denominated Ecomorphological attributes , representing standards that express characteristics of the individual in relation to its environment, and can be interpreted as indicators of life habits or adaptations suffered due its occupation of different habitats. This work aims to contribute for the knowledge of the ecomorphology of the Brazilian marine ichthyofauna, specifically from Galinhos, located at Rio Grande do Norte state. 10 different species of fish were studied, belonging the families Gerreidae (Eucinostomus argenteus), Haemulidae (Orthopristis ruber,Pomadasyscorvinaeformis,Haemulonaurolineatum,Haemulonplumieri,Haemulonsteindachneri), Lutjanidae (Lutjanus synagris), Paralichthyidae (Syaciummicrurum), Bothidae (Bothus ocellatus) and Tetraodontidae (Sphoeroidestestudineus), which were obtained during five collections, in the period time of September/2004 to April/2005, utilizing three special nets. The ecomorphological study was performed at the laboratory. Eight to ten samples of each fish specie were measured. Fifteen morphological aspects were considered to calculate twelve ecomorphological attributes. Multivariate statistical analysis methods such as Principal Component Analysis (PCA) and Cluster Analysis were done to identify ecmorphological patterns to describe the data set obtained. As results, H.aurolineatumwas the most abundant specie found (23,03%) and S.testudineusthe less one with 0,23%. The 1st Principal component showed variation of 60,03% with influence of the ecomorphological attribute related to body morphology, while the 2nd PC with 23,25% variation had influence of the ecomorphological attribute related to oral morphology. The Cluster Analiysis promoted the identification of three distinct groups Perciformes, Pleuronectiformes and Tetraodontiformes. Based on the obtained data, considering morphological characters differences among the species studied, we suggest that all of them live at the medium (E.argenteus,O.rubber, P.corvinaeformis,H.aurolineatum,H.plumieri,H.steindachneri,L.synagris) and bottom (S.micrurum,B.ocellatus,S.testudineus) region of column water.
Resumo:
Produced water is characterized as one of the most common wastes generated during exploration and production of oil. This work aims to develop methodologies based on comparative statistical processes of hydrogeochemical analysis of production zones in order to minimize types of high-cost interventions to perform identification test fluids - TIF. For the study, 27 samples were collected from five different production zones were measured a total of 50 chemical species. After the chemical analysis was applied the statistical data, using the R Statistical Software, version 2.11.1. Statistical analysis was performed in three steps. In the first stage, the objective was to investigate the behavior of chemical species under study in each area of production through the descriptive graphical analysis. The second step was to identify a function that classify production zones from each sample, using discriminant analysis. In the training stage, the rate of correct classification function of discriminant analysis was 85.19%. The next stage of processing of the data used for Principal Component Analysis, by reducing the number of variables obtained from the linear combination of chemical species, try to improve the discriminant function obtained in the second stage and increase the discrimination power of the data, but the result was not satisfactory. In Profile Analysis curves were obtained for each production area, based on the characteristics of the chemical species present in each zone. With this study it was possible to develop a method using hydrochemistry and statistical analysis that can be used to distinguish the water produced in mature fields of oil, so that it is possible to identify the zone of production that is contributing to the excessive elevation of the water volume.
Resumo:
The increase in ultraviolet radiation (UV) at surface, the high incidence of non-melanoma skin cancer (NMSC) in coast of Northeast of Brazil (NEB) and reduction of total ozone were the motivation for the present study. The overall objective was to identify and understand the variability of UV or Index Ultraviolet Radiation (UV Index) in the capitals of the east coast of the NEB and adjust stochastic models to time series of UV index aiming make predictions (interpolations) and forecasts / projections (extrapolations) followed by trend analysis. The methodology consisted of applying multivariate analysis (principal component analysis and cluster analysis), Predictive Mean Matching method for filling gaps in the data, autoregressive distributed lag (ADL) and Mann-Kendal. The modeling via the ADL consisted of parameter estimation, diagnostics, residuals analysis and evaluation of the quality of the predictions and forecasts via mean squared error and Pearson correlation coefficient. The research results indicated that the annual variability of UV in the capital of Rio Grande do Norte (Natal) has a feature in the months of September and October that consisting of a stabilization / reduction of UV index because of the greater annual concentration total ozone. The increased amount of aerosol during this period contributes in lesser intensity for this event. The increased amount of aerosol during this period contributes in lesser intensity for this event. The application of cluster analysis on the east coast of the NEB showed that this event also occurs in the capitals of Paraiba (João Pessoa) and Pernambuco (Recife). Extreme events of UV in NEB were analyzed from the city of Natal and were associated with absence of cloud cover and levels below the annual average of total ozone and did not occurring in the entire region because of the uneven spatial distribution of these variables. The ADL (4, 1) model, adjusted with data of the UV index and total ozone to period 2001-2012 made a the projection / extrapolation for the next 30 years (2013-2043) indicating in end of that period an increase to the UV index of one unit (approximately), case total ozone maintain the downward trend observed in study period
Resumo:
The dyslipidemia and excess weight in adolescents, when combined, suggest a progression of risk factors for cardiovascular disease (CVD). Besides these, the dietary habits and lifestyle have also been considered unsuitable impacting the development of chronic diseases. The study objectives were: (1) estimate the prevalence of lipid profile and correlate with body mass index (BMI), waist circumference (WC) and waist / height ratio (WHR) in adolescents, considering the maturation sexual, (2) know the sources of variance in the diet and the number of days needed to estimate the usual diet of adolescents and (3) describe the dietary patterns and lifestyle of adolescents, family history of CVD and age correlates them with the patterns of risk for CVD, adjusted for sexual maturation. A cross-sectional study was performed with 432 adolescents, aged 10-19 years from public schools of the Natal city, Brazil. The dyslipidemias were evaluated considering the lipid profile, the index of I Castelli (TC / HDL) and II (LDL / HDL) and non-HDL cholesterol. Anthropometric indicators were BMI, WC and WHR. The intake of energy, nutrients including fiber, fatty acids and cholesterol was estimated from two 24-hour recalls (24HR). The variables of lipid profile, anthropometric and clinical data were used in the models of Pearson correlation and linear regression, considering the sexual maturation. The variance ratio of the diet was calculated from the component-person variance, determined by analysis of variance (ANOVA). The definition of the number of days to estimate the usual intake of each nutrient was obtained by taking the hypothetical correlation (r) ≥ 0.9, between nutrient intake and the true observed. We used the principal component analysis as a method of extracting factors that 129 accounted for the dependent variables and known cardiovascular risk obtained from the lipid profile, the index for Castelli I and II, non-HDL cholesterol, BMI, and WC the WHR. Dietary patterns and lifestyle were obtained from the independent variables, based on nutrients consumed and physical activity weekly. In the study of principal component analysis (PCA) was investigated associations between the patterns of cardiovascular risk factors in dietary patterns and lifestyle, age and positive family history of CVD, through bivariate and multiple logistic regression adjusted for sexual maturation. The low HDL-C dyslipidemia was most prevalent (50.5%) for adolescents. Significant correlations were observed between hypercholesterolemia and positive family history of CVD (r = 0.19, p <0.01) and hypertriglyceridemia with BMI (r = 0.30, p <0.01), with the CC (r = 0.32, p <0.01) and WHR (r = 0.33, p <0.01). The linear model constructed with sexual maturation, age and BMI explained about 1 to 10.4% of the variation in the lipid profile. The sources of variance between individuals were greater for all nutrients in both sexes. The reasons for variances were 1 for all nutrients were higher in females. The results suggest that to assess the diet of adolescents with greater precision, 2 days would be enough to R24h consumption of energy, carbohydrates, fiber, saturated and monounsaturated fatty acids. In contrast, 3 days would be recommended for protein, lipid, polyunsaturated fatty acids and cholesterol. Two cardiovascular risk factors as have been extracted in the ACP, referring to the dependent variables: the standard lipid profile (HDL-C and non-HDL cholesterol) and "standard anthropometric index (BMI, WC, WHR) with a power explaining 75% of the variance of the original data. The factors are representative of two independent variables led to dietary patterns, "pattern 130 western diet" and "pattern protein diet", and one on the lifestyle, "pattern energy balance". Together, these patterns provide an explanation power of 67%. Made adjustment for sexual maturation in males remained significant variables: the associations between puberty and be pattern anthropometric indicator (OR = 3.32, CI 1.34 to 8.17%), and between family history of CVD and the pattern lipid profile (OR = 2.62, CI 1.20 to 5.72%). In females adolescents, associations were identified between age after the first stage of puberty with anthropometric pattern (OR = 3.59, CI 1.58 to 8.17%) and lipid profile (OR = 0.33, CI 0.15 to 0.75%). Conclusions: The low HDL-C was the most prevalent dyslipidemia independent of sex and nutritional status of adolescents. Hypercholesterolemia was influenced by family history of CVD and sexual maturation, in turn, hypertriglyceridemia was closely associated with anthropometric indicators. The variance between the diets was greater for all nutrients. This fact reflected in a variance ratio less than 1 and consequently in a lower number of days requerid to estimate the usual diet of adolescents considering gender. The two dietary patterns were extracted and the pattern considered unhealthy lifestyle as healthy. The associations were found between the patterns of CVD risk with age and family history of CVD in the studied adolescents
Resumo:
The demographic growth press environments that are more susceptible to perturbations, like riparian areas, without knowing about the effects of replacing these natural environments by different land uses on soil quality and, consequently, on watershed. The study of soil quality has evolved as an important tool for soil sustainable management of this component of the biosphere that affects aquatic and terrestrial ecosystems functions. Thus, physical and chemical soil proprieties were measured to assess soil quality under different land uses (agricultural, pasture, urban, industrial and natural vegetation,) in the riparian zone of Extremoz Lake, an important human water source, evaluating whether the soil offers potential risk to water pollution. Data were subjected to descriptive statistics and Principal Component Analysis (PCA). The results showed negative changes in soil quality such as alkalinization and increase in P, Pb, Mn and Zn contents in most anthropized areas. The sandy texture and low organic matter content in all soils showed the fragility of the soil to erosion and leaching of elements in excess to water bodies, evidencing that this soils has potential to diffuse contaminants. Conservative management of soil is necessary to provide an adequate ecological state in riparian zones of the Extremoz Lake, thus allowing controlling and buffering diffuse sources of pollution to this important water supply source
Resumo:
The problems of combinatory optimization have involved a large number of researchers in search of approximative solutions for them, since it is generally accepted that they are unsolvable in polynomial time. Initially, these solutions were focused on heuristics. Currently, metaheuristics are used more for this task, especially those based on evolutionary algorithms. The two main contributions of this work are: the creation of what is called an -Operon- heuristic, for the construction of the information chains necessary for the implementation of transgenetic (evolutionary) algorithms, mainly using statistical methodology - the Cluster Analysis and the Principal Component Analysis; and the utilization of statistical analyses that are adequate for the evaluation of the performance of the algorithms that are developed to solve these problems. The aim of the Operon is to construct good quality dynamic information chains to promote an -intelligent- search in the space of solutions. The Traveling Salesman Problem (TSP) is intended for applications based on a transgenetic algorithmic known as ProtoG. A strategy is also proposed for the renovation of part of the chromosome population indicated by adopting a minimum limit in the coefficient of variation of the adequation function of the individuals, with calculations based on the population. Statistical methodology is used for the evaluation of the performance of four algorithms, as follows: the proposed ProtoG, two memetic algorithms and a Simulated Annealing algorithm. Three performance analyses of these algorithms are proposed. The first is accomplished through the Logistic Regression, based on the probability of finding an optimal solution for a TSP instance by the algorithm being tested. The second is accomplished through Survival Analysis, based on a probability of the time observed for its execution until an optimal solution is achieved. The third is accomplished by means of a non-parametric Analysis of Variance, considering the Percent Error of the Solution (PES) obtained by the percentage in which the solution found exceeds the best solution available in the literature. Six experiments have been conducted applied to sixty-one instances of Euclidean TSP with sizes of up to 1,655 cities. The first two experiments deal with the adjustments of four parameters used in the ProtoG algorithm in an attempt to improve its performance. The last four have been undertaken to evaluate the performance of the ProtoG in comparison to the three algorithms adopted. For these sixty-one instances, it has been concluded on the grounds of statistical tests that there is evidence that the ProtoG performs better than these three algorithms in fifty instances. In addition, for the thirty-six instances considered in the last three trials in which the performance of the algorithms was evaluated through PES, it was observed that the PES average obtained with the ProtoG was less than 1% in almost half of these instances, having reached the greatest average for one instance of 1,173 cities, with an PES average equal to 3.52%. Therefore, the ProtoG can be considered a competitive algorithm for solving the TSP, since it is not rare in the literature find PESs averages greater than 10% to be reported for instances of this size.
Resumo:
The exponential growth in the applications of radio frequency (RF) is accompanied by great challenges as more efficient use of spectrum as in the design of new architectures for multi-standard receivers or software defined radio (SDR) . The key challenge in designing architecture of the software defined radio is the implementation of a wide-band receiver, reconfigurable, low cost, low power consumption, higher level of integration and flexibility. As a new solution of SDR design, a direct demodulator architecture, based on fiveport technology, or multi-port demodulator, has been proposed. However, the use of the five-port as a direct-conversion receiver requires an I/Q calibration (or regeneration) procedure in order to generate the in-phase (I) and quadrature (Q) components of the transmitted baseband signal. In this work, we propose to evaluate the performance of a blind calibration technique without additional knowledge about training or pilot sequences of the transmitted signal based on independent component analysis for the regeneration of I/Q five-port downconversion, by exploiting the information on the statistical properties of the three output signals
Resumo:
Nowadays, where the market competition requires products with better quality and a constant search for cost savings and a better use of raw materials, the research for more efficient control strategies becomes vital. In Natural Gas Processin Units (NGPUs), as in the most chemical processes, the quality control is accomplished through their products composition. However, the chemical composition analysis has a long measurement time, even when performed by instruments such as gas chromatographs. This fact hinders the development of control strategies to provide a better process yield. The natural gas processing is one of the most important activities in the petroleum industry. The main economic product of a NGPU is the liquefied petroleum gas (LPG). The LPG is ideally composed by propane and butane, however, in practice, its composition has some contaminants, such as ethane and pentane. In this work is proposed an inferential system using neural networks to estimate the ethane and pentane mole fractions in LPG and the propane mole fraction in residual gas. The goal is to provide the values of these estimated variables in every minute using a single multilayer neural network, making it possibly to apply inferential control techniques in order to monitor the LPG quality and to reduce the propane loss in the process. To develop this work a NGPU was simulated in HYSYS R software, composed by two distillation collumns: deethanizer and debutanizer. The inference is performed through the process variables of the PID controllers present in the instrumentation of these columns. To reduce the complexity of the inferential neural network is used the statistical technique of principal component analysis to decrease the number of network inputs, thus forming a hybrid inferential system. It is also proposed in this work a simple strategy to correct the inferential system in real-time, based on measurements of the chromatographs which may exist in process under study
Resumo:
Blind Source Separation (BSS) refers to the problem of estimate original signals from observed linear mixtures with no knowledge about the sources or the mixing process. Independent Component Analysis (ICA) is a technique mainly applied to BSS problem and from the algorithms that implement this technique, FastICA is a high performance iterative algorithm of low computacional cost that uses nongaussianity measures based on high order statistics to estimate the original sources. The great number of applications where ICA has been found useful reects the need of the implementation of this technique in hardware and the natural paralelism of FastICA favors the implementation of this algorithm on digital hardware. This work proposes the implementation of FastICA on a reconfigurable hardware platform for the viability of it's use in blind source separation problems, more specifically in a hardware prototype embedded in a Field Programmable Gate Array (FPGA) board for the monitoring of beds in hospital environments. The implementations will be carried out by Simulink models and it's synthesizing will be done through the DSP Builder software from Altera Corporation.
Resumo:
Conventional methods to solve the problem of blind source separation nonlinear, in general, using series of restrictions to obtain the solution, often leading to an imperfect separation of the original sources and high computational cost. In this paper, we propose an alternative measure of independence based on information theory and uses the tools of artificial intelligence to solve problems of blind source separation linear and nonlinear later. In the linear model applies genetic algorithms and Rényi of negentropy as a measure of independence to find a separation matrix from linear mixtures of signals using linear form of waves, audio and images. A comparison with two types of algorithms for Independent Component Analysis widespread in the literature. Subsequently, we use the same measure of independence, as the cost function in the genetic algorithm to recover source signals were mixed by nonlinear functions from an artificial neural network of radial base type. Genetic algorithms are powerful tools for global search, and therefore well suited for use in problems of blind source separation. Tests and analysis are through computer simulations
Resumo:
Image compress consists in represent by small amount of data, without loss a visual quality. Data compression is important when large images are used, for example satellite image. Full color digital images typically use 24 bits to specify the color of each pixel of the Images with 8 bits for each of the primary components, red, green and blue (RGB). Compress an image with three or more bands (multispectral) is fundamental to reduce the transmission time, process time and record time. Because many applications need images, that compression image data is important: medical image, satellite image, sensor etc. In this work a new compression color images method is proposed. This method is based in measure of information of each band. This technique is called by Self-Adaptive Compression (S.A.C.) and each band of image is compressed with a different threshold, for preserve information with better result. SAC do a large compression in large redundancy bands, that is, lower information and soft compression to bands with bigger amount of information. Two image transforms are used in this technique: Discrete Cosine Transform (DCT) and Principal Component Analysis (PCA). Primary step is convert data to new bands without relationship, with PCA. Later Apply DCT in each band. Data Loss is doing when a threshold discarding any coefficients. This threshold is calculated with two elements: PCA result and a parameter user. Parameters user define a compression tax. The system produce three different thresholds, one to each band of image, that is proportional of amount information. For image reconstruction is realized DCT and PCA inverse. SAC was compared with JPEG (Joint Photographic Experts Group) standard and YIQ compression and better results are obtain, in MSE (Mean Square Root). Tests shown that SAC has better quality in hard compressions. With two advantages: (a) like is adaptive is sensible to image type, that is, presents good results to divers images kinds (synthetic, landscapes, people etc., and, (b) it need only one parameters user, that is, just letter human intervention is required