934 resultados para Methods: data analysis


Relevância:

90.00% 90.00%

Publicador:

Resumo:

Clustering techniques which can handle incomplete data have become increasingly important due to varied applications in marketing research, medical diagnosis and survey data analysis. Existing techniques cope up with missing values either by using data modification/imputation or by partial distance computation, often unreliable depending on the number of features available. In this paper, we propose a novel approach for clustering data with missing values, which performs the task by Symmetric Non-Negative Matrix Factorization (SNMF) of a complete pair-wise similarity matrix, computed from the given incomplete data. To accomplish this, we define a novel similarity measure based on Average Overlap similarity metric which can effectively handle missing values without modification of data. Further, the similarity measure is more reliable than partial distances and inherently possesses the properties required to perform SNMF. The experimental evaluation on real world datasets demonstrates that the proposed approach is efficient, scalable and shows significantly better performance compared to the existing techniques.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

A split-phase induction motor is fed from two three-phase voltage source inverters for speed control. This study analyses carrier-comparison based pulse width modulation (PWM) schemes for a split-phase motor drive, from a space-vector perspective. Sine-triangle PWM, one zero-sequence injection PWM where the same zero-sequence signal is used for both the inverters, and another zero-sequence injection PWM where different zero-sequence signals are employed for the two inverters are considered. The set of voltage vectors applied, the sequence in which the voltage vectors are applied, and the resulting current ripple vector are analysed for all the PWM methods. Besides all the PWM methods are compared in terms of dc bus utilisation. For the same three-phase sine reference, the PWM method with different zero-sequence signals for the two inverters is found to employ a set of vectors different from the other methods. Both analysis and experimental results show that this method results in lower total harmonic distortion and higher dc bus utilisation than the other two PWM methods.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Based on the theory of the pumping well test, the transient injection well test was suggested in this paper. The design method and the scope of application are discussed in detail. The mathematical models are developed for the short-time and long-time transient injection test respectively. A double logarithm type curve matching method was introduced for analyzing the field transient injection test data. A set of methods for the transient injection test design, experiment performance and data analysis were established. Some field tests were analyzed, and the results show that the test model and method are suitable for the transient injection test and can be used to deal with the real engineering problems.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Modern technology has allowed real-time data collection in a variety of domains, ranging from environmental monitoring to healthcare. Consequently, there is a growing need for algorithms capable of performing inferential tasks in an online manner, continuously revising their estimates to reflect the current status of the underlying process. In particular, we are interested in constructing online and temporally adaptive classifiers capable of handling the possibly drifting decision boundaries arising in streaming environments. We first make a quadratic approximation to the log-likelihood that yields a recursive algorithm for fitting logistic regression online. We then suggest a novel way of equipping this framework with self-tuning forgetting factors. The resulting scheme is capable of tracking changes in the underlying probability distribution, adapting the decision boundary appropriately and hence maintaining high classification accuracy in dynamic or unstable environments. We demonstrate the scheme's effectiveness in both real and simulated streaming environments. © Springer-Verlag 2009.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

In this paper methods are developed for enhancement and analysis of autoregressive moving average (ARMA) signals observed in additive noise which can be represented as mixtures of heavy-tailed non-Gaussian sources and a Gaussian background component. Such models find application in systems such as atmospheric communications channels or early sound recordings which are prone to intermittent impulse noise. Markov Chain Monte Carlo (MCMC) simulation techniques are applied to the joint problem of signal extraction, model parameter estimation and detection of impulses within a fully Bayesian framework. The algorithms require only simple linear iterations for all of the unknowns, including the MA parameters, which is in contrast with existing MCMC methods for analysis of noise-free ARMA models. The methods are illustrated using synthetic data and noise-degraded sound recordings.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Background: Malignancies arising in the large bowel cause the second largest number of deaths from cancer in the Western World. Despite progresses made during the last decades, colorectal cancer remains one of the most frequent and deadly neoplasias in the western countries. Methods: A genomic study of human colorectal cancer has been carried out on a total of 31 tumoral samples, corresponding to different stages of the disease, and 33 non-tumoral samples. The study was carried out by hybridisation of the tumour samples against a reference pool of non-tumoral samples using Agilent Human 1A 60- mer oligo microarrays. The results obtained were validated by qRT-PCR. In the subsequent bioinformatics analysis, gene networks by means of Bayesian classifiers, variable selection and bootstrap resampling were built. The consensus among all the induced models produced a hierarchy of dependences and, thus, of variables. Results: After an exhaustive process of pre-processing to ensure data quality–lost values imputation, probes quality, data smoothing and intraclass variability filtering–the final dataset comprised a total of 8, 104 probes. Next, a supervised classification approach and data analysis was carried out to obtain the most relevant genes. Two of them are directly involved in cancer progression and in particular in colorectal cancer. Finally, a supervised classifier was induced to classify new unseen samples. Conclusions: We have developed a tentative model for the diagnosis of colorectal cancer based on a biomarker panel. Our results indicate that the gene profile described herein can discriminate between non-cancerous and cancerous samples with 94.45% accuracy using different supervised classifiers (AUC values in the range of 0.997 and 0.955).

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Survival from out-of-hospital cardiac arrest depends largely on two factors: early cardiopulmonary resuscitation (CPR) and early defibrillation. CPR must be interrupted for a reliable automated rhythm analysis because chest compressions induce artifacts in the ECG. Unfortunately, interrupting CPR adversely affects survival. In the last twenty years, research has been focused on designing methods for analysis of ECG during chest compressions. Most approaches are based either on adaptive filters to remove the CPR artifact or on robust algorithms which directly diagnose the corrupted ECG. In general, all the methods report low specificity values when tested on short ECG segments, but how to evaluate the real impact on CPR delivery of continuous rhythm analysis during CPR is still unknown. Recently, researchers have proposed a new methodology to measure this impact. Moreover, new strategies for fast rhythm analysis during ventilation pauses or high-specificity algorithms have been reported. Our objective is to present a thorough review of the field as the starting point for these late developments and to underline the open questions and future lines of research to be explored in the following years.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Recent observations of the temperature anisotropies of the cosmic microwave background (CMB) favor an inflationary paradigm in which the scale factor of the universe inflated by many orders of magnitude at some very early time. Such a scenario would produce the observed large-scale isotropy and homogeneity of the universe, as well as the scale-invariant perturbations responsible for the observed (10 parts per million) anisotropies in the CMB. An inflationary epoch is also theorized to produce a background of gravitational waves (or tensor perturbations), the effects of which can be observed in the polarization of the CMB. The E-mode (or parity even) polarization of the CMB, which is produced by scalar perturbations, has now been measured with high significance. Con- trastingly, today the B-mode (or parity odd) polarization, which is sourced by tensor perturbations, has yet to be observed. A detection of the B-mode polarization of the CMB would provide strong evidence for an inflationary epoch early in the universe’s history.

In this work, we explore experimental techniques and analysis methods used to probe the B- mode polarization of the CMB. These experimental techniques have been used to build the Bicep2 telescope, which was deployed to the South Pole in 2009. After three years of observations, Bicep2 has acquired one of the deepest observations of the degree-scale polarization of the CMB to date. Similarly, this work describes analysis methods developed for the Bicep1 three-year data analysis, which includes the full data set acquired by Bicep1. This analysis has produced the tightest constraint on the B-mode polarization of the CMB to date, corresponding to a tensor-to-scalar ratio estimate of r = 0.04±0.32, or a Bayesian 95% credible interval of r < 0.70. These analysis methods, in addition to producing this new constraint, are directly applicable to future analyses of Bicep2 data. Taken together, the experimental techniques and analysis methods described herein promise to open a new observational window into the inflationary epoch and the initial conditions of our universe.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Seismic reflection methods have been extensively used to probe the Earth's crust and suggest the nature of its formative processes. The analysis of multi-offset seismic reflection data extends the technique from a reconnaissance method to a powerful scientific tool that can be applied to test specific hypotheses. The treatment of reflections at multiple offsets becomes tractable if the assumptions of high-frequency rays are valid for the problem being considered. Their validity can be tested by applying the methods of analysis to full wave synthetics.

Three studies illustrate the application of these principles to investigations of the nature of the crust in southern California. A survey shot by the COCORP consortium in 1977 across the San Andreas fault near Parkfield revealed events in the record sections whose arrival time decreased with offset. The reflectors generating these events are imaged using a multi-offset three-dimensional Kirchhoff migration. Migrations of full wave acoustic synthetics having the same limitations in geometric coverage as the field survey demonstrate the utility of this back projection process for imaging. The migrated depth sections show the locations of the major physical boundaries of the San Andreas fault zone. The zone is bounded on the southwest by a near-vertical fault juxtaposing a Tertiary sedimentary section against uplifted crystalline rocks of the fault zone block. On the northeast, the fault zone is bounded by a fault dipping into the San Andreas, which includes slices of serpentinized ultramafics, intersecting it at 3 km depth. These interpretations can be made despite complications introduced by lateral heterogeneities.

In 1985 the Calcrust consortium designed a survey in the eastern Mojave desert to image structures in both the shallow and the deep crust. Preliminary field experiments showed that the major geophysical acquisition problem to be solved was the poor penetration of seismic energy through a low-velocity surface layer. Its effects could be mitigated through special acquisition and processing techniques. Data obtained from industry showed that quality data could be obtained from areas having a deeper, older sedimentary cover, causing a re-definition of the geologic objectives. Long offset stationary arrays were designed to provide reversed, wider angle coverage of the deep crust over parts of the survey. The preliminary field tests and constant monitoring of data quality and parameter adjustment allowed 108 km of excellent crustal data to be obtained.

This dataset, along with two others from the central and western Mojave, was used to constrain rock properties and the physical condition of the crust. The multi-offset analysis proceeded in two steps. First, an increase in reflection peak frequency with offset is indicative of a thinly layered reflector. The thickness and velocity contrast of the layering can be calculated from the spectral dispersion, to discriminate between structures resulting from broad scale or local effects. Second, the amplitude effects at different offsets of P-P scattering from weak elastic heterogeneities indicate whether the signs of the changes in density, rigidity, and Lame's parameter at the reflector agree or are opposed. The effects of reflection generation and propagation in a heterogeneous, anisotropic crust were contained by the design of the experiment and the simplicity of the observed amplitude and frequency trends. Multi-offset spectra and amplitude trend stacks of the three Mojave Desert datasets suggest that the most reflective structures in the middle crust are strong Poisson's ratio (σ) contrasts. Porous zones or the juxtaposition of units of mutually distant origin are indicated. Heterogeneities in σ increase towards the top of a basal crustal zone at ~22 km depth. The transition to the basal zone and to the mantle include increases in σ. The Moho itself includes ~400 m layering having a velocity higher than that of the uppermost mantle. The Moho maintains the same configuration across the Mojave despite 5 km of crustal thinning near the Colorado River. This indicates that Miocene extension there either thinned just the basal zone, or that the basal zone developed regionally after the extensional event.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The cytochromes P450 (P450s) are a remarkable class of heme enzymes that catalyze the metabolism of xenobiotics and the biosynthesis of signaling molecules. Controlled electron flow into the thiolate-ligated heme active site allows P450s to activate molecular oxygen and hydroxylate aliphatic C–H bonds via the formation of high-valent metal-oxo intermediates (compounds I and II). Due to the reactive nature and short lifetimes of these intermediates, many of the fundamental steps in catalysis have not been observed directly. The Gray group and others have developed photochemical methods, known as “flash-quench,” for triggering electron transfer (ET) and generating redox intermediates in proteins in the absence of native ET partners. Photo-triggering affords a high degree of temporal precision for the gating of an ET event; the initial ET and subsequent reactions can be monitored on the nanosecond-to-second timescale using transient absorption (TA) spectroscopies. Chapter 1 catalogues critical aspects of P450 structure and mechanism, including the native pathway for formation of compound I, and outlines the development of photochemical processes that can be used to artificially trigger ET in proteins. Chapters 2 and 3 describe the development of these photochemical methods to establish electronic communication between a photosensitizer and the buried P450 heme. Chapter 2 describes the design and characterization of a Ru-P450-BM3 conjugate containing a ruthenium photosensitizer covalently tethered to the P450 surface, and nanosecond-to-second kinetics of the photo-triggered ET event are presented. By analyzing data at multiple wavelengths, we have identified the formation of multiple ET intermediates, including the catalytically relevant compound II; this intermediate is generated by oxidation of a bound water molecule in the ferric resting state enzyme. The work in Chapter 3 probes the role of a tryptophan residue situated between the photosensitizer and heme in the aforementioned Ru-P450 BM3 conjugate. Replacement of this tryptophan with histidine does not perturb the P450 structure, yet it completely eliminates the ET reactivity described in Chapter 2. The presence of an analogous tryptophan in Ru-P450 CYP119 conjugates also is necessary for observing oxidative ET, but the yield of heme oxidation is lower. Chapter 4 offers a basic description of the theoretical underpinnings required to analyze ET. Single-step ET theory is first presented, followed by extensions to multistep ET: electron “hopping.” The generation of “hopping maps” and use of a hopping map program to analyze the rate advantage of hopping over single-step ET is described, beginning with an established rhenium-tryptophan-azurin hopping system. This ET analysis is then applied to the Ru-tryptophan-P450 systems described in Chapter 2; this strongly supports the presence of hopping in Ru-P450 conjugates. Chapter 5 explores the implementation of flash-quench and other phototriggered methods to examine the native reductive ET and gas binding events that activate molecular oxygen. In particular, TA kinetics that demonstrate heme reduction on the microsecond timescale for four Ru-P450 conjugates are presented. In addition, we implement laser flash-photolysis of P450 ferrous–CO to study the rates of CO rebinding in the thermophilic P450 CYP119 at variable temperature. Chapter 6 describes the development and implementation of air-sensitive potentiometric redox titrations to determine the solution reduction potentials of a series of P450 BM3 mutants, which were designed for non-native cyclopropanation of styrene in vivo. An important conclusion from this work is that substitution of the axial cysteine for serine shifts the wild type reduction potential positive by 130 mV, facilitating reduction by biological redox cofactors in the presence of poorly-bound substrates. While this mutation abolishes oxygenation activity, these mutants are capable of catalyzing the cyclopropanation of styrene, even within the confines of an E. coli cell. Four appendices are also provided, including photochemical heme oxidation in ruthenium-modified nitric oxide synthase (Appendix A), general protocols (Appendix B), Chapter-specific notes (Appendix C) and Matlab scripts used for data analysis (Appendix D).

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Planetary atmospheres exist in a seemingly endless variety of physical and chemical environments. There are an equally diverse number of methods by which we can study and characterize atmospheric composition. In order to better understand the fundamental chemistry and physical processes underlying all planetary atmospheres, my research of the past four years has focused on two distinct topics. First, I focused on the data analysis and spectral retrieval of observations obtained by the Ultraviolet Imaging Spectrograph (UVIS) instrument onboard the Cassini spacecraft while in orbit around Saturn. These observations consisted of stellar occultation measurements of Titan's upper atmosphere, probing the chemical composition in the region 300 to 1500 km above Titan's surface. I examined the relative abundances of Titan's two most prevalent chemical species, nitrogen and methane. I also focused on the aerosols that are formed through chemistry involving these two major species, and determined the vertical profiles of aerosol particles as a function of time and latitude. Moving beyond our own solar system, my second topic of investigation involved analysis of infra-red light curves from the Spitzer space telescope, obtained as it measured the light from stars hosting planets of their own. I focused on both transit and eclipse modeling during Spitzer data reduction and analysis. In my initial work, I utilized the data to search for transits of planets a few Earth masses in size. In more recent research, I analyzed secondary eclipses of three exoplanets and constrained the range of possible temperatures and compositions of their atmospheres.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

O propósito desta Tese foi detectar e caracterizar áreas sob alto risco para leishmaniose visceral (LV) e descrever os padrões de ocorrência e difusão da doença, entre os anos de 1993 a 1996 e 2001 a 2006, em Teresina, Piauí, por meio de métodos estatísticos para análise de dados espaciais, sistemas de informações geográficas e imagens de sensoriamento remoto. Os resultados deste estudo são apresentados na forma de três manuscritos. O primeiro usou análise de dados espaciais para identificar as áreas com maior risco de LV na área urbana de Teresina entre 2001 e 2006. Os resultados utilizando razão de kernels demonstraram que as regiões periféricas da cidade foram mais fortemente afetadas ao longo do período analisado. A análise com indicadores locais de autocorrelação espacial mostrou que, no início do período de estudo, os agregados de alta incidência de LV localizavam-se principalmente na região sul e nordeste da cidade, mas nos anos seguintes os eles apareceram também na região norte da cidade, sugerindo que o padrão de ocorrência de LV não é estático e a doença pode se espalhar ocasionalmente para outras áreas do município. O segundo estudo teve como objetivo caracterizar e predizer territórios de alto risco para ocorrência da LV em Teresina, com base em indicadores socioeconômicos e dados ambientais, obtidos por sensoriamento remoto. Os resultados da classificação orientada a objeto apontam a expansão da área urbana para a periferia da cidade, onde antes havia maior cobertura de vegetação. O modelo desenvolvido foi capaz de discriminar 15 conjuntos de setores censitário (SC) com diferentes probabilidades de conterem SC com alto risco de ocorrência de LV. O subconjunto com maior probabilidade de conter SC com alto risco de LV (92%) englobou SC com percentual de chefes de família alfabetizados menor que a mediana (≤64,2%), com maior área coberta por vegetação densa, com percentual de até 3 moradores por domicílio acima do terceiro quartil (>31,6%). O modelo apresentou, respectivamente, na amostra de treinamento e validação, sensibilidade de 79% e 54%, especificidade de 74% e 71%, acurácia global de 75% e 67% e área sob a curva ROC de 83% e 66%. O terceiro manuscrito teve como objetivo avaliar a aplicabilidade da estratégia de classificação orientada a objeto na busca de possíveis indicadores de cobertura do solo relacionados com a ocorrência da LV em meio urbano. Os índices de acurácia foram altos em ambas as imagens (>90%). Na correlação da incidência da LV com os indicadores ambientais verificou-se correlações positivas com os indicadores Vegetação densa, Vegetação rasteira e Solo exposto e negativa com os indicadores Água, Urbana densa e Urbana verde, todos estatisticamente significantes. Os resultados desta tese revelam que a ocorrência da LV na periferia de Teresina está intensamente relacionada às condições socioeconômicas inadequadas e transformações ambientais decorrentes do processo de expansão urbana, favorecendo a ocorrência do vetor (Lutzomyia longipalpis) nestas regiões.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This work deals with two related areas: processing of visual information in the central nervous system, and the application of computer systems to research in neurophysiology.

Certain classes of interneurons in the brain and optic lobes of the blowfly Calliphora phaenicia were previously shown to be sensitive to the direction of motion of visual stimuli. These units were identified by visual field, preferred direction of motion, and anatomical location from which recorded. The present work is addressed to the questions: (1) is there interaction between pairs of these units, and (2) if such relationships can be found, what is their nature. To answer these questions, it is essential to record from two or more units simultaneously, and to use more than a single recording electrode if recording points are to be chosen independently. Accordingly, such techniques were developed and are described.

One must also have practical, convenient means for analyzing the large volumes of data so obtained. It is shown that use of an appropriately designed computer system is a profitable approach to this problem. Both hardware and software requirements for a suitable system are discussed and an approach to computer-aided data analysis developed. A description is given of members of a collection of application programs developed for analysis of neuro-physiological data and operated in the environment of and with support from an appropriate computer system. In particular, techniques developed for classification of multiple units recorded on the same electrode are illustrated as are methods for convenient graphical manipulation of data via a computer-driven display.

By means of multiple electrode techniques and the computer-aided data acquisition and analysis system, the path followed by one of the motion detection units was traced from open optic lobe through the brain and into the opposite lobe. It is further shown that this unit and its mirror image in the opposite lobe have a mutually inhibitory relationship. This relationship is investigated. The existence of interaction between other pairs of units is also shown. For pairs of units responding to motion in the same direction, the relationship is of an excitatory nature; for those responding to motion in opposed directions, it is inhibitory.

Experience gained from use of the computer system is discussed and a critical review of the current system is given. The most useful features of the system were found to be the fast response, the ability to go from one analysis technique to another rapidly and conveniently, and the interactive nature of the display system. The shortcomings of the system were problems in real-time use and the programming barrier—the fact that building new analysis techniques requires a high degree of programming knowledge and skill. It is concluded that computer system of the kind discussed will play an increasingly important role in studies of the central nervous system.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Amphibian declines and extinctions have been documented around the world, often in protected natural areas. Concern for this trend has prompted the U.S. Geological Survey and the National Park Service to document all species of amphibians that occur within U.S. National Parks and to search for any signs that amphibians may be declining. This study, an inventory of amphibian species in Big Cypress National Preserve, was conducted from 2002 to 2003. The goals of the project were to create a georeferenced inventory of amphibian species, use new analytical techniques to estimate proportion of sites occupied by each species, look for any signs of amphibian decline (missing species, disease, die-offs, and so forth.), and to establish a protocol that could be used for future monitoring efforts. Several sampling methods were used to accomplish these goals. Visual encounter surveys and anuran vocalization surveys were conducted in all habitats throughout the park to estimate the proportion of sites or proportion of area occupied (PAO) by each amphibian species in each habitat. Opportunistic collections, as well as limited drift fence data, were used to augment the visual encounter methods for highly aquatic or cryptic species. A total of 545 visits to 104 sites were conducted for standard sampling alone, and 2,358 individual amphibians and 374 reptiles were encountered. Data analysis was conducted in program PRESENCE to provide PAO estimates for each of the anuran species. All of the amphibian species historically found in Big Cypress National Preserve were detected during this project. At least one individual of each of the four salamander species was captured during sampling. Each of the anuran species in the preserve was adequately sampled using standard herpetological sampling methods, and PAO estimates were produced for each species of anuran by habitat. This information serves as an indicator of habitat associations of the species and relative abundance of sites occupied, but it will also be useful as a comparative baseline for future monitoring efforts. In addition to sampling for amphibians, all encounters with reptiles were documented. The sampling methods used for detecting amphibians are also appropriate for many reptile species. These reptile locations are included in this report, but the number of reptile observations was not sufficient to estimate PAO for reptile species. We encountered 35 of the 46 species of reptiles believed to be present in Big Cypress National Preserve during this study, and evidence exists of the presence of four other reptile species in the Preserve. This study found no evidence of amphibian decline in Big Cypress National Preserve. Although no evidence of decline was observed, several threats to amphibians were identified. Introduced species, especially the Cuban treefrog (Osteopilus septentrionalis), are predators and competitors with several native frog species. The recreational use of off-road vehicles has the potential to affect some amphibian populations, and a study on those potential impacts is currently underway. Also, interference by humans with the natural hydrologic cycle of south Florida has the potential to alter the amphibian community. Continued monitoring of the amphibian species in Big Cypress National Preserve is recommended. The methods used in this study were adequate to produce reliable estimates of the proportion of sites occupied by most anuran species, and are a cost-effective means of determining the status of their populations.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Phosalone is a non systematic, wide spectrum organophosphate pesticide which was discovered in 1961 in the laboratories of the Societe des Usines Chimique Rhone-Poulenc in France. It has been approved for commercial use since 1964 in France, in Australia since 1966, in the United Kingdom in 1967 and in many other countries including Japan, Egypt, USSR and the USA. This study provides a full literature review on all aspects of phosalone including its physical, biological and chemical characteristics, and analytical methods of analysis with particular reference to soils/sediments. Furthermore, it aims to develop a method for the determintion of phosalone in aquatic sediments and to determine the adsorption of phosalone onto kaolinite.