961 resultados para Local finance -- Accounting -- Data processing


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Multi-centre data repositories like the Alzheimer's Disease Neuroimaging Initiative (ADNI) offer a unique research platform, but pose questions concerning comparability of results when using a range of imaging protocols and data processing algorithms. The variability is mainly due to the non-quantitative character of the widely used structural T1-weighted magnetic resonance (MR) images. Although the stability of the main effect of Alzheimer's disease (AD) on brain structure across platforms and field strength has been addressed in previous studies using multi-site MR images, there are only sparse empirically-based recommendations for processing and analysis of pooled multi-centre structural MR data acquired at different magnetic field strengths (MFS). Aiming to minimise potential systematic bias when using ADNI data we investigate the specific contributions of spatial registration strategies and the impact of MFS on voxel-based morphometry in AD. We perform a whole-brain analysis within the framework of Statistical Parametric Mapping, testing for main effects of various diffeomorphic spatial registration strategies, of MFS and their interaction with disease status. Beyond the confirmation of medial temporal lobe volume loss in AD, we detect a significant impact of spatial registration strategy on estimation of AD related atrophy. Additionally, we report a significant effect of MFS on the assessment of brain anatomy (i) in the cerebellum, (ii) the precentral gyrus and (iii) the thalamus bilaterally, showing no interaction with the disease status. We provide empirical evidence in support of pooling data in multi-centre VBM studies irrespective of disease status or MFS.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This report is divided into two volumes. This volume (Volume I) summarizes a structural health monitoring (SHM) system that was developed for the Iowa DOT to remotely and continuously monitor fatigue critical bridges (FCB) to aid in the detection of crack formation. The developed FCB SHM system enables bridge owners to remotely monitor FCB for gradual or sudden damage formation. The SHM system utilizes fiber bragg grating (FBG) fiber optic sensors (FOSs) to measure strains at critical locations. The strain-based SHM system is trained with measured performance data to identify typical bridge response when subjected to ambient traffic loads, and that knowledge is used to evaluate newly collected data. At specified intervals, the SHM system autonomously generates evaluation reports that summarize the current behavior of the bridge. The evaluation reports are collected and distributed to the bridge owner for interpretation and decision making. Volume II summarizes the development and demonstration of an autonomous, continuous SHM system that can be used to monitor typical girder bridges. The developed SHM system can be grouped into two main categories: an office component and a field component. The office component is a structural analysis software program that can be used to generate thresholds which are used for identifying isolated events. The field component includes hardware and field monitoring software which performs data processing and evaluation. The hardware system consists of sensors, data acquisition equipment, and a communication system backbone. The field monitoring software has been developed such that, once started, it will operate autonomously with minimal user interaction. In general, the SHM system features two key uses. First, the system can be integrated into an active bridge management system that tracks usage and structural changes. Second, the system helps owners to identify damage and deterioration.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This work proposes a parallel architecture for a motion estimation algorithm. It is well known that image processing requires a huge amount of computation, mainly at low level processing where the algorithms are dealing with a great numbers of data-pixel. One of the solutions to estimate motions involves detection of the correspondences between two images. Due to its regular processing scheme, parallel implementation of correspondence problem can be an adequate approach to reduce the computation time. This work introduces parallel and real-time implementation of such low-level tasks to be carried out from the moment that the current image is acquired by the camera until the pairs of point-matchings are detected

Relevância:

100.00% 100.00%

Publicador:

Resumo:

L’objecte del present treball és la realització d’una aplicació que permeti portar a terme el control estadístic multivariable en línia d’una planta SBR.Aquesta eina ha de permetre realitzar un anàlisi estadístic multivariable complet del lot en procés, de l’últim lot finalitzat i de la resta de lots processats a la planta.L’aplicació s’ha de realitzar en l’entorn LabVIEW. L’elecció d’aquest programa vecondicionada per l’actualització del mòdul de monitorització de la planta que s’estàdesenvolupant en aquest mateix entorn

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Treball fi de carrera sobre les mètriques de productivitat de programari per a la gestió de projectes.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper addresses the application of a PCA analysis on categorical data prior to diagnose a patients data set using a Case-Based Reasoning (CBR) system. The particularity is that the standard PCA techniques are designed to deal with numerical attributes, but our medical data set contains many categorical data and alternative methods as RS-PCA are required. Thus, we propose to hybridize RS-PCA (Regular Simplex PCA) and a simple CBR. Results show how the hybrid system produces similar results when diagnosing a medical data set, that the ones obtained when using the original attributes. These results are quite promising since they allow to diagnose with less computation effort and memory storage

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Land plants have had the reputation of being problematic for DNA barcoding for two general reasons: (i) the standard DNA regions used in algae, animals and fungi have exceedingly low levels of variability and (ii) the typically used land plant plastid phylogenetic markers (e.g. rbcL, trnL-F, etc.) appear to have too little variation. However, no one has assessed how well current phylogenetic resources might work in the context of identification (versus phylogeny reconstruction). In this paper, we make such an assessment, particularly with two of the markers commonly sequenced in land plant phylogenetic studies, plastid rbcL and internal transcribed spacers of the large subunits of nuclear ribosomal DNA (ITS), and find that both of these DNA regions perform well even though the data currently available in GenBank/EBI were not produced to be used as barcodes and BLAST searches are not an ideal tool for this purpose. These results bode well for the use of even more variable regions of plastid DNA (such as, for example, psbA-trnH) as barcodes, once they have been widely sequenced. In the short term, efforts to bring land plant barcoding up to the standards being used now in other organisms should make swift progress. There are two categories of DNA barcode users, scientists in fields other than taxonomy and taxonomists. For the former, the use of mitochondrial and plastid DNA, the two most easily assessed genomes, is at least in the short term a useful tool that permits them to get on with their studies, which depend on knowing roughly which species or species groups they are dealing with, but these same DNA regions have important drawbacks for use in taxonomic studies (i.e. studies designed to elucidate species limits). For these purposes, DNA markers from uniparentally (usually maternally) inherited genomes can only provide half of the story required to improve taxonomic standards being used in DNA barcoding. In the long term, we will need to develop more sophisticated barcoding tools, which would be multiple, low-copy nuclear markers with sufficient genetic variability and PCR-reliability; these would permit the detection of hybrids and permit researchers to identify the 'genetic gaps' that are useful in assessing species limits.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

BACKGROUND: PCR has the potential to detect and precisely quantify specific DNA sequences, but it is not yet often used as a fully quantitative method. A number of data collection and processing strategies have been described for the implementation of quantitative PCR. However, they can be experimentally cumbersome, their relative performances have not been evaluated systematically, and they often remain poorly validated statistically and/or experimentally. In this study, we evaluated the performance of known methods, and compared them with newly developed data processing strategies in terms of resolution, precision and robustness. RESULTS: Our results indicate that simple methods that do not rely on the estimation of the efficiency of the PCR amplification may provide reproducible and sensitive data, but that they do not quantify DNA with precision. Other evaluated methods based on sigmoidal or exponential curve fitting were generally of both poor resolution and precision. A statistical analysis of the parameters that influence efficiency indicated that it depends mostly on the selected amplicon and to a lesser extent on the particular biological sample analyzed. Thus, we devised various strategies based on individual or averaged efficiency values, which were used to assess the regulated expression of several genes in response to a growth factor. CONCLUSION: Overall, qPCR data analysis methods differ significantly in their performance, and this analysis identifies methods that provide DNA quantification estimates of high precision, robustness and reliability. These methods allow reliable estimations of relative expression ratio of two-fold or higher, and our analysis provides an estimation of the number of biological samples that have to be analyzed to achieve a given precision.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Motivation: The comparative analysis of gene gain and loss rates is critical for understanding the role of natural selection and adaptation in shaping gene family sizes. Studying complete genome data from closely related species allows accurate estimation of gene family turnover rates. Current methods and software tools, however, are not well designed for dealing with certain kinds of functional elements, such as microRNAs or transcription factor binding sites. Results: Here, we describe BadiRate, a new software tool to estimate family turnover rates, as well as the number of elements in internal phylogenetic nodes, by likelihood-based methods and parsimony. It implements two stochastic population models, which provide the appropriate statistical framework for testing hypothesis, such as lineage-specific gene family expansions or contractions. We have assessed the accuracy of BadiRate by computer simulations, and have also illustrated its functionality by analyzing a representative empirical dataset.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Motivation: The comparative analysis of gene gain and loss rates is critical for understanding the role of natural selection and adaptation in shaping gene family sizes. Studying complete genome data from closely related species allows accurate estimation of gene family turnover rates. Current methods and software tools, however, are not well designed for dealing with certain kinds of functional elements, such as microRNAs or transcription factor binding sites. Results: Here, we describe BadiRate, a new software tool to estimate family turnover rates, as well as the number of elements in internal phylogenetic nodes, by likelihood-based methods and parsimony. It implements two stochastic population models, which provide the appropriate statistical framework for testing hypothesis, such as lineage-specific gene family expansions or contractions. We have assessed the accuracy of BadiRate by computer simulations, and have also illustrated its functionality by analyzing a representative empirical dataset.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

INTRODUCTION: Dietary supplement (DS) use increased rapidly in recent years. However, evidence of benefits of many DSs for healthy users is scarce and may not equate with known risks of overdose, drug interaction and recently discovered negative long-term effects. This exploratory study aimed to investigate the perceptions and motivations of DS users in Lausanne, Switzerland. METHOD: A convenience sample (n = 147) was recruited at the entrances of local sales points. Data were collected in on-site semistructured interviews that assessed dietary supplementation habits. RESULTS: The majority of DSs were all-in-one products, containing a mixture of minerals and vitamins, or products containing only minerals. Among the 147 users, 72 (49%) used one all-in-one product and 3 (2%) used two all-in-one products. Thirty-one (21%) consumers did not know for at least one product what the purpose of their DS use was. Seventy-five percent of participants thought that DS use presents no risk or nearly no risk. Only 49% of participants stated that their physicians were informed about their consumption. Although men searched more often for potential risks (p <0.001), they turned less frequently to health professionals to get this information (p = 0.007). DISCUSSION: As in other surveys performed elsewhere, our study shows that, in Lausanne (Switzerland), DSs are commonly used as mixed products. Risk perception seems generally low among DS users. Physicians should be trained to evaluate patients' perceived needs and DS consumption in order to provide good evidence-based information or to propose alternatives to DS use.

Relevância:

100.00% 100.00%

Publicador:

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Pärjätäkseen nykypäivän kiristyvässä kilpailussa yritysten on kuuluttava kilpailukykyiseen toimitusketjuun ja toimittava yhteistyössä muiden yritysten kanssa. Yksi tunnetuimmista yhteistyömenetelmistä on täydennysyhteistyö, jossa toimittaja vastaa toimitusten suunnittelusta ilman perinteistä tilaus-toimitus-prosessia. Täydennysyhteistyöstä käytetään eri nimityksiä toimialasta ja toimitusketjun portaasta riippuen, mutta yleisellä tasolla on kyse samoista periaatteista: toimittajalla on vastuu varaston täydentämisestä ja sillä on käytössään täydennyspäätöksiin tarvittavaa tietoa. Työssä tarkastellaan tapausta, jossa toimittaja on ottanut käyttöön uuden täydennysyhteistyöhön liittyvän toimintatavan ja täydennystoimitusten laskentaan käytettävän työkalun. Tavoitteena on selvittää uuden täydennysyhteistyömallin käyttöönoton vaikutukset toimittajan näkökulmasta. Kirjallisuuskatsauksen avulla kartoitetaan täydennystoimitusten suunnitteluun vaikuttavat tekijät. Työssä kuvataan vanha ja uusi täydennysyhteistyömalli ja niitä vertaillaan keskenään. Uuden tyädennysyhteistyömallin vaikutuksia tutkitaan kolmella tasolla: yhteistyön mahdollistajat, suunnittelu ja prosessit sekä suorituskyky. Tutkimuksen perusteella uuden täydennysyhteistyömallin etuja ovat muun muassa tiedon käsittelyn nopeus ja helppous. Puutteeksi koetaan toimitussuunnitelman muuttamisen hankaluus työkalussa. Vaikka lähes kaikkien tutkimusotoksen tuotteiden keskimääräiset varastotasot olivat alentuneet uuden täydennysyhteistyömallin käyttöönoton jälkeen, varastotavoitteiden saavuttaminen ei ollut parantunut. Tuloksiin vaikuttavat useattekijät, joista tärkeimpiä ovat kysynnän vaihtelu ja volyymi sekä varastotavoite. Lisäksi ennustetarkkuus vaikuttaa toimitusten suunnitteluun ja siten varastonsuorituskykyyn.