939 resultados para Statistical Language Model


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Evolutionary algorithms are suitable to solve damage identification problems in a multi-objective context. However, the performance of these methods can deteriorate quickly with increasing noise intensities originating numerous uncertainties. In this paper, a statistic structural damage detection method formulated in a multi-objective context is proposed. The statistic analysis is implemented to take into account the uncertainties existing in the structural model and measured structural modal parameters. The presented method is verified by a number of simulated damage scenarios. The effects of noise and damage levels on damage detection are investigated.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The present paper describes the advancement and evaluation of air quality-related impacts with the Atmospheric Evaluation and Research Integrated system for Spain (AERIS). In its current version, AERIS is able to provide estimates on the impacts of air quality over human health (PM2.5 and O3), crops and vegetation (O3). The modules that allow quantifying the before mentioned impacts were modeled by applying different approaches (mostly for the European context) present in scientific literature to the conditions of the Iberian Peninsula. This application was supported by reliable data sources, as well as by the good predictive capacity of AERIS for ambient concentrations. For validation purposes, the estimates of AERIS for impacts on human health (change in the statistical life expectancy-PM2.5) and vegetation (loss of wheat crops-O3) were compared against results from the SERCA project and GAINS estimates for two emission scenarios. In general, good results evidenced by reasonable correlation coefficients were obtained, therefore confirming the adequateness of the followed modeling approaches and the quality of AERIS predictions.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

En hidrodinámica, el fenómeno de Sloshing se puede definir como el movimiento de la superficie libre de un fluido dentro de un contenedor sometido a fuerzas y perturbaciones externas. El fluido en cuestión experimenta violentos movimientos con importantes deformaciones de su superficie libre. La dinámica del fluido puede llegar a generar cargas hidrodinámicas considerables las cuales pueden afectar la integridad estructural y/o comprometer la estabilidad del vehículo que transporta dicho contenedor. El fenómeno de Sloshing ha sido extensivamente investigado matemática, numérica y experimentalmente, siendo el enfoque experimental el más usado debido a la complejidad del problema, para el cual los modelos matemáticos y de simulación son aun incapaces de predecir con suficiente rapidez y precisión las cargas debidas a dicho fenómeno. El flujo generado por el Sloshing usualmente se caracteriza por la presencia de un fluido multifase (gas-liquido) y turbulencia. Reducir al máximo posible la complejidad del fenómeno de Sloshing sin perder la esencia del problema es el principal reto de esta tesis doctoral, donde un trabajo experimental enfocado en casos canónicos de Sloshing es presentado y documentado con el objetivo de aumentar la comprensión de dicho fenómeno y por tanto intentar proveer información valiosa para validaciones de códigos numéricos. El fenómeno de Sloshing juega un papel importante en la industria del transporte marítimo de gas licuado (LNG). El mercado de LNG en los últimos años ha reportado un crecimiento hasta tres veces mayor al de los mercados de petróleo y gas convencionales. Ingenieros en laboratorios de investigación e ingenieros adscritos a la industria del LNG trabajan continuamente buscando soluciones económicas y seguras para contener, transferir y transportar grandes volúmenes de LNG. Los buques transportadores de LNG (LNGC) han pasado de ser unos pocos buques con capacidad de 75000 m3 hace unos treinta años, a una amplia flota con una capacidad de 140000 m3 actualmente. En creciente número, hoy día se construyen buques con capacidades que oscilan entre 175000 m3 y 250000 m3. Recientemente un nuevo concepto de buque LNG ha salido al mercado y se le conoce como FLNG. Un FLNG es un buque de gran valor añadido que solventa los problemas de extracción, licuefacción y almacenamiento del LNG, ya que cuenta con equipos de extracción y licuefacción a bordo, eliminando por tanto las tareas de transvase de las estaciones de licuefacción en tierra hacia los buques LNGC. EL LNG por tanto puede ser transferido directamente desde el FLNG hacia los buques LNGC en mar abierto. Niveles de llenado intermedios en combinación con oleaje durante las operaciones de trasvase inducen movimientos en los buques que generan por tanto el fenómeno de Sloshing dentro de los tanques de los FLNG y los LNGC. El trabajo de esta tesis doctoral lidia con algunos de los problemas del Sloshing desde un punto de vista experimental y estadístico, para ello una serie de tareas, descritas a continuación, se han llevado a cabo : 1. Un dispositivo experimental de Sloshing ha sido configurado. Dicho dispositivo ha permitido ensayar secciones rectangulares de tanques LNGC a escala con movimientos angulares de un grado de libertad. El dispositivo experimental ha sido instrumentado para realizar mediciones de movimiento, presiones, vibraciones y temperatura, así como la grabación de imágenes y videos. 2. Los impactos de olas generadas dentro de una sección rectangular de un LNGC sujeto a movimientos regulares forzados han sido estudiados mediante la caracterización del fenómeno desde un punto de vista estadístico enfocado en la repetitividad y la ergodicidad del problema. 3. El estudio de los impactos provocados por movimientos regulares ha sido extendido a un escenario más realístico mediante el uso de movimientos irregulares forzados. 4. El acoplamiento del Sloshing generado por el fluido en movimiento dentro del tanque LNGC y la disipación de la energía mecánica de un sistema no forzado de un grado de libertad (movimiento angular) sujeto a una excitación externa ha sido investigado. 5. En la última sección de esta tesis doctoral, la interacción entre el Sloshing generado dentro en una sección rectangular de un tanque LNGC sujeto a una excitación regular y un cuerpo elástico solidario al tanque ha sido estudiado. Dicho estudio corresponde a un problema de interacción fluido-estructura. Abstract In hydrodynamics, we refer to sloshing as the motion of liquids in containers subjected to external forces with large free-surface deformations. The liquid motion dynamics can generate loads which may affect the structural integrity of the container and the stability of the vehicle that carries such container. The prediction of these dynamic loads is a major challenge for engineers around the world working on the design of both the container and the vehicle. The sloshing phenomenon has been extensively investigated mathematically, numerically and experimentally. The latter has been the most fruitful so far, due to the complexity of the problem, for which the numerical and mathematical models are still incapable of accurately predicting the sloshing loads. The sloshing flows are usually characterised by the presence of multiphase interaction and turbulence. Reducing as much as possible the complexity of the sloshing problem without losing its essence is the main challenge of this phd thesis, where experimental work on selected canonical cases are presented and documented in order to better understand the phenomenon and to serve, in some cases, as an useful information for numerical validations. Liquid sloshing plays a key roll in the liquified natural gas (LNG) maritime transportation. The LNG market growth is more than three times the rated growth of the oil and traditional gas markets. Engineers working in research laboratories and companies are continuously looking for efficient and safe ways for containing, transferring and transporting the liquified gas. LNG carrying vessels (LNGC) have evolved from a few 75000 m3 vessels thirty years ago to a huge fleet of ships with a capacity of 140000 m3 nowadays and increasing number of 175000 m3 and 250000 m3 units. The concept of FLNG (Floating Liquified Natural Gas) has appeared recently. A FLNG unit is a high value-added vessel which can solve the problems of production, treatment, liquefaction and storage of the LNG because the vessel is equipped with a extraction and liquefaction facility. The LNG is transferred from the FLNG to the LNGC in open sea. The combination of partial fillings and wave induced motions may generate sloshing flows inside both the LNGC and the FLNG tanks. This work has dealt with sloshing problems from a experimental and statistical point of view. A series of tasks have been carried out: 1. A sloshing rig has been set up. It allows for testing tanks with one degree of freedom angular motion. The rig has been instrumented to measure motions, pressure and conduct video and image recording. 2. Regular motion impacts inside a rectangular section LNGC tank model have been studied, with forced motion tests, in order to characterise the phenomenon from a statistical point of view by assessing the repeatability and practical ergodicity of the problem. 3. The regular motion analysis has been extended to an irregular motion framework in order to reproduce more realistic scenarios. 4. The coupled motion of a single degree of freedom angular motion system excited by an external moment and affected by the fluid moment and the mechanical energy dissipation induced by sloshing inside the tank has been investigated. 5. The last task of the thesis has been to conduct an experimental investigation focused on the strong interaction between a sloshing flow in a rectangular section of a LNGC tank subjected to regular excitation and an elastic body clamped to the tank. It is thus a fluid structure interaction problem.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This PhD dissertation is framed in the emergent fields of Reverse Logistics and ClosedLoop Supply Chain (CLSC) management. This subarea of supply chain management has gained researchers and practitioners' attention over the last 15 years to become a fully recognized subdiscipline of the Operations Management field. More specifically, among all the activities that are included within the CLSC area, the focus of this dissertation is centered in direct reuse aspects. The main contribution of this dissertation to current knowledge is twofold. First, a framework for the so-called reuse CLSC is developed. This conceptual model is grounded in a set of six case studies conducted by the author in real industrial settings. The model has also been contrasted with existing literature and with academic and professional experts on the topic as well. The framework encompasses four building blocks. In the first block, a typology for reusable articles is put forward, distinguishing between Returnable Transport Items (RTI), Reusable Packaging Materials (RPM), and Reusable Products (RP). In the second block, the common characteristics that render reuse CLSC difficult to manage from a logistical standpoint are identified, namely: fleet shrinkage, significant investment and limited visibility. In the third block, the main problems arising in the management of reuse CLSC are analyzed, such as: (1) define fleet size dimension, (2) control cycle time and promote articles rotation, (3) control return rate and prevent shrinkage, (4) define purchase policies for new articles, (5) plan and control reconditioning activities, and (6) balance inventory between depots. Finally, in the fourth block some solutions to those issues are developed. Firstly, problems (2) and (3) are addressed through the comparative analysis of alternative strategies for controlling cycle time and return rate. Secondly, a methodology for calculating the required fleet size is elaborated (problem (1)). This methodology is valid for different configurations of the physical flows in the reuse CLSC. Likewise, some directions are pointed out for further development of a similar method for defining purchase policies for new articles (problem (4)). The second main contribution of this dissertation is embedded in the solutions part (block 4) of the conceptual framework and comprises a two-level decision problem integrating two mixed integer linear programming (MILP) models that have been formulated and solved to optimality using AIMMS as modeling language, CPLEX as solver and Excel spreadsheet for data introduction and output presentation. The results obtained are analyzed in order to measure in a client-supplier system the economic impact of two alternative control strategies (recovery policies) in the context of reuse. In addition, the models support decision-making regarding the selection of the appropriate recovery policy against the characteristics of demand pattern and the structure of the relevant costs in the system. The triangulation of methods used in this thesis has enabled to address the same research topic with different approaches and thus, the robustness of the results obtained is strengthened.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We define a capacity reserve model to dimension passenger car service installations according to the demographic distribution of the area to be serviced by using hospital?s emergency room analogies. Usually, service facilities are designed applying empirical methods, but customers arrive under uncertain conditions not included in the original estimations, and there is a gap between customer?s real demand and the service?s capacity. Our research establishes a valid methodology and covers the absence of recent researches and the lack of statistical techniques implementation, integrating demand uncertainty in a unique model built in stages by implementing ARIMA forecasting, queuing theory, and Monte Carlo simulation to optimize the service capacity and occupancy, minimizing the implicit cost of the capacity that must be reserved to service unexpected customers. Our model has proved to be a useful tool for optimal decision making under uncertainty integrating the prediction of the cost implicit in the reserve capacity to serve unexpected demand and defining a set of new process indicators, such us capacity, occupancy, and cost of capacity reserve never studied before. The new indicators are intended to optimize the service operation. This set of new indicators could be implemented in the information systems used in the passenger car services.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

El objetivo de esta investigación consiste en definir un modelo de reserva de capacidad, por analogías con emergencias hospitalarias, que pueda ser implementado en el sector de servicios. Este está específicamente enfocado a su aplicación en talleres de servicio de automóviles. Nuestra investigación incorpora la incertidumbre de la demanda en un modelo singular diseñado en etapas que agrupa técnicas ARIMA, teoría de colas y simulación Monte Carlo para definir los conceptos de capacidad y ocupación de servicio, que serán utilizados para minimizar el coste implícito de la reserva capacidad necesaria para atender a clientes que carecen de cita previa. Habitualmente, las compañías automovilísticas estiman la capacidad de sus instalaciones de servicio empíricamente, pero los clientes pueden llegar bajo condiciones de incertidumbre que no se tienen en cuenta en dichas estimaciones, por lo que existe una diferencia entre lo que el cliente realmente demanda y la capacidad que ofrece el servicio. Nuestro enfoque define una metodología válida para el sector automovilístico que cubre la ausencia genérica de investigaciones recientes y la habitual falta de aplicación de técnicas estadísticas en el sector. La equivalencia con la gestión de urgencias hospitalarias se ha validado a lo largo de la investigación en la se definen nuevos indicadores de proceso (KPIs) Tal y como hacen los hospitales, aplicamos modelos estocásticos para dimensionar las instalaciones de servicio de acuerdo con la distribución demográfica del área de influencia. El modelo final propuesto integra la predicción del coste implícito en la reserva de capacidad para atender la demanda no prevista. Asimismo, se ha desarrollado un código en Matlab que puede integrarse como un módulo adicional a los sistemas de información (DMS) que se usan actualmente en el sector, con el fin de emplear los nuevos indicadores de proceso definidos en el modelo. Los resultados principales del modelo son nuevos indicadores de servicio, tales como la capacidad, ocupación y coste de reserva de capacidad, que nunca antes han sido objeto de estudio en la industria automovilística, y que están orientados a gestionar la operativa del servicio. ABSTRACT Our aim is to define a Capacity Reserve model to be implemented in the service sector by hospital's emergency room (ER) analogies, with a practical approach to passenger car services. A stochastic model has been implemented using R and a Monte Carlo simulation code written in Matlab and has proved a very useful tool for optimal decision making under uncertainty. The research integrates demand uncertainty in a unique model which is built in stages by implementing ARIMA forecasting, Queuing Theory and a Monte Carlo simulation to define the concepts of service capacity and occupancy, minimizing the implicit cost of the capacity that must be reserved to service unexpected customers. Usually, passenger car companies estimate their service facilities capacity using empirical methods, but customers arrive under uncertain conditions not included in the estimations. Thus, there is a gap between customer’s real demand and the dealer’s capacity. This research sets a valid methodology for the passenger car industry to cover the generic absence of recent researches and the generic lack of statistical techniques implementation. The hospital’s emergency room (ER) equalization has been confirmed to be valid for the passenger car industry and new process indicators have been defined to support the study. As hospitals do, we aim to apply stochastic models to dimension installations according to the demographic distribution of the area to be serviced. The proposed model integrates the prediction of the cost implicit in the reserve capacity to serve unexpected demand. The Matlab code could be implemented as part of the existing information technology systems (ITs) to support the existing service management tools, creating a set of new process indicators. Main model outputs are new indicators, such us Capacity, Occupancy and Cost of Capacity Reserve, never studied in the passenger car service industry before, and intended to manage the service operation.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

El uso de aritmética de punto fijo es una opción de diseño muy extendida en sistemas con fuertes restricciones de área, consumo o rendimiento. Para producir implementaciones donde los costes se minimicen sin impactar negativamente en la precisión de los resultados debemos llevar a cabo una asignación cuidadosa de anchuras de palabra. Encontrar la combinación óptima de anchuras de palabra en coma fija para un sistema dado es un problema combinatorio NP-hard al que los diseñadores dedican entre el 25 y el 50 % del ciclo de diseño. Las plataformas hardware reconfigurables, como son las FPGAs, también se benefician de las ventajas que ofrece la aritmética de coma fija, ya que éstas compensan las frecuencias de reloj más bajas y el uso más ineficiente del hardware que hacen estas plataformas respecto a los ASICs. A medida que las FPGAs se popularizan para su uso en computación científica los diseños aumentan de tamaño y complejidad hasta llegar al punto en que no pueden ser manejados eficientemente por las técnicas actuales de modelado de señal y ruido de cuantificación y de optimización de anchura de palabra. En esta Tesis Doctoral exploramos distintos aspectos del problema de la cuantificación y presentamos nuevas metodologías para cada uno de ellos: Las técnicas basadas en extensiones de intervalos han permitido obtener modelos de propagación de señal y ruido de cuantificación muy precisos en sistemas con operaciones no lineales. Nosotros llevamos esta aproximación un paso más allá introduciendo elementos de Multi-Element Generalized Polynomial Chaos (ME-gPC) y combinándolos con una técnica moderna basada en Modified Affine Arithmetic (MAA) estadístico para así modelar sistemas que contienen estructuras de control de flujo. Nuestra metodología genera los distintos caminos de ejecución automáticamente, determina las regiones del dominio de entrada que ejercitarán cada uno de ellos y extrae los momentos estadísticos del sistema a partir de dichas soluciones parciales. Utilizamos esta técnica para estimar tanto el rango dinámico como el ruido de redondeo en sistemas con las ya mencionadas estructuras de control de flujo y mostramos la precisión de nuestra aproximación, que en determinados casos de uso con operadores no lineales llega a tener tan solo una desviación del 0.04% con respecto a los valores de referencia obtenidos mediante simulación. Un inconveniente conocido de las técnicas basadas en extensiones de intervalos es la explosión combinacional de términos a medida que el tamaño de los sistemas a estudiar crece, lo cual conlleva problemas de escalabilidad. Para afrontar este problema presen tamos una técnica de inyección de ruidos agrupados que hace grupos con las señales del sistema, introduce las fuentes de ruido para cada uno de los grupos por separado y finalmente combina los resultados de cada uno de ellos. De esta forma, el número de fuentes de ruido queda controlado en cada momento y, debido a ello, la explosión combinatoria se minimiza. También presentamos un algoritmo de particionado multi-vía destinado a minimizar la desviación de los resultados a causa de la pérdida de correlación entre términos de ruido con el objetivo de mantener los resultados tan precisos como sea posible. La presente Tesis Doctoral también aborda el desarrollo de metodologías de optimización de anchura de palabra basadas en simulaciones de Monte-Cario que se ejecuten en tiempos razonables. Para ello presentamos dos nuevas técnicas que exploran la reducción del tiempo de ejecución desde distintos ángulos: En primer lugar, el método interpolativo aplica un interpolador sencillo pero preciso para estimar la sensibilidad de cada señal, y que es usado después durante la etapa de optimización. En segundo lugar, el método incremental gira en torno al hecho de que, aunque es estrictamente necesario mantener un intervalo de confianza dado para los resultados finales de nuestra búsqueda, podemos emplear niveles de confianza más relajados, lo cual deriva en un menor número de pruebas por simulación, en las etapas iniciales de la búsqueda, cuando todavía estamos lejos de las soluciones optimizadas. Mediante estas dos aproximaciones demostramos que podemos acelerar el tiempo de ejecución de los algoritmos clásicos de búsqueda voraz en factores de hasta x240 para problemas de tamaño pequeño/mediano. Finalmente, este libro presenta HOPLITE, una infraestructura de cuantificación automatizada, flexible y modular que incluye la implementación de las técnicas anteriores y se proporciona de forma pública. Su objetivo es ofrecer a desabolladores e investigadores un entorno común para prototipar y verificar nuevas metodologías de cuantificación de forma sencilla. Describimos el flujo de trabajo, justificamos las decisiones de diseño tomadas, explicamos su API pública y hacemos una demostración paso a paso de su funcionamiento. Además mostramos, a través de un ejemplo sencillo, la forma en que conectar nuevas extensiones a la herramienta con las interfaces ya existentes para poder así expandir y mejorar las capacidades de HOPLITE. ABSTRACT Using fixed-point arithmetic is one of the most common design choices for systems where area, power or throughput are heavily constrained. In order to produce implementations where the cost is minimized without negatively impacting the accuracy of the results, a careful assignment of word-lengths is required. The problem of finding the optimal combination of fixed-point word-lengths for a given system is a combinatorial NP-hard problem to which developers devote between 25 and 50% of the design-cycle time. Reconfigurable hardware platforms such as FPGAs also benefit of the advantages of fixed-point arithmetic, as it compensates for the slower clock frequencies and less efficient area utilization of the hardware platform with respect to ASICs. As FPGAs become commonly used for scientific computation, designs constantly grow larger and more complex, up to the point where they cannot be handled efficiently by current signal and quantization noise modelling and word-length optimization methodologies. In this Ph.D. Thesis we explore different aspects of the quantization problem and we present new methodologies for each of them: The techniques based on extensions of intervals have allowed to obtain accurate models of the signal and quantization noise propagation in systems with non-linear operations. We take this approach a step further by introducing elements of MultiElement Generalized Polynomial Chaos (ME-gPC) and combining them with an stateof- the-art Statistical Modified Affine Arithmetic (MAA) based methodology in order to model systems that contain control-flow structures. Our methodology produces the different execution paths automatically, determines the regions of the input domain that will exercise them, and extracts the system statistical moments from the partial results. We use this technique to estimate both the dynamic range and the round-off noise in systems with the aforementioned control-flow structures. We show the good accuracy of our approach, which in some case studies with non-linear operators shows a 0.04 % deviation respect to the simulation-based reference values. A known drawback of the techniques based on extensions of intervals is the combinatorial explosion of terms as the size of the targeted systems grows, which leads to scalability problems. To address this issue we present a clustered noise injection technique that groups the signals in the system, introduces the noise terms in each group independently and then combines the results at the end. In this way, the number of noise sources in the system at a given time is controlled and, because of this, the combinato rial explosion is minimized. We also present a multi-way partitioning algorithm aimed at minimizing the deviation of the results due to the loss of correlation between noise terms, in order to keep the results as accurate as possible. This Ph.D. Thesis also covers the development of methodologies for word-length optimization based on Monte-Carlo simulations in reasonable times. We do so by presenting two novel techniques that explore the reduction of the execution times approaching the problem in two different ways: First, the interpolative method applies a simple but precise interpolator to estimate the sensitivity of each signal, which is later used to guide the optimization effort. Second, the incremental method revolves on the fact that, although we strictly need to guarantee a certain confidence level in the simulations for the final results of the optimization process, we can do it with more relaxed levels, which in turn implies using a considerably smaller amount of samples, in the initial stages of the process, when we are still far from the optimized solution. Through these two approaches we demonstrate that the execution time of classical greedy techniques can be accelerated by factors of up to ×240 for small/medium sized problems. Finally, this book introduces HOPLITE, an automated, flexible and modular framework for quantization that includes the implementation of the previous techniques and is provided for public access. The aim is to offer a common ground for developers and researches for prototyping and verifying new techniques for system modelling and word-length optimization easily. We describe its work flow, justifying the taken design decisions, explain its public API and we do a step-by-step demonstration of its execution. We also show, through an example, the way new extensions to the flow should be connected to the existing interfaces in order to expand and improve the capabilities of HOPLITE.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Ocean energy is a promising resource for renewable electricity generation that presents many advantages, such as being more predictable than wind energy, but also some disadvantages such as large and slow amplitude variations in the generated power. This paper presents a hardware-in-the-loop prototype that allows the study of the electric power profile generated by a wave power plant based on the oscillating water column (OWC) principle. In particular, it facilitates the development of new solutions to improve the intermittent profile of the power fed into the grid or the test of the OWC behavior when facing a voltage dip. Also, to obtain a more realistic model behavior, statistical models of real waves have been implemented.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Several languages have been proposed for the task of describing networks of systems, either to help on managing, simulate or deploy testbeds for testing purposes. However, there is no one specifically designed to describe the honeynets, covering the specific characteristics in terms of applications and tools included in the honeypot systems that make the honeynet. In this paper, the requirements of honeynet description are studied and a survey of existing description languages is presented, concluding that a CIM (Common Information Model) match the basic requirements. Thus, a CIM like technology independent honeynet description language (TIHDL) is proposed. The language is defined being independent of the platform where the honeynet will be deployed later, and it can be translated, either using model-driven techniques or other translation mechanisms, into the description languages of honeynet deployment platforms and tools. This approach gives flexibility to allow the use of a combination of heterogeneous deployment platforms. Besides, a flexible virtual honeynet generation tool (HoneyGen) based on the approach and description language proposed and capable of deploying honeynets over VNX (Virtual Networks over LinuX) and Honeyd platforms is presented for validation purposes.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We present statistical methods for analyzing replicated cDNA microarray expression data and report the results of a controlled experiment. The study was conducted to investigate inherent variability in gene expression data and the extent to which replication in an experiment produces more consistent and reliable findings. We introduce a statistical model to describe the probability that mRNA is contained in the target sample tissue, converted to probe, and ultimately detected on the slide. We also introduce a method to analyze the combined data from all replicates. Of the 288 genes considered in this controlled experiment, 32 would be expected to produce strong hybridization signals because of the known presence of repetitive sequences within them. Results based on individual replicates, however, show that there are 55, 36, and 58 highly expressed genes in replicates 1, 2, and 3, respectively. On the other hand, an analysis by using the combined data from all 3 replicates reveals that only 2 of the 288 genes are incorrectly classified as expressed. Our experiment shows that any single microarray output is subject to substantial variability. By pooling data from replicates, we can provide a more reliable analysis of gene expression data. Therefore, we conclude that designing experiments with replications will greatly reduce misclassification rates. We recommend that at least three replicates be used in designing experiments by using cDNA microarrays, particularly when gene expression data from single specimens are being analyzed.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The availability of complete genome sequences and mRNA expression data for all genes creates new opportunities and challenges for identifying DNA sequence motifs that control gene expression. An algorithm, “MobyDick,” is presented that decomposes a set of DNA sequences into the most probable dictionary of motifs or words. This method is applicable to any set of DNA sequences: for example, all upstream regions in a genome or all genes expressed under certain conditions. Identification of words is based on a probabilistic segmentation model in which the significance of longer words is deduced from the frequency of shorter ones of various lengths, eliminating the need for a separate set of reference data to define probabilities. We have built a dictionary with 1,200 words for the 6,000 upstream regulatory regions in the yeast genome; the 500 most significant words (some with as few as 10 copies in all of the upstream regions) match 114 of 443 experimentally determined sites (a significance level of 18 standard deviations). When analyzing all of the genes up-regulated during sporulation as a group, we find many motifs in addition to the few previously identified by analyzing the subclusters individually to the expression subclusters. Applying MobyDick to the genes derepressed when the general repressor Tup1 is deleted, we find known as well as putative binding sites for its regulatory partners.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The present work develops and implements a biomathematical statement of how reciprocal connectivity drives stress-adaptive homeostasis in the corticotropic (hypothalamo-pituitary-adrenal) axis. In initial analyses with this interactive construct, we test six specific a priori hypotheses of mechanisms linking circadian (24-h) rhythmicity to pulsatile secretory output. This formulation offers a dynamic framework for later statistical estimation of unobserved in vivo neurohormone secretion and within-axis, dose-responsive interfaces in health and disease. Explication of the core dynamics of the stress-responsive corticotropic axis based on secure physiological precepts should help to unveil new biomedical hypotheses of stressor-specific system failure.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A statistical modeling approach is proposed for use in searching large microarray data sets for genes that have a transcriptional response to a stimulus. The approach is unrestricted with respect to the timing, magnitude or duration of the response, or the overall abundance of the transcript. The statistical model makes an accommodation for systematic heterogeneity in expression levels. Corresponding data analyses provide gene-specific information, and the approach provides a means for evaluating the statistical significance of such information. To illustrate this strategy we have derived a model to depict the profile expected for a periodically transcribed gene and used it to look for budding yeast transcripts that adhere to this profile. Using objective criteria, this method identifies 81% of the known periodic transcripts and 1,088 genes, which show significant periodicity in at least one of the three data sets analyzed. However, only one-quarter of these genes show significant oscillations in at least two data sets and can be classified as periodic with high confidence. The method provides estimates of the mean activation and deactivation times, induced and basal expression levels, and statistical measures of the precision of these estimates for each periodic transcript.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A model of interdependent decision making has been developed to understand group differences in socioeconomic behavior such as nonmarital fertility, school attendance, and drug use. The statistical mechanical structure of the model illustrates how the physical sciences contain useful tools for the study of socioeconomic phenomena.