937 resultados para Discrete time inventory models
Resumo:
This dissertation explores phase I dose-finding designs in cancer trials from three perspectives: the alternative Bayesian dose-escalation rules, a design based on a time-to-dose-limiting toxicity (DLT) model, and a design based on a discrete-time multi-state (DTMS) model. We list alternative Bayesian dose-escalation rules and perform a simulation study for the intra-rule and inter-rule comparisons based on two statistical models to identify the most appropriate rule under certain scenarios. We provide evidence that all the Bayesian rules outperform the traditional ``3+3'' design in the allocation of patients and selection of the maximum tolerated dose. The design based on a time-to-DLT model uses patients' DLT information over multiple treatment cycles in estimating the probability of DLT at the end of treatment cycle 1. Dose-escalation decisions are made whenever a cycle-1 DLT occurs, or two months after the previous check point. Compared to the design based on a logistic regression model, the new design shows more safety benefits for trials in which more late-onset toxicities are expected. As a trade-off, the new design requires more patients on average. The design based on a discrete-time multi-state (DTMS) model has three important attributes: (1) Toxicities are categorized over a distribution of severity levels, (2) Early toxicity may inform dose escalation, and (3) No suspension is required between accrual cohorts. The proposed model accounts for the difference in the importance of the toxicity severity levels and for transitions between toxicity levels. We compare the operating characteristics of the proposed design with those from a similar design based on a fully-evaluated model that directly models the maximum observed toxicity level within the patients' entire assessment window. We describe settings in which, under comparable power, the proposed design shortens the trial. The proposed design offers more benefit compared to the alternative design as patient accrual becomes slower.
Resumo:
We examine the time-series relationship between housing prices in eight Southern California metropolitan statistical areas (MSAs). First, we perform cointegration tests of the housing price indexes for the MSAs, finding seven cointegrating vectors. Thus, the evidence suggests that one common trend links the housing prices in these eight MSAs, a purchasing power parity finding for the housing prices in Southern California. Second, we perform temporal Granger causality tests revealing intertwined temporal relationships. The Santa Anna MSA leads the pack in temporally causing housing prices in six of the other seven MSAs, excluding only the San Luis Obispo MSA. The Oxnard MSA experienced the largest number of temporal effects from other MSAs, six of the seven, excluding only Los Angeles. The Santa Barbara MSA proved the most isolated in that it temporally caused housing prices in only two other MSAs (Los Angels and Oxnard) and housing prices in the Santa Anna MSA temporally caused prices in Santa Barbara. Third, we calculate out-of-sample forecasts in each MSA, using various vector autoregressive (VAR) and vector error-correction (VEC) models, as well as Bayesian, spatial, and causality versions of these models with various priors. Different specifications provide superior forecasts in the different MSAs. Finally, we consider the ability of theses time-series models to provide accurate out-of-sample predictions of turning points in housing prices that occurred in 2006:Q4. Recursive forecasts, where the sample is updated each quarter, provide reasonably good forecasts of turning points.
Resumo:
The infant mortality rate (IMR) is considered to be one of the most important indices of a country's well-being. Countries around the world and other health organizations like the World Health Organization are dedicating their resources, knowledge and energy to reduce the infant mortality rates. The well-known Millennium Development Goal 4 (MDG 4), whose aim is to archive a two thirds reduction of the under-five mortality rate between 1990 and 2015, is an example of the commitment. ^ In this study our goal is to model the trends of IMR between the 1950s to 2010s for selected countries. We would like to know how the IMR is changing overtime and how it differs across countries. ^ IMR data collected over time forms a time series. The repeated observations of IMR time series are not statistically independent. So in modeling the trend of IMR, it is necessary to account for these correlations. We proposed to use the generalized least squares method in general linear models setting to deal with the variance-covariance structure in our model. In order to estimate the variance-covariance matrix, we referred to the time-series models, especially the autoregressive and moving average models. Furthermore, we will compared results from general linear model with correlation structure to that from ordinary least squares method without taking into account the correlation structure to check how significantly the estimates change.^
Resumo:
Hospitals, like all organizations, have both a mission and a finite supply of resources with which to accomplish that mission. Because the inventory of therapeutic drugs is among the more expensive resources needed by a hospital to achieve its mission, a conceptual model of structure plus process equals outcome posits that adequate emphasis should be placed on optimization of the organization's investment in this important structural resource to provide highest quality outcomes. Therefore emphasis should be placed on the optimization of pharmacy inventory because lowering the financial investment in drug inventory and associated costs increases productive efficiency, a key element of quality. ^ In this study, a post-intervention analysis of a hospital pharmacy inventory management technology implementation at The University of Texas M.D. Anderson Cancer Center was conducted to determine if an intervention which reduced a hospital's financial investment in pharmaceutical inventory provided an opportunity to incrementally optimize the organization's mix of structural resources thereby improving quality of care. The results suggest that hospital pharmacies currently lacking technology to support automated purchasing logistics and perpetual, real-time inventory management for drugs may achieve measurable benefits from the careful implementation of such technology, enabling the hospital to lower its investment in on-hand inventory and, potentially, to reduce overall purchasing expenditures. ^ The importance of these savings to the hospital and potentially to the patient should not be underestimated for their ability to generate funding for previously unfunded public health programs or in their ability to provide financial relief to patients in the form of lower drug costs given the current climate of escalating healthcare costs and tightening reimbursements.^
Resumo:
Division of labor is a widely studied aspect of colony behavior of social insects. Division of labor models indicate how individuals distribute themselves in order to perform different tasks simultaneously. However, models that study division of labor from a dynamical system point of view cannot be found in the literature. In this paper, we define a division of labor model as a discrete-time dynamical system, in order to study the equilibrium points and their properties related to convergence and stability. By making use of this analytical model, an adaptive algorithm based on division of labor can be designed to satisfy dynamic criteria. In this way, we have designed and tested an algorithm that varies the response thresholds in order to modify the dynamic behavior of the system. This behavior modification allows the system to adapt to specific environmental and collective situations, making the algorithm a good candidate for distributed control applications. The variable threshold algorithm is based on specialization mechanisms. It is able to achieve an asymptotically stable behavior of the system in different environments and independently of the number of individuals. The algorithm has been successfully tested under several initial conditions and number of individuals.
Resumo:
The study area is La Colacha sub-basins from Arroyos Menores basins, natural areas at West and South of Río Cuarto in Province of Córdoba of Argentina, fertile with loess soils and monsoon temperate climate, but with soil erosions including regressive gullies that degrade them progressively. Cultivated gently since some hundred sixty years, coordinated action planning became necessary to conserve lands while keeping good agro-production. The authors had improved data on soils and on hydrology for the study area, evaluated systems of soil uses and actions to be recommended and applied Decision Support Systems (DSS) tools for that, and were conducted to use discrete multi-criteria models (MCDM) for the more global views about soil conservation and hydraulic management actions and about main types of use of soils. For that they used weighted PROMETHEE, ELECTRE, and AHP methods with a system of criteria grouped as environmental, economic and social, and criteria from their data on effects of criteria. The alternatives resulting offer indication for planning depending somehow on sub basins and on selections of weights, but actions for conservation of soils and water management measures are recommended to conserve the basins conditions, actually sensibly degrading, mainly keeping actual uses of the lands.
Resumo:
Esta tesis propone una completa formulación termo-mecánica para la simulación no-lineal de mecanismos flexibles basada en métodos libres de malla. El enfoque se basa en tres pilares principales: la formulación de Lagrangiano total para medios continuos, la discretización de Bubnov-Galerkin, y las funciones de forma libres de malla. Los métodos sin malla se caracterizan por la definición de un conjunto de funciones de forma en dominios solapados, junto con una malla de integración de las ecuaciones discretas de balance. Dos tipos de funciones de forma se han seleccionado como representación de las familias interpolantes (Funciones de Base Radial) y aproximantes (Mínimos Cuadrados Móviles). Su formulación se ha adaptado haciendo sus parámetros compatibles, y su ausencia de conectividad predefinida se ha aprovechado para interconectar múltiples dominios de manera automática, permitiendo el uso de mallas de fondo no conformes. Se propone una formulación generalizada de restricciones, juntas y contactos, válida para sólidos rígidos y flexibles, siendo estos últimos discretizados mediante elementos finitos (MEF) o libres de malla. La mayor ventaja de este enfoque reside en que independiza completamente el dominio con respecto de las uniones y acciones externas a cada sólido, permitiendo su definición incluso fuera del contorno. Al mismo tiempo, también se minimiza el número de ecuaciones de restricción necesarias para la definición de uniones realistas. Las diversas validaciones, ejemplos y comparaciones detalladas muestran como el enfoque propuesto es genérico y extensible a un gran número de sistemas. En concreto, las comparaciones con el MEF indican una importante reducción del error para igual número de nodos, tanto en simulaciones mecánicas, como térmicas y termo-mecánicas acopladas. A igualdad de error, la eficiencia numérica de los métodos libres de malla es mayor que la del MEF cuanto más grosera es la discretización. Finalmente, la formulación se aplica a un problema de diseño real sobre el mantenimiento de estructuras masivas en el interior de un reactor de fusión, demostrando su viabilidad en análisis de problemas reales, y a su vez mostrando su potencial para su uso en simulación en tiempo real de sistemas no-lineales. A new complete formulation is proposed for the simulation of nonlinear dynamic of multibody systems with thermo-mechanical behaviour. The approach is founded in three main pillars: total Lagrangian formulation, Bubnov-Galerkin discretization, and meshfree shape functions. Meshfree methods are characterized by the definition of a set of shape functions in overlapping domains, and a background grid for integration of the Galerkin discrete equations. Two different types of shape functions have been chosen as representatives of interpolation (Radial Basis Functions), and approximation (Moving Least Squares) families. Their formulation has been adapted to use compatible parameters, and their lack of predefined connectivity is used to interconnect different domains seamlessly, allowing the use of non-conforming meshes. A generalized formulation for constraints, joints, and contacts is proposed, which is valid for rigid and flexible solids, being the later discretized using either finite elements (FEM) or meshfree methods. The greatest advantage of this approach is that makes the domain completely independent of the external links and actions, allowing to even define them outside of the boundary. At the same time, the number of constraint equations needed for defining realistic joints is minimized. Validation, examples, and benchmarks are provided for the proposed formulation, demonstrating that the approach is generic and extensible to further problems. Comparisons with FEM show a much lower error for the same number of nodes, both for mechanical and thermal analyses. The numerical efficiency is also better when coarse discretizations are used. A final demonstration to a real problem for handling massive structures inside of a fusion reactor is presented. It demonstrates that the application of meshfree methods is feasible and can provide an advantage towards the definition of nonlinear real-time simulation models.
Resumo:
This paper revises mainstream economic models which include time use in an explicit and endogenous manner, suggesting a extended theory which escape from the main problem existing in the literature. In order to do it, we start by presenting in section 2 the mainstream time use models in economics, showing their main features. Once this is done, we introduce the reader in the main problems this kind of well established models imply, within section 3, being the most highlighted the problem of joint production. Subsequently, we propose an extended theory which solves the problem of joint production; this is extensively described in section 4. Last, but not least, we apply this model to offer a time use analysis of the effect of a policy which increases the retirement age in a life-cycle perspective for a representative individual.
Resumo:
A aquisição experimental de sinais neuronais é um dos principais avanços da neurociência. Por meio de observações da corrente e do potencial elétricos em uma região cerebral, é possível entender os processos fisiológicos envolvidos na geração do potencial de ação, e produzir modelos matemáticos capazes de simular o comportamento de uma célula neuronal. Uma prática comum nesse tipo de experimento é obter leituras a partir de um arranjo de eletrodos posicionado em um meio compartilhado por diversos neurônios, o que resulta em uma mistura de sinais neuronais em uma mesma série temporal. Este trabalho propõe um modelo linear de tempo discreto para o sinal produzido durante o disparo do neurônio. Os coeficientes desse modelo são calculados utilizando-se amostras reais dos sinais neuronais obtidas in vivo. O processo de modelagem concebido emprega técnicas de identificação de sistemas e processamento de sinais, e é dissociado de considerações sobre o funcionamento biofísico da célula, fornecendo uma alternativa de baixa complexidade para a modelagem do disparo neuronal. Além disso, a representação por meio de sistemas lineares permite idealizar um sistema inverso, cuja função é recuperar o sinal original de cada neurônio ativo em uma mistura extracelular. Nesse contexto, são discutidas algumas soluções baseadas em filtros adaptativos para a simulação do sistema inverso, introduzindo uma nova abordagem para o problema de separação de spikes neuronais.
Resumo:
Underwater video transects have become a common tool for quantitative analysis of the seafloor. However a major difficulty remains in the accurate determination of the area surveyed as underwater navigation can be unreliable and image scaling does not always compensate for distortions due to perspective and topography. Depending on the camera set-up and available instruments, different methods of surface measurement are applied, which make it difficult to compare data obtained by different vehicles. 3-D modelling of the seafloor based on 2-D video data and a reference scale can be used to compute subtransect dimensions. Focussing on the length of the subtransect, the data obtained from 3-D models created with the software PhotoModeler Scanner are compared with those determined from underwater acoustic positioning (ultra short baseline, USBL) and bottom tracking (Doppler velocity log, DVL). 3-D model building and scaling was successfully conducted on all three tested set-ups and the distortion of the reference scales due to substrate roughness was identified as the main source of imprecision. Acoustic positioning was generally inaccurate and bottom tracking unreliable on rough terrain. Subtransect lengths assessed with PhotoModeler were on average 20% longer than those derived from acoustic positioning due to the higher spatial resolution and the inclusion of slope. On a high relief wall bottom tracking and 3-D modelling yielded similar results. At present, 3-D modelling is the most powerful, albeit the most time-consuming, method for accurate determination of video subtransect dimensions.
Resumo:
Thesis (Master's)--University of Washington, 2016-06
Resumo:
Thesis (Master's)--University of Washington, 2016-06
Resumo:
The paper presents a computational system based upon formal principles to run spatial models for environmental processes. The simulator is named SimuMap because it is typically used to simulate spatial processes over a mapped representation of terrain. A model is formally represented in SimuMap as a set of coupled sub-models. The paper considers the situation where spatial processes operate at different time levels, but are still integrated. An example of such a situation commonly occurs in watershed hydrology where overland flow and stream channel flow have very different flow rates but are highly related as they are subject to the same terrain runoff processes. SimuMap is able to run a network of sub-models that express different time-space derivatives for water flow processes. Sub-models may be coded generically with a map algebra programming language that uses a surface data model. To address the problem of differing time levels in simulation, the paper: (i) reviews general approaches for numerical solvers, (ii) considers the constraints that need to be enforced to use more adaptive time steps in discrete time specified simulations, and (iii) scaling transfer rates in equations that use different time bases for time-space derivatives. A multistep scheme is proposed for SimuMap. This is presented along with a description of its visual programming interface, its modelling formalisms and future plans. (C) 2003 Elsevier Ltd. All rights reserved.
Resumo:
This study explores whether the introduction of selectively trained radiographers reporting Accident and Emergency (A&E) X-ray examinations or the appendicular skeleton affected the availability of reports for A&E and General Practitioner (GP) examinations at it typical district general hospital. This was achieved by analysing monthly data on A&E and GP examinations for 1993 1997 using structural time-series models. Parameters to capture stochastic seasonal effects and stochastic time trends were included ill the models. The main outcome measures were changes in the number, proportion and timeliness of A&E and GP examinations reported. Radiographer reporting X-ray examinations requested by A&E was associated with it 12% (p = 0.050) increase in the number of A&E examinations reported and it 37% (p
Resumo:
In this study. the authors examined the 2-, 3-, and 4-year outcomes of a school-based, universal approach to the prevention of adolescent depression. Despite initial short-term positive effects, these benefits were not maintained over time. Adolescents who completed the teacher-administered cognitive-behavioral intervention did not differ significantly from adolescents in the monitoring-control condition in terms of changes in depressive symptoms, problem solving, attributional style, or other indicators of psychopathology from preintervention to 4-year follow-up. Results were equivalent irrespective of initial level of depressive symptoms.