17 resultados para Stochastic processes -- Mathematical models
em QUB Research Portal - Research Directory and Institutional Repository for Queen's University Belfast
Resumo:
The degradation of resorbable polymeric devices often takes months to years. Accelerated testing at elevated temperatures is an attractive but controversial technique. The purposes of this paper include: (a) to provide a summary of the mathematical models required to analyse accelerated degradation data and to indicate the pitfalls of using these models; (b) to improve the model previously developed by Han and Pan; (c) to provide a simple version of the model of Han and Pan with an analytical solution that is convenient to use; (d) to demonstrate the application of the improved model in two different poly(lactic acid) systems. It is shown that the simple analytical relations between molecular weight and degradation time widely used in the literature can lead to inadequate conclusions. In more general situations the rate equations are only part of a complete degradation model. Together with previous works in the literature, our study calls for care in using the accelerated testing technique.
Resumo:
Extreme arid regions in the worlds' major deserts are typified by quartz pavement terrain. Cryptic hypolithic communities colonize the ventral surface of quartz rocks and this habitat is characterized by a relative lack of environmental and trophic complexity. Combined with readily identifiable major environmental stressors this provides a tractable model system for determining the relative role of stochastic and deterministic drivers in community assembly. Through analyzing an original, worldwide data set of 16S rRNA-gene defined bacterial communities from the most extreme deserts on the Earth, we show that functional assemblages within the communities were subject to different assembly influences. Null models applied to the photosynthetic assemblage revealed that stochastic processes exerted most effect on the assemblage, although the level of community dissimilarity varied between continents in a manner not always consistent with neutral models. The heterotrophic assemblages displayed signatures of niche processes across four continents, whereas in other cases they conformed to neutral predictions. Importantly, for continents where neutrality was either rejected or accepted, assembly drivers differed between the two functional groups. This study demonstrates that multi-trophic microbial systems may not be fully described by a single set of niche or neutral assembly rules and that stochasticity is likely a major determinant of such systems, with significant variation in the influence of these determinants on a global scale.
Resumo:
1. Ecologists are debating the relative role of deterministic and stochastic determinants of community structure. Although the high diversity and strong spatial structure of soil animal assemblages could provide ecologists with an ideal ecological scenario, surprisingly little information is available on these assemblages.
2. We studied species-rich soil oribatid mite assemblages from a Mediterranean beech forest and a grassland. We applied multivariate regression approaches and analysed spatial autocorrelation at multiple spatial scales using Moran's eigenvectors. Results were used to partition community variance in terms of the amount of variation uniquely accounted for by environmental correlates (e.g. organic matter) and geographical position. Estimated neutral diversity and immigration parameters were also applied to a soil animal group for the first time to simulate patterns of community dissimilarity expected under neutrality, thereby testing neutral predictions.
3. After accounting for spatial autocorrelation, the correlation between community structure and key environmental parameters disappeared: about 40% of community variation consisted of spatial patterns independent of measured environmental variables such as organic matter. Environmentally independent spatial patterns encompassed the entire range of scales accounted for by the sampling design (from tens of cm to 100 m). This spatial variation could be due to either unmeasured but spatially structured variables or stochastic drift mediated by dispersal. Observed levels of community dissimilarity were significantly different from those predicted by neutral models.
4. Oribatid mite assemblages are dominated by processes involving both deterministic and stochastic components and operating at multiple scales. Spatial patterns independent of the measured environmental variables are a prominent feature of the targeted assemblages, but patterns of community dissimilarity do not match neutral predictions. This suggests that either niche-mediated competition or environmental filtering or both are contributing to the core structure of the community. This study indicates new lines of investigation for understanding the mechanisms that determine the signature of the deterministic component of animal community assembly.
Resumo:
Self-compacting concrete (SCC) is generally designed with a relatively higher content of finer, which includes cement, and dosage of superplasticizer than the conventional concrete. The design of the current SCC leads to high compressive strength, which is already used in special applications, where the high cost of materials can be tolerated. Using SCC, which eliminates the need for vibration, leads to increased speed of casting and thus reduces labour requirement, energy consumption, construction time, and cost of equipment. In order to obtain and gain maximum benefit from SCC it has to be used for wider applications. The cost of materials will be decreased by reducing the cement content and using a minimum amount of admixtures. This paper reviews statistical models obtained from a factorial design which was carried out to determine the influence of four key parameters on filling ability, passing ability, segregation and compressive strength. These parameters are important for the successful development of medium strength self-compacting concrete (MS-SCC). The parameters considered in the study were the contents of cement and pulverised fuel ash (PFA), water-to-powder ratio (W/P), and dosage of superplasticizer (SP). The responses of the derived statistical models are slump flow, fluidity loss, rheological parameters, Orimet time, V-funnel time, L-box, JRing combined to Orimet, JRing combined to cone, fresh segregation, and compressive strength at 7, 28 and 90 days. The models are valid for mixes made with 0.38 to 0.72 W/P ratio, 60 to 216 kg/m3 of cement content, 183 to 317 kg/m3 of PFA and 0 to 1% of SP, by mass of powder. The utility of such models to optimize concrete mixes to achieve good balance between filling ability, passing ability, segregation, compressive strength, and cost is discussed. Examples highlighting the usefulness of the models are presented using isoresponse surfaces to demonstrate single and coupled effects of mix parameters on slump flow, loss of fluidity, flow resistance, segregation, JRing combined to Orimet, and compressive strength at 7 and 28 days. Cost analysis is carried out to show trade-offs between cost of materials and specified consistency levels and compressive strength at 7 and 28 days that can be used to identify economic mixes. The paper establishes the usefulness of the mathematical models as a tool to facilitate the test protocol required to optimise medium strength SCC.
Resumo:
Community structure depends on both deterministic and stochastic processes. However, patterns of community dissimilarity (e.g. difference in species composition) are difficult to interpret in terms of the relative roles of these processes. Local communities can be more dissimilar (divergence) than, less dissimilar (convergence) than, or as dissimilar as a hypothetical control based on either null or neutral models. However, several mechanisms may result in the same pattern, or act concurrently to generate a pattern, and much research has recently been focusing on unravelling these mechanisms and their relative contributions. Using a simulation approach, we addressed the effect of a complex but realistic spatial structure in the distribution of the niche axis and we analysed patterns of species co-occurrence and beta diversity as measured by dissimilarity indices (e.g. Jaccard index) using either expectations under a null model or neutral dynamics (i.e., based on switching off the niche effect). The strength of niche processes, dispersal, and environmental noise strongly interacted so that niche-driven dynamics may result in local communities that either diverge or converge depending on the combination of these factors. Thus, a fundamental result is that, in real systems, interacting processes of community assembly can be disentangled only by measuring traits such as niche breadth and dispersal. The ability to detect the signal of the niche was also dependent on the spatial resolution of the sampling strategy, which must account for the multiple scale spatial patterns in the niche axis. Notably, some of the patterns we observed correspond to patterns of community dissimilarities previously observed in the field and suggest mechanistic explanations for them or the data required to solve them. Our framework offers a synthesis of the patterns of community dissimilarity produced by the interaction of deterministic and stochastic determinants of community assembly in a spatially explicit and complex context.
Resumo:
Mathematical modelling has become an essential tool in the design of modern catalytic systems. Emissions legislation is becoming increasingly stringent, and so mathematical models of aftertreatment systems must become more accurate in order to provide confidence that a catalyst will convert pollutants over the required range of conditions.
Automotive catalytic converter models contain several sub-models that represent processes such as mass and heat transfer, and the rates at which the reactions proceed on the surface of the precious metal. Of these sub-models, the prediction of the surface reaction rates is by far the most challenging due to the complexity of the reaction system and the large number of gas species involved. The reaction rate sub-model uses global reaction kinetics to describe the surface reaction rate of the gas species and is based on the Langmuir Hinshelwood equation further developed by Voltz et al. [1] The reactions can be modelled using the pre-exponential and activation energies of the Arrhenius equations and the inhibition terms.
The reaction kinetic parameters of aftertreatment models are found from experimental data, where a measured light-off curve is compared against a predicted curve produced by a mathematical model. The kinetic parameters are usually manually tuned to minimize the error between the measured and predicted data. This process is most commonly long, laborious and prone to misinterpretation due to the large number of parameters and the risk of multiple sets of parameters giving acceptable fits. Moreover, the number of coefficients increases greatly with the number of reactions. Therefore, with the growing number of reactions, the task of manually tuning the coefficients is becoming increasingly challenging.
In the presented work, the authors have developed and implemented a multi-objective genetic algorithm to automatically optimize reaction parameters in AxiSuite®, [2] a commercial aftertreatment model. The genetic algorithm was developed and expanded from the code presented by Michalewicz et al. [3] and was linked to AxiSuite using the Simulink add-on for Matlab.
The default kinetic values stored within the AxiSuite model were used to generate a series of light-off curves under rich conditions for a number of gas species, including CO, NO, C3H8 and C3H6. These light-off curves were used to generate an objective function.
This objective function was used to generate a measure of fit for the kinetic parameters. The multi-objective genetic algorithm was subsequently used to search between specified limits to attempt to match the objective function. In total the pre-exponential factors and activation energies of ten reactions were simultaneously optimized.
The results reported here demonstrate that, given accurate experimental data, the optimization algorithm is successful and robust in defining the correct kinetic parameters of a global kinetic model describing aftertreatment processes.
Resumo:
All mammals lose their ability to produce lactase (β-galactosidase), the enzyme that cleaves lactose into galactose and glucose, after weaning. The prevalence of lactase deficiency (LD) spans from 2 to 15% among northern Europeans, to nearly 100% among Asians. Following lactose consumption, people with LD often experience gastrointestinal symptoms such as abdominal pain, bowel distension, cramps and flatulence, or even systemic problems such as headache, loss of concentration and muscle pain. These symptoms vary depending on the amount of lactose ingested, type of food and degree of intolerance. Although those affected can avoid the uptake of dairy products, in doing so, they lose a readily available source of calcium and protein. In this work, gels obtained by complexation of Tetronic 90R4 with α-cyclodextrin loaded with β-galactosidase are proposed as a way to administer the enzyme immediately before or with the lactose-containing meal. Both molecules are biocompatible, can form gels in situ, and show sustained erosion kinetics in aqueous media. The complex was characterized by FTIR that evidenced an inclusion complex between the polyethylene oxide block and α-cyclodextrin. The release profiles of β-galactosidase from two different matrices (gels and tablets) of the in situ hydrogels have been obtained. The influence of the percentage of Tetronic in media of different pH was evaluated. No differences were observed regarding the release rate from the gel matrices at pH 6 (t50 = 105 min). However, in the case of the tablets, the kinetics were faster and they released a greater amount of 90R4 (25%, t50 = 40–50 min). Also, the amount of enzyme released was higher for mixtures with 25% Tetronic. Using suitable mathematical models, the corresponding kinetic parameters have been calculated. In all cases, the release data fit quite well to the Peppas–Sahlin model equation, indicating that the release of β-galactosidase is governed by a combination of diffusion and erosion processes. It has been observed that the diffusion mechanism prevails over erosion during the first 50 minutes, followed by continued release of the enzyme due to the disintegration of the matrix.
Resumo:
This paper provides an overview of the basic theory underlying 1D unsteady gas dynamics, the computational method developed at Queen’s University Belfast (QUB), the use of CFD as an alternative and some experimental results that demonstrate the techniques used to develop the mathematical models.
Resumo:
Artifact removal from physiological signals is an essential component of the biosignal processing pipeline. The need for powerful and robust methods for this process has become particularly acute as healthcare technology deployment undergoes transition from the current hospital-centric setting toward a wearable and ubiquitous monitoring environment. Currently, determining the relative efficacy and performance of the multiple artifact removal techniques available on real world data can be problematic, due to incomplete information on the uncorrupted desired signal. The majority of techniques are presently evaluated using simulated data, and therefore, the quality of the conclusions is contingent on the fidelity of the model used. Consequently, in the biomedical signal processing community, there is considerable focus on the generation and validation of appropriate signal models for use in artifact suppression. Most approaches rely on mathematical models which capture suitable approximations to the signal dynamics or underlying physiology and, therefore, introduce some uncertainty to subsequent predictions of algorithm performance. This paper describes a more empirical approach to the modeling of the desired signal that we demonstrate for functional brain monitoring tasks which allows for the procurement of a ground truth signal which is highly correlated to a true desired signal that has been contaminated with artifacts. The availability of this ground truth, together with the corrupted signal, can then aid in determining the efficacy of selected artifact removal techniques. A number of commonly implemented artifact removal techniques were evaluated using the described methodology to validate the proposed novel test platform. © 2012 IEEE.
Resumo:
Purpose
Recent in vitro results have shown significant contributions to cell killing from signaling effects at doses that are typically used in radiation therapy. This study investigates whether these in vitro observations can be reconciled with in vivo knowledge and how signaling may have an impact on future developments in radiation therapy.
Methods and Materials
Prostate cancer treatment plans were generated for a series of 10 patients using 3-dimensional conformal therapy, intensity modulated radiation therapy (IMRT), and volumetric modulated arc therapy techniques. These plans were evaluated using mathematical models of survival following modulated radiation exposures that were developed from in vitro observations and incorporate the effects of intercellular signaling. The impact on dose-volume histograms and mean doses were evaluated by converting these survival levels into "signaling-adjusted doses" for comparison.
Results
Inclusion of intercellular communication leads to significant differences between the signalling-adjusted and physical doses across a large volume. Organs in low-dose regions near target volumes see the largest increases, with mean signaling-adjusted bladder doses increasing from 23 to 33 Gy in IMRT plans. By contrast, in high-dose regions, there is a small decrease in signaling-adjusted dose due to reduced contributions from neighboring cells, with planning target volume mean doses falling from 74 to 71 Gy in IMRT. Overall, however, the dose distributions remain broadly similar, and comparisons between the treatment modalities are largely unchanged whether physical or signaling-adjusted dose is compared. Conclusions Although incorporating cellular signaling significantly affects cell killing in low-dose regions and suggests a different interpretation for many phenomena, their effect in high-dose regions for typical planning techniques is comparatively small. This indicates that the significant signaling effects observed in vitro are not contradicted by comparison with clinical observations. Future investigations are needed to validate these effects in vivo and to quantify their ranges and potential impact on more advanced radiation therapy techniques.
Resumo:
Parasites play pivotal roles in structuring communities, often via indirect interactions with non-host species. These effects can be density-mediated (through mortality) or trait-mediated (behavioural, physiological and developmental), and may be crucial to population interactions, including biological invasions. For instance, parasitism can alter intraguild predation (IGP) between native and invasive crustaceans, reversing invasion outcomes. Here, we use mathematical models to examine how parasite-induced trait changes influence the population dynamics of hosts that interact via IGP. We show that trait-mediated indirect interactions impart keystone effects, promoting or inhibiting host coexistence. Parasites can thus have strong ecological impacts, even if they have negligible virulence, underscoring the need to consider trait-mediated effects when predicting effects of parasites on community structure in general and biological invasions in particular.
Resumo:
Mortality models used for forecasting are predominantly based on the statistical properties of time series and do not generally incorporate an understanding of the forces driving secular trends. This paper addresses three research questions: Can the factors found in stochastic mortality-forecasting models be associated with real-world trends in health-related variables? Does inclusion of health-related factors in models improve forecasts? Do resulting models give better forecasts than existing stochastic mortality models? We consider whether the space spanned by the latent factor structure in mortality data can be adequately described by developments in gross domestic product, health expenditure and lifestyle-related risk factors using statistical techniques developed in macroeconomics and finance. These covariates are then shown to improve forecasts when incorporated into a Bayesian hierarchical model. Results are comparable or better than benchmark stochastic mortality models.
Resumo:
Gels obtained by complexation of octablock star polyethylene oxide/polypropylene oxide copolymers (Tetronic 90R4) with -cyclodextrin (-CD) were evaluated as matrices for drug release. Both molecules are biocompatible so they can be potentially applied to drug delivery systems. Two different types of matrices of Tetronic 90R4 and -CD were evaluated: gels and tablets. These gels are capable to gelifying in situ and show sustained erosion kinetics in aqueous media. Tablets were prepared by freeze-drying and comprising the gels. Using these two different matrices, the release of two model molecules, L-tryptophan (Trp), and a protein, bovine serum albumin (BSA), was evaluated. The release profiles of these molecules from gels and tablets prove that they are suitable for sustained delivery. Mathematical models were applied to the release curves from tablets to elucidate the drug delivery mechanism. Good correlations were found for the fittings of the release curves to different equations. The results point that the release of Trp from different tablets is always governed by Fickian diffusion, whereas the release of BSA is governed by a combination of diffusion and tablet erosion.