7 resultados para parallel computation model

em Duke University


Relevância:

80.00% 80.00%

Publicador:

Resumo:

CD4+ T cells play a crucial in the adaptive immune system. They function as the central hub to orchestrate the rest of immunity: CD4+ T cells are essential governing machinery in antibacterial and antiviral responses by facilitating B cell affinity maturation and coordinating the innate and adaptive immune systems to boost the overall immune outcome; on the contrary, hyperactivation of the inflammatory lineages of CD4+ T cells, as well as the impairments of suppressive CD4+ regulatory T cells, are the etiology of various autoimmunity and inflammatory diseases. The broad role of CD4+ T cells in both physiological and pathological contexts prompted me to explore the modulation of CD4+ T cells on the molecular level.

microRNAs (miRNAs) are small RNA molecules capable of regulating gene expression post-transcriptionally. miRNAs have been shown to exert substantial regulatory effects on CD4+ T cell activation, differentiation and helper function. Specifically, my lab has previously established the function of the miR-17-92 cluster in Th1 differentiation and anti-tumor responses. Here, I further analyzed the role of this miRNA cluster in Th17 differentiation, specifically, in the context of autoimmune diseases. Using both gain- and loss-of-function approaches, I demonstrated that miRNAs in miR-17-92, specifically, miR-17 and miR-19b in this cluster, is a crucial promoter of Th17 differentiation. Consequently, loss of miR-17-92 expression in T cells mitigated the progression of experimental autoimmune encephalomyelitis and T cell-induced colitis. In combination with my previous data, the molecular dissection of this cluster establishes that miR-19b and miR-17 play a comprehensive role in promoting multiple aspects of inflammatory T cell responses, which underscore them as potential targets for oligonucleotide-based therapy in treating autoimmune diseases.

To systematically study miRNA regulation in effector CD4+ T cells, I devised a large-scale miRNAome profiling to track in vivo miRNA changes in antigen-specific CD4+ T cells activated by Listeria challenge. From this screening, I identified that miR-23a expression tightly correlates with CD4+ effector expansion. Ectopic expression and genetic deletion strategies validated that miR-23a was required for antigen-stimulated effector CD4+ T cell survival in vitro and in vivo. I further determined that miR-23a targets Ppif, a gatekeeper of mitochondrial reactive oxygen species (ROS) release that protects CD4+ T cells from necrosis. Necrosis is a type of cell death that provokes inflammation, and it is prominently triggered by ROS release and its consequent oxidative stress. My finding that miR-23a curbs ROS-mediated necrosis highlights the essential role of this miRNA in maintaining immune homeostasis.

A key feature of miRNAs is their ability to modulate different biological aspects in different cell populations. Previously, my lab found that miR-23a potently suppresses CD8+ T cell cytotoxicity by restricting BLIMP1 expression. Since BLIMP1 has been found to inhibit T follicular helper (Tfh) differentiation by antagonizing the master transcription factor BCL6, I investigated whether miR-23a is also involved in Tfh differentiation. However, I found that miR-23a does not target BLIMP1 in CD4+ T cells and loss of miR-23a even fostered Tfh differentiation. This data indicate that miR-23a may target other pathways in CD4+ T cells regarding the Tfh differentiation pathway.

Although the lineage identity and regulatory networks for Tfh cells have been defined, the differentiation path of Tfh cells remains elusive. Two models have been proposed to explain the differentiation process of Tfh cells: in the parallel differentiation model, the Tfh lineage is segregated from other effector lineages at the early stage of antigen activation; alternatively, the sequential differentiation model suggests that naïve CD4+ T cells first differentiate into various effector lineages, then further program into Tfh cells. To address this question, I developed a novel in vitro co-culture system that employed antigen-specific CD4+ T cells, naïve B cells presenting cognate T cell antigen and BAFF-producing feeder cells to mimic germinal center. Using this system, I were able to robustly generate GC-like B cells. Notably, well-differentiated Th1 or Th2 effector cells also quickly acquired Tfh phenotype and function during in vitro co-culture, which suggested a sequential differentiation path for Tfh cells. To examine this path in vivo, under conditions of classical Th1- or Th2-type immunizations, I employed a TCRβ repertoire sequencing technique to track the clonotype origin of Tfh cells. Under both Th1- and Th2- immunization conditions, I observed profound repertoire overlaps between the Teff and Tfh populations, which strongly supports the proposed sequential differentiation model. Therefore, my studies establish a new platform to conveniently study Tfh-GC B cell interactions and provide insights into Tfh differentiation processes.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The work presented in this dissertation is focused on applying engineering methods to develop and explore probabilistic survival models for the prediction of decompression sickness in US NAVY divers. Mathematical modeling, computational model development, and numerical optimization techniques were employed to formulate and evaluate the predictive quality of models fitted to empirical data. In Chapters 1 and 2 we present general background information relevant to the development of probabilistic models applied to predicting the incidence of decompression sickness. The remainder of the dissertation introduces techniques developed in an effort to improve the predictive quality of probabilistic decompression models and to reduce the difficulty of model parameter optimization.

The first project explored seventeen variations of the hazard function using a well-perfused parallel compartment model. Models were parametrically optimized using the maximum likelihood technique. Model performance was evaluated using both classical statistical methods and model selection techniques based on information theory. Optimized model parameters were overall similar to those of previously published Results indicated that a novel hazard function definition that included both ambient pressure scaling and individually fitted compartment exponent scaling terms.

We developed ten pharmacokinetic compartmental models that included explicit delay mechanics to determine if predictive quality could be improved through the inclusion of material transfer lags. A fitted discrete delay parameter augmented the inflow to the compartment systems from the environment. Based on the observation that symptoms are often reported after risk accumulation begins for many of our models, we hypothesized that the inclusion of delays might improve correlation between the model predictions and observed data. Model selection techniques identified two models as having the best overall performance, but comparison to the best performing model without delay and model selection using our best identified no delay pharmacokinetic model both indicated that the delay mechanism was not statistically justified and did not substantially improve model predictions.

Our final investigation explored parameter bounding techniques to identify parameter regions for which statistical model failure will not occur. When a model predicts a no probability of a diver experiencing decompression sickness for an exposure that is known to produce symptoms, statistical model failure occurs. Using a metric related to the instantaneous risk, we successfully identify regions where model failure will not occur and identify the boundaries of the region using a root bounding technique. Several models are used to demonstrate the techniques, which may be employed to reduce the difficulty of model optimization for future investigations.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The development of atherosclerosis in the aorta is associated with low and oscillatory wall shear stress for normal patients. Moreover, localized differences in wall shear stress heterogeneity have been correlated with the presence of complex plaques in the descending aorta. While it is known that coarctation of the aorta can influence indices of wall shear stress, it is unclear how the degree of narrowing influences resulting patterns. We hypothesized that the degree of coarctation would have a strong influence on focal heterogeneity of wall shear stress. To test this hypothesis, we modeled the fluid dynamics in a patient-specific aorta with varied degrees of coarctation. We first validated a massively parallel computational model against experimental results for the patient geometry and then evaluated local shear stress patterns for a range of degrees of coarctation. Wall shear stress patterns at two cross sectional slices prone to develop atherosclerotic plaques were evaluated. Levels at different focal regions were compared to the conventional measure of average circumferential shear stress to enable localized quantification of coarctation-induced shear stress alteration. We find that the coarctation degree causes highly heterogeneous changes in wall shear stress.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this study, we developed and improved the numerical mode matching (NMM) method which has previously been shown to be a fast and robust semi-analytical solver to investigate the propagation of electromagnetic (EM) waves in an isotropic layered medium. The applicable models, such as cylindrical waveguide, optical fiber, and borehole with earth geological formation, are generally modeled as an axisymmetric structure which is an orthogonal-plano-cylindrically layered (OPCL) medium consisting of materials stratified planarly and layered concentrically in the orthogonal directions.

In this report, several important improvements have been made to extend applications of this efficient solver to the anisotropic OCPL medium. The formulas for anisotropic media with three different diagonal elements in the cylindrical coordinate system are deduced to expand its application to more general materials. The perfectly matched layer (PML) is incorporated along the radial direction as an absorbing boundary condition (ABC) to make the NMM method more accurate and efficient for wave diffusion problems in unbounded media and applicable to scattering problems with lossless media. We manipulate the weak form of Maxwell's equations and impose the correct boundary conditions at the cylindrical axis to solve the singularity problem which is ignored by all previous researchers. The spectral element method (SEM) is introduced to more efficiently compute the eigenmodes of higher accuracy with less unknowns, achieving a faster mode matching procedure between different horizontal layers. We also prove the relationship of the field between opposite mode indices for different types of excitations, which can reduce the computational time by half. The formulas for computing EM fields excited by an electric or magnetic dipole located at any position with an arbitrary orientation are deduced. And the excitation are generalized to line and surface current sources which can extend the application of NMM to the simulations of controlled source electromagnetic techniques. Numerical simulations have demonstrated the efficiency and accuracy of this method.

Finally, the improved numerical mode matching (NMM) method is introduced to efficiently compute the electromagnetic response of the induction tool from orthogonal transverse hydraulic fractures in open or cased boreholes in hydrocarbon exploration. The hydraulic fracture is modeled as a slim circular disk which is symmetric with respect to the borehole axis and filled with electrically conductive or magnetic proppant. The NMM solver is first validated by comparing the normalized secondary field with experimental measurements and a commercial software. Then we analyze quantitatively the induction response sensitivity of the fracture with different parameters, such as length, conductivity and permeability of the filled proppant, to evaluate the effectiveness of the induction logging tool for fracture detection and mapping. Casings with different thicknesses, conductivities and permeabilities are modeled together with the fractures in boreholes to investigate their effects for fracture detection. It reveals that the normalized secondary field will not be weakened at low frequencies, ensuring the induction tool is still applicable for fracture detection, though the attenuation of electromagnetic field through the casing is significant. A hybrid approach combining the NMM method and BCGS-FFT solver based integral equation has been proposed to efficiently simulate the open or cased borehole with tilted fractures which is a non-axisymmetric model.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A class of multi-process models is developed for collections of time indexed count data. Autocorrelation in counts is achieved with dynamic models for the natural parameter of the binomial distribution. In addition to modeling binomial time series, the framework includes dynamic models for multinomial and Poisson time series. Markov chain Monte Carlo (MCMC) and Po ́lya-Gamma data augmentation (Polson et al., 2013) are critical for fitting multi-process models of counts. To facilitate computation when the counts are high, a Gaussian approximation to the P ́olya- Gamma random variable is developed.

Three applied analyses are presented to explore the utility and versatility of the framework. The first analysis develops a model for complex dynamic behavior of themes in collections of text documents. Documents are modeled as a “bag of words”, and the multinomial distribution is used to characterize uncertainty in the vocabulary terms appearing in each document. State-space models for the natural parameters of the multinomial distribution induce autocorrelation in themes and their proportional representation in the corpus over time.

The second analysis develops a dynamic mixed membership model for Poisson counts. The model is applied to a collection of time series which record neuron level firing patterns in rhesus monkeys. The monkey is exposed to two sounds simultaneously, and Gaussian processes are used to smoothly model the time-varying rate at which the neuron’s firing pattern fluctuates between features associated with each sound in isolation.

The third analysis presents a switching dynamic generalized linear model for the time-varying home run totals of professional baseball players. The model endows each player with an age specific latent natural ability class and a performance enhancing drug (PED) use indicator. As players age, they randomly transition through a sequence of ability classes in a manner consistent with traditional aging patterns. When the performance of the player significantly deviates from the expected aging pattern, he is identified as a player whose performance is consistent with PED use.

All three models provide a mechanism for sharing information across related series locally in time. The models are fit with variations on the P ́olya-Gamma Gibbs sampler, MCMC convergence diagnostics are developed, and reproducible inference is emphasized throughout the dissertation.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

OBJECTIVE: To demonstrate the application of causal inference methods to observational data in the obstetrics and gynecology field, particularly causal modeling and semi-parametric estimation. BACKGROUND: Human immunodeficiency virus (HIV)-positive women are at increased risk for cervical cancer and its treatable precursors. Determining whether potential risk factors such as hormonal contraception are true causes is critical for informing public health strategies as longevity increases among HIV-positive women in developing countries. METHODS: We developed a causal model of the factors related to combined oral contraceptive (COC) use and cervical intraepithelial neoplasia 2 or greater (CIN2+) and modified the model to fit the observed data, drawn from women in a cervical cancer screening program at HIV clinics in Kenya. Assumptions required for substantiation of a causal relationship were assessed. We estimated the population-level association using semi-parametric methods: g-computation, inverse probability of treatment weighting, and targeted maximum likelihood estimation. RESULTS: We identified 2 plausible causal paths from COC use to CIN2+: via HPV infection and via increased disease progression. Study data enabled estimation of the latter only with strong assumptions of no unmeasured confounding. Of 2,519 women under 50 screened per protocol, 219 (8.7%) were diagnosed with CIN2+. Marginal modeling suggested a 2.9% (95% confidence interval 0.1%, 6.9%) increase in prevalence of CIN2+ if all women under 50 were exposed to COC; the significance of this association was sensitive to method of estimation and exposure misclassification. CONCLUSION: Use of causal modeling enabled clear representation of the causal relationship of interest and the assumptions required to estimate that relationship from the observed data. Semi-parametric estimation methods provided flexibility and reduced reliance on correct model form. Although selected results suggest an increased prevalence of CIN2+ associated with COC, evidence is insufficient to conclude causality. Priority areas for future studies to better satisfy causal criteria are identified.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Fitting statistical models is computationally challenging when the sample size or the dimension of the dataset is huge. An attractive approach for down-scaling the problem size is to first partition the dataset into subsets and then fit using distributed algorithms. The dataset can be partitioned either horizontally (in the sample space) or vertically (in the feature space), and the challenge arise in defining an algorithm with low communication, theoretical guarantees and excellent practical performance in general settings. For sample space partitioning, I propose a MEdian Selection Subset AGgregation Estimator ({\em message}) algorithm for solving these issues. The algorithm applies feature selection in parallel for each subset using regularized regression or Bayesian variable selection method, calculates the `median' feature inclusion index, estimates coefficients for the selected features in parallel for each subset, and then averages these estimates. The algorithm is simple, involves very minimal communication, scales efficiently in sample size, and has theoretical guarantees. I provide extensive experiments to show excellent performance in feature selection, estimation, prediction, and computation time relative to usual competitors.

While sample space partitioning is useful in handling datasets with large sample size, feature space partitioning is more effective when the data dimension is high. Existing methods for partitioning features, however, are either vulnerable to high correlations or inefficient in reducing the model dimension. In the thesis, I propose a new embarrassingly parallel framework named {\em DECO} for distributed variable selection and parameter estimation. In {\em DECO}, variables are first partitioned and allocated to m distributed workers. The decorrelated subset data within each worker are then fitted via any algorithm designed for high-dimensional problems. We show that by incorporating the decorrelation step, DECO can achieve consistent variable selection and parameter estimation on each subset with (almost) no assumptions. In addition, the convergence rate is nearly minimax optimal for both sparse and weakly sparse models and does NOT depend on the partition number m. Extensive numerical experiments are provided to illustrate the performance of the new framework.

For datasets with both large sample sizes and high dimensionality, I propose a new "divided-and-conquer" framework {\em DEME} (DECO-message) by leveraging both the {\em DECO} and the {\em message} algorithm. The new framework first partitions the dataset in the sample space into row cubes using {\em message} and then partition the feature space of the cubes using {\em DECO}. This procedure is equivalent to partitioning the original data matrix into multiple small blocks, each with a feasible size that can be stored and fitted in a computer in parallel. The results are then synthezied via the {\em DECO} and {\em message} algorithm in a reverse order to produce the final output. The whole framework is extremely scalable.