942 resultados para Parallel programming model
Resumo:
The intestinal protozoan parasite Giardia lamblia causes diarrhoea in humans and animals. In the present study, we used the C57BL/6 inbred mouse model to assess the impact of a nematode (Trichinella spiralis) infection on the course of a G. lamblia (clone GS/M-83-H7) infection. Acute trichinellosis coincided with transient intestinal inflammation and generated an intestinal environment that strongly promoted growth of G. lamblia trophozoites although the local anti-Giardia immunoglobulin (Ig) A production was not affected. This increased G. lamblia infection intensity correlated with intestinal mast cell infiltration, mast cell degranulation, and total IgE production. Furthermore, a G. lamblia single-infection investigated in parallel also resulted in intestinal mast cell accumulation but severe infiltration was triggered in the absence of IgE. Recently, intestinal mast cells emerging during a G. lamblia infection were reported to be involved in those immunological mechanisms that control intestinal proliferation of the parasite in mice. This anti-giardial activity was assumed to be related to the capacity of mast cells to produce IL-6. However, this previous assumption was questioned by our present immunohistological findings indicating that murine intestinal mast cells, activated during a G. lamblia infection were IL-6-negative. In the present co-infection experiments, mast cells induced during acute trichinellosis were not able to control a concurrent G. lamblia infection. This observation makes it feasible that the T. spiralis infection created an immunological and physiological environment that superimposed the anti-giardial effect of mast cells and thus favoured intestinal growth of G. lamblia trophozoites in double-infected mice. Furthermore, our findings raise the possibility that intestinal inflammation e.g. as a consequence of a 'pre-existing' nematode infection is a factor which contributes to increased susceptibility of a host to a G. lamblia infection. The phenomenon of a 'pre-existing' nematode infection prior to a G. lamblia infection is a frequent constellation in endemic areas of giardiasis and may therefore have a direct impact on the epidemiological situation of the disease.
Resumo:
BACKGROUND: Drugs are routinely combined in anesthesia and pain management to obtain an enhancement of the desired effects. However, a parallel enhancement of the undesired effects might take place as well, resulting in a limited therapeutic usefulness. Therefore, when addressing the question of optimal drug combinations, side effects must be taken into account. METHODS: By extension of a previously published interaction model, the authors propose a method to study drug interactions considering also their side effects. A general outcome parameter identified as patient's well-being is defined by superposition of positive and negative effects. Well-being response surfaces are computed and analyzed for varying drugs pharmacodynamics and interaction types. In particular, the existence of multiple maxima and of optimal drug combinations is investigated for the combination of two drugs. RESULTS: Both drug pharmacodynamics and interaction type affect the well-being surface and the deriving optimal combinations. The effect of the interaction parameters can be explained in terms of synergy and antagonism and remains unchanged for varying pharmacodynamics. For all simulations performed for the combination of two drugs, the presence of more than one maximum was never observed. CONCLUSIONS: The model is consistent with clinical knowledge and supports previously published experimental results on optimal drug combinations. This new framework improves understanding of the characteristics of drug combinations used in clinical practice and can be used in clinical research to identify optimal drug dosing.
Resumo:
Objectives: The goal of the present study was to elucidate the contribution of the newly recognized virulence factor choline to the pathogenesis of Streptococcus pneumoniae in an animal model of meningitis. Results: The choline containing strain D39Cho(-) and its isogenic choline-free derivative D39Cho(-)licA64 -each expressing the capsule polysaccharide 2 - were introduced intracisternally at an inoculum size of 10(3) CFU into 11 days old Wistar rats. During the first 8 h post infection both strains multiplied and stimulated a similar immune response that involved expression of high levels of proinflammatory cytokines, the matrix metalloproteinase 9 (MMP-9), IL-10, and the influx of white blood cells into the CSF. Virtually identical immune response was also elicited by intracisternal inoculation of 10(7) CFU equivalents of either choline-containing or choline-free cell walls. At sampling times past 8 h strain D39Cho(-) continued to replicate accompanied by an intense inflammatory response and strong granulocytic pleiocytosis. Animals infected with D39Cho(-) died within 20 h and histopathology revealed brain damage in the cerebral cortex and hippocampus. In contrast, the initial immune response generated by the choline-free strain D39Cho(-)licA64 began to decline after the first 8 h accompanied by elimination of the bacteria from the CSF in parallel with a strong WBC response peaking at 8 h after infection. All animals survived and there was no evidence for brain damage. Conclusion: Choline in the cell wall is essential for pneumococci to remain highly virulent and survive within the host and establish pneumococcal meningitis.
Resumo:
An important problem in computational biology is finding the longest common subsequence (LCS) of two nucleotide sequences. This paper examines the correctness and performance of a recently proposed parallel LCS algorithm that uses successor tables and pruning rules to construct a list of sets from which an LCS can be easily reconstructed. Counterexamples are given for two pruning rules that were given with the original algorithm. Because of these errors, performance measurements originally reported cannot be validated. The work presented here shows that speedup can be reliably achieved by an implementation in Unified Parallel C that runs on an Infiniband cluster. This performance is partly facilitated by exploiting the software cache of the MuPC runtime system. In addition, this implementation achieved speedup without bulk memory copy operations and the associated programming complexity of message passing.
Resumo:
Linear programs, or LPs, are often used in optimization problems, such as improving manufacturing efficiency of maximizing the yield from limited resources. The most common method for solving LPs is the Simplex Method, which will yield a solution, if one exists, but over the real numbers. From a purely numerical standpoint, it will be an optimal solution, but quite often we desire an optimal integer solution. A linear program in which the variables are also constrained to be integers is called an integer linear program or ILP. It is the focus of this report to present a parallel algorithm for solving ILPs. We discuss a serial algorithm using a breadth-first branch-and-bound search to check the feasible solution space, and then extend it into a parallel algorithm using a client-server model. In the parallel mode, the search may not be truly breadth-first, depending on the solution time for each node in the solution tree. Our search takes advantage of pruning, often resulting in super-linear improvements in solution time. Finally, we present results from sample ILPs, describe a few modifications to enhance the algorithm and improve solution time, and offer suggestions for future work.
Resumo:
As an important Civil Engineering material, asphalt concrete (AC) is commonly used to build road surfaces, airports, and parking lots. With traditional laboratory tests and theoretical equations, it is a challenge to fully understand such a random composite material. Based on the discrete element method (DEM), this research seeks to develop and implement computer models as research approaches for improving understandings of AC microstructure-based mechanics. In this research, three categories of approaches were developed or employed to simulate microstructures of AC materials, namely the randomly-generated models, the idealized models, and image-based models. The image-based models were recommended for accurately predicting AC performance, while the other models were recommended as research tools to obtain deep insight into the AC microstructure-based mechanics. A viscoelastic micromechanical model was developed to capture viscoelastic interactions within the AC microstructure. Four types of constitutive models were built to address the four categories of interactions within an AC specimen. Each of the constitutive models consists of three parts which represent three different interaction behaviors: a stiffness model (force-displace relation), a bonding model (shear and tensile strengths), and a slip model (frictional property). Three techniques were developed to reduce the computational time for AC viscoelastic simulations. It was found that the computational time was significantly reduced to days or hours from years or months for typical three-dimensional models. Dynamic modulus and creep stiffness tests were simulated and methodologies were developed to determine the viscoelastic parameters. It was found that the DE models could successfully predict dynamic modulus, phase angles, and creep stiffness in a wide range of frequencies, temperatures, and time spans. Mineral aggregate morphology characteristics (sphericity, orientation, and angularity) were studied to investigate their impacts on AC creep stiffness. It was found that aggregate characteristics significantly impact creep stiffness. Pavement responses and pavement-vehicle interactions were investigated by simulating pavement sections under a rolling wheel. It was found that wheel acceleration, steadily moving, and deceleration significantly impact contact forces. Additionally, summary and recommendations were provided in the last chapter and part of computer programming codes wree provided in the appendixes.
Resumo:
A range of societal issues have been caused by fossil fuel consumption in the transportation sector in the United States (U.S.), including health related air pollution, climate change, the dependence on imported oil, and other oil related national security concerns. Biofuels production from various lignocellulosic biomass types such as wood, forest residues, and agriculture residues have the potential to replace a substantial portion of the total fossil fuel consumption. This research focuses on locating biofuel facilities and designing the biofuel supply chain to minimize the overall cost. For this purpose an integrated methodology was proposed by combining the GIS technology with simulation and optimization modeling methods. The GIS based methodology was used as a precursor for selecting biofuel facility locations by employing a series of decision factors. The resulted candidate sites for biofuel production served as inputs for simulation and optimization modeling. As a precursor to simulation or optimization modeling, the GIS-based methodology was used to preselect potential biofuel facility locations for biofuel production from forest biomass. Candidate locations were selected based on a set of evaluation criteria, including: county boundaries, a railroad transportation network, a state/federal road transportation network, water body (rivers, lakes, etc.) dispersion, city and village dispersion, a population census, biomass production, and no co-location with co-fired power plants. The simulation and optimization models were built around key supply activities including biomass harvesting/forwarding, transportation and storage. The built onsite storage served for spring breakup period where road restrictions were in place and truck transportation on certain roads was limited. Both models were evaluated using multiple performance indicators, including cost (consisting of the delivered feedstock cost, and inventory holding cost), energy consumption, and GHG emissions. The impact of energy consumption and GHG emissions were expressed in monetary terms to keep consistent with cost. Compared with the optimization model, the simulation model represents a more dynamic look at a 20-year operation by considering the impacts associated with building inventory at the biorefinery to address the limited availability of biomass feedstock during the spring breakup period. The number of trucks required per day was estimated and the inventory level all year around was tracked. Through the exchange of information across different procedures (harvesting, transportation, and biomass feedstock processing procedures), a smooth flow of biomass from harvesting areas to a biofuel facility was implemented. The optimization model was developed to address issues related to locating multiple biofuel facilities simultaneously. The size of the potential biofuel facility is set up with an upper bound of 50 MGY and a lower bound of 30 MGY. The optimization model is a static, Mathematical Programming Language (MPL)-based application which allows for sensitivity analysis by changing inputs to evaluate different scenarios. It was found that annual biofuel demand and biomass availability impacts the optimal results of biofuel facility locations and sizes.
DESIGN AND IMPLEMENT DYNAMIC PROGRAMMING BASED DISCRETE POWER LEVEL SMART HOME SCHEDULING USING FPGA
Resumo:
With the development and capabilities of the Smart Home system, people today are entering an era in which household appliances are no longer just controlled by people, but also operated by a Smart System. This results in a more efficient, convenient, comfortable, and environmentally friendly living environment. A critical part of the Smart Home system is Home Automation, which means that there is a Micro-Controller Unit (MCU) to control all the household appliances and schedule their operating times. This reduces electricity bills by shifting amounts of power consumption from the on-peak hour consumption to the off-peak hour consumption, in terms of different “hour price”. In this paper, we propose an algorithm for scheduling multi-user power consumption and implement it on an FPGA board, using it as the MCU. This algorithm for discrete power level tasks scheduling is based on dynamic programming, which could find a scheduling solution close to the optimal one. We chose FPGA as our system’s controller because FPGA has low complexity, parallel processing capability, a large amount of I/O interface for further development and is programmable on both software and hardware. In conclusion, it costs little time running on FPGA board and the solution obtained is good enough for the consumers.
Resumo:
Software must be constantly adapted to changing requirements. The time scale, abstraction level and granularity of adaptations may vary from short-term, fine-grained adaptation to long-term, coarse-grained evolution. Fine-grained, dynamic and context-dependent adaptations can be particularly difficult to realize in long-lived, large-scale software systems. We argue that, in order to effectively and efficiently deploy such changes, adaptive applications must be built on an infrastructure that is not just model-driven, but is both model-centric and context-aware. Specifically, this means that high-level, causally-connected models of the application and the software infrastructure itself should be available at run-time, and that changes may need to be scoped to the run-time execution context. We first review the dimensions of software adaptation and evolution, and then we show how model-centric design can address the adaptation needs of a variety of applications that span these dimensions. We demonstrate through concrete examples how model-centric and context-aware designs work at the level of application interface, programming language and runtime. We then propose a research agenda for a model-centric development environment that supports dynamic software adaptation and evolution.
Resumo:
One of the main roles of the Neural Open Markup Language, NeuroML, is to facilitate cooperation in building, simulating, testing and publishing models of channels, neurons and networks of neurons. MorphML, which was developed as a common format for exchange of neural morphology data, is distributed as part of NeuroML but can be used as a stand-alone application. In this collection of tutorials and workshop summary, we provide an overview of these XML schemas and provide examples of their use in down-stream applications. We also summarize plans for the further development of XML specifications for modeling channels, channel distributions, and network connectivity.
Resumo:
The three-step test is central to the regulation of copyright limitations at the international level. Delineating the room for exemptions with abstract criteria, the three-step test is by far the most important and comprehensive basis for the introduction of national use privileges. It is an essential, flexible element in the international limitation infrastructure that allows national law makers to satisfy domestic social, cultural, and economic needs. Given the universal field of application that follows from the test’s open-ended wording, the provision creates much more breathing space than the more specific exceptions recognized in international copyright law. EC copyright legislation, however, fails to take advantage of the flexibility inherent in the three-step test. Instead of using the international provision as a means to open up the closed EC catalogue of permissible exceptions, offer sufficient breathing space for social, cultural, and economic needs, and enable EC copyright law to keep pace with the rapid development of the Internet, the Copyright Directive 2001/29/EC encourages the application of the three-step test to further restrict statutory exceptions that are often defined narrowly in national legislation anyway. In the current online environment, however, enhanced flexibility in the field of copyright limitations is indispensable. From a social and cultural perspective, the web 2.0 promotes and enhances freedom of expression and information with its advanced search engine services, interactive platforms, and various forms of user-generated content. From an economic perspective, it creates a parallel universe of traditional content providers relying on copyright protection, and emerging Internet industries whose further development depends on robust copyright limita- tions. In particular, the newcomers in the online market – social networking sites, video forums, and virtual worlds – promise a remarkable potential for economic growth that has already attracted the attention of the OECD. Against this background, the time is ripe to debate the introduction of an EC fair use doctrine on the basis of the three-step test. Otherwise, EC copyright law is likely to frustrate important opportunities for cultural, social, and economic development. To lay groundwork for the debate, the differences between the continental European and the Anglo-American approach to copyright limitations (section 1), and the specific merits of these two distinct approaches (section 2), will be discussed first. An analysis of current problems that have arisen under the present dysfunctional EC system (section 3) will then serve as a starting point for proposing an EC fair use doctrine based on the three-step test (section 4). Drawing conclusions, the international dimension of this fair use proposal will be considered (section 5).
Resumo:
27-Channel EEG potential map series were recorded from 12 normals with closed and open eyes. Intracerebral dipole model source locations in the frequency domain were computed. Eye opening (visual input) caused centralization (convergence and elevation) of the source locations of the seven frequency bands, indicative of generalized activity; especially, there was clear anteriorization of α-2 (10.5–12 Hz) and β-2 (18.5–21 Hz) sources (α-2 also to the left). Complexity of the map series' trajectories in state space (assessed by Global Dimensional Complexity and Global OMEGA Complexity) increased significantly with eye opening, indicative of more independent, parallel, active processes. Contrary to PET and fMRI, these results suggest that brain activity is more distributed and independent during visual input than after eye closing (when it is more localized and more posterior).
Resumo:
Insults during the fetal period predispose the offspring to systemic cardiovascular disease, but little is known about the pulmonary circulation and the underlying mechanisms. Maternal undernutrition during pregnancy may represent a model to investigate underlying mechanisms, because it is associated with systemic vascular dysfunction in the offspring in animals and humans. In rats, restrictive diet during pregnancy (RDP) increases oxidative stress in the placenta. Oxygen species are known to induce epigenetic alterations and may cross the placental barrier. We hypothesized that RDP in mice induces pulmonary vascular dysfunction in the offspring that is related to an epigenetic mechanism. To test this hypothesis, we assessed pulmonary vascular function and lung DNA methylation in offspring of RDP and in control mice at the end of a 2-wk exposure to hypoxia. We found that endothelium-dependent pulmonary artery vasodilation in vitro was impaired and hypoxia-induced pulmonary hypertension and right ventricular hypertrophy in vivo were exaggerated in offspring of RDP. This pulmonary vascular dysfunction was associated with altered lung DNA methylation. Administration of the histone deacetylase inhibitors butyrate and trichostatin A to offspring of RDP normalized pulmonary DNA methylation and vascular function. Finally, administration of the nitroxide Tempol to the mother during RDP prevented vascular dysfunction and dysmethylation in the offspring. These findings demonstrate that in mice undernutrition during gestation induces pulmonary vascular dysfunction in the offspring by an epigenetic mechanism. A similar mechanism may be involved in the fetal programming of vascular dysfunction in humans.
Resumo:
The Earth's bow shock is very efficient in accelerating ions out of the incident solar wind distribution to high energies (≈ 200 keV/e). Fluxes of energetic ions accelerated at the quasi-parallel bow shock, also known as diffuse ions, are best represented by exponential spectra in energy/charge, which require additional assumptions to be incorporated into these model spectra. One of these assumptions is a so-called "free escape boundary" along the interplanetary magnetic field into the upstream direction. Locations along the IBEX orbit are ideally suited for in situ measurements to investigate the existence of an upstream free escape boundary for bow shock accelerated ions. In this study we use 2 years of ion measurements from the background monitor on the IBEX spacecraft, supported by ACE solar wind observations. The IBEX Background Monitor is sensitive to protons > 14 keV, which includes the energy of the maximum flux for diffuse ions. With increasing distance from the bow shock along the interplanetary magnetic field, the count rates for diffuse ions stay constant for ions streaming away from the bow shock, while count rates for diffuse ions streaming toward the shock gradually decrease from a maximum value to ~1/e at distances of about 10 RE to 14 RE. These observations of a gradual decrease support the transition to a free escape continuum for ions of energy >14 keV at distances from 10 RE to 14 RE from the bow shock.
Resumo:
PURPOSE Segmentation of the proximal femur in digital antero-posterior (AP) pelvic radiographs is required to create a three-dimensional model of the hip joint for use in planning and treatment. However, manually extracting the femoral contour is tedious and prone to subjective bias, while automatic segmentation must accommodate poor image quality, anatomical structure overlap, and femur deformity. A new method was developed for femur segmentation in AP pelvic radiographs. METHODS Using manual annotations on 100 AP pelvic radiographs, a statistical shape model (SSM) and a statistical appearance model (SAM) of the femur contour were constructed. The SSM and SAM were used to segment new AP pelvic radiographs with a three-stage approach. At initialization, the mean SSM model is coarsely registered to the femur in the AP radiograph through a scaled rigid registration. Mahalanobis distance defined on the SAM is employed as the search criteria for each annotated suggested landmark location. Dynamic programming was used to eliminate ambiguities. After all landmarks are assigned, a regularized non-rigid registration method deforms the current mean shape of SSM to produce a new segmentation of proximal femur. The second and third stages are iteratively executed to convergence. RESULTS A set of 100 clinical AP pelvic radiographs (not used for training) were evaluated. The mean segmentation error was [Formula: see text], requiring [Formula: see text] s per case when implemented with Matlab. The influence of the initialization on segmentation results was tested by six clinicians, demonstrating no significance difference. CONCLUSIONS A fast, robust and accurate method for femur segmentation in digital AP pelvic radiographs was developed by combining SSM and SAM with dynamic programming. This method can be extended to segmentation of other bony structures such as the pelvis.