887 resultados para Specific theories and interaction models


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Immunoglobulin E (IgE) and mast cells are believed to play important roles in allergic inflammation. However, their contributions to the pathogenesis of human asthma have not been clearly established. Significant progress has been made recently in our understanding of airway inflammation and airway hyperresponsiveness through studies of murine models of asthma and genetically engineered mice. Some of the studies have provided significant insights into the role of IgE and mast cells in the allergic airway response. In these models mice are immunized systemically with soluble protein antigens and then receive an antigen challenge through the airways. Bronchoalveolar lavage fluid from mice with allergic airway inflammation contains significant amounts of IgE. The IgE can capture the antigen presented to the airways and the immune complexes so formed can augment allergic airway response in a high-affinity IgE receptor (FcepsilonRI)-dependent manner. Previously, there were conflicting reports regarding the role of mast cells in murine models of asthma, based on studies of mast cell-deficient mice. More recent studies have suggested that the extent to which mast cells contribute to murine models of asthma depends on the experimental conditions employed to generate the airway response. This conclusion was further supported by studies using FcepsilonRI-deficient mice. Therefore, IgE-dependent activation of mast cells plays an important role in the development of allergic airway inflammation and airway hyperresponsiveness in mice under specific conditions. The murine models used should be of value for testing inhibitors of IgE or mast cells for the development of therapeutic agents for human asthma.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Human plasma kallikrein, a serine proteinase, plays a key role in intrinsic blood clotting, in the kallikrein-kinin system, and in fibrinolysis. The proteolytic enzymes involved in these processes are usually controlled by specific inhibitors and may be influenced by several factors including glycosaminoglycans, as recently demonstrated by our group. The aim of the present study was to investigate the effect of glycosaminoglycans (30 to 250 µg/ml) on kallikrein activity on plasminogen and factor XII and on the inhibition of kallikrein by the plasma proteins C1-inhibitor and antithrombin. Almost all available glycosaminoglycans (heparin, heparan sulfate, bovine and tuna dermatan sulfate, chondroitin 4- and 6-sulfates) reduced (1.2 to 3.0 times) the catalytic efficiency of kallikrein (in a nanomolar range) on the hydrolysis of plasminogen (0.3 to 1.8 µM) and increased (1.9 to 7.7 times) the enzyme efficiency in factor XII (0.1 to 10 µM) activation. On the other hand, heparin, heparan sulfate, and bovine and tuna dermatan sulfate improved (1.2 to 3.4 times) kallikrein inhibition by antithrombin (1.4 µM), while chondroitin 4- and 6-sulfates reduced it (1.3 times). Heparin and heparan sulfate increased (1.4 times) the enzyme inhibition by the C1-inhibitor (150 nM).

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Meandering rivers have been perceived to evolve rather similarly around the world independently of the location or size of the river. Despite the many consistent processes and characteristics they have also been noted to show complex and unique sets of fluviomorphological processes in which local factors play important role. These complex interactions of flow and morphology affect notably the development of the river. Comprehensive and fundamental field, flume and theoretically based studies of fluviomorphological processes in meandering rivers have been carried out especially during the latter part of the 20th century. However, as these studies have been carried out with traditional field measurements techniques their spatial and temporal resolution is not competitive to the level achievable today. The hypothesis of this study is that, by exploiting e increased spatial and temporal resolution of the data, achieved by combining conventional field measurements with a range of modern technologies, will provide new insights to the spatial patterns of the flow-sediment interaction in meandering streams, which have perceived to show notable variation in space and time. This thesis shows how the modern technologies can be combined to derive very high spatial and temporal resolution data on fluvio-morphological processes over meander bends. The flow structure over the bends is recorded in situ using acoustic Doppler current profiler (ADCP) and the spatial and temporal resolution of the flow data is enhanced using 2D and 3D CFD over various meander bends. The CFD are also exploited to simulate sediment transport. Multi-temporal terrestrial laser scanning (TLS), mobile laser scanning (MLS) and echo sounding data are used to measure the flow-based changes and formations over meander bends and to build the computational models. The spatial patterns of erosion and deposition over meander bends are analysed relative to the measured and modelled flow field and sediment transport. The results are compared with the classic theories of the processes in meander bends. Mainly, the results of this study follow well the existing theories and results of previous studies. However, some new insights regarding to the spatial and temporal patterns of the flow-sediment interaction in a natural sand-bed meander bend are provided. The results of this study show the advantages of the rapid and detailed measurements techniques and the achieved spatial and temporal resolution provided by CFD, unachievable with field measurements. The thesis also discusses the limitations which remain in the measurement and modelling methods and in understanding of fluvial geomorphology of meander bends. Further, the hydro- and morphodynamic models’ sensitivity to user-defined parameters is tested, and the modelling results are assessed against detailed field measurement. The study is implemented in the meandering sub-Arctic Pulmanki River in Finland. The river is unregulated and sand-bed and major morphological changes occur annually on the meander point bars, which are inundated only during the snow-melt-induced spring floods. The outcome of this study applies to sandbed meandering rivers in regions where normally one significant flood event occurs annually, such as Arctic areas with snow-melt induced spring floods, and where the point bars of the meander bends are inundated only during the flood events.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The investments have always been considered as an essential backbone and so-called ‘locomotive’ for the competitive economies. However, in various countries, the state has been put under tight budget constraints for the investments in capital intensive projects. In response to this situation, the cooperation between public and private sector has grown based on public-private mechanism. The promotion of favorable arrangement for collaboration between public and private sectors for the provision of policies, services, and infrastructure in Russia can help to address the problems of dry ports development that neither municipalities nor the private sector can solve alone. Especially, the stimulation of public-private collaboration is significant under the exposure to externalities that affect the magnitude of the risks during all phases of project realization. In these circumstances, the risk in the projects also is becoming increasingly a part of joint research and risk management practice, which is viewed as a key approach, aiming to take active actions on existing global and specific factors of uncertainties. Meanwhile, a relatively little progress has been made on the inclusion of the resilience aspects into the planning process of a dry ports construction that would instruct the capacity planner, on how to mitigate the occurrence of disruptions that may lead to million dollars of losses due to the deviation of the future cash flows from the expected financial flows on the project. The current experience shows that the existing methodological base is developed fragmentary within separate steps of supply chain risk management (SCRM) processes: risk identification, risk evaluation, risk mitigation, risk monitoring and control phases. The lack of the systematic approach hinders the solution of the problem of risk management processes of dry port implementation. Therefore, management of various risks during the investments phases of dry port projects still presents a considerable challenge from the practical and theoretical points of view. In this regard, the given research became a logical continuation of fundamental research, existing in the financial models and theories (e.g., capital asset pricing model and real option theory), as well as provided a complementation for the portfolio theory. The goal of the current study is in the design of methods and models for the facilitation of dry port implementation through the mechanism of public-private partnership on the national market that implies the necessity to mitigate, first and foremost, the shortage of the investments and consequences of risks. The problem of the research was formulated on the ground of the identified contradictions. They rose as a continuation of the trade-off between the opportunities that the investors can gain from the development of terminal business in Russia (i.e. dry port implementation) and risks. As a rule, the higher the investment risk, the greater should be their expected return. However, investors have a different tolerance for the risks. That is why it would be advisable to find an optimum investment. In the given study, the optimum relates to the search for the efficient portfolio, which can provide satisfaction to the investor, depending on its degree of risk aversion. There are many theories and methods in finance, concerning investment choices. Nevertheless, the appropriateness and effectiveness of particular methods should be considered with the allowance of the specifics of the investment projects. For example, the investments in dry ports imply not only the lump sum of financial inflows, but also the long-term payback periods. As a result, capital intensity and longevity of their construction determine the necessity from investors to ensure the return on investment (profitability), along with the rapid return on investment (liquidity), without precluding the fact that the stochastic nature of the project environment is hardly described by the formula-based approach. The current theoretical base for the economic appraisals of the dry port projects more often perceives net present value (NPV) as a technique superior to other decision-making criteria. For example, the portfolio theory, which considers different risk preference of an investor and structures of utility, defines net present value as a better criterion of project appraisal than discounted payback period (DPP). Meanwhile, in business practice, the DPP is more popular. Knowing that the NPV is based on the assumptions of certainty of project life, it cannot be an accurate appraisal approach alone to determine whether or not the project should be accepted for the approval in the environment that is not without of uncertainties. In order to reflect the period or the project’s useful life that is exposed to risks due to changes in political, operational, and financial factors, the second capital budgeting criterion – discounted payback period is profoundly important, particularly for the Russian environment. Those statements represent contradictions that exist in the theory and practice of the applied science. Therefore, it would be desirable to relax the assumptions of portfolio theory and regard DPP as not fewer relevant appraisal approach for the assessment of the investment and risk measure. At the same time, the rationality of the use of both project performance criteria depends on the methods and models, with the help of which these appraisal approaches are calculated in feasibility studies. The deterministic methods cannot ensure the required precision of the results, while the stochastic models guarantee the sufficient level of the accuracy and reliability of the obtained results, providing that the risks are properly identified, evaluated, and mitigated. Otherwise, the project performance indicators may not be confirmed during the phase of project realization. For instance, the economic and political instability can result in the undoing of hard-earned gains, leading to the need for the attraction of the additional finances for the project. The sources of the alternative investments, as well as supportive mitigation strategies, can be studied during the initial phases of project development. During this period, the effectiveness of the investments undertakings can also be improved by the inclusion of the various investors, e.g. Russian Railways’ enterprises and other private companies in the dry port projects. However, the evaluation of the effectiveness of the participation of different investors in the project lack the methods and models that would permit doing the particular feasibility study, foreseeing the quantitative characteristics of risks and their mitigation strategies, which can meet the tolerance of the investors to the risks. For this reason, the research proposes a combination of Monte Carlo method, discounted cash flow technique, the theory of real options, and portfolio theory via a system dynamics simulation approach. The use of this methodology allows for comprehensive risk management process of dry port development to cover all aspects of risk identification, risk evaluation, risk mitigation, risk monitoring, and control phases. A designed system dynamics model can be recommended for the decision-makers on the dry port projects that are financed via a public-private partnership. It permits investors to make a decision appraisal based on random variables of net present value and discounted payback period, depending on different risks factors, e.g. revenue risks, land acquisition risks, traffic volume risks, construction hazards, and political risks. In this case, the statistical mean is used for the explication of the expected value of the DPP and NPV; the standard deviation is proposed as a characteristic of risks, while the elasticity coefficient is applied for rating of risks. Additionally, the risk of failure of project investments and guaranteed recoupment of capital investment can be considered with the help of the model. On the whole, the application of these modern methods of simulation creates preconditions for the controlling of the process of dry port development, i.e. making managerial changes and identifying the most stable parameters that contribute to the optimal alternative scenarios of the project realization in the uncertain environment. System dynamics model allows analyzing the interactions in the most complex mechanism of risk management process of the dry ports development and making proposals for the improvement of the effectiveness of the investments via an estimation of different risk management strategies. For the comparison and ranking of these alternatives in their order of preference to the investor, the proposed indicators of the efficiency of the investments, concerning the NPV, DPP, and coefficient of variation, can be used. Thus, rational investors, who averse to taking increased risks unless they are compensated by the commensurate increase in the expected utility of a risky prospect of dry port development, can be guided by the deduced marginal utility of investments. It is computed on the ground of the results from the system dynamics model. In conclusion, the outlined theoretical and practical implications for the management of risks, which are the key characteristics of public-private partnerships, can help analysts and planning managers in budget decision-making, substantially alleviating the effect from various risks and avoiding unnecessary cost overruns in dry port projects.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Implicit theories of shyness refer to a beUef that shyness is a fixed trait versus the belief that shyness is changeable and controllable. In this study, I explored the association between overall shyness and children's implicit self-theories of shyness, as well as between implicit self-theories of shyness and children's other shyness-related beliefs (perceptions of others' theories of shyness, shyness as a perceived problem, and ideas about treatment for shyness). Forty-six 10-12- year- old children (M = 10.74, SD = .88) were interviewed individually, filled out a set of questionnaires, and completed a computer-presented task. ' "^ As was expected, in ambiguous social situations, children perceived others' theories of shyness in a way that confirmed their own theories. The hypothesized curvilinear relation between shy and implicit self-theories of shyness was not found; instead, a linear positive relationship between these two variables emerged. Although implicit self-theories of shyness were not effective in predicting either the children's views of shyness as a perceived problem or children's ideas about treatment for shyness, some interesting results were found. Specifically, children's motivation to change their shyness correlated with their views of shyness as a problem for children in general and their perceptions of others' theories of shyness. Specific agents and strategies were regarded by children as having different effectiveness in their potential to change shyness. The theoretical and practical implications of these findings were discussed. Suggestions for future research were provided.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Depuis l’introduction de la mécanique quantique, plusieurs mystères de la nature ont trouvé leurs explications. De plus en plus, les concepts de la mécanique quantique se sont entremêlés avec d’autres de la théorie de la complexité du calcul. De nouvelles idées et solutions ont été découvertes et élaborées dans le but de résoudre ces problèmes informatiques. En particulier, la mécanique quantique a secoué plusieurs preuves de sécurité de protocoles classiques. Dans ce m´emoire, nous faisons un étalage de résultats récents de l’implication de la mécanique quantique sur la complexité du calcul, et cela plus précisément dans le cas de classes avec interaction. Nous présentons ces travaux de recherches avec la nomenclature des jeux à information imparfaite avec coopération. Nous exposons les différences entre les théories classiques, quantiques et non-signalantes et les démontrons par l’exemple du jeu à cycle impair. Nous centralisons notre attention autour de deux grands thèmes : l’effet sur un jeu de l’ajout de joueurs et de la répétition parallèle. Nous observons que l’effet de ces modifications a des conséquences très différentes en fonction de la théorie physique considérée.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This thesis is concerned with the interaction between literature and abstract thought. More specifically, it studies the epistemological charge of the literary, the type of knowledge that is carried by elements proper to fictional narratives into different disciplines. By concentrating on two different theoretical methods, the creation of thought experiments and the framing of possible worlds, methods which were elaborated and are still used today in spheres as varied as modal logics, analytic philosophy and physics, and by following their reinsertion within literary theory, the research develops the theory that both thought experiments and possible worlds are in fact short narrative stories that inform knowledge through literary means. By using two novels, Abbott’s Flatland and Vonnegut’s The Sirens of Titan, that describe extra-dimensional existence in radically different ways, respectively as a phenomenologically unknowable space and as an outward perspective on time, it becomes clear that literature is constitutive of the way in which worlds, fictive, real or otherwise, are constructed and understood. Thus dimensions, established through extensional analogies as either experimental knowledge or modal possibility for a given world, generate new directions for thought, which can then take part in the inductive/deductive process of scientia. By contrasting the dimensions of narrative with the way that dimensions were historically constituted, the research also establishes that the literary opens up an infinite potential of abstract space-time domains, defined by their specific rules and limits, and that these different experimental folds are themselves partaking in a dimensional process responsible for new forms of understanding. Over against science fiction literary theories of speculation that posit an equation between the fictive and the real, this thesis examines the complex structure of many overlapping possibilities that can organise themselves around larger compossible wholes, thus offering a theory of reading that is both non-mimetic and non-causal. It consequently examines the a dynamic process whereby literature is always reconceived through possibilities actualised by reading while never defining how the reader will ultimately understand the overarching structure. In this context, the thesis argues that a causal story can be construed out of any one interaction with a given narrative—underscoring, for example, the divinatory strength of a particular vision of the future—even as this narrative represents only a fraction of the potential knowledge of any particular literary text. Ultimately, the study concludes by tracing out how novel comprehensions of the literary, framed by the material conditions of their own space and time, endlessly renew themselves through multiple interactions, generating analogies and speculations that facilitate the creation of new knowledge.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The high level of realism and interaction in many computer graphic applications requires techniques for processing complex geometric models. First, we present a method that provides an accurate low-resolution approximation from a multi-chart textured model that guarantees geometric fidelity and correct preservation of the appearance attributes. Then, we introduce a mesh structure called Compact Model that approximates dense triangular meshes while preserving sharp features, allowing adaptive reconstructions and supporting textured models. Next, we design a new space deformation technique called *Cages based on a multi-level system of cages that preserves the smoothness of the mesh between neighbouring cages and is extremely versatile, allowing the use of heterogeneous sets of coordinates and different levels of deformation. Finally, we propose a hybrid method that allows to apply any deformation technique on large models obtaining high quality results with a reduced memory footprint and a high performance.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The aim of this thesis is to narrow the gap between two different control techniques: the continuous control and the discrete event control techniques DES. This gap can be reduced by the study of Hybrid systems, and by interpreting as Hybrid systems the majority of large-scale systems. In particular, when looking deeply into a process, it is often possible to identify interaction between discrete and continuous signals. Hybrid systems are systems that have both continuous, and discrete signals. Continuous signals are generally supposed continuous and differentiable in time, since discrete signals are neither continuous nor differentiable in time due to their abrupt changes in time. Continuous signals often represent the measure of natural physical magnitudes such as temperature, pressure etc. The discrete signals are normally artificial signals, operated by human artefacts as current, voltage, light etc. Typical processes modelled as Hybrid systems are production systems, chemical process, or continuos production when time and continuous measures interacts with the transport, and stock inventory system. Complex systems as manufacturing lines are hybrid in a global sense. They can be decomposed into several subsystems, and their links. Another motivation for the study of Hybrid systems is the tools developed by other research domains. These tools benefit from the use of temporal logic for the analysis of several properties of Hybrid systems model, and use it to design systems and controllers, which satisfies physical or imposed restrictions. This thesis is focused in particular types of systems with discrete and continuous signals in interaction. That can be modelled hard non-linealities, such as hysteresis, jumps in the state, limit cycles, etc. and their possible non-deterministic future behaviour expressed by an interpretable model description. The Hybrid systems treated in this work are systems with several discrete states, always less than thirty states (it can arrive to NP hard problem), and continuous dynamics evolving with expression: with Ki ¡ Rn constant vectors or matrices for X components vector. In several states the continuous evolution can be several of them Ki = 0. In this formulation, the mathematics can express Time invariant linear system. By the use of this expression for a local part, the combination of several local linear models is possible to represent non-linear systems. And with the interaction with discrete events of the system the model can compose non-linear Hybrid systems. Especially multistage processes with high continuous dynamics are well represented by the proposed methodology. Sate vectors with more than two components, as third order models or higher is well approximated by the proposed approximation. Flexible belt transmission, chemical reactions with initial start-up and mobile robots with important friction are several physical systems, which profits from the benefits of proposed methodology (accuracy). The motivation of this thesis is to obtain a solution that can control and drive the Hybrid systems from the origin or starting point to the goal. How to obtain this solution, and which is the best solution in terms of one cost function subject to the physical restrictions and control actions is analysed. Hybrid systems that have several possible states, different ways to drive the system to the goal and different continuous control signals are problems that motivate this research. The requirements of the system on which we work is: a model that can represent the behaviour of the non-linear systems, and that possibilities the prediction of possible future behaviour for the model, in order to apply an supervisor which decides the optimal and secure action to drive the system toward the goal. Specific problems can be determined by the use of this kind of hybrid models are: - The unity of order. - Control the system along a reachable path. - Control the system in a safe path. - Optimise the cost function. - Modularity of control The proposed model solves the specified problems in the switching models problem, the initial condition calculus and the unity of the order models. Continuous and discrete phenomena are represented in Linear hybrid models, defined with defined eighth-tuple parameters to model different types of hybrid phenomena. Applying a transformation over the state vector : for LTI system we obtain from a two-dimensional SS a single parameter, alpha, which still maintains the dynamical information. Combining this parameter with the system output, a complete description of the system is obtained in a form of a graph in polar representation. Using Tagaki-Sugeno type III is a fuzzy model which include linear time invariant LTI models for each local model, the fuzzyfication of different LTI local model gives as a result a non-linear time invariant model. In our case the output and the alpha measure govern the membership function. Hybrid systems control is a huge task, the processes need to be guided from the Starting point to the desired End point, passing a through of different specific states and points in the trajectory. The system can be structured in different levels of abstraction and the control in three layers for the Hybrid systems from planning the process to produce the actions, these are the planning, the process and control layer. In this case the algorithms will be applied to robotics ¡V a domain where improvements are well accepted ¡V it is expected to find a simple repetitive processes for which the extra effort in complexity can be compensated by some cost reductions. It may be also interesting to implement some control optimisation to processes such as fuel injection, DC-DC converters etc. In order to apply the RW theory of discrete event systems on a Hybrid system, we must abstract the continuous signals and to project the events generated for these signals, to obtain new sets of observable and controllable events. Ramadge & Wonham¡¦s theory along with the TCT software give a Controllable Sublanguage of the legal language generated for a Discrete Event System (DES). Continuous abstraction transforms predicates over continuous variables into controllable or uncontrollable events, and modifies the set of uncontrollable, controllable observable and unobservable events. Continuous signals produce into the system virtual events, when this crosses the bound limits. If this event is deterministic, they can be projected. It is necessary to determine the controllability of this event, in order to assign this to the corresponding set, , controllable, uncontrollable, observable and unobservable set of events. Find optimal trajectories in order to minimise some cost function is the goal of the modelling procedure. Mathematical model for the system allows the user to apply mathematical techniques over this expression. These possibilities are, to minimise a specific cost function, to obtain optimal controllers and to approximate a specific trajectory. The combination of the Dynamic Programming with Bellman Principle of optimality, give us the procedure to solve the minimum time trajectory for Hybrid systems. The problem is greater when there exists interaction between adjacent states. In Hybrid systems the problem is to determine the partial set points to be applied at the local models. Optimal controller can be implemented in each local model in order to assure the minimisation of the local costs. The solution of this problem needs to give us the trajectory to follow the system. Trajectory marked by a set of set points to force the system to passing over them. Several ways are possible to drive the system from the Starting point Xi to the End point Xf. Different ways are interesting in: dynamic sense, minimum states, approximation at set points, etc. These ways need to be safe and viable and RchW. And only one of them must to be applied, normally the best, which minimises the proposed cost function. A Reachable Way, this means the controllable way and safe, will be evaluated in order to obtain which one minimises the cost function. Contribution of this work is a complete framework to work with the majority Hybrid systems, the procedures to model, control and supervise are defined and explained and its use is demonstrated. Also explained is the procedure to model the systems to be analysed for automatic verification. Great improvements were obtained by using this methodology in comparison to using other piecewise linear approximations. It is demonstrated in particular cases this methodology can provide best approximation. The most important contribution of this work, is the Alpha approximation for non-linear systems with high dynamics While this kind of process is not typical, but in this case the Alpha approximation is the best linear approximation to use, and give a compact representation.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Space weather effects on technological systems originate with energy carried from the Sun to the terrestrial environment by the solar wind. In this study, we present results of modeling of solar corona-heliosphere processes to predict solar wind conditions at the L1 Lagrangian point upstream of Earth. In particular we calculate performance metrics for (1) empirical, (2) hybrid empirical/physics-based, and (3) full physics-based coupled corona-heliosphere models over an 8-year period (1995–2002). L1 measurements of the radial solar wind speed are the primary basis for validation of the coronal and heliosphere models studied, though other solar wind parameters are also considered. The models are from the Center for Integrated Space-Weather Modeling (CISM) which has developed a coupled model of the whole Sun-to-Earth system, from the solar photosphere to the terrestrial thermosphere. Simple point-by-point analysis techniques, such as mean-square-error and correlation coefficients, indicate that the empirical coronal-heliosphere model currently gives the best forecast of solar wind speed at 1 AU. A more detailed analysis shows that errors in the physics-based models are predominately the result of small timing offsets to solar wind structures and that the large-scale features of the solar wind are actually well modeled. We suggest that additional “tuning” of the coupling between the coronal and heliosphere models could lead to a significant improvement of their accuracy. Furthermore, we note that the physics-based models accurately capture dynamic effects at solar wind stream interaction regions, such as magnetic field compression, flow deflection, and density buildup, which the empirical scheme cannot.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Blumeria graminis is an economically important obligate plant-pathogenic fungus, whose entire genome was recently sequenced and manually annotated using ab initio in silico predictions [7]. Employing large scale proteogenomic analysis we are now able to verify independently the existence of proteins predicted by 24% of open reading frame models. We compared the haustoria and sporulating hyphae proteomes and identified 71 proteins exclusively in haustoria, the feeding and effector-delivery organs of the pathogen. These proteins are ‘significantly smaller than the rest of the protein pool and predicted to be secreted. Most do not share any similarities with Swiss–Prot or Trembl entries nor possess any identifiable Pfam domains. We used a novel automated prediction pipeline to model the 3D structures of the proteins, identify putative ligand binding sites and predict regions of intrinsic disorder. This revealed that the protein set found exclusively in haustoria is significantly less disordered than the rest of the identified Blumeria proteins or random (and representative) protein sets generated from the yeast proteome. For most of the haustorial proteins with unknown functions no good templates could be found, from which to generate high quality models. Thus, these unknown proteins present potentially new protein folds that can be specific to the interaction of the pathogen with its host.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Climate modeling is a complex process, requiring accurate and complete metadata in order to identify, assess and use climate data stored in digital repositories. The preservation of such data is increasingly important given the development of ever-increasingly complex models to predict the effects of global climate change. The EU METAFOR project has developed a Common Information Model (CIM) to describe climate data and the models and modelling environments that produce this data. There is a wide degree of variability between different climate models and modelling groups. To accommodate this, the CIM has been designed to be highly generic and flexible, with extensibility built in. METAFOR describes the climate modelling process simply as "an activity undertaken using software on computers to produce data." This process has been described as separate UML packages (and, ultimately, XML schemas). This fairly generic structure canbe paired with more specific "controlled vocabularies" in order to restrict the range of valid CIM instances. The CIM will aid digital preservation of climate models as it will provide an accepted standard structure for the model metadata. Tools to write and manage CIM instances, and to allow convenient and powerful searches of CIM databases,. Are also under development. Community buy-in of the CIM has been achieved through a continual process of consultation with the climate modelling community, and through the METAFOR team’s development of a questionnaire that will be used to collect the metadata for the Intergovernmental Panel on Climate Change’s (IPCC) Coupled Model Intercomparison Project Phase 5 (CMIP5) model runs.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The first genome-wide association study for BMI identified a polymorphism, rs7566605, 10 kb upstream of the insulin-induced gene 2 (INSIG2) transcription start site, as the most significantly associated variant in children and adults. Subsequent studies, however, showed inconsistent association of this polymorphism with obesity traits. This polymorphism has been hypothesized to alter INSIG2 expression leading to inhibition of fatty acid and cholesterol synthesis. Hence, we investigated the association of the INSIG2 rs7566605 polymorphism with obesity- and lipid-related traits in Danish and Estonian children (930 boys and 1,073 girls) from the European Youth Heart Study (EYHS), a school-based, cross-sectional study of pre- and early pubertal children. The association between the polymorphism and obesity traits was tested using additive and recessive models adjusted for age, age-group, gender, maturity and country. Interactions were tested by including the interaction terms in the model. Despite having sufficient power (98%) to detect the previously reported effect size for association with BMI, we did not find significant effects of rs7566605 on BMI (additive, P = 0.68; recessive, P = 0.24). Accordingly, the polymorphism was not associated with overweight (P = 0.87) or obesity (P = 0.34). We also did not find association with waist circumference (WC), sum of four skinfolds, or with total cholesterol, triglycerides, low-density lipoprotein, or high-density lipoprotein. There were no gender-specific (P = 0.55), age-group-specific (P = 0.63) or country-specific (P = 0.56) effects. There was also no evidence of interaction between genotype and physical activity (P = 0.95). Despite an adequately powered study, our findings suggest that rs7566605 is not associated with obesity-related traits and lipids in the EYHS.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

As weather and climate models move toward higher resolution, there is growing excitement about potential future improvements in the understanding and prediction of atmospheric convection and its interaction with larger-scale phenomena. A meeting in January 2013 in Dartington, Devon was convened to address the best way to maximise these improvements, specifically in a UK context but with international relevance. Specific recommendations included increased convective-scale observations, high-resolution virtual laboratories, and a system of parameterization test beds with a range of complexities. The main recommendation was to facilitate the development of physically based convective parameterizations that are scale-aware, non-local, non-equilibrium, and stochastic.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The kinematic expansion history of the universe is investigated by using the 307 supernovae type Ia from the Union Compilation set. Three simple model parameterizations for the deceleration parameter ( constant, linear and abrupt transition) and two different models that are explicitly parametrized by the cosmic jerk parameter ( constant and variable) are considered. Likelihood and Bayesian analyses are employed to find best fit parameters and compare models among themselves and with the flat Lambda CDM model. Analytical expressions and estimates for the deceleration and cosmic jerk parameters today (q(0) and j(0)) and for the transition redshift (z(t)) between a past phase of cosmic deceleration to a current phase of acceleration are given. All models characterize an accelerated expansion for the universe today and largely indicate that it was decelerating in the past, having a transition redshift around 0.5. The cosmic jerk is not strongly constrained by the present supernovae data. For the most realistic kinematic models the 1 sigma confidence limits imply the following ranges of values: q(0) is an element of [-0.96, -0.46], j(0) is an element of [-3.2,-0.3] and z(t) is an element of [0.36, 0.84], which are compatible with the Lambda CDM predictions, q(0) = -0.57 +/- 0.04, j(0) = -1 and z(t) = 0.71 +/- 0.08. We find that even very simple kinematic models are equally good to describe the data compared to the concordance Lambda CDM model, and that the current observations are not powerful enough to discriminate among all of them.