956 resultados para dynamic response optimization


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Dans le sillage de la récession mondiale de 2008-09, plusieurs questions ont été soulevées dans la littérature économique sur les effets à court et à long terme de la politique budgétaire sur l’activité économique par rapport à son signe, sa taille et sa durée. Ceux-ci ont des implications importantes pour mieux comprendre les canaux de transmission et l’efficacité des politiques budgétaires, avec la politique monétaire étant poursuivi, ainsi que pour leurs retombées économiques. Cette thèse fait partie de ce regain d’intérêt de la littérature d’examiner comment les changements dans la politique budgétaire affectent l’activité économique. Elle repose alors sur trois essais: les effets macroéconomiques des chocs de dépenses publiques et des recettes fiscales, les résultats macroéconomiques de l’interaction entre les politiques budgétaire et monétaire et le lien entre la politique budgétaire et la répartition des revenus. Le premier chapitre examine les effets des chocs de politique budgétaire (chocs de dépenses publiques et chocs de recettes fiscales) sur l’économie canadienne au cours de la période 1970-2010, en s’appuyant sur la méthode d’identification des restrictions de signe développée par Mountford et Uhlig [2009]. En réponse à la récession mondiale, les autorités fiscales dans les économies avancées, dont le Canada ont généralement mis en oeuvre une approche en deux phases pour la politique budgétaire. Tout d’abord, ils ont introduit des plans de relance sans précédent pour relancer leurs économies. Par exemple, les mesures de relance au Canada, introduites à travers le Plan d’action économique du Canada, ont été projetées à 3.2 pour cent du PIB dans le budget fédéral de 2009 tandis que l’ "American Recovery and Reinvestment Act"(ARRA) a été estimé à 7 pour cent du PIB. Par la suite, ils ont mis en place des plans d’ajustement en vue de réduire la dette publique et en assurer la soutenabilité à long terme. Dans ce contexte, évaluer les effets multiplicateurs de la politique budgétaire est important en vue d’informer sur l'efficacité de telles mesures dans la relance ou non de l'activité économique. Les résultats montrent que les multiplicateurs d'impôt varient entre 0.2 et 0.5, tandis que les multiplicateurs de dépenses varient entre 0.2 et 1.1. Les multiplicateurs des dépenses ont tendance à être plus grand que les multiplicateurs des recettes fiscales au cours des deux dernières décennies. Comme implications de politique économique, ces résultats tendent à suggérer que les ajustements budgétaires par le biais de grandes réductions de dépenses publiques pourraient être plus dommageable pour l'économie que des ajustements budgétaires par la hausse des impôts. Le deuxième chapitre, co-écrit avec Constant Lonkeng Ngouana, estime les effets multiplicateurs des dépenses publiques aux Etats-Unis en fonction du cycle de la politique monétaire. Les chocs de dépenses publiques sont identifiés comme étant des erreurs de prévision du taux de croissance des dépenses publiques à partir des données d'Enquêtes des prévisionnistes professionnels et des informations contenues dans le "Greenbook". L'état de la politique monétaire est déduite à partir de la déviation du taux des fonds fédéraux du taux cible de la Réserve Fédérale, en faisant recours à une fonction lisse de transition. L'application de la méthode des «projections locales» aux données trimestrielles américaines au cours de la période 1965-2012 suggère que les effets multiplicateurs des dépenses fédérales sont sensiblement plus élevées quand la politique monétaire est accommodante que lorsqu'elle ne l'est pas. Les résultats suggèrent aussi que les dépenses fédérales peuvent stimuler ou non la consommation privée, dépendamment du degré d’accommodation de la politique monétaire. Ce dernier résultat réconcilie ainsi, sur la base d’un cadre unifié des résultats autrement contradictoires à première vue dans la littérature. Ces résultats ont d'importantes implications de politique économique. Ils suggèrent globalement que la politique budgétaire est plus efficace lorsqu'on en a le plus besoin (par exemple, lorsque le taux de chômage est élevé), si elle est soutenue par la politique monétaire. Ils ont également des implications pour la normalisation des conditions monétaires dans les pays avancés: la sortie des politiques monétaires non-conventionnelles conduirait à des multiplicateurs de dépenses fédérales beaucoup plus faibles qu'autrement, même si le niveau de chômage restait élevé. Ceci renforce la nécessité d'une calibration prudente du calendrier de sortie des politiques monétaires non-conventionnelles. Le troisième chapitre examine l'impact des mesures d'expansion et de contraction budgétaire sur la distribution des revenus dans un panel de 18 pays d'Amérique latine au cours de la période 1990-2010, avec un accent sur les deniers 40 pour cent. Il explore alors comment ces mesures fiscales ainsi que leur composition affectent la croissance des revenus des dernier 40 pour cent, la croissance de leur part de revenu ainsi que la croissance économique. Les mesures d'expansion et de contraction budgétaire sont identifiées par des périodes au cours desquels il existe une variation significative du déficit primaire corrigé des variations conjoncturelles en pourcentage du PIB. Les résultats montrent qu'en moyenne l'expansion budgétaire par la hausse des dépenses publiques est plus favorable à la croissance des revenus des moins bien-nantis que celle par la baisse des impôts. Ce résultat est principalement soutenu par la hausse des dépenses gouvernementales de consommation courante, les transferts et subventions. En outre ces mesures d’expansion budgétaire sont favorables à la réduction des inégalités car elles permettent d'améliorer la part des revenus des moins bien-nantis tout en réduisant la part des revenus des mieux-nantis de la distribution des revenus. En outre ces mesures d’expansion budgétaire sont favorables à la réduction des inégalités car elles permettent d'améliorer la part des revenus des moins bien-nantis tout en réduisant la part des revenus des mieux-nantis de la distribution des revenus. Cependant, l'expansion budgétaire pourrait soit n'avoir aucun effet sur la croissance économique ou entraver cette dernière à travers la hausse des dépenses en capital. Les résultats relatifs à la contraction budgétaire sont quelque peu mitigés. Parfois, les mesures de contraction budgétaire sont associées à une baisse de la croissance des revenus des moins bien nantis et à une hausse des inégalités, parfois l'impact de ces mesures est non significatif. Par ailleurs, aucune des mesures n’affecte de manière significative la croissance du PIB. Comme implications de politique économique, les pays avec une certaine marge de manœuvre budgétaire pourraient entamer ou continuer à mettre en œuvre des programmes de "filets de sauvetage"--par exemple les programmes de transfert monétaire conditionnel--permettant aux segments vulnérables de la population de faire face à des chocs négatifs et aussi d'améliorer leur conditions de vie. Avec un potentiel de stimuler l'emploi peu qualifié, une relance budgétaire sage par les dépenses publique courantes pourrait également jouer un rôle important pour la réduction des inégalités. Aussi, pour éviter que les dépenses en capital freinent la croissance économique, les projets d'investissements publics efficients devraient être prioritaires dans le processus d'élaboration des politiques. Ce qui passe par la mise en œuvre des projets d'investissement avec une productivité plus élevée capable de générer la croissance économique nécessaire pour réduire les inégalités.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The main source of protein for human and animal consumption is from the agricultural sector, where the production is vulnerable to diseases, fluctuations in climatic conditions and deteriorating hydrological conditions due to water pollution. Therefore Single Cell Protein (SCP) production has evolved as an excellent alternative. Among all sources of microbial protein, yeast has attained global acceptability and has been preferred for SCP production. The screening and evaluation of nutritional and other culture variables of microorganisms are very important in the development of a bioprocess for SCP production. The application of statistical experimental design in bioprocess development can result in improved product yields, reduced process variability, closer confirmation of the output response to target requirements and reduced development time and overall cost.The present work was undertaken to develop a bioprocess technology for the mass production of a marine yeast, Candida sp.S27. Yeasts isolated from the offshore waters of the South west coast of India and maintained in the Microbiology Laboratory were subjected to various tests for the selection of a potent strain for biomass production. The selected marine yeast was identified based on ITS sequencing. Biochemical/nutritional characterization of Candida sp.S27 was carried out. Using Response Surface Methodology (RSM) the process parameters (pH, temperature and salinity) were optimized. For mass production of yeast biomass, a chemically defined medium (Barnett and Ingram, 1955) and a crude medium (Molasses-Yeast extract) were optimized using RSM. Scale up of biomass production was done in a Bench top Fermenter using these two optimized media. Comparative efficacy of the defined and crude media were estimated besides nutritional evaluation of the biomass developed using these two optimized media.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

To ensure quality of machined products at minimum machining costs and maximum machining effectiveness, it is very important to select optimum parameters when metal cutting machine tools are employed. Traditionally, the experience of the operator plays a major role in the selection of optimum metal cutting conditions. However, attaining optimum values each time by even a skilled operator is difficult. The non-linear nature of the machining process has compelled engineers to search for more effective methods to attain optimization. The design objective preceding most engineering design activities is simply to minimize the cost of production or to maximize the production efficiency. The main aim of research work reported here is to build robust optimization algorithms by exploiting ideas that nature has to offer from its backyard and using it to solve real world optimization problems in manufacturing processes.In this thesis, after conducting an exhaustive literature review, several optimization techniques used in various manufacturing processes have been identified. The selection of optimal cutting parameters, like depth of cut, feed and speed is a very important issue for every machining process. Experiments have been designed using Taguchi technique and dry turning of SS420 has been performed on Kirlosker turn master 35 lathe. Analysis using S/N and ANOVA were performed to find the optimum level and percentage of contribution of each parameter. By using S/N analysis the optimum machining parameters from the experimentation is obtained.Optimization algorithms begin with one or more design solutions supplied by the user and then iteratively check new design solutions, relative search spaces in order to achieve the true optimum solution. A mathematical model has been developed using response surface analysis for surface roughness and the model was validated using published results from literature.Methodologies in optimization such as Simulated annealing (SA), Particle Swarm Optimization (PSO), Conventional Genetic Algorithm (CGA) and Improved Genetic Algorithm (IGA) are applied to optimize machining parameters while dry turning of SS420 material. All the above algorithms were tested for their efficiency, robustness and accuracy and observe how they often outperform conventional optimization method applied to difficult real world problems. The SA, PSO, CGA and IGA codes were developed using MATLAB. For each evolutionary algorithmic method, optimum cutting conditions are provided to achieve better surface finish.The computational results using SA clearly demonstrated that the proposed solution procedure is quite capable in solving such complicated problems effectively and efficiently. Particle Swarm Optimization (PSO) is a relatively recent heuristic search method whose mechanics are inspired by the swarming or collaborative behavior of biological populations. From the results it has been observed that PSO provides better results and also more computationally efficient.Based on the results obtained using CGA and IGA for the optimization of machining process, the proposed IGA provides better results than the conventional GA. The improved genetic algorithm incorporating a stochastic crossover technique and an artificial initial population scheme is developed to provide a faster search mechanism. Finally, a comparison among these algorithms were made for the specific example of dry turning of SS 420 material and arriving at optimum machining parameters of feed, cutting speed, depth of cut and tool nose radius for minimum surface roughness as the criterion. To summarize, the research work fills in conspicuous gaps between research prototypes and industry requirements, by simulating evolutionary procedures seen in nature that optimize its own systems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The proliferation of wireless sensor networks in a large spectrum of applications had been spurered by the rapid advances in MEMS(micro-electro mechanical systems )based sensor technology coupled with low power,Low cost digital signal processors and radio frequency circuits.A sensor network is composed of thousands of low cost and portable devices bearing large sensing computing and wireless communication capabilities. This large collection of tiny sensors can form a robust data computing and communication distributed system for automated information gathering and distributed sensing.The main attractive feature is that such a sensor network can be deployed in remote areas.Since the sensor node is battery powered,all the sensor nodes should collaborate together to form a fault tolerant network so as toprovide an efficient utilization of precious network resources like wireless channel,memory and battery capacity.The most crucial constraint is the energy consumption which has become the prime challenge for the design of long lived sensor nodes.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A marine isolate of jáÅêçÅçÅÅìë MCCB 104 has been identified as an aquaculture probiotic antagonistic to sáÄêáç. In the present study different carbon and nitrogen sources and growth factors in a mineral base medium were optimized for enhanced biomass production and antagonistic activity against the target pathogen, sáÄêáç=Ü~êîÉóá, following response surface methodology (RSM). Accordingly the minimum and maximum limits of the selected variables were determined and a set of fifty experiments programmed employing central composite design (CCD) of RSM for the final optimization. The response surface plots of biomass showed similar pattern with that of antagonistic activity, which indicated a strong correlation between the biomass and antagonism. The optimum concentration of the carbon sources, nitrogen sources, and growth factors for both biomass and antagonistic activity were glucose (17.4 g/L), lactose (17 g/L), sodium chloride (16.9 g/L), ammonium chloride (3.3 g/L), and mineral salts solution (18.3 mL/L). © KSBB

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Coded OFDM is a transmission technique that is used in many practical communication systems. In a coded OFDM system, source data are coded, interleaved and multiplexed for transmission over many frequency sub-channels. In a conventional coded OFDM system, the transmission power of each subcarrier is the same regardless of the channel condition. However, some subcarrier can suffer deep fading with multi-paths and the power allocated to the faded subcarrier is likely to be wasted. In this paper, we compute the FER and BER bounds of a coded OFDM system given as convex functions for a given channel coder, inter-leaver and channel response. The power optimization is shown to be a convex optimization problem that can be solved numerically with great efficiency. With the proposed power optimization scheme, near-optimum power allocation for a given coded OFDM system and channel response to minimize FER or BER under a constant transmission power constraint is obtained

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This study was undertaken to isolate ligninase-producing white-rot fungi for use in the extraction of fibre from pineapple leaf agriwaste. Fifteen fungal strains were isolated from dead tree trunks and leaf litter. Ligninolytic enzymes (lignin peroxidase (LiP), manganese peroxidase (MnP), and laccase (Lac)), were produced by solid-state fermentation (SSF) using pineapple leaves as the substrate. Of the isolated strains, the one showing maximum production of ligninolytic enzymes was identified to be Ganoderma lucidum by 18S ribotyping. Single parameter optimization and response surface methodology of different process variables were carried out for enzyme production. Incubation period, agitation, and Tween-80 were identified to be the most significant variables through Plackett-Burman design. These variables were further optimized by Box-Behnken design. The overall maximum yield of ligninolytic enzymes was achieved by experimental analysis under these optimal conditions. Quantitative lignin analysis of pineapple leaves by Klason lignin method showed significant degradation of lignin by Ganoderma lucidum under SSF

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The traditional task of a central bank is to preserve price stability and, in doing so, not to impair the real economy more than necessary. To meet this challenge, it is of great relevance whether inflation is only driven by inflation expectations and the current output gap or whether it is, in addition, influenced by past inflation. In the former case, as described by the New Keynesian Phillips curve, the central bank can immediately and simultaneously achieve price stability and equilibrium output, the so-called ‘divine coincidence’ (Blanchard and Galí 2007). In the latter case, the achievement of price stability is costly in terms of output and will be pursued over several periods. Similarly, it is important to distinguish this latter case, which describes ‘intrinsic’ inflation persistence, from that of ‘extrinsic’ inflation persistence, where the sluggishness of inflation is not a ‘structural’ feature of the economy but merely ‘inherited’ from the sluggishness of the other driving forces, inflation expectations and output. ‘Extrinsic’ inflation persistence is usually considered to be the less challenging case, as policy-makers are supposed to fight against the persistence in the driving forces, especially to reduce the stickiness of inflation expectations by a credible monetary policy, in order to reestablish the ‘divine coincidence’. The scope of this dissertation is to contribute to the vast literature and ongoing discussion on inflation persistence: Chapter 1 describes the policy consequences of inflation persistence and summarizes the empirical and theoretical literature. Chapter 2 compares two models of staggered price setting, one with a fixed two-period duration and the other with a stochastic duration of prices. I show that in an economy with a timeless optimizing central bank the model with the two-period alternating price-setting (for most parameter values) leads to more persistent inflation than the model with stochastic price duration. This result amends earlier work by Kiley (2002) who found that the model with stochastic price duration generates more persistent inflation in response to an exogenous monetary shock. Chapter 3 extends the two-period alternating price-setting model to the case of 3- and 4-period price durations. This results in a more complex Phillips curve with a negative impact of past inflation on current inflation. As simulations show, this multi-period Phillips curve generates a too low degree of autocorrelation and too early turnings points of inflation and is outperformed by a simple Hybrid Phillips curve. Chapter 4 starts from the critique of Driscoll and Holden (2003) on the relative real-wage model of Fuhrer and Moore (1995). While taking the critique seriously that Fuhrer and Moore’s model will collapse to a much simpler one without intrinsic inflation persistence if one takes their arguments literally, I extend the model by a term for inequality aversion. This model extension is not only in line with experimental evidence but results in a Hybrid Phillips curve with inflation persistence that is observably equivalent to that presented by Fuhrer and Moore (1995). In chapter 5, I present a model that especially allows to study the relationship between fairness attitudes and time preference (impatience). In the model, two individuals take decisions in two subsequent periods. In period 1, both individuals are endowed with resources and are able to donate a share of their resources to the other individual. In period 2, the two individuals might join in a common production after having bargained on the split of its output. The size of the production output depends on the relative share of resources at the end of period 1 as the human capital of the individuals, which is built by means of their resources, cannot fully be substituted one against each other. Therefore, it might be rational for a well-endowed individual in period 1 to act in a seemingly ‘fair’ manner and to donate own resources to its poorer counterpart. This decision also depends on the individuals’ impatience which is induced by the small but positive probability that production is not possible in period 2. As a general result, the individuals in the model economy are more likely to behave in a ‘fair’ manner, i.e., to donate resources to the other individual, the lower their own impatience and the higher the productivity of the other individual. As the (seemingly) ‘fair’ behavior is modelled as an endogenous outcome and as it is related to the aspect of time preference, the presented framework might help to further integrate behavioral economics and macroeconomics.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Web services from different partners can be combined to applications that realize a more complex business goal. Such applications built as Web service compositions define how interactions between Web services take place in order to implement the business logic. Web service compositions not only have to provide the desired functionality but also have to comply with certain Quality of Service (QoS) levels. Maximizing the users' satisfaction, also reflected as Quality of Experience (QoE), is a primary goal to be achieved in a Service-Oriented Architecture (SOA). Unfortunately, in a dynamic environment like SOA unforeseen situations might appear like services not being available or not responding in the desired time frame. In such situations, appropriate actions need to be triggered in order to avoid the violation of QoS and QoE constraints. In this thesis, proper solutions are developed to manage Web services and Web service compositions with regard to QoS and QoE requirements. The Business Process Rules Language (BPRules) was developed to manage Web service compositions when undesired QoS or QoE values are detected. BPRules provides a rich set of management actions that may be triggered for controlling the service composition and for improving its quality behavior. Regarding the quality properties, BPRules allows to distinguish between the QoS values as they are promised by the service providers, QoE values that were assigned by end-users, the monitored QoS as measured by our BPR framework, and the predicted QoS and QoE values. BPRules facilitates the specification of certain user groups characterized by different context properties and allows triggering a personalized, context-aware service selection tailored for the specified user groups. In a service market where a multitude of services with the same functionality and different quality values are available, the right services need to be selected for realizing the service composition. We developed new and efficient heuristic algorithms that are applied to choose high quality services for the composition. BPRules offers the possibility to integrate multiple service selection algorithms. The selection algorithms are applicable also for non-linear objective functions and constraints. The BPR framework includes new approaches for context-aware service selection and quality property predictions. We consider the location information of users and services as context dimension for the prediction of response time and throughput. The BPR framework combines all new features and contributions to a comprehensive management solution. Furthermore, it facilitates flexible monitoring of QoS properties without having to modify the description of the service composition. We show how the different modules of the BPR framework work together in order to execute the management rules. We evaluate how our selection algorithms outperform a genetic algorithm from related research. The evaluation reveals how context data can be used for a personalized prediction of response time and throughput.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This research aims to understand the fundamental dynamic behavior of servo-controlled machinery in response to various types of sensory feedback. As an example of such a system, we study robot force control, a scheme which promises to greatly expand the capabilities of industrial robots by allowing manipulators to interact with uncertain and dynamic tasks. Dynamic models are developed which allow the effects of actuator dynamics, structural flexibility, and workpiece interaction to be explored in the frequency and time domains. The models are used first to explain the causes of robot force control instability, and then to find methods of improving this performance.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

La coordinació i assignació de tasques en entorns distribuïts ha estat un punt important de la recerca en els últims anys i aquests temes són el cor dels sistemes multi-agent. Els agents en aquests sistemes necessiten cooperar i considerar els altres agents en les seves accions i decisions. A més a més, els agents han de coordinar-se ells mateixos per complir tasques complexes que necessiten més d'un agent per ser complerta. Aquestes tasques poden ser tan complexes que els agents poden no saber la ubicació de les tasques o el temps que resta abans de que les tasques quedin obsoletes. Els agents poden necessitar utilitzar la comunicació amb l'objectiu de conèixer la tasca en l'entorn, en cas contrari, poden perdre molt de temps per trobar la tasca dins de l'escenari. De forma similar, el procés de presa de decisions distribuït pot ser encara més complexa si l'entorn és dinàmic, amb incertesa i en temps real. En aquesta dissertació, considerem entorns amb sistemes multi-agent amb restriccions i cooperatius (dinàmics, amb incertesa i en temps real). En aquest sentit es proposen dues aproximacions que permeten la coordinació dels agents. La primera és un mecanisme semi-centralitzat basat en tècniques de subhastes combinatòries i la idea principal es minimitzar el cost de les tasques assignades des de l'agent central cap als equips d'agents. Aquest algoritme té en compte les preferències dels agents sobre les tasques. Aquestes preferències estan incloses en el bid enviat per l'agent. La segona és un aproximació d'scheduling totalment descentralitzat. Això permet als agents assignar les seves tasques tenint en compte les preferències temporals sobre les tasques dels agents. En aquest cas, el rendiment del sistema no només depèn de la maximització o del criteri d'optimització, sinó que també depèn de la capacitat dels agents per adaptar les seves assignacions eficientment. Addicionalment, en un entorn dinàmic, els errors d'execució poden succeir a qualsevol pla degut a la incertesa i error de accions individuals. A més, una part indispensable d'un sistema de planificació és la capacitat de re-planificar. Aquesta dissertació també proveeix una aproximació amb re-planificació amb l'objectiu de permetre als agent re-coordinar els seus plans quan els problemes en l'entorn no permeti la execució del pla. Totes aquestes aproximacions s'han portat a terme per permetre als agents assignar i coordinar de forma eficient totes les tasques complexes en un entorn multi-agent cooperatiu, dinàmic i amb incertesa. Totes aquestes aproximacions han demostrat la seva eficiència en experiments duts a terme en l'entorn de simulació RoboCup Rescue.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A number of recent experiments suggest that, at a given wetting speed, the dynamic contact angle formed by an advancing liquid-gas interface with a solid substrate depends on the flow field and geometry near the moving contact line. In the present work, this effect is investigated in the framework of an earlier developed theory that was based on the fact that dynamic wetting is, by its very name, a process of formation of a new liquid-solid interface (newly “wetted” solid surface) and hence should be considered not as a singular problem but as a particular case from a general class of flows with forming or/and disappearing interfaces. The results demonstrate that, in the flow configuration of curtain coating, where a liquid sheet (“curtain”) impinges onto a moving solid substrate, the actual dynamic contact angle indeed depends not only on the wetting speed and material constants of the contacting media, as in the so-called slip models, but also on the inlet velocity of the curtain, its height, and the angle between the falling curtain and the solid surface. In other words, for the same wetting speed the dynamic contact angle can be varied by manipulating the flow field and geometry near the moving contact line. The obtained results have important experimental implications: given that the dynamic contact angle is determined by the values of the surface tensions at the contact line and hence depends on the distributions of the surface parameters along the interfaces, which can be influenced by the flow field, one can use the overall flow conditions and the contact angle as a macroscopic multiparametric signal-response pair that probes the dynamics of the liquid-solid interface. This approach would allow one to investigate experimentally such properties of the interface as, for example, its equation of state and the rheological properties involved in the interface’s response to an external torque, and would help to measure its parameters, such as the coefficient of sliding friction, the surface-tension relaxation time, and so on.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Here, we identify the Arabidopsis thaliana ortholog of the mammalian DEAD box helicase, eIF4A-III, the putative anchor protein of exon junction complex (EJC) on mRNA. Arabidopsis eIF4A-III interacts with an ortholog of the core EJC component, ALY/Ref, and colocalizes with other EJC components, such as Mago, Y14, and RNPS1, suggesting a similar function in EJC assembly to animal eIF4A-III. A green fluorescent protein (GFP)-eIF4A-III fusion protein showed localization to several subnuclear domains: to the nucleoplasm during normal growth and to the nucleolus and splicing speckles in response to hypoxia. Treatment with the respiratory inhibitor sodium azide produced an identical response to the hypoxia stress. Treatment with the proteasome inhibitor MG132 led to accumulation of GFP-eIF4A-III mainly in the nucleolus, suggesting that transition of eIF4A-III between subnuclear domains and/or accumulation in nuclear speckles is controlled by proteolysis-labile factors. As revealed by fluorescence recovery after photobleaching analysis, the nucleoplasmic fraction was highly mobile, while the speckles were the least mobile fractions, and the nucleolar fraction had an intermediate mobility. Sequestration of eIF4A-III into nuclear pools with different mobility is likely to reflect the transcriptional and mRNA processing state of the cell.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND: Exposure of macrophages to bacterial products such as lipopolysaccharide (LPS) results in activation of the NF-kappaB transcription factor, which orchestrates a gene expression programme that underpins the macrophage-dependent immune response. These changes include the induction or repression of a wide range of genes that regulate inflammation, cell proliferation, migration and cell survival. This process is tightly regulated and loss of control is associated with conditions such as septic shock, inflammatory diseases and cancer. To study this response, it is important to have in vitro model systems that reflect the behaviour of cells in vivo. In addition, it is necessary to understand the natural differences that can occur between individuals. In this report, we have investigated and compared the LPS response in macrophage derived cell lines and peripheral blood mononuclear cell (PBMC) derived macrophages. RESULTS: Gene expression profiles were determined following LPS treatment of THP-1 cells for 1 and 4 hours. LPS significantly induced or repressed 72 out of 465 genes selected as being known or putative NF-kappaB target genes, which exhibited 4 temporal patterns of expression. Results for 34 of these genes, including several genes not previously identified as LPS target genes, were validated using real time PCR. A high correlation between microarray and real time PCR data was found. Significantly, the LPS induced expression profile of THP-1 cells, as determined using real time PCR, was found to be very similar to that of human PBMC derived macrophages. Interestingly, some differences were observed in the LPS response between the two donor PBMC macrophage populations. Surprisingly, we found that the LPS response in U937 cells was dramatically different to both THP-1 and PBMC derived macrophages. CONCLUSION: This study revealed a dynamic and diverse transcriptional response to LPS in macrophages, involving both the induction and repression of gene expression in a time dependent manner. Moreover, we demonstrated that the LPS induced transcriptional response in the THP-1 cell line is very similar to primary PBMC derived macrophages. Therefore, THP-1 cells represent a good model system for studying the mechanisms of LPS and NF-kappaB dependent gene expression.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We estimate the body sizes of direct ancestors of extant carnivores, and examine selected aspects of life history as a function not only of species' current size, but also of recent changes in size. Carnivore species that have undergone marked recent evolutionary size change show life history characteristics typically associated with species closer to the ancestral body size. Thus, phyletic giants tend to mature earlier and have larger litters of smaller offspring at shorter intervals than do species of the same body size that are not phyletic giants. Phyletic dwarfs, by contrast, have slower life histories than nondwarf species of the same body size. We discuss two possible mechanisms for the legacy of recent size change: lag (in which life history variables cannot evolve as quickly as body size, leading to species having the 'wrong' life history for their body size) and body size optimization (in which life history and hence body size evolve in response to changes in energy availability); at present, we cannot distinguish between these alternatives. Our finding that recent body size changes help explain residual variation around life history allometries shows that a more dynamic view of character change enables comparative studies to make more precise predictions about species traits in the context of their evolutionary background.