809 resultados para Performance model
Resumo:
Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)
Resumo:
Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)
Resumo:
Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)
Resumo:
Nowadays, many forms of reproduction coexist in nature: Asexual, Sexual, apomictic and meiotic parthenogenesis, hermaphroditism and parasex. The mechanisms of their evolution and what made them successful reproductive alternatives are very challenging and debated questions. Here, using a simple evolutionary aging model, we give I possible scenario. By studying the performance of Populations where individuals may have diverse characteristics-different ploidies, sex with or without crossing over, as well as the absence of sex-we find all evolution sequence that may explain why there are actually two major or leading groups: Sexual and asexual. We also investigate the dependence of these characteristics on different conditions of fertility and deleterious mutations. Finally, if the primeval organisms oil Earth were, in fact, asexual individuals we conjecture that the sexual form of reproduction could have more easily been set and found its niche during a period of low-intensity mutations. (c) 2005 Elsevier B.V. All rights reserved.
Resumo:
CMS is a general purpose experiment, designed to study the physics of pp collisions at 14 TeV at the Large Hadron Collider ( LHC). It currently involves more than 2000 physicists from more than 150 institutes and 37 countries. The LHC will provide extraordinary opportunities for particle physics based on its unprecedented collision energy and luminosity when it begins operation in 2007. The principal aim of this report is to present the strategy of CMS to explore the rich physics programme offered by the LHC. This volume demonstrates the physics capability of the CMS experiment. The prime goals of CMS are to explore physics at the TeV scale and to study the mechanism of electroweak symmetry breaking - through the discovery of the Higgs particle or otherwise. To carry out this task, CMS must be prepared to search for new particles, such as the Higgs boson or supersymmetric partners of the Standard Model particles, from the start- up of the LHC since new physics at the TeV scale may manifest itself with modest data samples of the order of a few fb(-1) or less. The analysis tools that have been developed are applied to study in great detail and with all the methodology of performing an analysis on CMS data specific benchmark processes upon which to gauge the performance of CMS. These processes cover several Higgs boson decay channels, the production and decay of new particles such as Z' and supersymmetric particles, B-s production and processes in heavy ion collisions. The simulation of these benchmark processes includes subtle effects such as possible detector miscalibration and misalignment. Besides these benchmark processes, the physics reach of CMS is studied for a large number of signatures arising in the Standard Model and also in theories beyond the Standard Model for integrated luminosities ranging from 1 fb(-1) to 30 fb(-1). The Standard Model processes include QCD, B-physics, diffraction, detailed studies of the top quark properties, and electroweak physics topics such as the W and Z(0) boson properties. The production and decay of the Higgs particle is studied for many observable decays, and the precision with which the Higgs boson properties can be derived is determined. About ten different supersymmetry benchmark points are analysed using full simulation. The CMS discovery reach is evaluated in the SUSY parameter space covering a large variety of decay signatures. Furthermore, the discovery reach for a plethora of alternative models for new physics is explored, notably extra dimensions, new vector boson high mass states, little Higgs models, technicolour and others. Methods to discriminate between models have been investigated. This report is organized as follows. Chapter 1, the Introduction, describes the context of this document. Chapters 2-6 describe examples of full analyses, with photons, electrons, muons, jets, missing E-T, B-mesons and tau's, and for quarkonia in heavy ion collisions. Chapters 7-15 describe the physics reach for Standard Model processes, Higgs discovery and searches for new physics beyond the Standard Model.
Resumo:
Due to the wide range of design possibilities, simple manufactured, low maintenance and low cost, cross-flow heat exchangers are extensively used in the petroleum, petrochemical, air conditioning, food storage, and others industries. In this paper a mathematical model for cross-flow heat exchangers with complex flow arrangements for determining epsilon -NTU relations is presented. The model is based on the tube element approach, according to which the heat exchanger outlet temperatures are obtained by discretizing the coil along the tube fluid path. In each cross section of the element, tube-side fluid temperature is assumed to be constant because the heat capacity rate ratio C*=Cmin/Cmax tends toward zero in the element. Thus temperature is controlled by effectiveness of a local element corresponding to an evaporator or a condenser-type element. The model is validated through comparison with theoretical algebraic relations for single-pass cross-flow arrangements with one or more rows. Very small relative errors are obtained showing the accuracy of the present model. epsilon -NTU curves for several complex circuit arrangements are presented. The model developed represents a useful research tool for theoretical and experimental studies on heat exchangers performance.
Resumo:
This work presents a methodology for elastic-plastic fracture reliability analysis of plane and axisymmetric structures. The structural reliability analysis is accomplished by means of the FORM analytical method. The virtual crack extension technique based on a direct minimization of potencial energy is utililized for the calculation of the energy release rate. Results are presented to illustrate the performance of the adopted methodology.
Resumo:
A boundary element method (BEM) formulation to predict the behavior of solids exhibiting displacement (strong) discontinuity is presented. In this formulation, the effects of the displacement jump of a discontinuity interface embedded in an internal cell are reproduced by an equivalent strain field over the cell. To compute the stresses, this equivalent strain field is assumed as the inelastic part of the total strain. As a consequence, the non-linear BEM integral equations that result from the proposed approach are similar to those of the implicit BEM based on initial strains. Since discontinuity interfaces can be introduced inside the cell independently on the cell boundaries, the proposed BEM formulation, combined with a tracking scheme to trace the discontinuity path during the analysis, allows for arbitrary discontinuity propagation using a fixed mesh. A simple technique to track the crack path is outlined. This technique is based on the construction of a polygonal line formed by segments inside the cells, in which the assumed failure criterion is reached. Two experimental concrete fracture tests were analyzed to assess the performance of the proposed formulation.
Resumo:
Objective: Nitroxides have strong antioxidant capacity but their effectiveness is limited by their rapid intracellular inactivation. Poly nitroxyl-Albumin (PNA) is capable of regenerating inactivated nitroxide. We tested the effect of PNA against reperfusion injury in heart transplantation. Methods: Pig hearts were transplanted orthotopically. In the control group (n = 9) reperfusion was performed without reperfusion modifications. In the experimental group (n = 10) 1 ml/kg PNA was given before cross-clamp release. Results: Hemodynamic performance was impaired after transplantation in both groups without significant intergroup differences. Plasma malonedialdehyde levels were significantly diminished in the PNA group as compared to the controls. CK-MB levels in both groups were increased within the first 2 h of reperfusion without significant intergroup differences. In contrast, there were found significant higher values of myocardial specific lactate dehydrogenase (LD1) in the controls versus PNA group. Conclusions: PNA was able to reduce lipid peroxidation and attenuate free radical activity. Contractile dysfunction could no be improved, indicating that (a) the radical scavenging effect was to weak or (b) other mechanisms than free oxygen radicals are responsible for myocardial damage in this experimental model. (C) 2001 Elsevier B.V. B.V. All rights reserved.
Resumo:
The paper presents a constructive heuristic algorithm (CHA) for solving directly the long-term transmission-network-expansion-planning (LTTNEP) problem using the DC model. The LTTNEP is a very complex mixed-integer nonlinear-programming problem and presents a combinatorial growth in the search space. The CHA is used to find a solution for the LTTNEP problem of good quality. A sensitivity index is used in each step of the CHA to add circuits to the system. This sensitivity index is obtained by solving the relaxed problem of LTTNEP, i.e. considering the number of circuits to be added as a continuous variable. The relaxed problem is a large and complex nonlinear-programming problem and was solved through the interior-point method (IPM). Tests were performed using Garver's system, the modified IEEE 24-Bus system and the Southern Brazilian reduced system. The results presented show the good performance of IPM inside the CHA.
Resumo:
Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)
Assessing the uncertainties of model estimates of primary productivity in the tropical Pacific Ocean
Resumo:
Depth-integrated primary productivity (PP) estimates obtained from satellite ocean color-based models (SatPPMs) and those generated from biogeochemical ocean general circulation models (BCGCMs) represent a key resource for biogeochemical and ecological studies at global as well as regional scales. Calibration and validation of these PP models are not straightforward, however, and comparative studies show large differences between model estimates. The goal of this paper is to compare PP estimates obtained from 30 different models (21 SatPPMs and 9 BOGCMs) to a tropical Pacific PP database consisting of similar to 1000 C-14 measurements spanning more than a decade (1983-1996). Primary findings include: skill varied significantly between models, but performance was not a function of model complexity or type (i.e. SatPPM vs. BOGCM); nearly all models underestimated the observed variance of PR specifically yielding too few low PP (< 0.2 g Cm-2 d(-1)) values; more than half of the total root-mean-squared model-data differences associated with the satellite-based PP models might be accounted for by uncertainties in the input variables and/or the PP data; and the tropical Pacific database captures a broad scale shift from low biomassnormalized productivity in the 1980s to higher biomass-normalized productivity in the 1990s, which was not successfully captured by any of the models. This latter result suggests that interdecadal and global changes will be a significant challenge for both SatPPMs and BOGCMs. Finally, average root-mean-squared differences between in situ PP data on the equator at 140 degrees W and PP estimates from the satellite-based productivity models were 58% lower than analogous values computed in a previous PP model comparison 6 years ago. The success of these types of comparison exercises is illustrated by the continual modification and improvement of the participating models and the resulting increase in model skill. (C) 2008 Elsevier BY. All rights reserved.
Resumo:
Nearly half of the earth's photosynthetically fixed carbon derives from the oceans. To determine global and region specific rates, we rely on models that estimate marine net primary productivity (NPP) thus it is essential that these models are evaluated to determine their accuracy. Here we assessed the skill of 21 ocean color models by comparing their estimates of depth-integrated NPP to 1156 in situ C-14 measurements encompassing ten marine regions including the Sargasso Sea, pelagic North Atlantic, coastal Northeast Atlantic, Black Sea, Mediterranean Sea, Arabian Sea, subtropical North Pacific, Ross Sea, West Antarctic Peninsula, and the Antarctic Polar Frontal Zone. Average model skill, as determined by root-mean square difference calculations, was lowest in the Black and Mediterranean Seas, highest in the pelagic North Atlantic and the Antarctic Polar Frontal Zone, and intermediate in the other six regions. The maximum fraction of model skill that may be attributable to uncertainties in both the input variables and in situ NPP measurements was nearly 72%. on average, the simplest depth/wavelength integrated models performed no worse than the more complex depth/wavelength resolved models. Ocean color models were not highly challenged in extreme conditions of surface chlorophyll-a and sea surface temperature, nor in high-nitrate low-chlorophyll waters. Water column depth was the primary influence on ocean color model performance such that average skill was significantly higher at depths greater than 250 m, suggesting that ocean color models are more challenged in Case-2 waters (coastal) than in Case-1 (pelagic) waters. Given that in situ chlorophyll-a data was used as input data, algorithm improvement is required to eliminate the poor performance of ocean color NPP models in Case-2 waters that are close to coastlines. Finally, ocean color chlorophyll-a algorithms are challenged by optically complex Case-2 waters, thus using satellite-derived chlorophyll-a to estimate NPP in coastal areas would likely further reduce the skill of ocean color models.