920 resultados para Wolff, Jorge
Resumo:
Agrobacterium is widely considered to be the only bacterial genus capable of transferring genes to plants. When suitably modified, Agrobacterium has become the most effective vector for gene transfer in plant biotechnology1. However, the complexity of the patent landscape2 has created both real and perceived obstacles to the effective use of this technology for agricultural improvements by many public and private organizations worldwide. Here we show that several species of bacteria outside the Agrobacterium genus can be modified to mediate gene transfer to a number of diverse plants. These plant-associated symbiotic bacteria were made competent for gene transfer by acquisition of both a disarmed Ti plasmid and a suitable binary vector. This alternative to Agrobacterium-mediated technology for crop improvement, in addition to affording a versatile ‘open source’ platform for plant biotechnology, may lead to new uses of natural bacteria– plant interactions to achieve plant transformation.
Resumo:
This paper demonstrates the capabilities of wavelet transform (WT) for analyzing important features related to bottleneck activations and traffic oscillations in congested traffic in a systematic manner. In particular, the analysis of loop detector data from a freeway shows that the use of wavelet-based energy can effectively identify the location of an active bottleneck, the arrival time of the resulting queue at each upstream sensor location, and the start and end of a transition during the onset of a queue. Vehicle trajectories were also analyzed using WT and our analysis shows that the wavelet-based energies of individual vehicles can effectively detect the origins of deceleration waves and shed light on possible triggers (e.g., lane-changing). The spatiotemporal propagations of oscillations identified by tracing wavelet-based energy peaks from vehicle to vehicle enable analysis of oscillation amplitude, duration and intensity.
Resumo:
In this paper we identify the origins of stop-and-go (or slow-and-go) driving and measure microscopic features of their propagations by analyzing vehicle trajectories via Wavelet Transform. Based on 53 oscillation cases analyzed, we find that oscillations can be originated by either lane-changing maneuvers (LCMs) or car-following behavior (CF). LCMs were predominantly responsible for oscillation formations in the absence of considerable horizontal or vertical curves, whereas oscillations formed spontaneously near roadside work on an uphill segment. Regardless of the trigger, the features of oscillation propagations were similar in terms of propagation speed, oscillation duration, and amplitude. All observed cases initially exhibited a precursor phase, in which slow-and-go motions were localized. Some of them eventually transitioned into a well developed phase, in which oscillations propagated upstream in queue. LCMs were primarily responsible for the transition, although some transitions occurred without LCMs. Our findings also suggest that an oscillation has a regressive effect on car following behavior: a deceleration wave of an oscillation affects a timid driver (with larger response time and minimum spacing) to become less timid and an aggressive driver less aggressive, although this change may be short-lived. An extended framework of Newell’s CF is able to describe the regressive effects with two additional parameters with reasonable accuracy, as verified using vehicle trajectory data.
Resumo:
Teaching is emotional work. This is especially the case in the first years of teaching when new teachers are particularly vulnerable. By understanding changes in teacher emotions in the early years of teaching we hope to identify strategies that might ultimately reduce teacher attrition. As part of a larger study of the transition of new teachers to the profession, this ethnographic case study explores how a new science teacher produced and reproduced positive emotional interaction rituals with her students in her first year of teaching. We show how dialogical interactions were positive and satisfying experiences for the teacher, and how they were reproduced successfully in different contexts. We also illustrate how both teacher and students used humor to create a structure for dialogical interactions. During these successful interactions the students used shared resources to satisfy their teacher that they were engaging in the relevant science content. The implications of what we have learned for the professional development of new teachers are discussed in relation to an expanded understanding of teacher emotions.
Resumo:
Laughter is a fundamental human phenomenon. Yet there is little educational research on the potential functions of laughter on the enacted (lived) curriculum. In this study, we identify the functions of laughter in a beginning science teacher’s classroom throughout her first year of teaching. Our study shows that laughter is more than a gratuitous phenomenon. It is the result of a collective interactive achievement of the classroom participants that offsets the seriousness of science as a discipline. Laughter, whereas it challenges the seriousness of science, also includes the dialectical inversion of the challenge: it simultaneously reinforces the idea of science as serious business.
Resumo:
Asset health inspections can produce two types of indicators: (1) direct indicators (e.g. the thickness of a brake pad, and the crack depth on a gear) which directly relate to a failure mechanism; and (2) indirect indicators (e.g. the indicators extracted from vibration signals and oil analysis data) which can only partially reveal a failure mechanism. While direct indicators enable more precise references to asset health condition, they are often more difficult to obtain than indirect indicators. The state space model provides an efficient approach to estimating direct indicators by using indirect indicators. However, existing state space models to estimate direct indicators largely depend on assumptions such as, discrete time, discrete state, linearity, and Gaussianity. The discrete time assumption requires fixed inspection intervals. The discrete state assumption entails discretising continuous degradation indicators, which often introduces additional errors. The linear and Gaussian assumptions are not consistent with nonlinear and irreversible degradation processes in most engineering assets. This paper proposes a state space model without these assumptions. Monte Carlo-based algorithms are developed to estimate the model parameters and the remaining useful life. These algorithms are evaluated for performance using numerical simulations through MATLAB. The result shows that both the parameters and the remaining useful life are estimated accurately. Finally, the new state space model is used to process vibration and crack depth data from an accelerated test of a gearbox. During this application, the new state space model shows a better fitness result than the state space model with linear and Gaussian assumption.
Resumo:
Due to the limitation of current condition monitoring technologies, the estimates of asset health states may contain some uncertainties. A maintenance strategy ignoring this uncertainty of asset health state can cause additional costs or downtime. The partially observable Markov decision process (POMDP) is a commonly used approach to derive optimal maintenance strategies when asset health inspections are imperfect. However, existing applications of the POMDP to maintenance decision-making largely adopt the discrete time and state assumptions. The discrete-time assumption requires the health state transitions and maintenance activities only happen at discrete epochs, which cannot model the failure time accurately and is not cost-effective. The discrete health state assumption, on the other hand, may not be elaborate enough to improve the effectiveness of maintenance. To address these limitations, this paper proposes a continuous state partially observable semi-Markov decision process (POSMDP). An algorithm that combines the Monte Carlo-based density projection method and the policy iteration is developed to solve the POSMDP. Different types of maintenance activities (i.e., inspections, replacement, and imperfect maintenance) are considered in this paper. The next maintenance action and the corresponding waiting durations are optimized jointly to minimize the long-run expected cost per unit time and availability. The result of simulation studies shows that the proposed maintenance optimization approach is more cost-effective than maintenance strategies derived by another two approximate methods, when regular inspection intervals are adopted. The simulation study also shows that the maintenance cost can be further reduced by developing maintenance strategies with state-dependent maintenance intervals using the POSMDP. In addition, during the simulation studies the proposed POSMDP shows the ability to adopt a cost-effective strategy structure when multiple types of maintenance activities are involved.
Resumo:
Abstract As regional and continental carbon balances of terrestrial ecosystems become available, it becomes clear that the soils are the largest source of uncertainty. Repeated inventories of soil organic carbon (SOC) organized in soil monitoring networks (SMN) are being implemented in a number of countries. This paper reviews the concepts and design of SMNs in ten countries, and discusses the contribution of such networks to reducing the uncertainty of soil carbon balances. Some SMNs are designed to estimate country-specific land use or management effects on SOC stocks, while others collect soil carbon and ancillary data to provide a nationally consistent assessment of soil carbon condition across the major land-use/soil type combinations. The former use a single sampling campaign of paired sites, while for the latter both systematic (usually grid based) and stratified repeated sampling campaigns (5–10 years interval) are used with densities of one site per 10–1,040 km². For paired sites, multiple samples at each site are taken in order to allow statistical analysis, while for the single sites, composite samples are taken. In both cases, fixed depth increments together with samples for bulk density and stone content are recommended. Samples should be archived to allow for re-measurement purposes using updated techniques. Information on land management, and where possible, land use history should be systematically recorded for each site. A case study of the agricultural frontier in Brazil is presented in which land use effect factors are calculated in order to quantify the CO2 fluxes from national land use/management conversion matrices. Process-based SOC models can be run for the individual points of the SMN, provided detailed land management records are available. These studies are still rare, as most SMNs have been implemented recently or are in progress. Examples from the USA and Belgium show that uncertainties in SOC change range from 1.6–6.5 Mg C ha−1 for the prediction of SOC stock changes on individual sites to 11.72 Mg C ha−1 or 34% of the median SOC change for soil/land use/climate units. For national SOC monitoring, stratified sampling sites appears to be the most straightforward attribution of SOC values to units with similar soil/land use/climate conditions (i.e. a spatially implicit upscaling approach). Keywords Soil monitoring networks - Soil organic carbon - Modeling - Sampling design
Resumo:
This panel discusses the impact of Green IT on information systems and how information systems can meet environmental challenges and ensure sustainability. We wish to highlight the role of green business processes, and specifically the contributions that the management of these processes can play in leveraging the transformative power of IS in order to create an environmentally sustainable society. The management of business processes has typically been thought of in terms of business improvement alongside the dimensions time, cost, quality, or flexibility – the so-called ‘devil’s quadrangle’. Contemporary organizations, however, increasingly become aware of the need to create more sustainable, IT-enabled business processes that are also successful in terms of their economic, ecological, as well as social impact. Exemplary ecological key performance indicators that increasingly find their way into the agenda of managers include carbon emissions, data center energy, or renewable energy consumption (SAP 2010). The key challenge, therefore, is to extend the devil’s quadrangle to a devil’s pentagon, including sustainability as an important fifth dimension in process change.
Resumo:
OBJECTIVE: This paper reviews the epidemiological evidence on the relationship between ambient temperature and morbidity. It assesses the methodological issues in previous studies, and proposes future research directions. DATA SOURCES AND DATA EXTRACTION: We searched the PubMed database for epidemiological studies on ambient temperature and morbidity of non-communicable diseases published in refereed English journals prior to June 2010. 40 relevant studies were identified. Of these, 24 examined the relationship between ambient temperature and morbidity, 15 investigated the short-term effects of heatwave on morbidity, and 1 assessed both temperature and heatwave effects. DATA SYNTHESIS: Descriptive and time-series studies were the two main research designs used to investigate the temperature–morbidity relationship. Measurements of temperature exposure and health outcomes used in these studies differed widely. The majority of studies reported a significant relationship between ambient temperature and total or cause-specific morbidities. However, there were some inconsistencies in the direction and magnitude of non-linear lag effects. The lag effect of hot temperature on morbidity was shorter (several days) compared to that of cold temperature (up to a few weeks). The temperature–morbidity relationship may be confounded and/or modified by socio-demographic factors and air pollution. CONCLUSIONS: There is a significant short-term effect of ambient temperature on total and cause-specific morbidities. However, further research is needed to determine an appropriate temperature measure, consider a diverse range of morbidities, and to use consistent methodology to make different studies more comparable.
Resumo:
In an effort to evaluate and improve their practices to ensure the future excellence of the Texas highway system, the Texas Department of Transportation (TxDOT) sought a forum in which experts from other state departments of transportation could share their expertise. Thus, the Peer State Review of TxDOT Maintenance Practices project was organized and conducted for TxDOT by the Center for Transportation Research (CTR) at The University of Texas at Austin. The goal of the project was to conduct a workshop at CTR and in the Austin District that would educate the visiting peers on TxDOT’s maintenance practices and invite their feedback. CTR and TxDOT arranged the participation of the following directors of maintenance: Steve Takigawa, CA; Roy Rissky, KS; Eric Pitts, GA; Jim Carney, MO; Jennifer Brandenburg, NC; and David Bierschbach, WA. One of the means used to capture the peer reviewers’ opinions was a carefully designed booklet of 15 questions. The peers provided TxDOT with written responses to these questions, and the oral comments made during the workshop were also captured. This information was then compiled and summarized in the following report. An examination of the peers’ comments suggests that TxDOT should use a more holistic, statewide approach to funding and planning rather than funding and planning for each district separately. Additionally, the peers stressed the importance of allocating funds based on the actual conditions of the roadways instead of on inventory. The visiting directors of maintenance also recommended continuing and proliferating programs that enhance communication, such as peer review workshops.
Resumo:
Process mining techniques are able to extract knowledge from event logs commonly available in today’s information systems. These techniques provide new means to discover, monitor, and improve processes in a variety of application domains. There are two main drivers for the growing interest in process mining. On the one hand, more and more events are being recorded, thus, providing detailed information about the history of processes. On the other hand, there is a need to improve and support business processes in competitive and rapidly changing environments. This manifesto is created by the IEEE Task Force on Process Mining and aims to promote the topic of process mining. Moreover, by defining a set of guiding principles and listing important challenges, this manifesto hopes to serve as a guide for software developers, scientists, consultants, business managers, and end-users. The goal is to increase the maturity of process mining as a new tool to improve the (re)design, control, and support of operational business processes.