928 resultados para temperature-based models


Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper develops analytical distributions of temperature indices on which temperature derivatives are written. If the deviations of daily temperatures from their expected values are modelled as an Ornstein-Uhlenbeck process with timevarying variance, then the distributions of the temperature index on which the derivative is written is the sum of truncated, correlated Gaussian deviates. The key result of this paper is to provide an analytical approximation to the distribution of this sum, thus allowing the accurate computation of payoffs without the need for any simulation. A data set comprising average daily temperature spanning over a hundred years for four Australian cities is used to demonstrate the efficacy of this approach for estimating the payoffs to temperature derivatives. It is demonstrated that expected payoffs computed directly from historical records are a particularly poor approach to the problem when there are trends in underlying average daily temperature. It is shown that the proposed analytical approach is superior to historical pricing.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The experimental solubilities of the mixture of nitrophenol (m- and p-) isomers were determined at 308, 318 and 328 K over a pressure range of 10-17.55 MPa. Compared to the binary solubilities, the ternary solubilities of m-nitrophenol increased at 308, 318 and 328 K. The ternary solubilities of p-nitrophenol increased at 308 K, while the ternary solubilities decreased at lower pressures and increased at higher pressure at 318 and 328 K. The solubilities of the solid mixtures in supercritical carbon dioxide (SCCO2) were correlated with solution models by incorporating the non-idealities using activity coefficient based models. The Wilson and NRTL activity coefficient models were applied to determine the nature of the interactions between the molecules. The equation developed by using the NRTL model has three parameters and correlates mixture solubilities of solid solutes in terms of temperature and cosolute composition. The equation derived from the Wilson model contains five parameters and correlates solubilities in terms of temperature, density and cosolute composition. These two new equations developed in this work were used to correlate the solubilities of 25 binary solid mixtures including the current data. The average AARDs of the model equations derived using the NRTL and Wilson models for the solid mixtures were found to be 7% and 4%, respectively. (C) 2012 Elsevier B.V. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Li-ion batteries have been widely used in the EVs, and the battery thermal management is a key but challenging part of the battery management system. For EV batteries, only the battery surface temperature can be measured in real-time. However, it is the battery internal temperature that directly affects the battery performance, and large temperature difference may exist between surface and internal temperatures, especially in high power demand applications. In this paper, an online battery internal temperature estimation method is proposed based on a novel simplified thermoelectric model. The battery thermal behaviour is first described by a simplified thermal model, and battery electrical behaviour by an electric model. Then, these two models are interrelated to capture the interactions between battery thermal and electrical behaviours, thus offer a comprehensive description of the battery behaviour that is useful for battery management. Finally, based on the developed model, the battery internal temperature is estimated using an extended Kalman filter. The experimental results confirm the efficacy of the proposed method, and it can be used for online internal temperature estimation which is a key indicator for better real-time battery thermal management.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Summary 1. Agent-based models (ABMs) are widely used to predict how populations respond to changing environments. As the availability of food varies in space and time, individuals should have their own energy budgets, but there is no consensus as to how these should be modelled. Here, we use knowledge of physiological ecology to identify major issues confronting the modeller and to make recommendations about how energy budgets for use in ABMs should be constructed. 2. Our proposal is that modelled animals forage as necessary to supply their energy needs for maintenance, growth and reproduction. If there is sufficient energy intake, an animal allocates the energy obtained in the order: maintenance, growth, reproduction, energy storage, until its energy stores reach an optimal level. If there is a shortfall, the priorities for maintenance and growth/reproduction remain the same until reserves fall to a critical threshold below which all are allocated to maintenance. Rates of ingestion and allocation depend on body mass and temperature. We make suggestions for how each of these processes should be modelled mathematically. 3. Mortality rates vary with body mass and temperature according to known relationships, and these can be used to obtain estimates of background mortality rate. 4. If parameter values cannot be obtained directly, then values may provisionally be obtained by parameter borrowing, pattern-oriented modelling, artificial evolution or from allometric equations. 5. The development of ABMs incorporating individual energy budgets is essential for realistic modelling of populations affected by food availability. Such ABMs are already being used to guide conservation planning of nature reserves and shell fisheries, to assess environmental impacts of building proposals including wind farms and highways and to assess the effects on nontarget organisms of chemicals for the control of agricultural pests. Keywords: bioenergetics; energy budget; individual-based models; population dynamics.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Body surface temperature can be used to evaluate thermal equilibrium in animals. The bodies of broiler chickens, like those of all birds, are partially covered by feathers. Thus, the heat flow at the boundary layer between broilers' bodies and the environment differs between feathered and featherless areas. The aim of this investigation was to use linear regression models incorporating environmental parameters and age to predict the surface temperatures of the feathered and featherless areas of broiler chickens. The trial was conducted in a climate chamber, and 576 broilers were distributed in two groups. In the first trial, 288 broilers were monitored after exposure to comfortable or stressful conditions during a 6-week rearing period. Another 288 broilers were measured under the same conditions to test the predictive power of the models. Sensible heat flow was calculated, and for the regions covered by feathers, sensible heat flow was predicted based on the estimated surface temperatures. The surface temperatures of the feathered and featherless areas can be predicted based on air, black globe or operative temperatures. According to the sensible heat flow model, the broilers' ability to maintain thermal equilibrium by convection and radiation decreased during the rearing period. Sensible heat flow estimated based on estimated surface temperatures can be used to predict animal responses to comfortable and stressful conditions. © 2013 ISB.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Estimates of evapotranspiration on a local scale is important information for agricultural and hydrological practices. However, equations to estimate potential evapotranspiration based only on temperature data, which are simple to use, are usually less trustworthy than the Food and Agriculture Organization (FAO)Penman-Monteith standard method. The present work describes two correction procedures for potential evapotranspiration estimates by temperature, making the results more reliable. Initially, the standard FAO-Penman-Monteith method was evaluated with a complete climatologic data set for the period between 2002 and 2006. Then temperature-based estimates by Camargo and Jensen-Haise methods have been adjusted by error autocorrelation evaluated in biweekly and monthly periods. In a second adjustment, simple linear regression was applied. The adjusted equations have been validated with climatic data available for the Year 2001. Both proposed methodologies showed good agreement with the standard method indicating that the methodology can be used for local potential evapotranspiration estimates.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The identification of the primary drivers of stock returns has been of great interest to both financial practitioners and academics alike for many decades. Influenced by classical financial theories such as the CAPM (Sharp, 1964; Lintner, 1965) and APT (Ross, 1976), a linear relationship is conventionally assumed between company characteristics as derived from their financial accounts and forward returns. Whilst this assumption may be a fair approximation to the underlying structural relationship, it is often adopted for the purpose of convenience. It is actually quite rare that the assumptions of distributional normality and a linear relationship are explicitly assessed in advance even though this information would help to inform the appropriate choice of modelling technique. Non-linear models have nevertheless been applied successfully to the task of stock selection in the past (Sorensen et al, 2000). However, their take-up by the investment community has been limited despite the fact that researchers in other fields have found them to be a useful way to express knowledge and aid decision-making...

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Assessing and prioritising cost-effective strategies to mitigate the impacts of traffic incidents and accidents on non-recurrent congestion on major roads represents a significant challenge for road network managers. This research examines the influence of numerous factors associated with incidents of various types on their duration. It presents a comprehensive traffic incident data mining and analysis by developing an incident duration model based on twelve months of incident data obtained from the Australian freeway network. Parametric accelerated failure time (AFT) survival models of incident duration were developed, including log-logistic, lognormal, and Weibul-considering both fixed and random parameters, as well as a Weibull model with gamma heterogeneity. The Weibull AFT models with random parameters were appropriate for modelling incident duration arising from crashes and hazards. A Weibull model with gamma heterogeneity was most suitable for modelling incident duration of stationary vehicles. Significant variables affecting incident duration include characteristics of the incidents (severity, type, towing requirements, etc.), and location, time of day, and traffic characteristics of the incident. Moreover, the findings reveal no significant effects of infrastructure and weather on incident duration. A significant and unique contribution of this paper is that the durations of each type of incident are uniquely different and respond to different factors. The results of this study are useful for traffic incident management agencies to implement strategies to reduce incident duration, leading to reduced congestion, secondary incidents, and the associated human and economic losses.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Agent-based modelling (ABM), like other modelling techniques, is used to answer specific questions from real world systems that could otherwise be expensive or impractical. Its recent gain in popularity can be attributed to some degree to its capacity to use information at a fine level of detail of the system, both geographically and temporally, and generate information at a higher level, where emerging patterns can be observed. This technique is data-intensive, as explicit data at a fine level of detail is used and it is computer-intensive as many interactions between agents, which can learn and have a goal, are required. With the growing availability of data and the increase in computer power, these concerns are however fading. Nonetheless, being able to update or extend the model as more information becomes available can become problematic, because of the tight coupling of the agents and their dependence on the data, especially when modelling very large systems. One large system to which ABM is currently applied is the electricity distribution where thousands of agents representing the network and the consumers’ behaviours are interacting with one another. A framework that aims at answering a range of questions regarding the potential evolution of the grid has been developed and is presented here. It uses agent-based modelling to represent the engineering infrastructure of the distribution network and has been built with flexibility and extensibility in mind. What distinguishes the method presented here from the usual ABMs is that this ABM has been developed in a compositional manner. This encompasses not only the software tool, which core is named MODAM (MODular Agent-based Model) but the model itself. Using such approach enables the model to be extended as more information becomes available or modified as the electricity system evolves, leading to an adaptable model. Two well-known modularity principles in the software engineering domain are information hiding and separation of concerns. These principles were used to develop the agent-based model on top of OSGi and Eclipse plugins which have good support for modularity. Information regarding the model entities was separated into a) assets which describe the entities’ physical characteristics, and b) agents which describe their behaviour according to their goal and previous learning experiences. This approach diverges from the traditional approach where both aspects are often conflated. It has many advantages in terms of reusability of one or the other aspect for different purposes as well as composability when building simulations. For example, the way an asset is used on a network can greatly vary while its physical characteristics are the same – this is the case for two identical battery systems which usage will vary depending on the purpose of their installation. While any battery can be described by its physical properties (e.g. capacity, lifetime, and depth of discharge), its behaviour will vary depending on who is using it and what their aim is. The model is populated using data describing both aspects (physical characteristics and behaviour) and can be updated as required depending on what simulation is to be run. For example, data can be used to describe the environment to which the agents respond to – e.g. weather for solar panels, or to describe the assets and their relation to one another – e.g. the network assets. Finally, when running a simulation, MODAM calls on its module manager that coordinates the different plugins, automates the creation of the assets and agents using factories, and schedules their execution which can be done sequentially or in parallel for faster execution. Building agent-based models in this way has proven fast when adding new complex behaviours, as well as new types of assets. Simulations have been run to understand the potential impact of changes on the network in terms of assets (e.g. installation of decentralised generators) or behaviours (e.g. response to different management aims). While this platform has been developed within the context of a project focussing on the electricity domain, the core of the software, MODAM, can be extended to other domains such as transport which is part of future work with the addition of electric vehicles.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Social networking sites (SNSs), with their large numbers of users and large information base, seem to be perfect breeding grounds for exploiting the vulnerabilities of people, the weakest link in security. Deceiving, persuading, or influencing people to provide information or to perform an action that will benefit the attacker is known as “social engineering.” While technology-based security has been addressed by research and may be well understood, social engineering is more challenging to understand and manage, especially in new environments such as SNSs, owing to some factors of SNSs that reduce the ability of users to detect the attack and increase the ability of attackers to launch it. This work will contribute to the knowledge of social engineering by presenting the first two conceptual models of social engineering attacks in SNSs. Phase-based and source-based models are presented, along with an intensive and comprehensive overview of different aspects of social engineering threats in SNSs.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Agent-based modeling and simulation (ABMS) may fit well with entrepreneurship research and practice because the core concepts and basic premises of entrepreneurship coincide with the characteristics of ABMS. However, it is difficult to find cases where ABMS is applied to entrepreneurship research. To apply ABMS to entrepreneurship and organization studies, designing a conceptual model is important; thus to effectively design a conceptual model, various mixed method approaches are being attempted. As a new mixed method approach to ABMS, this study proposes a bibliometric approach to designing agent based models, which establishes and analyzes a domain corpus. This study presents an example on the venture creation process using the bibliometric approach. This example shows us that the results of the multi-agent simulations on the venturing process based on the bibliometric approach are close to each nation’s surveyed data on the venturing activities. In conclusion, by the bibliometric approach proposed in this study, all the agents and the agents’ behaviors related to a phenomenon can be extracted effectively, and a conceptual model for ABMS can be designed with the agents and their behaviors. This study contributes to the entrepreneurship and organization studies by promoting the application of ABMS.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

PURPOSE: This paper describes dynamic agent composition, used to support the development of flexible and extensible large-scale agent-based models (ABMs). This approach was motivated by a need to extend and modify, with ease, an ABM with an underlying networked structure as more information becomes available. Flexibility was also sought after so that simulations are set up with ease, without the need to program. METHODS: The dynamic agent composition approach consists in having agents, whose implementation has been broken into atomic units, come together at runtime to form the complex system representation on which simulations are run. These components capture information at a fine level of detail and provide a vast range of combinations and options for a modeller to create ABMs. RESULTS: A description of the dynamic agent composition is given in this paper, as well as details about its implementation within MODAM (MODular Agent-based Model), a software framework which is applied to the planning of the electricity distribution network. Illustrations of the implementation of the dynamic agent composition are consequently given for that domain throughout the paper. It is however expected that this approach will be beneficial to other problem domains, especially those with a networked structure, such as water or gas networks. CONCLUSIONS: Dynamic agent composition has many advantages over the way agent-based models are traditionally built for the users, the developers, as well as for agent-based modelling as a scientific approach. Developers can extend the model without the need to access or modify previously written code; they can develop groups of entities independently and add them to those already defined to extend the model. Users can mix-and-match already implemented components to form large-scales ABMs, allowing them to quickly setup simulations and easily compare scenarios without the need to program. The dynamic agent composition provides a natural simulation space over which ABMs of networked structures are represented, facilitating their implementation; and verification and validation of models is facilitated by quickly setting up alternative simulations.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This thesis presents a novel approach to building large-scale agent-based models of networked physical systems using a compositional approach to provide extensibility and flexibility in building the models and simulations. A software framework (MODAM - MODular Agent-based Model) was implemented for this purpose, and validated through simulations. These simulations allow assessment of the impact of technological change on the electricity distribution network looking at the trajectories of electricity consumption at key locations over many years.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Laboratory-based relationships that model the phytotoxicity of metals using soil properties have been developed. This paper presents the first field-based phytotoxicity relationships. Wheat(Triticum aestivum L) was grown at 11 Australian field sites at which soil was spiked with copper (Cu) and zinc (Zn) salts. Toxicity was measured as inhibition of plant growth at 8 weeks and grain yield at harvest. The added Cu and Zn EC10 values for both endpoints ranged from approximately 3 to 4760 mg/kg. There were no relationships between field-based 8-week biomass and grain yield toxicity values for either metal. Cu toxicity was best modelled using pH and organic carbon content while Zn toxicity was best modelled using pH and the cation exchange capacity. The best relationships estimated toxicity within a factor of two of measured values. Laboratory-based phytotoxicity relationships could not accurately predict field-based phytotoxicity responses.