889 resultados para Multiple-trait model
Resumo:
A mathematical model incorporating many of the important processes at work in the crystallization of emulsions is presented. The model describes nucleation within the discontinuous domain of an emulsion, precipitation in the continuous domain, transport of monomers between the two domains, and formation and subsequent growth of crystals in both domains. The model is formulated as an autonomous system of nonlinear, coupled ordinary differential equations. The description of nucleation and precipitation is based upon the Becker–Döring equations of classical nucleation theory. A particular feature of the model is that the number of particles of all species present is explicitly conserved; this differs from work that employs Arrhenius descriptions of nucleation rate. Since the model includes many physical effects, it is analyzed in stages so that the role of each process may be understood. When precipitation occurs in the continuous domain, the concentration of monomers falls below the equilibrium concentration at the surface of the drops of the discontinuous domain. This leads to a transport of monomers from the drops into the continuous domain that are then incorporated into crystals and nuclei. Since the formation of crystals is irreversible and their subsequent growth inevitable, crystals forming in the continuous domain effectively act as a sink for monomers “sucking” monomers from the drops. In this case, numerical calculations are presented which are consistent with experimental observations. In the case in which critical crystal formation does not occur, the stationary solution is found and a linear stability analysis is performed. Bifurcation diagrams describing the loci of stationary solutions, which may be multiple, are numerically calculated.
Resumo:
This paper aims to design a collaboration model for a Knowledge Community - SSMEnetUK. The research identifies SSMEnetUK as a socio-technical system and uses the core concepts of Service Science to explore the subject domain. The paper is positioned within the concept of Knowledge Management (KM) and utilising Web 2.0 tools for collaboration. A qualitative case study method was adopted and multiple data sources were used. In achieving that, the degree of co-relation between knowledge management activities and Web 2.0 tools for collaboration in the scenario are pitted against the concept of value propositions offered by both customer/user and service provider. The proposed model provides a better understanding of how Knowledge Management and Web 2.0 tools can enable effective collaboration within SSMEnetUK. This research is relevant to the wider service design and innovation community because it provides a basis for building a service-centric collaboration platform for the benefit of both customer/user and service provider.
Resumo:
In this paper ensembles of forecasts (of up to six hours) are studied from a convection-permitting model with a representation of model error due to unresolved processes. The ensemble prediction system (EPS) used is an experimental convection-permitting version of the UK Met Office’s 24- member Global and Regional Ensemble Prediction System (MOGREPS). The method of representing model error variability, which perturbs parameters within the model’s parameterisation schemes, has been modified and we investigate the impact of applying this scheme in different ways. These are: a control ensemble where all ensemble members have the same parameter values; an ensemble where the parameters are different between members, but fixed in time; and ensembles where the parameters are updated randomly every 30 or 60 min. The choice of parameters and their ranges of variability have been determined from expert opinion and parameter sensitivity tests. A case of frontal rain over the southern UK has been chosen, which has a multi-banded rainfall structure. The consequences of including model error variability in the case studied are mixed and are summarised as follows. The multiple banding, evident in the radar, is not captured for any single member. However, the single band is positioned in some members where a secondary band is present in the radar. This is found for all ensembles studied. Adding model error variability with fixed parameters in time does increase the ensemble spread for near-surface variables like wind and temperature, but can actually decrease the spread of the rainfall. Perturbing the parameters periodically throughout the forecast does not further increase the spread and exhibits “jumpiness” in the spread at times when the parameters are perturbed. Adding model error variability gives an improvement in forecast skill after the first 2–3 h of the forecast for near-surface temperature and relative humidity. For precipitation skill scores, adding model error variability has the effect of improving the skill in the first 1–2 h of the forecast, but then of reducing the skill after that. Complementary experiments were performed where the only difference between members was the set of parameter values (i.e. no initial condition variability). The resulting spread was found to be significantly less than the spread from initial condition variability alone.
Resumo:
The Lincoln–Petersen estimator is one of the most popular estimators used in capture–recapture studies. It was developed for a sampling situation in which two sources independently identify members of a target population. For each of the two sources, it is determined if a unit of the target population is identified or not. This leads to a 2 × 2 table with frequencies f11, f10, f01, f00 indicating the number of units identified by both sources, by the first but not the second source, by the second but not the first source and not identified by any of the two sources, respectively. However, f00 is unobserved so that the 2 × 2 table is incomplete and the Lincoln–Petersen estimator provides an estimate for f00. In this paper, we consider a generalization of this situation for which one source provides not only a binary identification outcome but also a count outcome of how many times a unit has been identified. Using a truncated Poisson count model, truncating multiple identifications larger than two, we propose a maximum likelihood estimator of the Poisson parameter and, ultimately, of the population size. This estimator shows benefits, in comparison with Lincoln–Petersen’s, in terms of bias and efficiency. It is possible to test the homogeneity assumption that is not testable in the Lincoln–Petersen framework. The approach is applied to surveillance data on syphilis from Izmir, Turkey.
Resumo:
There has been recent interest in sensory systems that are able to display a response which is proportional to a fold change in stimulus concentration, a feature referred to as fold-change detection (FCD). Here, we demonstrate FCD in a recent whole-pathway mathematical model of Escherichia coli chemotaxis. FCD is shown to hold for each protein in the signalling cascade and to be robust to kinetic rate and protein concentration variation. Using a sensitivity analysis, we find that only variations in the number of receptors within a signalling team lead to the model not exhibiting FCD. We also discuss the ability of a cell with multiple receptor types to display FCD and explain how a particular receptor configuration may be used to elucidate the two experimentally determined regimes of FCD behaviour. All findings are discussed in respect of the experimental literature.
Resumo:
Unorganized traffic is a generalized form of travel wherein vehicles do not adhere to any predefined lanes and can travel in-between lanes. Such travel is visible in a number of countries e.g. India, wherein it enables a higher traffic bandwidth, more overtaking and more efficient travel. These advantages are visible when the vehicles vary considerably in size and speed, in the absence of which the predefined lanes are near-optimal. Motion planning for multiple autonomous vehicles in unorganized traffic deals with deciding on the manner in which every vehicle travels, ensuring no collision either with each other or with static obstacles. In this paper the notion of predefined lanes is generalized to model unorganized travel for the purpose of planning vehicles travel. A uniform cost search is used for finding the optimal motion strategy of a vehicle, amidst the known travel plans of the other vehicles. The aim is to maximize the separation between the vehicles and static obstacles. The search is responsible for defining an optimal lane distribution among vehicles in the planning scenario. Clothoid curves are used for maintaining a lane or changing lanes. Experiments are performed by simulation over a set of challenging scenarios with a complex grid of obstacles. Additionally behaviours of overtaking, waiting for a vehicle to cross and following another vehicle are exhibited.
Resumo:
he first international urban land surface model comparison was designed to identify three aspects of the urban surface-atmosphere interactions: (1) the dominant physical processes, (2) the level of complexity required to model these, and 3) the parameter requirements for such a model. Offline simulations from 32 land surface schemes, with varying complexity, contributed to the comparison. Model results were analysed within a framework of physical classifications and over four stages. The results show that the following are important urban processes; (i) multiple reflections of shortwave radiation within street canyons, (ii) reduction in the amount of visible sky from within the canyon, which impacts on the net long-wave radiation, iii) the contrast in surface temperatures between building roofs and street canyons, and (iv) evaporation from vegetation. Models that use an appropriate bulk albedo based on multiple solar reflections, represent building roof surfaces separately from street canyons and include a representation of vegetation demonstrate more skill, but require parameter information on the albedo, height of the buildings relative to the width of the streets (height to width ratio), the fraction of building roofs compared to street canyons from a plan view (plan area fraction) and the fraction of the surface that is vegetated. These results, whilst based on a single site and less than 18 months of data, have implications for the future design of urban land surface models, the data that need to be measured in urban observational campaigns, and what needs to be included in initiatives for regional and global parameter databases.
Resumo:
Earthworms are significant ecosystem engineers and are an important component of the diet of many vertebrates and invertebrates, so the ability to predict their distribution and abundance would have wide application in ecology, conservation and land management. Earthworm viability is known to be affected by the availability and quality of food resources, soil water conditions and temperature, but has not yet been modelled mechanistically to link effects on individuals to field population responses. Here we present a novel model capable of predicting the effects of land management and environmental conditions on the distribution and abundance of Aporrectodea caliginosa, the dominant earthworm species in agroecosystems. Our process-based approach uses individual based modelling (IBM), in which each individual has its own energy budget. Individual earthworm energy budgets follow established principles of physiological ecology and are parameterised for A. caliginosa from experimental measurements under optimal conditions. Under suboptimal conditions (e.g. food limitation, low soil temperatures and water contents) reproduction is prioritised over growth. Good model agreement to independent laboratory data on individual cocoon production and growth of body mass, under variable feeding and temperature conditions support our representation of A. caliginosa physiology through energy budgets. Our mechanistic model is able to accurately predict A. caliginosa distribution and abundance in spatially heterogeneous soil profiles representative of field study conditions. Essential here is the explicit modelling of earthworm behaviour in the soil profile. Local earthworm movement responds to a trade-off between food availability and soil water conditions, and this determines the spatiotemporal distribution of the population in the soil profile. Importantly, multiple environmental variables can be manipulated simultaneously in the model to explore earthworm population exposure and effects to combinations of stressors. Potential applications include prediction of the population-level effects of pesticides and changes in soil management e.g. conservation tillage and climate change.
Resumo:
Abstract We present a refined parametric model for forecasting electricity demand which performed particularly well in the recent Global Energy Forecasting Competition (GEFCom 2012). We begin by motivating and presenting a simple parametric model, treating the electricity demand as a function of the temperature and day of the data. We then set out a series of refinements of the model, explaining the rationale for each, and using the competition scores to demonstrate that each successive refinement step increases the accuracy of the model’s predictions. These refinements include combining models from multiple weather stations, removing outliers from the historical data, and special treatments of public holidays.
Resumo:
Many studies evaluating model boundary-layer schemes focus either on near-surface parameters or on short-term observational campaigns. This reflects the observational datasets that are widely available for use in model evaluation. In this paper we show how surface and long-term Doppler lidar observations, combined in a way to match model representation of the boundary layer as closely as possible, can be used to evaluate the skill of boundary-layer forecasts. We use a 2-year observational dataset from a rural site in the UK to evaluate a climatology of boundary layer type forecast by the UK Met Office Unified Model. In addition, we demonstrate the use of a binary skill score (Symmetric Extremal Dependence Index) to investigate the dependence of forecast skill on season, horizontal resolution and forecast leadtime. A clear diurnal and seasonal cycle can be seen in the climatology of both the model and observations, with the main discrepancies being the model overpredicting cumulus capped and decoupled stratocumulus capped boundary-layers and underpredicting well mixed boundary-layers. Using the SEDI skill score the model is most skillful at predicting the surface stability. The skill of the model in predicting cumulus capped and stratocumulus capped stable boundary layer forecasts is low but greater than a 24 hr persistence forecast. In contrast, the prediction of decoupled boundary-layers and boundary-layers with multiple cloud layers is lower than persistence. This process based evaluation approach has the potential to be applied to other boundary-layer parameterisation schemes with similar decision structures.
Resumo:
We present a new parameterisation that relates surface mass balance (SMB: the sum of surface accumulation and surface ablation) to changes in surface elevation of the Greenland ice sheet (GrIS) for the MAR (Modèle Atmosphérique Régional: Fettweis, 2007) regional climate model. The motivation is to dynamically adjust SMB as the GrIS evolves, allowing us to force ice sheet models with SMB simulated by MAR while incorporating the SMB–elevation feedback, without the substantial technical challenges of coupling ice sheet and climate models. This also allows us to assess the effect of elevation feedback uncertainty on the GrIS contribution to sea level, using multiple global climate and ice sheet models, without the need for additional, expensive MAR simulations. We estimate this relationship separately below and above the equilibrium line altitude (ELA, separating negative and positive SMB) and for regions north and south of 77� N, from a set of MAR simulations in which we alter the ice sheet surface elevation. These give four “SMB lapse rates”, gradients that relate SMB changes to elevation changes. We assess uncertainties within a Bayesian framework, estimating probability distributions for each gradient from which we present best estimates and credibility intervals (CI) that bound 95% of the probability. Below the ELA our gradient estimates are mostly positive, because SMB usually increases with elevation: 0.56 (95% CI: −0.22 to 1.33) kgm−3 a−1 for the north, and 1.91 (1.03 to 2.61) kgm−3 a−1 for the south. Above the ELA, the gradients are much smaller in magnitude: 0.09 (−0.03 to 0.23) kgm−3 a−1 in the north, and 0.07 (−0.07 to 0.59) kgm−3 a−1 in the south, because SMB can either increase or decrease in response to increased elevation. Our statistically founded approach allows us to make probabilistic assessments for the effect of elevation feedback uncertainty on sea level projections (Edwards et al., 2014).
Resumo:
In this paper we report coordinated multispacecraft and ground-based observations of a double substorm onset close to Scandinavia on November 17, 1996. The Wind and the Geotail spacecraft, which were located in the solar wind and the subsolar magnetosheath, respectively, recorded two periods of southward directed interplanetary magnetic field (IMF). These periods were separated by a short northward IMF excursion associated with a solar wind pressure pulse, which compressed the magnetosphere to such a degree that Geotail for a short period was located outside the bow shock. The first period of southward IMF initiated a substorm growth. phase, which was clearly detected by an array of ground-based instrumentation and by Interball in the northern tail lobe. A first substorm onset occurred in close relation to the solar wind pressure pulse impinging on the magnetopause and almost simultaneously with the northward turning of the IMF. However, this substorm did not fully develop. In clear association with the expansion of the magnetosphere at the end of the pressure pulse, the auroral expansion was stopped, and the northern sky cleared. We will present evidence that the change in the solar wind dynamic pressure actively quenched the energy available for any further substorm expansion. Directly after this period, the magnetometer network detected signatures of a renewed substorm growth phase, which was initiated by the second southward turning of the IMF and which finally lead to a second, and this time complete, substorm intensification. We have used our multipoint observations in order to understand the solar wind control of the substorm onset and substorm quenching. The relative timings between the observations on the various satellites and on the ground were used to infer a possible causal relationship between the solar wind pressure variations and consequent substorm development. Furthermore, using a relatively simple algorithm to model the tail lobe field and the total tail flux, we show that there indeed exists a close relationship between the relaxation of a solar wind pressure pulse, the reduction of the tail lobe field, and the quenching of the initial substorm.
Resumo:
We present a simple, generic model of annual tree growth, called "T". This model accepts input from a first-principles light-use efficiency model (the "P" model). The P model provides values for gross primary production (GPP) per unit of absorbed photosynthetically active radiation (PAR). Absorbed PAR is estimated from the current leaf area. GPP is allocated to foliage, transport tissue, and fine-root production and respiration in such a way as to satisfy well-understood dimensional and functional relationships. Our approach thereby integrates two modelling approaches separately developed in the global carbon-cycle and forest-science literature. The T model can represent both ontogenetic effects (the impact of ageing) and the effects of environmental variations and trends (climate and CO2) on growth. Driven by local climate records, the model was applied to simulate ring widths during the period 1958–2006 for multiple trees of Pinus koraiensis from the Changbai Mountains in northeastern China. Each tree was initialised at its actual diameter at the time when local climate records started. The model produces realistic simulations of the interannual variability in ring width for different age cohorts (young, mature, and old). Both the simulations and observations show a significant positive response of tree-ring width to growing-season total photosynthetically active radiation (PAR0) and the ratio of actual to potential evapotranspiration (α), and a significant negative response to mean annual temperature (MAT). The slopes of the simulated and observed relationships with PAR0 and α are similar; the negative response to MAT is underestimated by the model. Comparison of simulations with fixed and changing atmospheric CO2 concentration shows that CO2 fertilisation over the past 50 years is too small to be distinguished in the ring-width data, given ontogenetic trends and interannual variability in climate.
Resumo:
Traditional dictionary learning algorithms are used for finding a sparse representation on high dimensional data by transforming samples into a one-dimensional (1D) vector. This 1D model loses the inherent spatial structure property of data. An alternative solution is to employ Tensor Decomposition for dictionary learning on their original structural form —a tensor— by learning multiple dictionaries along each mode and the corresponding sparse representation in respect to the Kronecker product of these dictionaries. To learn tensor dictionaries along each mode, all the existing methods update each dictionary iteratively in an alternating manner. Because atoms from each mode dictionary jointly make contributions to the sparsity of tensor, existing works ignore atoms correlations between different mode dictionaries by treating each mode dictionary independently. In this paper, we propose a joint multiple dictionary learning method for tensor sparse coding, which explores atom correlations for sparse representation and updates multiple atoms from each mode dictionary simultaneously. In this algorithm, the Frequent-Pattern Tree (FP-tree) mining algorithm is employed to exploit frequent atom patterns in the sparse representation. Inspired by the idea of K-SVD, we develop a new dictionary update method that jointly updates elements in each pattern. Experimental results demonstrate our method outperforms other tensor based dictionary learning algorithms.
Resumo:
Break crops and multi-crop rotations are common in arable farm management, and the soil quality inherited from a previous crop is one of the parameters that determine the gross margin that is achieved with a given crop from a given parcel of land. In previous work we developed a dynamic economic model to calculate the potential yield and gross margin of a set of crops grown in a selection of typical rotation scenarios, and we reported use of the model to calculate coexistence costs for GM maize grown in a crop rotation. The model predicts economic effects of pest and weed pressures in monthly time steps. Validation of the model in respect of specific traits is proceeding as data from trials with novel crop varieties is published. Alongside this aspect of the validation process, we are able to incorporate data representing the economic impact of abiotic stresses on conventional crops, and then use the model to predict the cumulative gross margin achievable from a sequence of conventional crops grown at varying levels of abiotic stress. We report new progress with this aspect of model validation. In this paper, we report the further development of the model to take account of abiotic stress arising from drought, flood, heat or frost; such stresses being introduced in addition to variable pest and weed pressure. The main purpose is to assess the economic incentive for arable farmers to adopt novel crop varieties having multiple ‘stacked’ traits introduced by means of various biotechnological tools available to crop breeders.