112 resultados para transformation parameter

em CentAUR: Central Archive University of Reading - UK


Relevância:

40.00% 40.00%

Publicador:

Resumo:

The work reported in this paper is motivated by biomimetic inspiration - the transformation of patterns. The major issue addressed is the development of feasible methods for transformation based on a macroscopic tool. The general requirement for the feasibility of the transformation method is determined by classifying pattern formation approaches an their characteristics. A formal definition for pattern transformation is provided and four special cases namely, elementary and geometric transformation based on repositioning all and some robotic agents are introduced. A feasible method for transforming patterns geometrically, based on the macroscopic parameter operation of a swarm is considered. The transformation method is applied to a swarm model which lends itself to the transformation technique. Simulation studies are developed to validate the feasibility of the approach, and do indeed confirm the approach.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper identifies the major challenges in the area of pattern formation. The work is also motivated by the need for development of a single framework to surmount these challenges. A framework based on the control of macroscopic parameters is proposed. The issue of transformation of patterns is specifically considered. A definition for transformation and four special cases, namely elementary and geometrical transformations by repositioning all or some robots in the pattern are provided. Two feasible tools for pattern transformation namely, a macroscopic parameter method and a mathematical tool - Moebius transformation also known as the linear fractional transformation are introduced. The realization of the unifying framework considering planning and communication is reported.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The ITCT-Lagrangian-2K4 (Intercontinental Transport and Chemical Transformation) experiment was conceived with an aim to quantify the effects of photochemistry and mixing on the transformation of air masses in the free troposphere away from emissions. To this end, attempts were made to intercept and sample air masses several times during their journey across the North Atlantic using four aircraft based in New Hampshire (USA), Faial (Azores) and Creil (France). This article begins by describing forecasts from two Lagrangian models that were used to direct the aircraft into target air masses. A novel technique then identifies Lagrangian matches between flight segments. Two independent searches are conducted: for Lagrangian model matches and for pairs of whole air samples with matching hydrocarbon fingerprints. The information is filtered further by searching for matching hydrocarbon samples that are linked by matching trajectories. The quality of these "coincident matches'' is assessed using temperature, humidity and tracer observations. The technique pulls out five clear Lagrangian cases covering a variety of situations and these are examined in detail. The matching trajectories and hydrocarbon fingerprints are shown, and the downwind minus upwind differences in tracers are discussed.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Bayesian inference has been used to determine rigorous estimates of hydroxyl radical concentrations () and air mass dilution rates (K) averaged following air masses between linked observations of nonmethane hydrocarbons (NMHCs) spanning the North Atlantic during the Intercontinental Transport and Chemical Transformation (ITCT)-Lagrangian-2K4 experiment. The Bayesian technique obtains a refined (posterior) distribution of a parameter given data related to the parameter through a model and prior beliefs about the parameter distribution. Here, the model describes hydrocarbon loss through OH reaction and mixing with a background concentration at rate K. The Lagrangian experiment provides direct observations of hydrocarbons at two time points, removing assumptions regarding composition or sources upstream of a single observation. The estimates are sharpened by using many hydrocarbons with different reactivities and accounting for their variability and measurement uncertainty. A novel technique is used to construct prior background distributions of many species, described by variation of a single parameter . This exploits the high correlation of species, related by the first principal component of many NMHC samples. The Bayesian method obtains posterior estimates of , K and following each air mass. Median values are typically between 0.5 and 2.0 × 106 molecules cm−3, but are elevated to between 2.5 and 3.5 × 106 molecules cm−3, in low-level pollution. A comparison of estimates from absolute NMHC concentrations and NMHC ratios assuming zero background (the “photochemical clock” method) shows similar distributions but reveals systematic high bias in the estimates from ratios. Estimates of K are ∼0.1 day−1 but show more sensitivity to the prior distribution assumed.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Data assimilation is a sophisticated mathematical technique for combining observational data with model predictions to produce state and parameter estimates that most accurately approximate the current and future states of the true system. The technique is commonly used in atmospheric and oceanic modelling, combining empirical observations with model predictions to produce more accurate and well-calibrated forecasts. Here, we consider a novel application within a coastal environment and describe how the method can also be used to deliver improved estimates of uncertain morphodynamic model parameters. This is achieved using a technique known as state augmentation. Earlier applications of state augmentation have typically employed the 4D-Var, Kalman filter or ensemble Kalman filter assimilation schemes. Our new method is based on a computationally inexpensive 3D-Var scheme, where the specification of the error covariance matrices is crucial for success. A simple 1D model of bed-form propagation is used to demonstrate the method. The scheme is capable of recovering near-perfect parameter values and, therefore, improves the capability of our model to predict future bathymetry. Such positive results suggest the potential for application to more complex morphodynamic models.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A new dynamic model of water quality, Q(2), has recently been developed, capable of simulating large branched river systems. This paper describes the application of a generalized sensitivity analysis (GSA) to Q(2) for single reaches of the River Thames in southern England. Focusing on the simulation of dissolved oxygen (DO) (since this may be regarded as a proxy for the overall health of a river); the GSA is used to identify key parameters controlling model behavior and provide a probabilistic procedure for model calibration. It is shown that, in the River Thames at least, it is more important to obtain high quality forcing functions than to obtain improved parameter estimates once approximate values have been estimated. Furthermore, there is a need to ensure reasonable simulation of a range of water quality determinands, since a focus only on DO increases predictive uncertainty in the DO simulations. The Q(2) model has been applied here to the River Thames, but it has a broad utility for evaluating other systems in Europe and around the world.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

There are now considerable expectations that semi-distributed models are useful tools for supporting catchment water quality management. However, insufficient attention has been given to evaluating the uncertainties inherent to this type of model, especially those associated with the spatial disaggregation of the catchment. The Integrated Nitrogen in Catchments model (INCA) is subjected to an extensive regionalised sensitivity analysis in application to the River Kennet, part of the groundwater-dominated upper Thames catchment, UK The main results are: (1) model output was generally insensitive to land-phase parameters, very sensitive to groundwater parameters, including initial conditions, and significantly sensitive to in-river parameters; (2) INCA was able to produce good fits simultaneously to the available flow, nitrate and ammonium in-river data sets; (3) representing parameters as heterogeneous over the catchment (206 calibrated parameters) rather than homogeneous (24 calibrated parameters) produced a significant improvement in fit to nitrate but no significant improvement to flow and caused a deterioration in ammonium performance; (4) the analysis indicated that calibrating the flow-related parameters first, then calibrating the remaining parameters (as opposed to calibrating all parameters together) was not a sensible strategy in this case; (5) even the parameters to which the model output was most sensitive suffered from high uncertainty due to spatial inconsistencies in the estimated optimum values, parameter equifinality and the sampling error associated with the calibration method; (6) soil and groundwater nutrient and flow data are needed to reduce. uncertainty in initial conditions, residence times and nitrogen transformation parameters, and long-term historic data are needed so that key responses to changes in land-use management can be assimilated. The results indicate the general, difficulty of reconciling the questions which catchment nutrient models are expected to answer with typically limited data sets and limited knowledge about suitable model structures. The results demonstrate the importance of analysing semi-distributed model uncertainties prior to model application, and illustrate the value and limitations of using Monte Carlo-based methods for doing so. (c) 2005 Elsevier B.V. All rights reserved.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

How do organizations previously dominated by the state develop dynamic capabilities that would support their growth in a competitive market economy? We develop a theoretical framework of organizational transformation that explains the processes by which organizations learn and develop dynamic capabilities in transition economies. Specifically, the framework theorizes about the importance of, and inter-relationships between, leadership, organizational learning, dynamic capabilities, and performance over three stages of transformation. Propositions derived from this framework explain the pre-conditions enabling organizational learning, the linkages between types of learning and functions of dynamic capabilities, and the feedback from dynamic capabilities to organizational learning that allows firms in transition economies to regain their footing and build long-term competitive advantage. We focus on transition contexts, where these processes have been magnified and thus offer new insights into strategizing in radically altered environments.