978 resultados para Mixed binary nonlinear programming
Resumo:
1. Establishing biological control agents in the field is a major step in any classical biocontrol programme, yet there are few general guidelines to help the practitioner decide what factors might enhance the establishment of such agents. 2. A stochastic dynamic programming (SDP) approach, linked to a metapopulation model, was used to find optimal release strategies (number and size of releases), given constraints on time and the number of biocontrol agents available. By modelling within a decision-making framework we derived rules of thumb that will enable biocontrol workers to choose between management options, depending on the current state of the system. 3. When there are few well-established sites, making a few large releases is the optimal strategy. For other states of the system, the optimal strategy ranges from a few large releases, through a mixed strategy (a variety of release sizes), to many small releases, as the probability of establishment of smaller inocula increases. 4. Given that the probability of establishment is rarely a known entity, we also strongly recommend a mixed strategy in the early stages of a release programme, to accelerate learning and improve the chances of finding the optimal approach.
Resumo:
We examine a problem with n players each facing the same binary choice. One choice is superior to the other. The simple assumption of competition - that an individual's payoff falls with a rise in the number of players making the same choice, guarantees the existence of a unique symmetric equilibrium (involving mixed strategies). As n increases, there are two opposing effects. First, events in the middle of the distribution - where a player finds itself having made the same choice as many others - become more likely, but the payoffs in these events fall. In opposition, events in the tails of the distribution - where a player finds itself having made the same choice as few others - become less likely, but the payoffs in these events remain high. We provide a sufficient condition (strong competition) under which an increase in the number of players leads to a reduction in the equilibrium probability that the superior choice is made.
Resumo:
In population pharmacokinetic studies, the precision of parameter estimates is dependent on the population design. Methods based on the Fisher information matrix have been developed and extended to population studies to evaluate and optimize designs. In this paper we propose simple programming tools to evaluate population pharmacokinetic designs. This involved the development of an expression for the Fisher information matrix for nonlinear mixed-effects models, including estimation of the variance of the residual error. We implemented this expression as a generic function for two software applications: S-PLUS and MATLAB. The evaluation of population designs based on two pharmacokinetic examples from the literature is shown to illustrate the efficiency and the simplicity of this theoretic approach. Although no optimization method of the design is provided, these functions can be used to select and compare population designs among a large set of possible designs, avoiding a lot of simulations.
Resumo:
The mechanism underlying segregation in liquid fluidized beds is investigated in this paper, A binary fluidized bed system not at a stable equilibrium condition. is modelled in the literature as forming a mixed part-corresponding to stable mixture-at the bottom of the bed and a pure layer of excess components always floating on the mixed part. On the basis of this model: (0 comprehensive criteria for binary particles of any type to mix/segregate, and (ii) mixing, segregation regime map in terms of size ratio and density ratio of the particles for a given fluidizing medium, are established in this work. Therefore, knowing the properties of given particles, a second type of particles can be chosen in order to avoid or to promote segregation according to the particular process requirements. The model is then advanced for multicomponent fluidized beds and validated against experimental results observed for ternary fluidized beds. (C) 2002 Elsevier Science B.V. All rights reserved.
Resumo:
A model for binary mixture adsorption accounting for energetic heterogeneity and intermolecular interactions is proposed in this paper. The model is based on statistical thermodynamics, and it is able to describe molecular rearrangement of a mixture in a nonuniform adsorption field inside a cavity. The Helmholtz free energy obtained in the framework of this approach has upper and lower limits, which define a permissible range in which all possible solutions will be found. One limit corresponds to a completely chaotic distribution of molecules within a cavity, while the other corresponds to a maximum ordered molecular structure. Comparison of the nearly ideal O-2-N-2-zeolite NaX system at ambient temperature with the system Of O-2-N-2-zeolite CaX at 144 K has shown that a decrease of temperature leads to a molecular rearrangement in the cavity volume, which results from the difference in the fluid-solid interactions. The model is able to describe this behavior and therefore allows predicting mixture adsorption more accurately compared to those assuming energetic uniformity of the adsorption volume. Another feature of the model is its ability to correctly describe the negative deviations from Raoult's law exhibited by the O-2-N-2-CaX system at 144 K. Analysis of the highly nonideal CO2-C2H6-zeolite NaX system has shown that the spatial molecular rearrangement in separate cavities is induced by not only the ion-quadrupole interaction of the CO2 molecule but also the significant difference in molecular size and the difference between the intermolecular interactions of molecules of the same species and those of molecules of different species. This leads to the highly ordered structure of this system.
Resumo:
We compare Bayesian methodology utilizing free-ware BUGS (Bayesian Inference Using Gibbs Sampling) with the traditional structural equation modelling approach based on another free-ware package, Mx. Dichotomous and ordinal (three category) twin data were simulated according to different additive genetic and common environment models for phenotypic variation. Practical issues are discussed in using Gibbs sampling as implemented by BUGS to fit subject-specific Bayesian generalized linear models, where the components of variation may be estimated directly. The simulation study (based on 2000 twin pairs) indicated that there is a consistent advantage in using the Bayesian method to detect a correct model under certain specifications of additive genetics and common environmental effects. For binary data, both methods had difficulty in detecting the correct model when the additive genetic effect was low (between 10 and 20%) or of moderate range (between 20 and 40%). Furthermore, neither method could adequately detect a correct model that included a modest common environmental effect (20%) even when the additive genetic effect was large (50%). Power was significantly improved with ordinal data for most scenarios, except for the case of low heritability under a true ACE model. We illustrate and compare both methods using data from 1239 twin pairs over the age of 50 years, who were registered with the Australian National Health and Medical Research Council Twin Registry (ATR) and presented symptoms associated with osteoarthritis occurring in joints of the hand.
Resumo:
We analyse the relationship between the privatization of a public firm and government preferences for tax revenue, by considering a (sequential) Stackelberg duopoly with the public firm as the leader. We assume that the government payoff is given by a weighted sum of tax revenue and the sum of consumer and producer surplus. We get that if the government puts a sufficiently larger weight on tax revenue than on the sum of both surpluses, it will not privatize the public firm. In contrast, if the government puts a moderately larger weight on tax revenue than on the sum of both surpluses, it will privatize the public firm. Furthermore, we compare our results with the ones previously published by an other author obtained in a (simultaneous) Cournot duopoly.
Resumo:
Binary operations on commutative Jordan algebras, CJA, can be used to study interactions between sets of factors belonging to a pair of models in which one nests the other. It should be noted that from two CJA we can, through these binary operations, build CJA. So when we nest the treatments from one model in each treatment of another model, we can study the interactions between sets of factors of the first and the second models.
Resumo:
A methodology to increase the probability of delivering power to any load point through the identification of new investments in distribution network components is proposed in this paper. The method minimizes the investment cost as well as the cost of energy not supplied in the network. A DC optimization model based on mixed integer non-linear programming is developed considering the Pareto front technique in order to identify the adequate investments in distribution networks components which allow increasing the probability of delivering power for any customer in the distribution system at the minimum possible cost for the system operator, while minimizing the energy not supplied cost. Thus, a multi-objective problem is formulated. To illustrate the application of the proposed methodology, the paper includes a case study which considers a 180 bus distribution network
Resumo:
L'objectiu principal d'aquest projecte és ampliar la llibreria BinaryCodes, iniciada al 2007, que ens permet construir i manipular codis binaris lineals i no lineals. Per aquest motiu, s'han desenvolupat una sèrie de funcions, amb els seus corresponents tests i exemples, en l'entorn de programació matemàtica Magma. Aquestes funcions consisteixen bàsicament en la construcció de nous codis a partir d'altres ja existents.
Resumo:
Large projects evaluation rises well known difficulties because -by definition- they modify the current price system; their public evaluation presents additional difficulties because they modify too existing shadow prices without the project. This paper analyzes -first- the basic methodologies applied until late 80s., based on the integration of projects in optimization models or, alternatively, based on iterative procedures with information exchange between two organizational levels. New methodologies applied afterwards are based on variational inequalities, bilevel programming and linear or nonlinear complementarity. Their foundations and different applications related with project evaluation are explored. As a matter of fact, these new tools are closely related among them and can treat more complex cases involving -for example- the reaction of agents to policies or the existence of multiple agents in an environment characterized by common functions representing demands or constraints on polluting emissions.
Resumo:
We present a new technique for audio signal comparison based on tonal subsequence alignment and its application to detect cover versions (i.e., different performances of the same underlying musical piece). Cover song identification is a task whose popularity has increased in the Music Information Retrieval (MIR) community along in the past, as it provides a direct and objective way to evaluate music similarity algorithms.This article first presents a series of experiments carried outwith two state-of-the-art methods for cover song identification.We have studied several components of these (such as chroma resolution and similarity, transposition, beat tracking or Dynamic Time Warping constraints), in order to discover which characteristics would be desirable for a competitive cover song identifier. After analyzing many cross-validated results, the importance of these characteristics is discussed, and the best-performing ones are finally applied to the newly proposed method. Multipleevaluations of this one confirm a large increase in identificationaccuracy when comparing it with alternative state-of-the-artapproaches.
Resumo:
This paper aims to estimate a translog stochastic frontier production function in the analysis of a panel of 150 mixed Catalan farms in the period 1989-1993, in order to attempt to measure and explain variation in technical inefficiency scores with a one-stage approach. The model uses gross value added as the output aggregate measure. Total employment, fixed capital, current assets, specific costs and overhead costs are introduced into the model as inputs. Stochasticfrontier estimates are compared with those obtained using a linear programming method using a two-stage approach. The specification of the translog stochastic frontier model appears as an appropriate representation of the data, technical change was rejected and the technical inefficiency effects were statistically significant. The mean technical efficiency in the period analyzed was estimated to be 64.0%. Farm inefficiency levels were found significantly at 5%level and positively correlated with the number of economic size units.
Resumo:
X-ray diffraction analyses of the pure components n-tricosane and n-pentacosane and of their binary mixed samples have enabled us to characterize the crystalline phases observed at low temperature. On the contrary to what was announced in literature on the structural behavior of mixed samples in odd-odd binary systems with D n = 2, the three domains are not all orthorhombic. This work has enabled us to show that two of the domains are, in fact, monoclinic, (Aa, Z = 4) and the other one is orthorhombic (Pca21, Z = 4). The conclusions drawn in this work can be easily transposed to other binary systems of n-alkanes.
Resumo:
Interfacial hydrodynamic instabilities arise in a range of chemical systems. One mechanism for instability is the occurrence of unstable density gradients due to the accumulation of reaction products. In this paper we conduct two-dimensional nonlinear numerical simulations for a member of this class of system: the methylene-blue¿glucose reaction. The result of these reactions is the oxidation of glucose to a relatively, but marginally, dense product, gluconic acid, that accumulates at oxygen permeable interfaces, such as the surface open to the atmosphere. The reaction is catalyzed by methylene-blue. We show that simulations help to disassemble the mechanisms responsible for the onset of instability and evolution of patterns, and we demonstrate that some of the results are remarkably consistent with experiments. We probe the impact of the upper oxygen boundary condition, for fixed flux, fixed concentration, or mixed boundary conditions, and find significant qualitative differences in solution behavior; structures either attract or repel one another depending on the boundary condition imposed. We suggest that measurement of the form of the boundary condition is possible via observation of oxygen penetration, and improved product yields may be obtained via proper control of boundary conditions in an engineering setting. We also investigate the dependence on parameters such as the Rayleigh number and depth. Finally, we find that pseudo-steady linear and weakly nonlinear techniques described elsewhere are useful tools for predicting the behavior of instabilities beyond their formal range of validity, as good agreement is obtained with the simulations.