934 resultados para Box constrained minimization
Resumo:
Incomplete understanding of three aspects of the climate system—equilibrium climate sensitivity, rate of ocean heat uptake and historical aerosol forcing—and the physical processes underlying them lead to uncertainties in our assessment of the global-mean temperature evolution in the twenty-first century1,2. Explorations of these uncertainties have so far relied on scaling approaches3,4, large ensembles of simplified climate models1,2, or small ensembles of complex coupled atmosphere–ocean general circulation models5,6 which under-represent uncertainties in key climate system properties derived from independent sources7–9. Here we present results from a multi-thousand-member perturbed-physics ensemble of transient coupled atmosphere–ocean general circulation model simulations. We find that model versions that reproduce observed surface temperature changes over the past 50 years show global-mean temperature increases of 1.4–3 K by 2050, relative to 1961–1990, under a mid-range forcing scenario. This range of warming is broadly consistent with the expert assessment provided by the Intergovernmental Panel on Climate Change Fourth Assessment Report10, but extends towards larger warming than observed in ensemblesof-opportunity5 typically used for climate impact assessments. From our simulations, we conclude that warming by the middle of the twenty-first century that is stronger than earlier estimates is consistent with recent observed temperature changes and a mid-range ‘no mitigation’ scenario for greenhouse-gas emissions.
Resumo:
In recent years several methodologies have been developed to combine and interpret ensembles of climate models with the aim of quantifying uncertainties in climate projections. Constrained climate model forecasts have been generated by combining various choices of metrics used to weight individual ensemble members, with diverse approaches to sampling the ensemble. The forecasts obtained are often significantly different, even when based on the same model output. Therefore, a climate model forecast classification system can serve two roles: to provide a way for forecast producers to self-classify their forecasts; and to provide information on the methodological assumptions underlying the forecast generation and its uncertainty when forecasts are used for impacts studies. In this review we propose a possible classification system based on choices of metrics and sampling strategies. We illustrate the impact of some of the possible choices in the uncertainty quantification of large scale projections of temperature and precipitation changes, and briefly discuss possible connections between climate forecast uncertainty quantification and decision making approaches in the climate change context.
Resumo:
We present results from experimental price-setting oligopolies in which green firms undertake different levels of energy-saving investments motivated by public subsidies and demand-side advantages. We find that consumers reveal higher willingness to pay for greener sellers’ products. This observation in conjunction to the fact that greener sellers set higher prices is compatible with the use and interpretation of energy-saving behaviour as a differentiation strategy. However, sellers do not exploit the resulting advantage through sufficiently high price-cost margins, because they seem trapped into “run to stay still” competition. Regarding the use of public subsidies to energy-saving sellers we uncover an undesirable crowding-out effect of consumers’ intrinsic tendency to support green manufacturers. Namely, consumers may be less willing to support a green seller whose energy-saving strategy yields a direct financial benefit. Finally, we disentangle two alternative motivations for consumer’s attractions to pro-social firms; first, the self-interested recognition of the firm’s contribution to the public and private welfare and, second, the need to compensate a firm for the cost entailed in each pro-social action. Our results show the prevalence of the former over the latter.
Resumo:
Periocular recognition has recently become an active topic in biometrics. Typically it uses 2D image data of the periocular region. This paper is the first description of combining 3D shape structure with 2D texture. A simple and effective technique using iterative closest point (ICP) was applied for 3D periocular region matching. It proved its strength for relatively unconstrained eye region capture, and does not require any training. Local binary patterns (LBP) were applied for 2D image based periocular matching. The two modalities were combined at the score-level. This approach was evaluated using the Bosphorus 3D face database, which contains large variations in facial expressions, head poses and occlusions. The rank-1 accuracy achieved from the 3D data (80%) was better than that for 2D (58%), and the best accuracy (83%) was achieved by fusing the two types of data. This suggests that significant improvements to periocular recognition systems could be achieved using the 3D structure information that is now available from small and inexpensive sensors.
Resumo:
In this paper, we develop a novel constrained recursive least squares algorithm for adaptively combining a set of given multiple models. With data available in an online fashion, the linear combination coefficients of submodels are adapted via the proposed algorithm.We propose to minimize the mean square error with a forgetting factor, and apply the sum to one constraint to the combination parameters. Moreover an l1-norm constraint to the combination parameters is also applied with the aim to achieve sparsity of multiple models so that only a subset of models may be selected into the final model. Then a weighted l2-norm is applied as an approximation to the l1-norm term. As such at each time step, a closed solution of the model combination parameters is available. The contribution of this paper is to derive the proposed constrained recursive least squares algorithm that is computational efficient by exploiting matrix theory. The effectiveness of the approach has been demonstrated using both simulated and real time series examples.
Resumo:
Field observations of new particle formation and the subsequent particle growth are typically only possible at a fixed measurement location, and hence do not follow the temporal evolution of an air parcel in a Lagrangian sense. Standard analysis for determining formation and growth rates requires that the time-dependent formation rate and growth rate of the particles are spatially invariant; air parcel advection means that the observed temporal evolution of the particle size distribution at a fixed measurement location may not represent the true evolution if there are spatial variations in the formation and growth rates. Here we present a zero-dimensional aerosol box model coupled with one-dimensional atmospheric flow to describe the impact of advection on the evolution of simulated new particle formation events. Wind speed, particle formation rates and growth rates are input parameters that can vary as a function of time and location, using wind speed to connect location to time. The output simulates measurements at a fixed location; formation and growth rates of the particle mode can then be calculated from the simulated observations at a stationary point for different scenarios and be compared with the ‘true’ input parameters. Hence, we can investigate how spatial variations in the formation and growth rates of new particles would appear in observations of particle number size distributions at a fixed measurement site. We show that the particle size distribution and growth rate at a fixed location is dependent on the formation and growth parameters upwind, even if local conditions do not vary. We also show that different input parameters used may result in very similar simulated measurements. Erroneous interpretation of observations in terms of particle formation and growth rates, and the time span and areal extent of new particle formation, is possible if the spatial effects are not accounted for.
Resumo:
A new species of cubozoan jellyfish has been discovered in shallow waters of Bonaire, Netherlands ( Dutch Caribbean). Thus far, approximately 50 sightings of the species, known commonly as the Bonaire banded box jelly, are recorded, and three specimens have been collected. Three physical encounters between humans and the species have been reported. Available evidence suggests that a serious sting is inflicted by this medusa. To increase awareness of the scientific disciplines of systematics and taxonomy, the public has been involved in naming this new species. The Bonaire banded box jelly, Tamoya ohboya, n. sp., can be distinguished from its close relatives T. haplonema from Brazil and T. sp. from the southeastern United States by differences in tentacle coloration, cnidome, and mitochondrial gene sequences. Tamoya ohboya n. sp. possesses striking dark brown to reddish-orange banded tentacles, nematocyst warts that densely cover the animal, and a deep stomach. We provide a detailed comparison of nematocyst data from Tamoya ohboya n. sp., T. haplonema from Brazil, and T. sp. from the Gulf of Mexico.
Resumo:
The constrained compartmentalized knapsack problem can be seen as an extension of the constrained knapsack problem. However, the items are grouped into different classes so that the overall knapsack has to be divided into compartments, and each compartment is loaded with items from the same class. Moreover, building a compartment incurs a fixed cost and a fixed loss of the capacity in the original knapsack, and the compartments are lower and upper bounded. The objective is to maximize the total value of the items loaded in the overall knapsack minus the cost of the compartments. This problem has been formulated as an integer non-linear program, and in this paper, we reformulate the non-linear model as an integer linear master problem with a large number of variables. Some heuristics based on the solution of the restricted master problem are investigated. A new and more compact integer linear model is also presented, which can be solved by a branch-and-bound commercial solver that found most of the optimal solutions for the constrained compartmentalized knapsack problem. On the other hand, heuristics provide good solutions with low computational effort. (C) 2011 Elsevier BM. All rights reserved.
Resumo:
Increasing efforts exist in integrating different levels of detail in models of the cardiovascular system. For instance, one-dimensional representations are employed to model the systemic circulation. In this context, effective and black-box-type decomposition strategies for one-dimensional networks are needed, so as to: (i) employ domain decomposition strategies for large systemic models (1D-1D coupling) and (ii) provide the conceptual basis for dimensionally-heterogeneous representations (1D-3D coupling, among various possibilities). The strategy proposed in this article works for both of these two scenarios, though the several applications shown to illustrate its performance focus on the 1D-1D coupling case. A one-dimensional network is decomposed in such a way that each coupling point connects two (and not more) of the sub-networks. At each of the M connection points two unknowns are defined: the flow rate and pressure. These 2M unknowns are determined by 2M equations, since each sub-network provides one (non-linear) equation per coupling point. It is shown how to build the 2M x 2M non-linear system with arbitrary and independent choice of boundary conditions for each of the sub-networks. The idea is then to solve this non-linear system until convergence, which guarantees strong coupling of the complete network. In other words, if the non-linear solver converges at each time step, the solution coincides with what would be obtained by monolithically modeling the whole network. The decomposition thus imposes no stability restriction on the choice of the time step size. Effective iterative strategies for the non-linear system that preserve the black-box character of the decomposition are then explored. Several variants of matrix-free Broyden`s and Newton-GMRES algorithms are assessed as numerical solvers by comparing their performance on sub-critical wave propagation problems which range from academic test cases to realistic cardiovascular applications. A specific variant of Broyden`s algorithm is identified and recommended on the basis of its computer cost and reliability. (C) 2010 Elsevier B.V. All rights reserved.
Resumo:
In this work we reported the synthesis and evaluation of the analgesic, anti-inflammatory, and platelet anti-aggregating properties of new 3-(arylideneamino)-2-methyl-6,7-methylenedioxy-quinazolin-4 (3H)-one derivatives (3a-j), designed as conformationally constrained analogues of analgesic 1,3- benzodioxolyl-N- acylhydrazones (1) previously developed at LASSBio. Target compounds were synthesized in very good yields exploiting abundant Brazilian natural product safrole (2) as starting material. The pharmacological assays lead us to identify compounds LASSBio-1240 (3b) and LASSBio-1272 (3d) as new analgesic prototypes, presenting an antinociceptive pro. le more potent and effective than dipyrone and indomethacin used, respectively, as standards in AcOH-induced abdominal constrictions assay and in the formalin test. These results confirmed the success in the exploitation of conformation restriction strategy for identification of novel cyclic N-acylhydrazone analogues with optimized analgesic profile (C) 2009 Elsevier Ltd. All rights reserved.
Resumo:
Given an algorithm A for solving some mathematical problem based on the iterative solution of simpler subproblems, an outer trust-region (OTR) modification of A is the result of adding a trust-region constraint to each subproblem. The trust-region size is adaptively updated according to the behavior of crucial variables. The new subproblems should not be more complex than the original ones, and the convergence properties of the OTR algorithm should be the same as those of Algorithm A. In the present work, the OTR approach is exploited in connection with the ""greediness phenomenon"" of nonlinear programming. Convergence results for an OTR version of an augmented Lagrangian method for nonconvex constrained optimization are proved, and numerical experiments are presented.
Resumo:
Arbetet består av en närläsning av karaktären Kurt Hummel i tv-serien Glee med fokus på stereotyper för homosexuella män, och hur Kurt förhåller sig till dem stereotyperna. Studien är gjord utifrån ett queerteoretiskt perspektiv och visar att Kurt i viss mån framställs stereotypt i seriens tidigare avsnitt, men att han efter hand uppvisar ett djup och komplexitet som man inte ofta ser hos homosexuella karaktärer på tv.
Resumo:
The context of this report and the IRIDIA laboratory are described in the preface. Evolutionary Robotics and the box-pushing task are presented in the introduction.The building of a test system supporting Evolutionary Robotics experiments is then detailed. This system is made of a robot simulator and a Genetic Algorithm. It is used to explore the possibility of evolving box-pushing behaviours. The bootstrapping problem is explained, and a novel approach for dealing with it is proposed, with results presented.Finally, ideas for extending this approach are presented in the conclusion.
Resumo:
In this paper, Finite Element method and full-scale experiments have been used to study a hot forging method for fabri-cation of a spindle using reduced initial stock size. The forging sequence is carried out in two stages. In the first stage, the hot rolled cylindrical billet is pre-formed and pierced in a closed die using a spherical nosed punch to within 20 mm of its base. This process of piercing or impact extrusion leads to high strains within the work piece but requires high press loads. In the second stage, the resulting cylinder is placed in a die with a flange chamber and upset forged to form a flange. The stock mass is optimized for complete die filling. Process parameters such as effective strain distribution, material flow and forging load in different stages of the process are analyzed. It is concluded from the simulations that minor modifications of piercing punch geometry to reduce contact between the punch and emerging vertical walls of the cylinder appreciably reduces the piercing load. In the flange chamber, a die surfaces angle of 52° instead of 45° is pro-posed to ensure effective material flow and exert sufficient tool pressure to achieve complete cavity filling. In order to achieve better compression, it is also proposed to shorten both the length of the inserted punch and the die “tongues” by a few mm.