53 resultados para Probabilistic load flow
em Consorci de Serveis Universitaris de Catalunya (CSUC), Spain
Resumo:
Material throughput is a means of measuring the so-called social metabolism, or physical dimensions of a society’s consumption, and can be taken as an indirect and approximate indicator of sustainability. Material flow accounting can be used to test the dematerialisation hypothesis, the idea that technological progress causes a decrease in total material used (strong dematerialisation) or material used per monetary unit of output (weak dematerialisation). This paper sets out the results of a material flow analysis for Spain for the period from 1980 to 2000. The analysis reveals that neither strong nor weak dematerialisation took place during the period analysed. Although the population did not increase considerably, materials mobilised by the Spanish economy (DMI) increased by 85% in absolute terms, surpassing GDP growth. In addition, Spain became more dependent on external trade in physical terms. In fact, its imports are more than twice the amount of its exports in terms of weight.
Resumo:
This technical background paper describes the methods applied and data sources used in the compilation of the 1980-2003 data set for material flow accounts of the Mexican economy and presents the data set. It is organised in four parts: the first part gives an overview of the Material Flow Accounting (MFA) methodology. The second part presents the main material flows of the Mexican economy including biomass, fossil fuels, metal ores, industrial minerals and, construction minerals. The aim of this part is to explain the procedures and methods followed, the data sources used as well as providing a brief evaluation of the quality and reliability of the information used and the accounts established. Finally, some conclusions will be provided.
Resumo:
In this paper we compare the resource flows of Chile, Ecuador, Mexico and Peru between 1980 and 2000. In this time span, the domestic extraction of materials increased in the four countries, mainly due to the mining sector in Chile and Peru, biomass and oil in Ecuador and construction minerals in Mexico. Imports and exports increased too, due to the increasing integration in the international markets, prompted by the liberalization policies undertaken by the four countries between the late 1970s and the late 1990s. The four countries had a negative physical trade balance for most of the period analyzed, meaning that their exports exceeded their imports in terms of weight. However, the increase of imports reduced the physical deficit in Chile, Mexico and Peru. Ecuador’s physical deficit was the highest and did not decrease in the period analyzed. Also, a diversification of exports away from bulk commodities could be observed in Chile and Mexico, and to a lesser extent in Peru, whereas in Ecuador the export sector remained mainly based on oil and biomass. More research is needed to explore the environmental effects of this phenomenon. Also, the indirect flows associated to the direct physical flows deserve to be subject to further analysis.
Resumo:
Existeixen creixents evidències què la resposta dels limfòcits T CD8+ alpha beta citotòxics (CTLs) és un element fonamental en la infecció produïda pel VIH. Les CTLs VIH especifiques es consideren molt importants en la reducció de la càrrega viral i en la contenció de la infecció. Encara que la combinació dels antiretrovirals (HAART) ha suposat una millora considerable en la lluita contra el VIH induint una important reducció de la càrrega viral i augmentant el nombre de cèl•lules T CD4+, diverses complicacions han fet ressaltar la necessitat de noves alternatives terapèutiques. Les complicacions inclouen: manca de recuperació d’una resposta immune sòlida contra el VIH, toxicitat a llarg termini de la teràpia i el descobriment que les cèl•lules T CD4+ constitueixen un reservori pel virus. Les noves alternatives controlaran la replicació viral i reconstituiran la immunitat. L’eficàcia de la immunoteràpia cel•lular amb transferència adoptiva de CTLs virals específics s’ha provat en diferents infeccions virals humanes, incloent el VIH. Proposem una modificació de la immunoteràpia adoptiva redirigint l’especificitat de les cèl•lules T contra el VIH mitjançant la transfecció dels gens del TCR. En aquest assaig preclínic, ens aprofitarem de la tecnologia dels animals transgènics per les molècules de HLA, amb la finalitat de generar TCRs d’alta afinitat dirigits contra epitops del VIH restringits per la molècula HLA. Aquests TCRs seran induïts in vivo i seleccionats in vitro. Les cadenes alpha i beta dels TCRs VIH específics procedents de les CTLs seran clonades mitjançant tècniques de biologia molecular. Aquests TCRs VIH específics seran transferits a cèl•lules T CD8+ humanes i la seva especificitat i capacitat citolítica contra cèl•lules diana que presentin antígens de VIH-1 s’estudiaran mitjançant la combinació de diverses tècniques noves (FCC, transfecció mitjançant Nucleoefector). Finalment, una construcció retroviral adient per la seva transducció en cèl•lules T humanes s’establirà amb un TCR òptim seleccionat.
Resumo:
This paper aims at providing a Bayesian parametric framework to tackle the accessibility problem across space in urban theory. Adopting continuous variables in a probabilistic setting we are able to associate with the distribution density to the Kendall's tau index and replicate the general issues related to the role of proximity in a more general context. In addition, by referring to the Beta and Gamma distribution, we are able to introduce a differentiation feature in each spatial unit without incurring in any a-priori definition of territorial units. We are also providing an empirical application of our theoretical setting to study the density distribution of the population across Massachusetts.
Resumo:
We evaluate the performance of different optimization techniques developed in the context of optical flowcomputation with different variational models. In particular, based on truncated Newton methods (TN) that have been an effective approach for large-scale unconstrained optimization, we develop the use of efficient multilevel schemes for computing the optical flow. More precisely, we evaluate the performance of a standard unidirectional multilevel algorithm - called multiresolution optimization (MR/OPT), to a bidrectional multilevel algorithm - called full multigrid optimization (FMG/OPT). The FMG/OPT algorithm treats the coarse grid correction as an optimization search direction and eventually scales it using a line search. Experimental results on different image sequences using four models of optical flow computation show that the FMG/OPT algorithm outperforms both the TN and MR/OPT algorithms in terms of the computational work and the quality of the optical flow estimation.
Resumo:
The work in this paper deals with the development of momentum and thermal boundary layers when a power law fluid flows over a flat plate. At the plate we impose either constant temperature, constant flux or a Newton cooling condition. The problem is analysed using similarity solutions, integral momentum and energy equations and an approximation technique which is a form of the Heat Balance Integral Method. The fluid properties are assumed to be independent of temperature, hence the momentum equation uncouples from the thermal problem. We first derive the similarity equations for the velocity and present exact solutions for the case where the power law index n = 2. The similarity solutions are used to validate the new approximation method. This new technique is then applied to the thermal boundary layer, where a similarity solution can only be obtained for the case n = 1.
Resumo:
Systematic asymptotic methods are used to formulate a model for the extensional flow of a thin sheet of nematic liquid crystal. With no external body forces applied, the model is found to be equivalent to the so-called Trouton model for Newtonian sheets (and fi bers), albeit with a modi fied "Trouton ratio". However, with a symmetry-breaking electric field gradient applied, behavior deviates from the Newtonian case, and the sheet can undergo fi nite-time breakup if a suitable destabilizing field is applied. Some simple exact solutions are presented to illustrate the results in certain idealized limits, as well as sample numerical results to the full model equations.
Resumo:
This paper proposes MSISpIC, a probabilistic sonar scan matching algorithm for the localization of an autonomous underwater vehicle (AUV). The technique uses range scans gathered with a Mechanical Scanning Imaging Sonar (MSIS), the robot displacement estimated through dead-reckoning using a Doppler velocity log (DVL) and a motion reference unit (MRU). The proposed method is an extension of the pIC algorithm. An extended Kalman filter (EKF) is used to estimate the robot-path during the scan in order to reference all the range and bearing measurements as well as their uncertainty to a scan fixed frame before registering. The major contribution consists of experimentally proving that probabilistic sonar scan matching techniques have the potential to improve the DVL-based navigation. The algorithm has been tested on an AUV guided along a 600 m path within an abandoned marina underwater environment with satisfactory results
Resumo:
This paper focuses on one of the methods for bandwidth allocation in an ATM network: the convolution approach. The convolution approach permits an accurate study of the system load in statistical terms by accumulated calculations, since probabilistic results of the bandwidth allocation can be obtained. Nevertheless, the convolution approach has a high cost in terms of calculation and storage requirements. This aspect makes real-time calculations difficult, so many authors do not consider this approach. With the aim of reducing the cost we propose to use the multinomial distribution function: the enhanced convolution approach (ECA). This permits direct computation of the associated probabilities of the instantaneous bandwidth requirements and makes a simple deconvolution process possible. The ECA is used in connection acceptance control, and some results are presented
Resumo:
The paper discusses maintenance challenges of organisations with a huge number of devices and proposes the use of probabilistic models to assist monitoring and maintenance planning. The proposal assumes connectivity of instruments to report relevant features for monitoring. Also, the existence of enough historical registers with diagnosed breakdowns is required to make probabilistic models reliable and useful for predictive maintenance strategies based on them. Regular Markov models based on estimated failure and repair rates are proposed to calculate the availability of the instruments and Dynamic Bayesian Networks are proposed to model cause-effect relationships to trigger predictive maintenance services based on the influence between observed features and previously documented diagnostics
Resumo:
This study explores whether firms have differential price-earnings multiples associated with their means of achieving a sequential pattern of increasing positive earnings. Our main findings show that market participants assign higher price-earnings multiples to firms when their pattern of increasing earnings is supported by the same pattern of increasing cash flows. Market participants assign lower price-earnings multiples to firms suspect of having engaged in accrual-based earnings management, sales manipulation, and overproduction to achieve the earnings pattern. We find, however, that market participants do not penalize firms suspect of having achieved the earnings pattern through the opportunistic reduction of discretionary expenses.
Resumo:
The identification and integration of reusable and customizable CSCL (Computer Supported Collaborative Learning) may benefit from the capture of best practices in collaborative learning structuring. The authors have proposed CLFPs (Collaborative Learning Flow Patterns) as a way of collecting these best practices. To facilitate the process of CLFPs by software systems, the paper proposes to specify these patterns using IMS Learning Design (IMS-LD). Thus, teachers without technical knowledge can particularize and integrate CSCL tools. Nevertheless, the support of IMS-LD for describing collaborative learning activities has some deficiencies: the collaborative tools that can be defined in these activities are limited. Thus, this paper proposes and discusses an extension to IMS-LD that enables to specify several characteristics of the use of tools that mediate collaboration. In order to obtain a Unit of Learning based on a CLFP, a three stage process is also proposed. A CLFP-based Unit of Learning example is used to illustrate the process and the need of the proposed extension.
Resumo:
From a managerial point of view, the more effcient, simple, and parameter-free (ESP) an algorithm is, the more likely it will be used in practice for solving real-life problems. Following this principle, an ESP algorithm for solving the Permutation Flowshop Sequencing Problem (PFSP) is proposed in this article. Using an Iterated Local Search (ILS) framework, the so-called ILS-ESP algorithm is able to compete in performance with other well-known ILS-based approaches, which are considered among the most effcient algorithms for the PFSP. However, while other similar approaches still employ several parameters that can affect their performance if not properly chosen, our algorithm does not require any particular fine-tuning process since it uses basic "common sense" rules for the local search, perturbation, and acceptance criterion stages of the ILS metaheuristic. Our approach defines a new operator for the ILS perturbation process, a new acceptance criterion based on extremely simple and transparent rules, and a biased randomization process of the initial solution to randomly generate different alternative initial solutions of similar quality -which is attained by applying a biased randomization to a classical PFSP heuristic. This diversification of the initial solution aims at avoiding poorly designed starting points and, thus, allows the methodology to take advantage of current trends in parallel and distributed computing. A set of extensive tests, based on literature benchmarks, has been carried out in order to validate our algorithm and compare it against other approaches. These tests show that our parameter-free algorithm is able to compete with state-of-the-art metaheuristics for the PFSP. Also, the experiments show that, when using parallel computing, it is possible to improve the top ILS-based metaheuristic by just incorporating to it our biased randomization process with a high-quality pseudo-random number generator.