966 resultados para dynamic methods
Resumo:
After outlining some comparative features of poverty in India, this article reviews critically recent literature on the dynamics of poverty. On economic efficiency grounds, it rejects the view that the chronically poor are more deserving than the non-chronic poor of poverty assistance. Mechanisms of households and communities for coping with poverty are discussed. The possibility is raised that where poverty has been persistent that rational methods for coping with it are likely to be well established, and less suffering may occur than for households and communities thrown temporarily into poverty. However, situations can also be envisaged where such rational behaviours deepen the poverty trap and create unfavourable externalities for poverty alleviation. Conflict can arise between programmes to alleviate poverty in poor communities and the sustainability of these communities and their local cultures. Problems posed by this are discussed. Furthermore, the impact of market extension on poor landholders is considered. In contrast to the prevailing view that increased market extension and liberalisation is favourable to poor farmers, it is argued that inescapable market transaction cost makes it difficult for the poor to survive as landholders in a fluid and changing market system. The likelihood of poor landholders joining the landless poor rises, and if they migrate from the countryside to the city they face further adjustment hurdles. Consequently, poor landholders may be poorer after the extension of the market system and only their offspring may reap benefits from market reforms.
Resumo:
A decision theory framework can be a powerful technique to derive optimal management decisions for endangered species. We built a spatially realistic stochastic metapopulation model for the Mount Lofty Ranges Southern Emu-wren (Stipiturus malachurus intermedius), a critically endangered Australian bird. Using diserete-time Markov,chains to describe the dynamics of a metapopulation and stochastic dynamic programming (SDP) to find optimal solutions, we evaluated the following different management decisions: enlarging existing patches, linking patches via corridors, and creating a new patch. This is the first application of SDP to optimal landscape reconstruction and one of the few times that landscape reconstruction dynamics have been integrated with population dynamics. SDP is a powerful tool that has advantages over standard Monte Carlo simulation methods because it can give the exact optimal strategy for every landscape configuration (combination of patch areas and presence of corridors) and pattern of metapopulation occupancy, as well as a trajectory of strategies. It is useful when a sequence of management actions can be performed over a given time horizon, as is the case for many endangered species recovery programs, where only fixed amounts of resources are available in each time step. However, it is generally limited by computational constraints to rather small networks of patches. The model shows that optimal metapopulation, management decisions depend greatly on the current state of the metapopulation,. and there is no strategy that is universally the best. The extinction probability over 30 yr for the optimal state-dependent management actions is 50-80% better than no management, whereas the best fixed state-independent sets of strategies are only 30% better than no management. This highlights the advantages of using a decision theory tool to investigate conservation strategies for metapopulations. It is clear from these results that the sequence of management actions is critical, and this can only be effectively derived from stochastic dynamic programming. The model illustrates the underlying difficulty in determining simple rules of thumb for the sequence of management actions for a metapopulation. This use of a decision theory framework extends the capacity of population viability analysis (PVA) to manage threatened species.
Resumo:
Reinforcement Learning is an area of Machine Learning that deals with how an agent should take actions in an environment such as to maximize the notion of accumulated reward. This type of learning is inspired by the way humans learn and has led to the creation of various algorithms for reinforcement learning. These algorithms focus on the way in which an agent’s behaviour can be improved, assuming independence as to their surroundings. The current work studies the application of reinforcement learning methods to solve the inverted pendulum problem. The importance of the variability of the environment (factors that are external to the agent) on the execution of reinforcement learning agents is studied by using a model that seeks to obtain equilibrium (stability) through dynamism – a Cart-Pole system or inverted pendulum. We sought to improve the behaviour of the autonomous agents by changing the information passed to them, while maintaining the agent’s internal parameters constant (learning rate, discount factors, decay rate, etc.), instead of the classical approach of tuning the agent’s internal parameters. The influence of changes on the state set and the action set on an agent’s capability to solve the Cart-pole problem was studied. We have studied typical behaviour of reinforcement learning agents applied to the classic BOXES model and a new form of characterizing the environment was proposed using the notion of convergence towards a reference value. We demonstrate the gain in performance of this new method applied to a Q-Learning agent.
Resumo:
O documento em anexo encontra-se na versão post-print (versão corrigida pelo editor).
Resumo:
The main purpose of this paper is to propose a Multi-Agent Autonomic and Bio-Inspired based framework with selfmanaging capabilities to solve complex scheduling problems using cooperative negotiation. Scheduling resolution requires the intervention of highly skilled human problem-solvers. This is a very hard and challenging domain because current systems are becoming more and more complex, distributed, interconnected and subject to rapidly changing. A natural Autonomic Computing (AC) evolution in relation to Current Computing is to provide systems with Self-Managing ability with a minimum human interference.
Resumo:
Introduction: Although relative uptake values aren’t the most important objective of a 99mTc-DMSA scan, they are important quantitative information. In most of the dynamic renal scintigraphies attenuation correction is essential if one wants to obtain a reliable result of the quantification process. Although in DMSA scans the absent of significant background and the lesser attenuation in pediatric patients, makes that this attenuation correction techniques are actually not applied. The geometric mean is the most common method, but that includes the acquisition of an anterior (extra) projection, which it is not acquired by a large number of NM departments. This method and the attenuation factors proposed by Tonnesen will be correlated with the absence of attenuation correction procedures. Material and Methods: Images from 20 individuals (aged 3 years +/- 2) were used and the two attenuation correction methods applied. The mean time of acquisition (time post DMSA administration) was 3.5 hours +/- 0.8h. Results: The absence of attenuation correction showed a good correlation with both attenuation methods (r=0.73 +/- 0.11) and the mean difference verified on the uptake values between the different methods were 4 +/- 3. The correlation was higher when the age was lower. The attenuation correction methods correlation was higher between them two than with the “no attenuation correction” method (r=0.82 +/- 0.8), and the mean differences of the uptake values were 2 +/- 2. Conclusion: The decision of not doing any kind of attenuation correction method can be justified by the minor differences verified on the relative kidney uptake values. Nevertheless, if it is recognized that there is a need for an accurate value of the relative kidney uptake, then an attenuation correction method should be used. Attenuation correction factors proposed by Tonnesen can be easily implemented and so become a practical and easy to implement alternative, namely when the anterior projection - needed for the geometric mean methodology – is not acquired.
Resumo:
Dissertação apresentada na Faculdade de Ciências e Tecnologia da Universidade Nova de Lisboa para a obtenção do grau de Mestre em Engenharia do Ambiente
Resumo:
ABSTRACT OBJECTIVE To develop an assessment tool to evaluate the efficiency of federal university general hospitals. METHODS Data envelopment analysis, a linear programming technique, creates a best practice frontier by comparing observed production given the amount of resources used. The model is output-oriented and considers variable returns to scale. Network data envelopment analysis considers link variables belonging to more than one dimension (in the model, medical residents, adjusted admissions, and research projects). Dynamic network data envelopment analysis uses carry-over variables (in the model, financing budget) to analyze frontier shift in subsequent years. Data were gathered from the information system of the Brazilian Ministry of Education (MEC), 2010-2013. RESULTS The mean scores for health care, teaching and research over the period were 58.0%, 86.0%, and 61.0%, respectively. In 2012, the best performance year, for all units to reach the frontier it would be necessary to have a mean increase of 65.0% in outpatient visits; 34.0% in admissions; 12.0% in undergraduate students; 13.0% in multi-professional residents; 48.0% in graduate students; 7.0% in research projects; besides a decrease of 9.0% in medical residents. In the same year, an increase of 0.9% in financing budget would be necessary to improve the care output frontier. In the dynamic evaluation, there was progress in teaching efficiency, oscillation in medical care and no variation in research. CONCLUSIONS The proposed model generates public health planning and programming parameters by estimating efficiency scores and making projections to reach the best practice frontier.
Resumo:
Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia de Electrónica e telecomunicações
Resumo:
Trabalho apresentado no âmbito do Mestrado em Engenharia Informática, como requisito parcial para obtenção do grau de Mestre em Engenharia Informática
Resumo:
Enterprise and Work Innovation Studies,6,IET, pp.9-51
Resumo:
One of today's biggest concerns is the increase of energetic needs, especially in the developed countries. Among various clean energies, wind energy is one of the technologies that assume greater importance on the sustainable development of humanity. Despite wind turbines had been developed and studied over the years, there are phenomena that haven't been yet fully understood. This work studies the soil-structure interaction that occurs on a wind turbine's foundation composed by a group of piles that is under dynamic loads caused by wind. This problem assumes special importance when the foundation is implemented on locations where safety criteria are very demanding, like the case of a foundation mounted on a dike. To the phenomenon of interaction between two piles and the soil between them it's given the name of pile-soil-pile interaction. It is known that such behavior is frequency dependent, and therefore, on this work evaluation of relevant frequencies for the intended analysis is held. During the development of this thesis, two methods were selected in order to assess pile-soil-pile interaction, being one of analytical nature and the other of numerical origin. The analytical solution was recently developed and its called Generalized pile-soil-pile theory, while for the numerical method the commercial nite element software PLAXIS 3D was used. A study of applicability of the numerical method is also done comparing the given solution by the nite element methods with a rigorous solution widely accepted by the majority of the authors.
Resumo:
Amyotrophic Lateral Sclerosis (ALS) is a neurodegenerative disease characterized by motor neurons degeneration, which reduces muscular force, being very difficult to diagnose. Mathematical methods are used in order to analyze the surface electromiographic signal’s dynamic behavior (Fractal Dimension (FD) and Multiscale Entropy (MSE)), evaluate different muscle group’s synchronization (Coherence and Phase Locking Factor (PLF)) and to evaluate the signal’s complexity (Lempel-Ziv (LZ) techniques and Detrended Fluctuation Analysis (DFA)). Surface electromiographic signal acquisitions were performed in upper limb muscles, being the analysis executed for instants of contraction for ipsilateral acquisitions for patients and control groups. Results from LZ, DFA and MSE analysis present capability to distinguish between the patient group and the control group, whereas coherence, PLF and FD algorithms present results very similar for both groups. LZ, DFA and MSE algorithms appear then to be a good measure of corticospinal pathways integrity. A classification algorithm was applied to the results in combination with extracted features from the surface electromiographic signal, with an accuracy percentage higher than 70% for 118 combinations for at least one classifier. The classification results demonstrate capability to distinguish members between patients and control groups. These results can demonstrate a major importance in the disease diagnose, once surface electromyography (sEMG) may be used as an auxiliary diagnose method.
Resumo:
INTRODUCTION: Sylvatic yellow fever (SYF) is enzootic in Brazil, causing periodic outbreaks in humans living near forest borders or in rural areas. In this study, the cycling patterns of this arbovirosis were analyzed. METHODS: Spectral Fourier analysis was used to capture the periodicity patterns of SYF in time series. RESULTS: SYF outbreaks have not increased in frequency, only in the number of cases. There are two dominant cycles in SYF outbreaks, a seven year cycle for the central-western region and a 14 year cycle for the northern region. Most of the variance was concentrated in the central-western region and dominated the entire endemic region. CONCLUSIONS: The seven year cycle is predominant in the endemic region of the disease due the greater contribution of variance in the central-western region; however, it was possible identify a 14 cycle that governs SYF outbreaks in the northern region. No periodicities were identified for the remaining geographical regions.
Resumo:
A Work Project, presented as part of the requirements for the Award of a Master’s Double Degree in Finance from Maastricht University and NOVA – School of Business and Economics