90 resultados para Agent observer

em Queensland University of Technology - ePrints Archive


Relevância:

20.00% 20.00%

Publicador:

Resumo:

The New Zealand green lipped mussel preparation Lyprinol is available without a prescription from a supermarket, pharmacy or Web. The Food and Drug Administration have recently warned Lyprinol USA about their extravagant anti-inflammatory claims for Lyprinol appearing on the web. These claims are put to thorough review. Lyprinol does have anti-inflammatory mechanisms, and has anti-inflammatory effects in some animal models of inflammation. Lyprinol may have benefits in dogs with arthritis. There are design problems with the clinical trials of Lyprinol in humans as an anti-inflammatory agent in osteoarthritis and rheumatoid arthritis, making it difficult to give a definite answer to how effective Lyprinol is in these conditions, but any benefit is small. Lyprinol also has a small benefit in atopic allergy. As anti-inflammatory agents, there is little to choose between Lyprinol and fish oil. No adverse effects have been reported with Lyprinol. Thus, although it is difficult to conclude whether Lyprinol does much good, it can be concluded that Lyprinol probably does no major harm.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The load–frequency control (LFC) problem has been one of the major subjects in a power system. In practice, LFC systems use proportional–integral (PI) controllers. However since these controllers are designed using a linear model, the non-linearities of the system are not accounted for and they are incapable of gaining good dynamical performance for a wide range of operating conditions in a multi-area power system. A strategy for solving this problem because of the distributed nature of a multi-area power system is presented by using a multi-agent reinforcement learning (MARL) approach. It consists of two agents in each power area; the estimator agent provides the area control error (ACE) signal based on the frequency bias estimation and the controller agent uses reinforcement learning to control the power system in which genetic algorithm optimisation is used to tune its parameters. This method does not depend on any knowledge of the system and it admits considerable flexibility in defining the control objective. Also, by finding the ACE signal based on the frequency bias estimation the LFC performance is improved and by using the MARL parallel, computation is realised, leading to a high degree of scalability. Here, to illustrate the accuracy of the proposed approach, a three-area power system example is given with two scenarios.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Alexithymia is characterised by deficits in emotional insight and self reflection, that impact on the efficacy of psychological treatments. Given the high prevalence of alexithymia in Alcohol Use Disorders, valid assessment tools are critical. The majority of research on the relationship between alexithymia and alcohol-dependence has employed the self-administered Toronto Alexithymia Scale (TAS-20). The Observer Alexithymia Scale (OAS) has also been recommended. The aim of the present study was to assess the validity and reliability of the OAS and the TAS-20 in an alcohol-dependent sample. Two hundred and ten alcohol-dependent participants in an outpatient Cognitive Behavioral Treatment program were administered the TAS-20 at assessment and upon treatment completion at 12 weeks. Clinical psychologists provided observer assessment data for a subsample of 159 patients. The findings confirmed acceptable internal consistency, test-retest reliability and scale homogeneity for both the OAS and TAS-20, except for the low internal consistency of the TAS-20 EOT scale. The TAS-20 was more strongly associated with alcohol problems than the OAS.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

An adaptive agent improves its performance by learning from experience. This paper describes an approach to adaptation based on modelling dynamic elements of the environment in order to make predictions of likely future state. This approach is akin to an elite sports player being able to “read the play”, allowing for decisions to be made based on predictions of likely future outcomes. Modelling of the agent‟s likely future state is performed using Markov Chains and a technique called “Motion and Occupancy Grids”. The experiments in this paper compare the performance of the planning system with and without the use of this predictive model. The results of the study demonstrate a surprising decrease in performance when using the predictions of agent occupancy. The results are derived from statistical analysis of the agent‟s performance in a high fidelity simulation of a world leading real robot soccer team.