788 resultados para agent-based simulation
Resumo:
This introduction to the Virtual Special Issue surveys the development of spatial housing economics from its roots in neo-classical theory, through more recent developments in social interactions modelling, and touching on the role of institutions, path dependence and economic history. The survey also points to some of the more promising future directions for the subject that are beginning to appear in the literature. The survey covers elements hedonic models, spatial econometrics, neighbourhood models, housing market areas, housing supply, models of segregation, migration, housing tenure, sub-national house price modelling including the so-called ripple effect, and agent-based models. Possible future directions are set in the context of a selection of recent papers that have appeared in Urban Studies. Nevertheless, there are still important gaps in the literature that merit further attention, arising at least partly from emerging policy problems. These include more research on housing and biodiversity, the relationship between housing and civil unrest, the effects of changing age distributions - notably housing for the elderly - and the impact of different international institutional structures. Methodologically, developments in Big Data provide an exciting framework for future work.
Resumo:
We study the relationship between the sentiment levels of Twitter users and the evolving network structure that the users created by @-mentioning each other. We use a large dataset of tweets to which we apply three sentiment scoring algorithms, including the open source SentiStrength program. Specifically we make three contributions. Firstly we find that people who have potentially the largest communication reach (according to a dynamic centrality measure) use sentiment differently than the average user: for example they use positive sentiment more often and negative sentiment less often. Secondly we find that when we follow structurally stable Twitter communities over a period of months, their sentiment levels are also stable, and sudden changes in community sentiment from one day to the next can in most cases be traced to external events affecting the community. Thirdly, based on our findings, we create and calibrate a simple agent-based model that is capable of reproducing measures of emotive response comparable to those obtained from our empirical dataset.
Resumo:
The main objective for this degree project is to implement an Application Availability Monitoring (AAM) system named Softek EnView for Fujitsu Services. The aim of implementing the AAM system is to proactively identify end user performance problems, such as application and site performance, before the actual end users experience them. No matter how well applications and sites are designed and nomatter how well they meet business requirements, they are useless to the end users if the performance is slow and/or unreliable. It is important for the customers to find out whether the end user problems are caused by the network or application malfunction. The Softek EnView was comprised of the following EnView components: Robot, Monitor, Reporter, Collector and Repository. The implemented system, however, is designed to use only some of these EnView elements: Robot, Reporter and depository. Robots can be placed at any key user location and are dedicated to customers, which means that when the number of customers increases, at the sametime the amount of Robots will increase. To make the AAM system ideal for the company to use, it was integrated with Fujitsu Services’ centralised monitoring system, BMC PATROL Enterprise Manager (PEM). That was actually the reason for deciding to drop the EnView Monitor element. After the system was fully implemented, the AAM system was ready for production. Transactions were (and are) written and deployed on Robots to simulate typical end user actions. These transactions are configured to run with certain intervals, which are defined collectively with customers. While they are driven against customers’ applicationsautomatically, transactions collect availability data and response time data all the time. In case of a failure in transactions, the robot immediately quits the transactionand writes detailed information to a log file about what went wrong and which element failed while going through an application. Then an alert is generated by a BMC PATROL Agent based on this data and is sent to the BMC PEM. Fujitsu Services’ monitoring room receives the alert, reacts to it according to the incident management process in ITIL and by alerting system specialists on critical incidents to resolve problems. As a result of the data gathered by the Robots, weekly reports, which contain detailed statistics and trend analyses of ongoing quality of IT services, is provided for the Customers.
Resumo:
The Grid is a large-scale computer system that is capable of coordinating resources that are not subject to centralised control, whilst using standard, open, general-purpose protocols and interfaces, and delivering non-trivial qualities of service. In this chapter, we argue that Grid applications very strongly suggest the use of agent-based computing, and we review key uses of agent technologies in Grids: user agents, able to customize and personalise data; agent communication languages offering a generic and portable communication medium; and negotiation allowing multiple distributed entities to reach service level agreements. In the second part of the chapter, we focus on Grid service discovery, which we have identified as a prime candidate for use of agent technologies: we show that Grid-services need to be located via personalised, semantic-rich discovery processes, which must rely on the storage of arbitrary metadata about services that originates from both service providers and service users. We present UDDI-MT, an extension to the standard UDDI service directory approach that supports the storage of such metadata via a tunnelling technique that ties the metadata store to the original UDDI directory. The outcome is a flexible service registry which is compatible with existing standards and also provides metadata-enhanced service discovery.
Resumo:
MyGrid is an e-Science Grid project that aims to help biologists and bioinformaticians to perform workflow-based in silico experiments, and help them to automate the management of such workflows through personalisation, notification of change and publication of experiments. In this paper, we describe the architecture of myGrid and how it will be used by the scientist. We then show how myGrid can benefit from agents technologies. We have identified three key uses of agent technologies in myGrid: user agents, able to customize and personalise data, agent communication languages offering a generic and portable communication medium, and negotiation allowing multiple distributed entities to reach service level agreements.
Resumo:
In order to facilitate the development of agent-based software, several agent programming languages and architectures, have been created. Plans in these architectures are often self-contained procedures with an associated triggering event and a context condition, while any further information about the consequences of executing a plan is absent. However, agents designed using such an approach have limited flexibility at runtime, and rely on the designer’s ability to foresee all relevant situations an agent might have to handle. In order to overcome this limitation, we have created AgentSpeak(PL), an interpreter capable of performing state-space planning to generate new high-level plans. As the planning module creates new plans, the plan library is expanded, improving performance over time. However, for new plans to be useful in the long run, it is critical that the context condition associated with new plans is carefully generated. In this paper we describe a plan reuse technique aimed at improving an agent’s runtime performance by deriving optimal context conditions for new plans, allowing an agent to reuse generated plans as much as possible.
Resumo:
Expressing contractual agreements electronically potentially allows agents to automatically perform functions surrounding contract use: establishment, fulfilment, renegotiation etc. For such automation to be used for real business concerns, there needs to be a high level of trust in the agent-based system. While there has been much research on simulating trust between agents, there are areas where such trust is harder to establish. In particular, contract proposals may come from parties that an agent has had no prior interaction with and, in competitive business-to-business environments, little reputation information may be available. In human practice, trust in a proposed contract is determined in part from the content of the proposal itself, and the similarity of the content to that of prior contracts, executed to varying degrees of success. In this paper, we argue that such analysis is also appropriate in automated systems, and to provide it we need systems to record salient details of prior contract use and algorithms for assessing proposals on their content. We use provenance technology to provide the former and detail algorithms for measuring contract success and similarity for the latter, applying them to an aerospace case study.
Resumo:
Small and medium-sized companies and other enterprises (SMEs) around the world are exposed to flood risk and many of the 4.5 million in the UK are at risk. As SMEs represent almost half of total business turnover in the UK, their protection is a vital part of the drive for greater climate change resilience. However, few have measures in place to ensure the continuity of their activities during a flood and its aftermath. The SESAME project aims to develop tools that encourage businesses to discover ways of becoming more resilient to floods and to appreciate how much better off they will be once they have adapted to the ongoing risk. By taking some of the mystery out of flooding and flood risk, it aims to make it susceptible to the same business acumen that enables the UK’s SMEs to deal with the many other challenges they face. In this paper we will report on the different aspects of the research in the project Understanding behaviour Changing behaviour Modelling impacts Economic impacts Through the above the project will advise government, local authorities and other public bodies on how to improve their responses to floods and will enable them to recommend ways to improve the guidelines provided to SMEs in flood risk areas.
Resumo:
Serious games are a category of games which are designed for a specific purpose other than for pure entertainment. It is not a new concept but serious games using real data, coupled with real time modelling and combining model results with social and economic factors opens up a new paradigm for active stakeholder participation. DHI and UNEP-DHI Centre initiated a project called Aqua Republica where a virtual world is developed which allows participants to develop a river basin and visualise the consequences of their decisions. The aim of this project is to raise awareness of the interconnectivity of water and educate on integrated water resources management. Aqua Republica combines a game layer with a water allocation model, MIKE BASIN, to create an interactive, realistic virtual environment where players play the role of a catchment manager of an undeveloped river catchment. Their main objective is to develop the river catchment to be as prosperous as it can be. To achieve that, they will need to generate a good economy in the catchment to provide the funds needed for development, have a steady food supply for their population and enough energy and water for the catchment. Through these actions by the player, a meaningful play is established to engage players and to educate them about the complex relationships between developmental actions in a river basin and the natural environment as well as their consequences. The game layer also consists of a reward system to encourage learning. People can play and replay the game, get rewarded from performing the right principles and penalised from failures in the game. This abstract will explain the concept of the game and how it has been used in a stakeholder participation environment.
Resumo:
Devido à necessidade de permanecer competitiva e rentável, a indústriada construção civil tem mudado a forma de conduzir seus negócios e adaptado filosofias gerenciais para a suarealidade, como é o casodo Just-in-Time (JIT)e dos Sistemas da Oualidade (por exemplo: TOM e TOC). Dentro deste contexto, encontra-se a aplicação da Teoria das Restrições (TR) no processo de Planejamento e Controle da Produção (PCP) de obras de edificação. A Teoria das Restrições é uma filosofia gerencial fundamentada no conceito de restrição e em como gerenciar as restrições identificadas, fornecendo: (a) um procedimento para os responsáveis pela produção identificarem onde e como concentrar esforços e (b) um processo de melhoria contínua para o sistema produtivo como um todo, buscando sempre o ótimo globalantes do local. O desconhecimento dos possíveis benefícios e dificuldades decorrentes da adaptação e aplicação desta teoria na indústria da construção civil fez com que a aplicação da Teoria das Restrições fosse simulada, usando a técnica STROBOSCOPE (Stateand Resource Based Simulation of Construction Processes), em um empreendimento para analisar os impactos que prazos e custos podem sofrer durante as aplicações em casos reais. Estas simulações mostraram que a Teoria das Restrições e suas técnicas e ferramentas são aplicáveis ao processo de PCP, reduzindo os atrasos e os custos por atraso que um empreendimento pode sofrer.
Resumo:
Neste artigo, discute-se os desafios da implementação da política de atendimento socioeducativo em torno dos entes federados, proposta pela Lei 12.594/ 2012– Sistema Nacional de Atendimento Socioeducativo – SINASE. Para isso, realizou-se análise comparativa das normativas de Minas Gerais e Rio Grande do Sul a partir do olhar sobre a regulamentação do exercício profissional do agente de segurança socioeducativa, tendo por base os instrumentos institucionais: as políticas estaduais e o regimento da função de segurança socioeducativa. Identificou-se a prevalência de aspectos de segurança aos de socioeducação, o que torna a efetiva implementação do novo paradigma ainda um desafio. Apontam-se mudanças institucionais como propostas de aperfeiçoamento da política nos entes estaduais.
Resumo:
O aumento da complexidade do mercado financeiro tem sido relatado por Rajan (2005), Gorton (2008) e Haldane e May (2011) como um dos principais fatores responsáveis pelo incremento do risco sistêmico que culminou na crise financeira de 2007/08. O Bank for International Settlements (2013) aborda a questão da complexidade no contexto da regulação bancária e discute a comparabilidade da adequação de capital entre os bancos e entre jurisdições. No entanto, as definições dos conceitos de complexidade e de sistemas adaptativos complexos são suprimidas das principais discussões. Este artigo esclarece alguns conceitos relacionados às teorias da Complexidade, como se dá a emergência deste fenômeno, como os conceitos podem ser aplicados ao mercado financeiro. São discutidas duas ferramentas que podem ser utilizadas no contexto de sistemas adaptativos complexos: Agent Based Models (ABMs) e entropia e comparadas com ferramentas tradicionais. Concluímos que ainda que a linha de pesquisa da complexidade deixe lacunas, certamente esta contribui com a agenda de pesquisa econômica para se compreender os mecanismos que desencadeiam riscos sistêmicos, bem como adiciona ferramentas que possibilitam modelar agentes heterogêneos que interagem, de forma a permitir o surgimento de fenômenos emergentes no sistema. Hipóteses de pesquisa são sugeridas para aprofundamento posterior.
Resumo:
Oil wells subjected to cyclic steam injection present important challenges for the development of well cementing systems, mainly due to tensile stresses caused by thermal gradients during its useful life. Cement sheath failures in wells using conventional high compressive strength systems lead to the use of cement systems that are more flexible and/or ductile, with emphasis on Portland cement systems with latex addition. Recent research efforts have presented geopolymeric systems as alternatives. These cementing systems are based on alkaline activation of amorphous aluminosilicates such as metakaolin or fly ash and display advantageous properties such as high compressive strength, fast setting and thermal stability. Basic geopolymeric formulations can be found in the literature, which meet basic oil industry specifications such as rheology, compressive strength and thickening time. In this work, new geopolymeric formulations were developed, based on metakaolin, potassium silicate, potassium hydroxide, silica fume and mineral fiber, using the state of the art in chemical composition, mixture modeling and additivation to optimize the most relevant properties for oil well cementing. Starting from molar ratios considered ideal in the literature (SiO2/Al2O3 = 3.8 e K2O/Al2O3 = 1.0), a study of dry mixtures was performed,based on the compressive packing model, resulting in an optimal volume of 6% for the added solid material. This material (silica fume and mineral fiber) works both as an additional silica source (in the case of silica fume) and as mechanical reinforcement, especially in the case of mineral fiber, which incremented the tensile strength. The first triaxial mechanical study of this class of materials was performed. For comparison, a mechanical study of conventional latex-based cementing systems was also carried out. Regardless of differences in the failure mode (brittle for geopolymers, ductile for latex-based systems), the superior uniaxial compressive strength (37 MPa for the geopolymeric slurry P5 versus 18 MPa for the conventional slurry P2), similar triaxial behavior (friction angle 21° for P5 and P2) and lower stifness (in the elastic region 5.1 GPa for P5 versus 6.8 GPa for P2) of the geopolymeric systems allowed them to withstand a similar amount of mechanical energy (155 kJ/m3 for P5 versus 208 kJ/m3 for P2), noting that geopolymers work in the elastic regime, without the microcracking present in the case of latex-based systems. Therefore, the geopolymers studied on this work must be designed for application in the elastic region to avoid brittle failure. Finally, the tensile strength of geopolymers is originally poor (1.3 MPa for the geopolymeric slurry P3) due to its brittle structure. However, after additivation with mineral fiber, the tensile strength became equivalent to that of latex-based systems (2.3 MPa for P5 and 2.1 MPa for P2). The technical viability of conventional and proposed formulations was evaluated for the whole well life, including stresses due to cyclic steam injection. This analysis was performed using finite element-based simulation software. It was verified that conventional slurries are viable up to 204ºF (400ºC) and geopolymeric slurries are viable above 500ºF (260ºC)
Resumo:
This paper aim to check a hypothesis that assumes several behaviors related to social work norm´s obeying as a phenomenon that can be explained by actor´s social network structure and the rational choice processes related to the social norm inside that network, principally the payoff´s analysis received by the closest actors, or neighbors, at a social situation. Taking the sociological paradigm of rational action theory as a basis, the focus is on a debate about the logic of social norms, from Émile Durkheim´s method to Jon Elster´s theory, but also including social network analysis´s variables according to Robert Hanneman; and also Vilfredo Pareto´s constants related to human sociability, at the aim to detect elements that can help the scholars to develop an agent based model which could explain the sociological problem of deviance by a better way than the common sense´s view about morality and ethics at a social work environment