846 resultados para Computer-based simulation


Relevância:

80.00% 80.00%

Publicador:

Resumo:

Devido à necessidade de permanecer competitiva e rentável, a indústriada construção civil tem mudado a forma de conduzir seus negócios e adaptado filosofias gerenciais para a suarealidade, como é o casodo Just-in-Time (JIT)e dos Sistemas da Oualidade (por exemplo: TOM e TOC). Dentro deste contexto, encontra-se a aplicação da Teoria das Restrições (TR) no processo de Planejamento e Controle da Produção (PCP) de obras de edificação. A Teoria das Restrições é uma filosofia gerencial fundamentada no conceito de restrição e em como gerenciar as restrições identificadas, fornecendo: (a) um procedimento para os responsáveis pela produção identificarem onde e como concentrar esforços e (b) um processo de melhoria contínua para o sistema produtivo como um todo, buscando sempre o ótimo globalantes do local. O desconhecimento dos possíveis benefícios e dificuldades decorrentes da adaptação e aplicação desta teoria na indústria da construção civil fez com que a aplicação da Teoria das Restrições fosse simulada, usando a técnica STROBOSCOPE (Stateand Resource Based Simulation of Construction Processes), em um empreendimento para analisar os impactos que prazos e custos podem sofrer durante as aplicações em casos reais. Estas simulações mostraram que a Teoria das Restrições e suas técnicas e ferramentas são aplicáveis ao processo de PCP, reduzindo os atrasos e os custos por atraso que um empreendimento pode sofrer.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

O crescimento do papel da tecnologia nas interações entre empresa clientes, bem como do número de serviços baseados em tecnologia, vem alterando bruscamente as relações interpessoais e, em algumas instâncias, eliminando-as por completo. Por outro lado, tecnologia vem possibilitando, de forma drástica, aumento no número de encontros entre os consumidores e as empresas, que faz crescer importância do tema. presente trabalho tem como objetivo estudar insatisfação dos consumidores no atendimento, através da decepção sofrida, abordada quando esta ocorre em dois níveis de interação destes com as empresas de serviços, descrevendo comparando estas ocorrências no atendimento pessoal impessoal neste último, através das novas tecnologias disponibilizadas pelas empresas prestadoras de serviços. Os dados analisados foram coletados através de uma netnografia, ou etnografia na internet, adaptada ao estudo de comunidades online, uma nova metodologia em pesquisa qualitativa que tem por finalidade estudar as culturas comunidades que estão surgindo através das comunicações via computador, conforme proposto por Kozinets (2002). Através da análise de conteúdo dos relatos, buscou-se analisar os antecedentes conseqüentes da insatisfação decepção dos consumidores, analisando-os, descrevendo-os comparando-os. Conclui-se que as falhas são maior fonte de insatisfação para ambos os tipos de consumidores, além de outras similaridades observadas; que os consumidores que buscam atendimento pessoal exigem um tratamento mais caloroso; aqueles que buscam atendimento via novas tecnologias são menos tolerantes às falhas; que ambos recorrem meios diferentes daquele que foi fonte da decepção na tentativa de resolver seus problemas.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Simulations based on cognitively rich agents can become a very intensive computing task, especially when the simulated environment represents a complex system. This situation becomes worse when time constraints are present. This kind of simulations would benefit from a mechanism that improves the way agents perceive and react to changes in these types of environments. In other worlds, an approach to improve the efficiency (performance and accuracy) in the decision process of autonomous agents in a simulation would be useful. In complex environments, and full of variables, it is possible that not every information available to the agent is necessary for its decision-making process, depending indeed, on the task being performed. Then, the agent would need to filter the coming perceptions in the same as we do with our attentions focus. By using a focus of attention, only the information that really matters to the agent running context are perceived (cognitively processed), which can improve the decision making process. The architecture proposed herein presents a structure for cognitive agents divided into two parts: 1) the main part contains the reasoning / planning process, knowledge and affective state of the agent, and 2) a set of behaviors that are triggered by planning in order to achieve the agent s goals. Each of these behaviors has a runtime dynamically adjustable focus of attention, adjusted according to the variation of the agent s affective state. The focus of each behavior is divided into a qualitative focus, which is responsible for the quality of the perceived data, and a quantitative focus, which is responsible for the quantity of the perceived data. Thus, the behavior will be able to filter the information sent by the agent sensors, and build a list of perceived elements containing only the information necessary to the agent, according to the context of the behavior that is currently running. Based on the human attention focus, the agent is also dotted of a affective state. The agent s affective state is based on theories of human emotion, mood and personality. This model serves as a basis for the mechanism of continuous adjustment of the agent s attention focus, both the qualitative and the quantative focus. With this mechanism, the agent can adjust its focus of attention during the execution of the behavior, in order to become more efficient in the face of environmental changes. The proposed architecture can be used in a very flexibly way. The focus of attention can work in a fixed way (neither the qualitative focus nor the quantitaive focus one changes), as well as using different combinations for the qualitative and quantitative foci variation. The architecture was built on a platform for BDI agents, but its design allows it to be used in any other type of agents, since the implementation is made only in the perception level layer of the agent. In order to evaluate the contribution proposed in this work, an extensive series of experiments were conducted on an agent-based simulation over a fire-growing scenario. In the simulations, the agents using the architecture proposed in this work are compared with similar agents (with the same reasoning model), but able to process all the information sent by the environment. Intuitively, it is expected that the omniscient agent would be more efficient, since they can handle all the possible option before taking a decision. However, the experiments showed that attention-focus based agents can be as efficient as the omniscient ones, with the advantage of being able to solve the same problems in a significantly reduced time. Thus, the experiments indicate the efficiency of the proposed architecture

Relevância:

80.00% 80.00%

Publicador:

Resumo:

A chemical process optimization and control is strongly correlated with the quantity of information can be obtained from the system. In biotechnological processes, where the transforming agent is a cell, many variables can interfere in the process, leading to changes in the microorganism metabolism and affecting the quantity and quality of final product. Therefore, the continuously monitoring of the variables that interfere in the bioprocess, is crucial to be able to act on certain variables of the system, keeping it under desirable operational conditions and control. In general, during a fermentation process, the analysis of important parameters such as substrate, product and cells concentration, is done off-line, requiring sampling, pretreatment and analytical procedures. Therefore, this steps require a significant run time and the use of high purity chemical reagents to be done. In order to implement a real time monitoring system for a benchtop bioreactor, these study was conducted in two steps: (i) The development of a software that presents a communication interface between bioreactor and computer based on data acquisition and process variables data recording, that are pH, temperature, dissolved oxygen, level, foam level, agitation frequency and the input setpoints of the operational parameters of the bioreactor control unit; (ii) The development of an analytical method using near-infrared spectroscopy (NIRS) in order to enable substrate, products and cells concentration monitoring during a fermentation process for ethanol production using the yeast Saccharomyces cerevisiae. Three fermentation runs were conducted (F1, F2 and F3) that were monitored by NIRS and subsequent sampling for analytical characterization. The data obtained were used for calibration and validation, where pre-treatments combined or not with smoothing filters were applied to spectrum data. The most satisfactory results were obtained when the calibration models were constructed from real samples of culture medium removed from the fermentation assays F1, F2 and F3, showing that the analytical method based on NIRS can be used as a fast and effective method to quantify cells, substrate and products concentration what enables the implementation of insitu real time monitoring of fermentation processes

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The search for better performance in the structural systems has been taken to more refined models, involving the analysis of a growing number of details, which should be correctly formulated aiming at defining a representative model of the real system. Representative models demand a great detailing of the project and search for new techniques of evaluation and analysis. Model updating is one of this technologies, it can be used to improve the predictive capabilities of computer-based models. This paper presents a FRF-based finite element model updating procedure whose the updating variables are physical parameters of the model. It includes the damping effects in the updating procedure assuming proportional and non proportional damping mechanism. The updating parameters are defined at an element level or macro regions of the model. So, the parameters are adjusted locally, facilitating the physical interpretation of the adjusting of the model. Different tests for simulated and experimental data are discussed aiming at evaluating the characteristics and potentialities of the methodology.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The search for better performance in the structural systems has been taken to more refined models, involving the analysis of a growing number of details, which should be correctly formulated aiming at defining a representative model of the real system. Representative models demand a great detailing of the project and search for new techniques of evaluation and analysis. Model updating is one of this technologies, it can be used to improve the predictive capabilities of computer-based models. This paper presents a FRF-based finite element model updating procedure whose the updating variables are physical parameters of the model. It includes the damping effects in the updating procedure assuming proportional and none proportional damping mechanism. The updating parameters are defined at an element level or macro regions of the model. So, the parameters are adjusted locally, facilitating the physical interpretation of the adjusting of the model. Different tests for simulated and experimental data are discussed aiming at defining the characteristics and potentialities of the methodology.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The analysis of large amounts of data is better performed by humans when represented in a graphical format. Therefore, a new research area called the Visual Data Mining is being developed endeavoring to use the number crunching power of computers to prepare data for visualization, allied to the ability of humans to interpret data presented graphically.This work presents the results of applying a visual data mining tool, called FastMapDB to detect the behavioral pattern exhibited by a dataset of clinical information about hemoglobinopathies known as thalassemia. FastMapDB is a visual data mining tool that get tabular data stored in a relational database such as dates, numbers and texts, and by considering them as points in a multidimensional space, maps them to a three-dimensional space. The intuitive three-dimensional representation of objects enables a data analyst to see the behavior of the characteristics from abnormal forms of hemoglobin, highlighting the differences when compared to data from a group without alteration.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Direct expressions for the design of a lead-lag continuous compensator using the root locus method and the procedure described in the 1970 and 1990 books by Ogata are presented. These results are useful in the Ogata design method because they avoid the geometrical determination of poles and zeros, making it easier to create a computer-based design.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Despite the frequent use of stepping motors in robotics, automation, and a variety of precision instruments, they can hardly be found in rotational viscometers. This paper proposes the use of a stepping motor to drive a conventional constant-shear-rate laboratory rotational viscometer to avoid the use of velocity sensor and gearbox and, thus, simplify the instrument design. To investigate this driving technique, a commercial rotating viscometer has been adapted to be driven by a bipolar stepping motor, which is controlled via a personal computer. Special circuitry has been added to microstep the stepping motor at selectable step sizes and to condition the torque signal. Tests have been carried out using the prototype to produce flow curves for two standard Newtonian fluids (920 and 12 560 mPa (.) s, both at 25 degrees C). The flow curves have been obtained by employing several distinct microstep sizes within the shear rate range of 50-500 s(-1). The results indicate the feasibility of the proposed driving technique.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)