99 resultados para multi-wave mixing
Resumo:
A novel agent-based approach to Meta-Heuristics self-configuration is proposed in this work. Meta-heuristics are examples of algorithms where parameters need to be set up as efficient as possible in order to unsure its performance. This paper presents a learning module for self-parameterization of Meta-heuristics (MHs) in a Multi-Agent System (MAS) for resolution of scheduling problems. The learning is based on Case-based Reasoning (CBR) and two different integration approaches are proposed. A computational study is made for comparing the two CBR integration perspectives. In the end, some conclusions are reached and future work outlined.
Resumo:
This paper describes a Multi-agent Scheduling System that assumes the existence of several Machines Agents (which are decision-making entities) distributed inside the Manufacturing System that interact and cooperate with other agents in order to obtain optimal or near-optimal global performances. Agents have to manage their internal behaviors and their relationships with other agents via cooperative negotiation in accordance with business policies defined by the user manager. Some Multi Agent Systems (MAS) organizational aspects are considered. An original Cooperation Mechanism for a Team-work based Architecture is proposed to address dynamic scheduling using Meta-Heuristics.
Resumo:
Emotion although being an important factor in our every day life it is many times forgotten in the development of systems to be used by persons. In this work we present an architecture for a ubiquitous group decision support system able to support persons in group decision processes. The system considers the emotional factors of the intervenient participants, as well as the argumentation between them. Particular attention will be taken to one of components of this system: the multi-agent simulator, modeling the human participants, considering emotional characteristics, and allowing the exchanges of hypothetic arguments among the participants.
Resumo:
With the increasing importance of large commerce across the Internet it is becoming increasingly evident that in a few years the Iternet will host a large number of interacting software agents. a vast number of them will be economically motivated, and will negociate a variety of goods and services. It is therefore important to consider the economic incentives and behaviours of economic software agents, and to use all available means to anticipate their collective interactions. This papers addresses this concern by presenting a multi-agent market simulator designed for analysing agent market strategies based on a complete understanding of buyer and seller behaviours, preference models and pricing algorithms, consideting risk preferences. The system includes agents that are capable of increasing their performance with their own experience, by adapting to the market conditions. The results of the negotiations between agents are analysed by data minig algorithms in order to extract rules that give agents feedback to imprive their strategies.
Resumo:
Mestrado em Engenharia Informática
Resumo:
Mestrado em Engenharia Informática
Resumo:
O mercado accionista, de uma forma global, tem-se revelado nos últimos tempos uma das principais fontes de incentivo ao mercado de valores mobiliários. O seu impacto junto do público em geral é enorme e a sua importância para as empresas é vital. Interessa, então, perceber como é que a teoria financeira tem obordado a avaliação e a compreensão do processo de formação de uma cotação. Desde os anos 50 até aos dias de hoje, interessa perceber como é que os diferentes autores têm tratado esta abordagem e quais os resultados deste confronto. Interessa sobretudo perceber o abordogem de Stephen Ross e a teoria do arbitragem. Na sequência desta obordagem e com o aparecimento do Multi Index Model, passou a ser possível extimar com maior precisão a evolução da cotação, na medida em que esta estaria dependente de um vasto conjunto de variavéis, que abragem uma vasta área de influência. O contributo de Ross é por isso decisivo. No final interessa reter a melhor técnica e teoria, que defende os interesses do investidor. Face o isto resta, então, saber qual a melhor técnica estatística para proceder a estes estudos empíricos.
Resumo:
Este trabalho visa apresentar um enquadramento da realidade económica e industrial do sector transformador de granitos ornamentais em Portugal e fazer uma análise do processo de serragem, com engenhos multi-lâminas e granalha de aço, na medida em que este é o método de seccionamento de blocos de granito mais utilizado pelas grandes indústrias do sector. Tendo em conta a importância económica desta operação produtiva na indústria em causa, foi definido como fito deste projecto a análise estatística dos custos de produção; a definição de fórmulas de cálculo que permitam prever o custo médio de serragem; e o estudo de soluções economicamente viáveis e ambientalmente sustentáveis para o problema das lamas resultantes do expurgo dos engenhos. Para a persecução deste projecto foi realizada uma recolha de dados implementando rotinas de controlo e registo dos mesmos, em quadros de produção normalizados e de fácil preenchimento, pelos operadores destes equipamentos. Esta recolha de dados permitiu isolar, quantificar e formular os factores de rentabilização do processo de serragem selecionando, dentro da amostra de estudo obtida, um conjunto de serragens com características similares e com valores próximos dos valores da média estatística. Apartir dos dados destas serragens foram geradas curvas de tendência polinomial com as quais se analisaram as variações provocadas no custo médio de serragem, pelas variações do factor em estudo. A formulação dos factores de rentabilização e os dados estatísticos obtidos permitiram depois o desenvolvimento de fórmulas de cálculo do custo médio de serragem que establecem o custo de produção diferenciado em função das espessuras com, ou sem, a incorporação dos factores de rentabilização. Como consequência do projecto realizado obteve-se um conjunto de conclusões util, para o sector industrial em causa, que evidencia a importancia da Ocupação dos engenhos e rentabilização de um espaço confinado, da Resistência oferecida à serragem pelos granitos, e da Diferença de altura entre os blocos de uma mesma carga, nos custos de transformação.
Resumo:
This chapter aims to demonstrate how PAOL - Unit for Innovation in Education, a project from ISCAP - School of Accounting and Administration of Oporto ....
Resumo:
A square-wave voltammetric (SWV) method using a hanging mercury drop electrode (HMDE) has been developed for determination of the herbicide molinate in a biodegradation process. The method is based on controlled adsorptive accumulation of molinate for 10 s at a potential of -0.8 V versus AgCl/Ag. An anodic peak, due to oxidation of the adsorbed pesticide, was observed in the cyclic voltammogram at ca. -0.320 V versus AgCl/Ag; a very small cathodic peak was also detected. The SWV calibration plot was established to be linear in the range 5.0x10-6 to 9.0x10-6 mol L-1; this corresponded to a detection limit of 3.5x10-8 mol L-1. This electroanalytical method was used to monitor the decrease of molinate concentration in river waters along a biodegradation process using a bacterial mixed culture. The results achieved with this voltammetric method were compared with those obtained by use of a chromatographic method (HPLC–UV) and no significant statistical differences were observed.
Resumo:
Electroanalytical methods based on square-wave adsorptive-stripping voltammetry (SWAdSV) and flow-injection analysis with square-wave adsorptive-stripping voltammetric detection (FIA-SWAdSV) were developed for the determination of fluoxetine (FXT). The methods were based on the reduction of FXT at a mercury drop electrode at -1.2 V versus Ag/AgCl, in a phosphate buffer of pH 12.0, and on the possibility of accumulating the compound at the electrode surface. The SWAdSV method was successfully applied in the quantification of FXT in pharmaceutical products, human serum samples, and in drug dissolution studies. Because the presence of dissolved oxygen did not interfere significantly with the analysis, it was possible to quantify FXT in several pharmaceutical products using FIA-SWAdSV. This method enables analysis of up to 120 samples per hour at reduced costs.
Resumo:
Solvent extraction is considered as a multi-criteria optimization problem, since several chemical species with similar extraction kinetic properties are frequently present in the aqueous phase and the selective extraction is not practicable. This optimization, applied to mixer–settler units, considers the best parameters and operating conditions, as well as the best structure or process flow-sheet. Global process optimization is performed for a specific flow-sheet and a comparison of Pareto curves for different flow-sheets is made. The positive weight sum approach linked to the sequential quadratic programming method is used to obtain the Pareto set. In all investigated structures, recovery increases with hold-up, residence time and agitation speed, while the purity has an opposite behaviour. For the same treatment capacity, counter-current arrangements are shown to promote recovery without significant impairment in purity. Recycling the aqueous phase is shown to be irrelevant, but organic recycling with as many stages as economically feasible clearly improves the design criteria and reduces the most efficient organic flow-rate.
Resumo:
Multi-objective particle swarm optimization (MOPSO) is a search algorithm based on social behavior. Most of the existing multi-objective particle swarm optimization schemes are based on Pareto optimality and aim to obtain a representative non-dominated Pareto front for a given problem. Several approaches have been proposed to study the convergence and performance of the algorithm, particularly by accessing the final results. In the present paper, a different approach is proposed, by using Shannon entropy to analyzethe MOPSO dynamics along the algorithm execution. The results indicate that Shannon entropy can be used as an indicator of diversity and convergence for MOPSO problems.
Resumo:
This study focused on the development of a sensitive enzymatic biosensor for the determination of pirimicarb pesticide based on the immobilization of laccase on composite carbon paste electrodes. Multi- walled carbon nanotubes(MWCNTs)paste electrode modified by dispersion of laccase(3%,w/w) within the optimum composite matrix(60:40%,w/w,MWCNTs and paraffin binder)showed the best performance, with excellent electron transfer kinetic and catalytic effects related to the redox process of the substrate4- aminophenol. No metal or anti-interference membrane was added. Based on the inhibition of laccase activity, pirimicarb can be determined in the range 9.90 ×10- 7 to 1.15 ×10- 5 molL 1 using 4- aminophenol as substrate at the optimum pH of 5.0, with acceptable repeatability and reproducibility (relative standard deviations lower than 5%).The limit of detection obtained was 1.8 × 10-7 molL 1 (0.04 mgkg 1 on a fresh weight vegetable basis).The high activity and catalytic properties of the laccase- based biosensor are retained during ca. one month. The optimized electroanalytical protocol coupled to the QuEChERS methodology were applied to tomato and lettuce samples spiked at three levels; recoveries ranging from 91.0±0.1% to 101.0 ± 0.3% were attained. No significant effects in the pirimicarb electro- analysis were observed by the presence of pro-vitamin A, vitamins B1 and C,and glucose in the vegetable extracts. The proposed biosensor- based pesticide residue methodology fulfills all requisites to be used in implementation of food safety programs.
Resumo:
We perform a comparison between the fractional iteration and decomposition methods applied to the wave equation on Cantor set. The operators are taken in the local sense. The results illustrate the significant features of the two methods which are both very effective and straightforward for solving the differential equations with local fractional derivative.