773 resultados para Collaborative Network Model


Relevância:

80.00% 80.00%

Publicador:

Resumo:

Antigen recognition by cytotoxic CD8 T cells is dependent upon a number of critical steps in MHC class I antigen processing including proteosomal cleavage, TAP transport into the endoplasmic reticulum, and MHC class 1 binding. Based on extensive experimental data relating to each of these steps there is now the capacity to model individual antigen processing steps with a high degree of accuracy. This paper demonstrates the potential to bring together models of individual antigen processing steps, for example proteosome cleavage, TAP transport, and MHC binding, to build highly informative models of functional pathways. In particular, we demonstrate how an artificial neural network model of TAP transport was used to mine a HLA-binding database so as to identify H LA-binding peptides transported by TAP. This integrated model of antigen processing provided the unique insight that HLA class I alleles apparently constitute two separate classes: those that are TAP-efficient for peptide loading (HLA-B27, -A3, and -A24) and those that are TAP-inefficient (HLA-A2, -B7, and -B8). Hence, using this integrated model we were able to generate novel hypotheses regarding antigen processing, and these hypotheses are now capable of being tested experimentally. This model confirms the feasibility of constructing a virtual immune system, whereby each additional step in antigen processing is incorporated into a single modular model. Accurate models of antigen processing have implications for the study of basic immunology as well as for the design of peptide-based vaccines and other immunotherapies. (C) 2004 Elsevier Inc. All rights reserved.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The present paper addresses two major concerns that were identified when developing neural network based prediction models and which can limit their wider applicability in the industry. The first problem is that it appears neural network models are not readily available to a corrosion engineer. Therefore the first part of this paper describes a neural network model of CO2 corrosion which was created using a standard commercial software package and simple modelling strategies. It was found that such a model was able to capture practically all of the trends noticed in the experimental data with acceptable accuracy. This exercise has proven that a corrosion engineer could readily develop a neural network model such as the one described below for any problem at hand, given that sufficient experimental data exist. This applies even in the cases when the understanding of the underlying processes is poor. The second problem arises from cases when all the required inputs for a model are not known or can be estimated with a limited degree of accuracy. It seems advantageous to have models that can take as input a range rather than a single value. One such model, based on the so-called Monte Carlo approach, is presented. A number of comparisons are shown which have illustrated how a corrosion engineer might use this approach to rapidly test the sensitivity of a model to the uncertainities associated with the input parameters. (C) 2001 Elsevier Science Ltd. All rights reserved.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In this paper, we develop a theory for diffusion and flow of pure sub-critical adsorbates in microporous activated carbon over a wide range of pressure, ranging from very low to high pressure, where capillary condensation is occurring. This theory does not require any fitting parameter. The only information needed for the prediction is the complete pore size distribution of activated carbon. The various interesting behaviors of permeability versus loading are observed such as the maximum permeability at high loading (occurred at about 0.8-0.9 relative pressure). The theory is tested with diffusion and flow of benzene through a commercial activated carbon, and the agreement is found to be very good in the light that there is no fitting parameter in the model. (C) 2001 Elsevier Science B.V. All rights reserved.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

A central problem in visual perception concerns how humans perceive stable and uniform object colors despite variable lighting conditions (i.e. color constancy). One solution is to 'discount' variations in lighting across object surfaces by encoding color contrasts, and utilize this information to 'fill in' properties of the entire object surface. Implicit in this solution is the caveat that the color contrasts defining object boundaries must be distinguished from the spurious color fringes that occur naturally along luminance-defined edges in the retinal image (i.e. optical chromatic aberration). In the present paper, we propose that the neural machinery underlying color constancy is complemented by an 'error-correction' procedure which compensates for chromatic aberration, and suggest that error-correction may be linked functionally to the experimentally induced illusory colored aftereffects known as McCollough effects (MEs). To test these proposals, we develop a neural network model which incorporates many of the receptive-field (RF) profiles of neurons in primate color vision. The model is composed of two parallel processing streams which encode complementary sets of stimulus features: one stream encodes color contrasts to facilitate filling-in and color constancy; the other stream selectively encodes (spurious) color fringes at luminance boundaries, and learns to inhibit the filling-in of these colors within the first stream. Computer simulations of the model illustrate how complementary color-spatial interactions between error-correction and filling-in operations (a) facilitate color constancy, (b) reveal functional links between color constancy and the ME, and (c) reconcile previously reported anomalies in the local (edge) and global (spreading) properties of the ME. We discuss the broader implications of these findings by considering the complementary functional roles performed by RFs mediating color-spatial interactions in the primate visual system. (C) 2002 Elsevier Science Ltd. All rights reserved.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia Mecânica

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Neste trabalho estuda-se a geração de trajectórias em tempo real de um robô quadrúpede. As trajectórias podem dividir-se em duas componentes: rítmica e discreta. A componente rítmica das trajectórias é modelada por uma rede de oito osciladores acoplados, com simetria 4 2 Z  Z . Cada oscilador é modelado matematicamente por um sistema de Equações Diferenciais Ordinárias. A referida rede foi proposta por Golubitsky, Stewart, Buono e Collins (1999, 2000), para gerar os passos locomotores de animais quadrúpedes. O trabalho constitui a primeira aplicação desta rede à geração de trajectórias de robôs quadrúpedes. A derivação deste modelo baseia-se na biologia, onde se crê que Geradores Centrais de Padrões de locomoção (CPGs), constituídos por redes neuronais, geram os ritmos associados aos passos locomotores dos animais. O modelo proposto gera soluções periódicas identificadas com os padrões locomotores quadrúpedes, como o andar, o saltar, o galopar, entre outros. A componente discreta das trajectórias dos robôs usa-se para ajustar a parte rítmica das trajectórias. Este tipo de abordagem é útil no controlo da locomoção em terrenos irregulares, em locomoção guiada (por exemplo, mover as pernas enquanto desempenha tarefas discretas para colocar as pernas em localizações específicas) e em percussão. Simulou-se numericamente o modelo de CPG usando o oscilador de Hopf para modelar a parte rítmica do movimento e um modelo inspirado no modelo VITE para modelar a parte discreta do movimento. Variou-se o parâmetro g e mediram-se a amplitude e a frequência das soluções periódicas identificadas com o passo locomotor quadrúpede Trot, para variação deste parâmetro. A parte discreta foi inserida na parte rítmica de duas formas distintas: (a) como um offset, (b) somada às equações que geram a parte rítmica. Os resultados obtidos para o caso (a), revelam que a amplitude e a frequência se mantêm constantes em função de g. Os resultados obtidos para o caso (b) revelam que a amplitude e a frequência aumentam até um determinado valor de g e depois diminuem à medida que o g aumenta, numa curva quase sinusoidal. A variação da amplitude das soluções periódicas traduz-se numa variação directamente proporcional na extensão do movimento do robô. A velocidade da locomoção do robô varia com a frequência das soluções periódicas, que são identificadas com passos locomotores quadrúpedes.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The demonstration proposal moves from the capabilities of a wireless biometric badge [4], which integrates a localization and tracking service along with an automatic personal identification mechanism, to show how a full system architecture is devised to enable the control of physical accesses to restricted areas. The system leverages on the availability of a novel IEEE 802.15.4/Zigbee Cluster Tree network model, on enhanced security levels and on the respect of all the users' privacy issues.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Este trabalho tem como objetivo a modelação hidráulica e de qualidade de água de parte de uma rede de distribuição a alta pressão do grande Porto. Após calibração de um modelo utilizado no software EPANet foi possível simular o decaimento do cloro livre num troço da rede de abastecimento. Foi ainda possível concluir que os valores dos parâmetros característicos do modelo de qualidade são uma constante de decaimento no seio do fluido de 1,78 dia-1 (0,001239 min-1) a cerca de 22 ºC e, no ramal 6244-6245 Ramalde – Cabanas – Pedrouços, uma constante de decaimento na parede da tubagem de forma generalizada de 0,28 m/dia. Não foi possível obter conclusões sobre a adutora 6261 Jovim-Nova Sintra 2, ficando explícita a necessidade de um maior controlo sobre a variável temperatura.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The prediction of the time and the efficiency of the remediation of contaminated soils using soil vapor extraction remain a difficult challenge to the scientific community and consultants. This work reports the development of multiple linear regression and artificial neural network models to predict the remediation time and efficiency of soil vapor extractions performed in soils contaminated separately with benzene, toluene, ethylbenzene, xylene, trichloroethylene, and perchloroethylene. The results demonstrated that the artificial neural network approach presents better performances when compared with multiple linear regression models. The artificial neural network model allowed an accurate prediction of remediation time and efficiency based on only soil and pollutants characteristics, and consequently allowing a simple and quick previous evaluation of the process viability.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Este relatório é fruto de um trabalho desenvolvido por um período de seis meses onde é apresentada uma breve introdução da empresa, bem como as políticas pelas quais ela funciona. São também descritas as formações obtidas e o plano de trabalhos seguido. Relativamente ao trabalho em si, o tema principal do trabalho desenvolvido, e tal como é apresentado no título deste relatório, é a modelação hidráulica de um sistema adutor em alta, isto é, onde foi elaborado um modelo para uma rede de abastecimento de água recorrendo aos registos de consumos fornecidos pela empresa Águas do Douro e Paiva, S.A.. É realizada uma descrição dos vários passos que levaram à sua elaboração e são apresentadas as conclusões relativas aos resultados obtidos. Uma vez que o trabalho incidiu sobre o funcionamento e aprendizagem de software apropriado, foram também retiradas conclusões/sugestões que poderão ser tomadas de forma a melhorar a experiência entre programas e utilizadores. Apesar do tempo disponível para a realização do plano de trabalhos ter demonstrado ser o suficiente para o cumprimento dos vários requisitos, foi também suficiente para permitir a realização de outros trabalhos que proporcionaram a aquisição de mais experiência no desenvolvimento de modelos em EPANET.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Objectives: Evaluate the production and the research collaborative network on Leishmaniasis in South America. Methods: A bibliometric research was carried out using SCOPUS database. The analysis unit was original research articles published from 2000 to 2011, that dealt with leishmaniasis and that included at least one South American author. The following items were obtained for each article: journal name, language, year of publication, number of authors, institutions, countries, and others variables. Results: 3,174 articles were published, 2,272 of them were original articles. 1,160 different institutional signatures, 58 different countries and 398 scientific journals were identified. Brazil was the country with more articles (60.7%) and Oswaldo Cruz Foundation (FIOCRUZ) had 18% of Brazilian production, which is the South American nucleus of the major scientific network in Leishmaniasis. Conclusions: South American scientific production on Leishmaniasis published in journals indexed in SCOPUS is focused on Brazilian research activity. It is necessary to strengthen the collaboration networks. The first step is to identify the institutions with higher production, in order to perform collaborative research according to the priorities of each country.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Dissertação para obtenção do Grau de Doutor em Informática

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This study analyses financial data using the result characterization of a self-organized neural network model. The goal was prototyping a tool that may help an economist or a market analyst to analyse stock market series. To reach this goal, the tool shows economic dependencies and statistics measures over stock market series. The neural network SOM (self-organizing maps) model was used to ex-tract behavioural patterns of the data analysed. Based on this model, it was de-veloped an application to analyse financial data. This application uses a portfo-lio of correlated markets or inverse-correlated markets as input. After the anal-ysis with SOM, the result is represented by micro clusters that are organized by its behaviour tendency. During the study appeared the need of a better analysis for SOM algo-rithm results. This problem was solved with a cluster solution technique, which groups the micro clusters from SOM U-Matrix analyses. The study showed that the correlation and inverse-correlation markets projects multiple clusters of data. These clusters represent multiple trend states that may be useful for technical professionals.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

La computación de altas prestaciones es una área de la informática que evoluciona rápidamente, en la que actualmente aparecen nuevos computadores que llegan a los petaflops. Al principio del trabajo, se estudian los distintos tipos de redes de interconexión y los modelos de red que se utilizan para medir su latencia. El objetivo de este trabajo, es el diseño, implementación y simulación de un modelo de red de interconexión basado en enlace, que tiene en cuenta la información de topología y enrutamiento de la red de interconexión. Teniendo en cuenta que los modelos son una abstracción del sistema, en éste trabajo se hace la verificación y validación del modelo, para asegurar que éste se aproxima a lo planteado en el diseño y también que se parece al sistema que se quiere modelar.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Graph pebbling is a network model for studying whether or not a given supply of discrete pebbles can satisfy a given demand via pebbling moves. A pebbling move across an edge of a graph takes two pebbles from one endpoint and places one pebble at the other endpoint; the other pebble is lost in transit as a toll. It has been shown that deciding whether a supply can meet a demand on a graph is NP-complete. The pebbling number of a graph is the smallest t such that every supply of t pebbles can satisfy every demand of one pebble. Deciding if the pebbling number is at most k is NP 2 -complete. In this paper we develop a tool, called theWeight Function Lemma, for computing upper bounds and sometimes exact values for pebbling numbers with the assistance of linear optimization. With this tool we are able to calculate the pebbling numbers of much larger graphs than in previous algorithms, and much more quickly as well. We also obtain results for many families of graphs, in many cases by hand, with much simpler and remarkably shorter proofs than given in previously existing arguments (certificates typically of size at most the number of vertices times the maximum degree), especially for highly symmetric graphs. Here we apply theWeight Function Lemma to several specific graphs, including the Petersen, Lemke, 4th weak Bruhat, Lemke squared, and two random graphs, as well as to a number of infinite families of graphs, such as trees, cycles, graph powers of cycles, cubes, and some generalized Petersen and Coxeter graphs. This partly answers a question of Pachter, et al., by computing the pebbling exponent of cycles to within an asymptotically small range. It is conceivable that this method yields an approximation algorithm for graph pebbling.