828 resultados para network models
Resumo:
O estudo dos diferentes fenômenos de separação tem sido cada vez mais importante para os diferentes ramos da indústria e ciência. Devido à grande capacidade computacional atual, é possível modelar e analisar os fenômenos cromatográficos a nível microscópico. Os modelos de rede vêm sendo cada vez mais utilizados, para representar processos de separação por cromatografia, pois através destes pode-se representar os aspectos topológicos e morfológicos dos diferentes materiais adsorventes disponíveis no mercado. Neste trabalho visamos o desenvolvimento de um modelo de rede tridimensional para representação de uma coluna cromatográfica, a nível microscópico, onde serão modelados os fenômenos de adsorção, dessorção e dispersão axial através de um método estocástico. Também foram utilizadas diferentes abordagens com relação ao impedimento estérico Os resultados obtidos foram comparados a resultados experimentais. Depois é utilizado um modelo de rede bidimensional para representar um sistema de adsorção do tipo batelada, mantendo-se a modelagem dos fenômenos de adsorção e dessorção, e comparados a sistemas reais posteriormente. Em ambos os sistemas modelados foram analisada as constantes de equilíbrio, parâmetro fundamental nos sistemas de adsorção, e por fim foram obtidas e analisadas isotermas de adsorção. Foi possível concluir que, para os modelos de rede, os fenômenos de adsorção e dessorção bastam para obter perfis de saída similares aos vistos experimentalmente, e que o fenômeno da dispersão axial influência menos que os fenômenos cinéticos em questão
Resumo:
Perceptual learning improves perception through training. Perceptual learning improves with most stimulus types but fails when . certain stimulus types are mixed during training (roving). This result is surprising because classical supervised and unsupervised neural network models can cope easily with roving conditions. What makes humans so inferior compared to these models? As experimental and conceptual work has shown, human perceptual learning is neither supervised nor unsupervised but reward-based learning. Reward-based learning suffers from the so-called unsupervised bias, i.e., to prevent synaptic " drift" , the . average reward has to be exactly estimated. However, this is impossible when two or more stimulus types with different rewards are presented during training (and the reward is estimated by a running average). For this reason, we propose no learning occurs in roving conditions. However, roving hinders perceptual learning only for combinations of similar stimulus types but not for dissimilar ones. In this latter case, we propose that a critic can estimate the reward for each stimulus type separately. One implication of our analysis is that the critic cannot be located in the visual system. © 2011 Elsevier Ltd.
Resumo:
This paper presents ongoing work on data collection and collation from a large number of laboratory cement-stabilization projects worldwide. The aim is to employ Artificial Neural Networks (ANN) to establish relationships between variables, which define the properties of cement-stabilized soils, and the two parameters determined by the Unconfined Compression Test, the Unconfined Compressive Strength (UCS), and stiffness, using E50 calculated from UCS results. Bayesian predictive neural network models are developed to predict the UCS values of cement-stabilized inorganic clays/silts, as well as sands as a function of selected soil mix variables, such as grain size distribution, water content, cement content and curing time. A model which can predict the stiffness values of cement-stabilized clays/silts is also developed and compared to the UCS model. The UCS model results emulate known trends better and provide more accurate estimates than the results from the E50 stiffness model. © 2013 American Society of Civil Engineers.
Resumo:
We compared nonlinear principal component analysis (NLPCA) with linear principal component analysis (LPCA) with the data of sea surface wind anomalies (SWA), surface height anomalies (SSHA), and sea surface temperature anomalies (SSTA), taken in the South China Sea (SCS) between 1993 and 2003. The SCS monthly data for SWA, SSHA and SSTA (i.e., the anomalies with climatological seasonal cycle removed) were pre-filtered by LPCA, with only three leading modes retained. The first three modes of SWA, SSHA, and SSTA of LPCA explained 86%, 71%, and 94% of the total variance in the original data, respectively. Thus, the three associated time coefficient functions (TCFs) were used as the input data for NLPCA network. The NLPCA was made based on feed-forward neural network models. Compared with classical linear PCA, the first NLPCA mode could explain more variance than linear PCA for the above data. The nonlinearity of SWA and SSHA were stronger in most areas of the SCS. The first mode of the NLPCA on the SWA and SSHA accounted for 67.26% of the variance versus 54.7%, and 60.24% versus 50.43%, respectively for the first LPCA mode. Conversely, the nonlinear SSTA, localized in the northern SCS and southern continental shelf region, resulted in little improvement in the explanation of the variance for the first NLPCA.
Resumo:
简要介绍了模糊petri网以及模糊产生式规则,给出了基于模糊petri网的专家系统的框架设计,并提出了模糊产生式规则和模糊petri网的详细设计,根据本设计方案开发了汽车变速箱故障诊断专家系统,证明设计方案简洁高效,扩充性和实用性好。
Resumo:
Financial time series convey the decisions and actions of a population of human actors over time. Econometric and regressive models have been developed in the past decades for analyzing these time series. More recently, biologically inspired artificial neural network models have been shown to overcome some of the main challenges of traditional techniques by better exploiting the non-linear, non-stationary, and oscillatory nature of noisy, chaotic human interactions. This review paper explores the options, benefits, and weaknesses of the various forms of artificial neural networks as compared with regression techniques in the field of financial time series analysis.
Resumo:
This article describes neural network models for adaptive control of arm movement trajectories during visually guided reaching and, more generally, a framework for unsupervised real-time error-based learning. The models clarify how a child, or untrained robot, can learn to reach for objects that it sees. Piaget has provided basic insights with his concept of a circular reaction: As an infant makes internally generated movements of its hand, the eyes automatically follow this motion. A transformation is learned between the visual representation of hand position and the motor representation of hand position. Learning of this transformation eventually enables the child to accurately reach for visually detected targets. Grossberg and Kuperstein have shown how the eye movement system can use visual error signals to correct movement parameters via cerebellar learning. Here it is shown how endogenously generated arm movements lead to adaptive tuning of arm control parameters. These movements also activate the target position representations that are used to learn the visuo-motor transformation that controls visually guided reaching. The AVITE model presented here is an adaptive neural circuit based on the Vector Integration to Endpoint (VITE) model for arm and speech trajectory generation of Bullock and Grossberg. In the VITE model, a Target Position Command (TPC) represents the location of the desired target. The Present Position Command (PPC) encodes the present hand-arm configuration. The Difference Vector (DV) population continuously.computes the difference between the PPC and the TPC. A speed-controlling GO signal multiplies DV output. The PPC integrates the (DV)·(GO) product and generates an outflow command to the arm. Integration at the PPC continues at a rate dependent on GO signal size until the DV reaches zero, at which time the PPC equals the TPC. The AVITE model explains how self-consistent TPC and PPC coordinates are autonomously generated and learned. Learning of AVITE parameters is regulated by activation of a self-regulating Endogenous Random Generator (ERG) of training vectors. Each vector is integrated at the PPC, giving rise to a movement command. The generation of each vector induces a complementary postural phase during which ERG output stops and learning occurs. Then a new vector is generated and the cycle is repeated. This cyclic, biphasic behavior is controlled by a specialized gated dipole circuit. ERG output autonomously stops in such a way that, across trials, a broad sample of workspace target positions is generated. When the ERG shuts off, a modulator gate opens, copying the PPC into the TPC. Learning of a transformation from TPC to PPC occurs using the DV as an error signal that is zeroed due to learning. This learning scheme is called a Vector Associative Map, or VAM. The VAM model is a general-purpose device for autonomous real-time error-based learning and performance of associative maps. The DV stage serves the dual function of reading out new TPCs during performance and reading in new adaptive weights during learning, without a disruption of real-time operation. YAMs thus provide an on-line unsupervised alternative to the off-line properties of supervised error-correction learning algorithms. YAMs and VAM cascades for learning motor-to-motor and spatial-to-motor maps are described. YAM models and Adaptive Resonance Theory (ART) models exhibit complementary matching, learning, and performance properties that together provide a foundation for designing a total sensory-cognitive and cognitive-motor autonomous system.
Resumo:
Neural network models of working memory, called Sustained Temporal Order REcurrent (STORE) models, are described. They encode the invariant temporal order of sequential events in short term memory (STM) in a way that mimics cognitive data about working memory, including primacy, recency, and bowed order and error gradients. As new items are presented, the pattern of previously stored items is invariant in the sense that, relative activations remain constant through time. This invariant temporal order code enables all possible groupings of sequential events to be stably learned and remembered in real time, even as new events perturb the system. Such a competence is needed to design self-organizing temporal recognition and planning systems in which any subsequence of events may need to be categorized in order to to control and predict future behavior or external events. STORE models show how arbitrary event sequences may be invariantly stored, including repeated events. A preprocessor interacts with the working memory to represent event repeats in spatially separate locations. It is shown why at least two processing levels are needed to invariantly store events presented with variable durations and interstimulus intervals. It is also shown how network parameters control the type and shape of primacy, recency, or bowed temporal order gradients that will be stored.
Resumo:
Multilevel algorithms are a successful class of optimisation techniques which address the mesh partitioning problem for distributing unstructured meshes onto parallel computers. They usually combine a graph contraction algorithm together with a local optimisation method which refines the partition at each graph level. To date these algorithms have been used almost exclusively to minimise the cut edge weight in the graph with the aim of minimising the parallel communication overhead, but recently there has been a perceived need to take into account the communications network of the parallel machine. For example the increasing use of SMP clusters (systems of multiprocessor compute nodes with very fast intra-node communications but relatively slow inter-node networks) suggest the use of hierarchical network models. Indeed this requirement is exacerbated in the early experiments with meta-computers (multiple supercomputers combined together, in extreme cases over inter-continental networks). In this paper therefore, we modify a multilevel algorithm in order to minimise a cost function based on a model of the communications network. Several network models and variants of the algorithm are tested and we establish that it is possible to successfully guide the optimisation to reflect the chosen architecture.
Resumo:
The diversity gains achievable in the generalised distributed antenna system with cooperative users (GDAS-CU) are considered. A GDAS-CU is comprised of M largely separated access points (APs) at one side of the link, and N geographically closed user terminals (UTs) at the other side. The UTs are collaborating together to enhance the system performance, where an idealised message sharing among the UTs is assumed. First, geometry-based network models are proposed to describe the topology of a GDAS-CU. The mean cross-correlation coefficients of signals received from non-collocated APs and UTs are calculated based on the network topology and the correlation models derived from the empirical data. The analysis is also extendable to more general scenarios where the APs are placed in a clustered form due to the constraints of street layout or building structure. Subsequently, a generalised signal attenuation model derived from several stochastic ray-tracing-based pathloss models is applied to describe the power-decaying pattern in urban built-up areas, where the GDAS-CU may be deployed. Armed with the cross-correlation and pathloss model preliminaries, an intrinsic measure of cooperative diversity obtainable from a GDAS-CU is then derived, which is the number of independent fading channels that can be averaged over to detect symbols. The proposed analytical framework would provide critical insight into the degree of possible performance improvement when combining multiple copies of the received signal in such systems.
Resumo:
A novel methodology is proposed for the development of neural network models for complex engineering systems exhibiting nonlinearity. This method performs neural network modeling by first establishing some fundamental nonlinear functions from a priori engineering knowledge, which are then constructed and coded into appropriate chromosome representations. Given a suitable fitness function, using evolutionary approaches such as genetic algorithms, a population of chromosomes evolves for a certain number of generations to finally produce a neural network model best fitting the system data. The objective is to improve the transparency of the neural networks, i.e. to produce physically meaningful
Resumo:
Neural network models have been explored for the prediction of the liquid-liquid equilibrium data and aromatic/aliphatic selectivity values. Four ternary systems composed of toluene, heptane, and the ionic liquids 1-ethyl-3-methylimidazolium ethylsulfate, or 1,3-dimethylimidazolium methylsulfate were investigated at 313.2 and 348.2 K.
Resumo:
The properties and characteristics of a recently proposed anisotropic metamaterial based upon layered arrays of tightly coupled pairs of "dogbone" shaped stripe conductors have been explored in detail. It has been found that a metamaterial composed of such stacked layers exhibits artificial magnetism and may support backward wave propagation. The equivalent network models of the constitutive conductor pairs arranged in the periodic array have been devised and applied to the identification of the specific types of resonances, and to the analysis of their contribution into the effective dielectric and magnetic properties of the artificial medium. The proposed "dogbone" configuration of conductor pairs has the advantage of being entirely realizable and assemblable in planar technology. It also appears more prospective than simple cut-wire or metal-plate pairs because the additional geometrical parameters provide an efficient control of separation between the electric and magnetic resonances that, in turn, makes it possible to obtain a fairly broadband left-handed behaviour of the structure at low frequencies.
Resumo:
Real-world graphs or networks tend to exhibit a well-known set of properties, such as heavy-tailed degree distributions, clustering and community formation. Much effort has been directed into creating realistic and tractable models for unlabelled graphs, which has yielded insights into graph structure and evolution. Recently, attention has moved to creating models for labelled graphs: many real-world graphs are labelled with both discrete and numeric attributes. In this paper, we presentAgwan (Attribute Graphs: Weighted and Numeric), a generative model for random graphs with discrete labels and weighted edges. The model is easily generalised to edges labelled with an arbitrary number of numeric attributes. We include algorithms for fitting the parameters of the Agwanmodel to real-world graphs and for generating random graphs from the model. Using real-world directed and undirected graphs as input, we compare our approach to state-of-the-art random labelled graph generators and draw conclusions about the contribution of discrete vertex labels and edge weights to graph structure.
Resumo:
Titanium alloy exhibits an excellent combination of bio-compatibility, corrosion resistance, strength and toughness. The microstructure of an alloy influences the properties. The microstructures depend mainly on alloying elements, method of production, mechanical, and thermal treatments. The relationships between these variables and final properties of the alloy are complex, non-linear in nature, which is the biggest hurdle in developing proper correlations between them by conventional methods. So, we developed artificial neural networks (ANN) models for solving these complex phenomena in titanium alloys.
In the present work, ANN models were used for the analysis and prediction of the correlation between the process parameters, the alloying elements, microstructural features, beta transus temperature and mechanical properties in titanium alloys. Sensitivity analysis of trained neural network models were studied which resulted a better understanding of relationships between inputs and outputs. The model predictions and the analysis are well in agreement with the experimental results. The simulation results show that the average output-prediction error by models are less than 5% of the prediction range in more than 95% of the cases, which is quite acceptable for all metallurgical purposes.