1000 resultados para Hopfield networks


Relevância:

100.00% 100.00%

Publicador:

Resumo:

When Recurrent Neural Networks (RNN) are going to be used as Pattern Recognition systems, the problem to be considered is how to impose prescribed prototype vectors ξ^1,ξ^2,...,ξ^p as fixed points. The synaptic matrix W should be interpreted as a sort of sign correlation matrix of the prototypes, In the classical approach. The weak point in this approach, comes from the fact that it does not have the appropriate tools to deal efficiently with the correlation between the state vectors and the prototype vectors The capacity of the net is very poor because one can only know if one given vector is adequately correlated with the prototypes or not and we are not able to know what its exact correlation degree. The interest of our approach lies precisely in the fact that it provides these tools. In this paper, a geometrical vision of the dynamic of states is explained. A fixed point is viewed as a point in the Euclidean plane R2. The retrieving procedure is analyzed trough statistical frequency distribution of the prototypes. The capacity of the net is improved and the spurious states are reduced. In order to clarify and corroborate the theoretical results, together with the formal theory, an application is presented

Relevância:

70.00% 70.00%

Publicador:

Resumo:

The Radius of Direct attraction of a discrete neural network is a measure of stability of the network. it is known that Hopfield networks designed using Hebb's Rule have a radius of direct attraction of Omega(n/p) where n is the size of the input patterns and p is the number of them. This lower bound is tight if p is no larger than 4. We construct a family of such networks with radius of direct attraction Omega(n/root plog p), for any p greater than or equal to 5. The techniques used to prove the result led us to the first polynomial-time algorithm for designing a neural network with maximum radius of direct attraction around arbitrary input patterns. The optimal synaptic matrix is computed using the ellipsoid method of linear programming in conjunction with an efficient separation oracle. Restrictions of symmetry and non-negative diagonal entries in the synaptic matrix can be accommodated within this scheme.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

This paper illustrates how internal model control of nonlinear processes can be achieved by recurrent neural networks, e.g. fully connected Hopfield networks. It is shown that using results developed by Kambhampati et al. (1995), that once a recurrent network model of a nonlinear system has been produced, a controller can be produced which consists of the network comprising the inverse of the model and a filter. Thus, the network providing control for the nonlinear system does not require any training after it has been trained to model the nonlinear system. Stability and other issues of importance for nonlinear control systems are also discussed.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

This paper presents an efficient approach based on recurrent neural network for solving nonlinear optimization. More specifically, a modified Hopfield network is developed and its internal parameters are computed using the valid subspace technique. These parameters guarantee the convergence of the network to the equilibrium points that represent an optimal feasible solution. The main advantage of the developed network is that it treats optimization and constraint terms in different stages with no interference with each other. Moreover, the proposed approach does not require specification of penalty and weighting parameters for its initialization. A study of the modified Hopfield model is also developed to analyze its stability and convergence. Simulation results are provided to demonstrate the performance of the proposed neural network. (c) 2005 Elsevier B.V. All rights reserved.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Economic dispatch (ED) problems have recently been solved by artificial neural network approaches. Systems based on artificial neural networks have high computational rates due to the use of a massive number of simple processing elements and the high degree of connectivity between these elements. The ability of neural networks to realize some complex non-linear function makes them attractive for system optimization. All ED models solved by neural approaches described in the literature fail to represent the transmission system. Therefore, such procedures may calculate dispatch policies, which do not take into account important active power constraints. Another drawback pointed out in the literature is that some of the neural approaches fail to converge efficiently toward feasible equilibrium points. A modified Hopfield approach designed to solve ED problems with transmission system representation is presented in this paper. The transmission system is represented through linear load flow equations and constraints on active power flows. The internal parameters of such modified Hopfield networks are computed using the valid-subspace technique. These parameters guarantee the network convergence to feasible equilibrium points, which represent the solution for the ED problem. Simulation results and a sensitivity analysis involving IEEE 14-bus test system are presented to illustrate efficiency of the proposed approach. (C) 2004 Elsevier Ltd. All rights reserved.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This paper presents an efficient approach based on a recurrent neural network for solving constrained nonlinear optimization. More specifically, a modified Hopfield network is developed, and its internal parameters are computed using the valid-subspace technique. These parameters guarantee the convergence of the network to the equilibrium points that represent an optimal feasible solution. The main advantage of the developed network is that it handles optimization and constraint terms in different stages with no interference from each other. Moreover, the proposed approach does not require specification for penalty and weighting parameters for its initialization. A study of the modified Hopfield model is also developed to analyse its stability and convergence. Simulation results are provided to demonstrate the performance of the proposed neural network.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Sistemas baseados em redes neurais artificiais fornecem altas taxas de computação devido ao uso de um número massivo de elementos processadores simples. Redes neurais com conexões realimentadas fornecem um modelo computacional capaz de resolver uma rica classe de problemas de otimização. Este artigo apresenta uma nova abordagem para resolver problemas de otimização restrita utilizando redes neurais artificiais. Mais especificamente, uma rede de Hopfield modificada é desenvolvida cujos parâmetros internos são calculados usando a técnica de subespaço válido de soluções. A partir da obtenção destes parâmetros a rede tende a convergir aos pontos de equilíbrio que representam as possíveis soluções para o problema. Exemplos de simulação são apresentados para justificar a validade da abordagem proposta.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Systems based on artificial neural networks have high computational rates due to the use of a massive number of simple processing elements and the high degree of connectivity between these elements. Neural networks with feedback connections provide a computing model capable of solving a large class of optimization problems. This paper presents a novel approach for solving dynamic programming problems using artificial neural networks. More specifically, a modified Hopfield network is developed and its internal parameters are computed using the valid-subspace technique. These parameters guarantee the convergence of the network to the equilibrium points which represent solutions (not necessarily optimal) for the dynamic programming problem. Simulated examples are presented and compared with other neural networks. The results demonstrate that proposed method gives a significant improvement.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This paper gives a condition for the global stability of a continuous-time hopfield neural network when its activation function maybe not monotonically increasing.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A number of neural network models, in which fixed-point and limit-cycle attractors of the underlying dynamics are used to store and associatively recall information, are described. In the first class of models, a hierarchical structure is used to store an exponentially large number of strongly correlated memories. The second class of models uses limit cycles to store and retrieve individual memories. A neurobiologically plausible network that generates low-amplitude periodic variations of activity, similar to the oscillations observed in electroencephalographic recordings, is also described. Results obtained from analytic and numerical studies of the properties of these networks are discussed.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A grating-lens combination unit is developed to form a scaling self-transform function that can self-image on scale. Then an array of many such grating-lens units is used for the optical interconnection of a two-dimensional neural network, and experiments are carried out. We find that our idea is feasible, the optical interconnection system is simple, and optical adjustment is easy. (C) 1998 Optical Society of America.