899 resultados para Synchronization algorithms
Resumo:
In subjects suffering from early onset strabismus, signals conveyed by the two eyes are not perceived simultaneously but in alternation. We exploited this phenomenon of interocular suppression to investigate the neuronal correlate of binocular rivalry in primary visual cortex of awake strabismic cats. Monocularly presented stimuli that were readily perceived by the animal evoked synchronized discharges with an oscillatory patterning in the γ-frequency range. Upon dichoptic stimulation, neurons responding to the stimulus that continued to be perceived increased the synchronicity and the regularity of their oscillatory patterning while the reverse was true for neurons responding to the stimulus that was no longer perceived. These differential changes were not associated with modifications of discharge rate, suggesting that at early stages of visual processing the degree of synchronicity rather than the amplitude of responses determines which signals are perceived and control behavioral responses.
Resumo:
Experimental and modeling efforts suggest that rhythms in the CA1 region of the hippocampus that are in the beta range (12–29 Hz) have a different dynamical structure than that of gamma (30–70 Hz). We use a simplified model to show that the different rhythms employ different dynamical mechanisms to synchronize, based on different ionic currents. The beta frequency is able to synchronize over long conduction delays (corresponding to signals traveling a significant distance in the brain) that apparently cannot be tolerated by gamma rhythms. The synchronization properties are consistent with data suggesting that gamma rhythms are used for relatively local computations whereas beta rhythms are used for higher level interactions involving more distant structures.
Resumo:
Mathematical analysis of the subthreshold oscillatory properties of inferior olivary neurons in vitro indicates that the oscillation is nonlinear and supports low dimensional chaotic dynamics. This property leads to the generation of complex functional states that can be attained rapidly via phase coherence that conform to the category of “generalized synchronization.” Functionally, this translates into neuronal ensemble properties that can support maximum functional permissiveness and that rapidly can transform into robustly determined multicellular coherence.
Resumo:
Increasing global competition, rapidly changing markets, and greater consumer awareness have altered the way in which corporations do business. To become more efficient, many industries have sought to model some operational aspects by gigantic optimization problems. It is not atypical to encounter models that capture 106 separate “yes” or “no” decisions to be made. Although one could, in principle, try all 2106 possible solutions to find the optimal one, such a method would be impractically slow. Unfortunately, for most of these models, no algorithms are known that find optimal solutions with reasonable computation times. Typically, industry must rely on solutions of unguaranteed quality that are constructed in an ad hoc manner. Fortunately, for some of these models there are good approximation algorithms: algorithms that produce solutions quickly that are provably close to optimal. Over the past 6 years, there has been a sequence of major breakthroughs in our understanding of the design of approximation algorithms and of limits to obtaining such performance guarantees; this area has been one of the most flourishing areas of discrete mathematics and theoretical computer science.
Resumo:
The collective behavior of interconnected spiking nerve cells is investigated. It is shown that a variety of model systems exhibit the same short-time behavior and rapidly converge to (approximately) periodic firing patterns with locally synchronized action potentials. The dynamics of one model can be described by a downhill motion on an abstract energy landscape. Since an energy landscape makes it possible to understand and program computation done by an attractor network, the results will extend our understanding of collective computation from models based on a firing-rate description to biologically more realistic systems with integrate-and-fire neurons.
Resumo:
Este trabalho apresenta uma análise de algoritmos computacionais aplicados à estimação de fasores elétricos em SEPs. A medição dos fasores é realizada por meio da alocação de Unidades de Medição Fasorial nestes sistemas e encontra diversas aplicações nas áreas de operação, controle, proteção e planejamento. Para que os fasores possam ser aplicados, são definidos padrões de medição, sincronização e comunicação, por meio da norma IEEE C37.118.1. A norma apresenta os padrões de mensagens, timetag, fasores, sistema de sincronização, e define testes para avaliar a estimação. Apesar de abranger todos esses critérios, a diretriz não define um algoritmo de estimação padrão, abrindo espaço para uso de diversos métodos, desde que a precisão seja atendida. Nesse contexto, o presente trabalho analisa alguns algoritmos de estimação de fasores definidos na literatura, avaliando o comportamento deles em determinados casos. Foram considerados, dessa forma, os métodos: Transformada Discreta de Fourier, Método dos Mínimos Quadrados e Transformada Wavelet Discreta, nas versões recursivas e não-recursivas. Esses métodos foram submetidos a sinais sintéticos, a fim de verificar o comportamento diante dos testes propostos pela norma, avaliando o Total Vector Error, tempo de resposta e atraso e overshoot. Os algoritmos também foram embarcados em um hardware, denominado PC104, e avaliados de acordo com os sinais medidos pelo equipamento na saída analógica de um simulador em tempo real (Real Time Digital Simulator).
Resumo:
Comunicación presentada en EVACES 2011, 4th International Conference on Experimental Vibration Analysis for Civil Engineering Structures, Varenna (Lecco), Italy, October 3-5, 2011.
Resumo:
Phase equilibrium data regression is an unavoidable task necessary to obtain the appropriate values for any model to be used in separation equipment design for chemical process simulation and optimization. The accuracy of this process depends on different factors such as the experimental data quality, the selected model and the calculation algorithm. The present paper summarizes the results and conclusions achieved in our research on the capabilities and limitations of the existing GE models and about strategies that can be included in the correlation algorithms to improve the convergence and avoid inconsistencies. The NRTL model has been selected as a representative local composition model. New capabilities of this model, but also several relevant limitations, have been identified and some examples of the application of a modified NRTL equation have been discussed. Furthermore, a regression algorithm has been developed that allows for the advisable simultaneous regression of all the condensed phase equilibrium regions that are present in ternary systems at constant T and P. It includes specific strategies designed to avoid some of the pitfalls frequently found in commercial regression tools for phase equilibrium calculations. Most of the proposed strategies are based on the geometrical interpretation of the lowest common tangent plane equilibrium criterion, which allows an unambiguous comprehension of the behavior of the mixtures. The paper aims to show all the work as a whole in order to reveal the necessary efforts that must be devoted to overcome the difficulties that still exist in the phase equilibrium data regression problem.
Resumo:
We present an algorithm to process images of reflected Placido rings captured by a commercial videokeratoscope. Raw data are obtained with no Cartesian-to-polar-coordinate conversion, thus avoiding interpolation and associated numerical artifacts. The method provides a characteristic equation for the device and is able to process around 6 times more corneal data than the commercial software. Our proposal allows complete control over the whole process from the capture of corneal images until the computation of curvature radii.
Resumo:
In this paper, parallel Relaxed and Extrapolated algorithms based on the Power method for accelerating the PageRank computation are presented. Different parallel implementations of the Power method and the proposed variants are analyzed using different data distribution strategies. The reported experiments show the behavior and effectiveness of the designed algorithms for realistic test data using either OpenMP, MPI or an hybrid OpenMP/MPI approach to exploit the benefits of shared memory inside the nodes of current SMP supercomputers.
Resumo:
Different kinds of algorithms can be chosen so as to compute elementary functions. Among all of them, it is worthwhile mentioning the shift-and-add algorithms due to the fact that they have been specifically designed to be very simple and to save computer resources. In fact, almost the only operations usually involved with these methods are additions and shifts, which can be easily and efficiently performed by a digital processor. Shift-and-add algorithms allow fairly good precision with low cost iterations. The most famous algorithm belonging to this type is CORDIC. CORDIC has the capability of approximating a wide variety of functions with only the help of a slight change in their iterations. In this paper, we will analyze the requirements of some engineering and industrial problems in terms of type of operands and functions to approximate. Then, we will propose the application of shift-and-add algorithms based on CORDIC to these problems. We will make a comparison between the different methods applied in terms of the precision of the results and the number of iterations required.
Resumo:
Background and objective: In this paper, we have tested the suitability of using different artificial intelligence-based algorithms for decision support when classifying the risk of congenital heart surgery. In this sense, classification of those surgical risks provides enormous benefits as the a priori estimation of surgical outcomes depending on either the type of disease or the type of repair, and other elements that influence the final result. This preventive estimation may help to avoid future complications, or even death. Methods: We have evaluated four machine learning algorithms to achieve our objective: multilayer perceptron, self-organizing map, radial basis function networks and decision trees. The architectures implemented have the aim of classifying among three types of surgical risk: low complexity, medium complexity and high complexity. Results: Accuracy outcomes achieved range between 80% and 99%, being the multilayer perceptron method the one that offered a higher hit ratio. Conclusions: According to the results, it is feasible to develop a clinical decision support system using the evaluated algorithms. Such system would help cardiology specialists, paediatricians and surgeons to forecast the level of risk related to a congenital heart disease surgery.
Resumo:
Software for video-based multi-point frequency measuring and mapping: http://hdl.handle.net/10045/53429
Resumo:
L'activité électrique du coeur est initiée par la génération spontanée de potentiels d'action venant des cellules pacemaker du noeud sinusal (SN). Toute dysfonction au niveau de cette région entraîne une instabilité électrique du coeur. La majorité des patients souffrant d'un noeud sinusal déficient nécessitent l'implantation chirurgicale d'un pacemaker électronique; cependant, les limitations de cette approche incitent à la recherche d'une alternative thérapeutique. La base moléculaire des courants ioniques jouant un rôle crucial dans l'activité du noeud sinusal sont de plus en plus connues. Une composante importante de l'activité des cellules pacemakers semble être le canal HCN, responsable du courant pacemaker If. Le facteur T-box 3 (Tbx3), un facteur de transcription conservé durant le processus de l'évolution, est nécessaire au développement du système de conduction cardiaque. De précédentes études ont démontré que dans différentes lignées cellulaires le Phorbol 12-myristate 13-acetate (PMA) active l'expression du gène codant Tbx3 via des réactions en cascade partant de la protéine kinase C (PKC). L'objectif principal de cette étude est de tester si le PMA peut augmenter la fréquence et la synchronisation de l'activité spontanée du pacemaker biologique en culture. Plus précisément, nous avons étudié les effets de l'exposition chronique au PMA sur l'expression du facteur de transcription Tbx3, sur HCN4 et l'activité spontanée chez des monocouches de culture de myocytes ventriculaires de rats néonataux (MVRN). Nos résultats démontrent que le PMA augmente significativement le facteur transcription de Tbx3 et l'expression ARNm de HCN4, favorisant ainsi l'augmentation du rythme et de la stabilité de l'activité autonome. De plus, une diminution significative de la vitesse de conduction a été relevée et est attribuée à la diminution du couplage intercellulaire. La diminution de la vitesse de conduction pourrait expliquer l'effet négatif du PMA sur la synchronisation de l'activité autonome du pacemaker biologique. Ces résultats ont été confirmés par un modèle mathématique multicellulaire suggérant que des fréquences et résistances intercellulaires plus élevée pourraient induire une activité plus stable et moins synchrone. Cette étude amène de nouvelles connaissances très importantes destinées à la production d'un pacemaker biologique efficient et robuste.
Resumo:
We analyse regional business cycle synchronization in the Euro Area, using Gross Value Added in 53 NUTS 1 regions for a period of thirty years (1975-2005), detrended by Hodrick-Prescott and the Christiano-Fitzgerald filters. We conclude that, on average, synchronization has increased for the period considered with exceptions during the eighties and the beginning of the nineties. Still, the correlation of the business cycle in some regions with the benchmark remained low or even decreased. Our findings also support the hypothesis of the existence of a ‘national border’ effect.