941 resultados para Grid search algorithm
Resumo:
Quantitative analysis of cine cardiac magnetic resonance (CMR) images for the assessment of global left ventricular morphology and function remains a routine task in clinical cardiology practice. To date, this process requires user interaction and therefore prolongs the examination (i.e. cost) and introduces observer variability. In this study, we sought to validate the feasibility, accuracy, and time efficiency of a novel framework for automatic quantification of left ventricular global function in a clinical setting.
Resumo:
The essay analises Sir Walter Raleigh's report (1595) of his search for Guiana, as well as his reflection on the New World and its inhabitants, which is, in many aspects, different from his contemporaries.
Resumo:
The reaction of nine vector species of Chagas' disease to infection by seven different Trypanosoma cruzi strains; Berenice, Y, FL, CL, S. Felipe, Colombiana and Gávea, are examined and compared. On the basis of the insects' ability to establish and maintain the infection, vector species could be divided into two distinct groups which differ in their reaction to an acute infection by T. cruzi. While the proportion of positive bugs was found to be low in Triatoma infestans and Triatoma dimidiata it was high, ranging from 96.9% to 100% in the group of wild (Rhodnius neglectus, Triatoma rubrovaria)and essentially sylvatic vectors in process of adaptation to human dwellings, maintained under control following successful insecticidal elimination of Triatoma infestans (Panstrongylus megistus, Triatoma sordida and Triatoma pseudomaculata). An intermediate position is held by Triatoma brasiliensis and Rhodnius prolixus. This latter has been found to interchange between domestic and sylvatic environments. The most important finding is the strikingly good reaction between each species of the sylvatic bugs and practically all T. cruzi strains herein studied, thus indicating that the factors responsible for the excellent reaction of P.megistus to infection by Y strain, as previously reported also come into operation in the reaction of the same vector species to acute infections by five of the remaining T.cruzi strains. Comparison or data reported by other investigators with those herein described form the basis of the discussion of Dipetalogaster maximus as regards its superiority as a xenodiagnostic agent.
Resumo:
Previous studies (1982,1987) have emphasized the superiority of sylvatic vector species over domestic species as xenodiagnostic agents in testing hosts with acute or chronic infections by T. cruzi "Y" stock. The present study, which is unique in that it contains data on both infectivity rates produced by the same stock in 11 different vector species and also the reaction of the same vector species to seven different parasite stocks, establishes the general validity of linking efficiency of xenodiagnosis to the biotope of its agent. For example, infectivity rates produced by "São Felipe" stock varied from 82.5% to 98.3% in sylvatic vectors but decreased to 42.5% to 71.3% in domestic species. "Colombiana" stock produced in the same sylvatic vectors infectivity rates ranging from 12.5% to 45%. These shrank to 5%-22.5% in domestic bugs. The functional role of the biotope in the vector-parasite interaction has not been eluddated. But since this phenomenon has been observed to be stable and easy to reproduce, it leads us to believe that the results obtained are valid. Data presented also provide increasing evidence that the infectivity rates exhibited by bugs from xenodiagnosis in chronic hosts, are parasite stock specific. For example, infectivity rates produced by "Berenice", "Y", "FL" and "CL" varied in R. neglectus from 26.3% to 75%; in P. megistus from 56.3% to 83.8%; in T. sordida from 28.8% to 58.8% in T. pseudomaculata from 41.3% to 66.3% and in T. rubrovaria from 48.8% to 85%. Data from xenodiagnosis in the same hosts, carrying acute infections by the same parasite stocks, gave the five sylvatic vectors a positive rating of approximately 100%, thus suggesting that the heavy loads of parasites circulating in the acute hosts obscured the characteristic interspecific differences for the parasite stock. Nonetheless these latter were revealed in the same hosts with chronic infections stimulated by very low numbers of the same parasite stocks. Certain observations here described lead us to speculate as to the possibility of further results from other parasite stocks, allowing the association of the infectivity rates produced in bugs by different parasite stocks with the isoenzymic patterns revealed by these stocks.
Resumo:
Lossless compression algorithms of the Lempel-Ziv (LZ) family are widely used nowadays. Regarding time and memory requirements, LZ encoding is much more demanding than decoding. In order to speed up the encoding process, efficient data structures, like suffix trees, have been used. In this paper, we explore the use of suffix arrays to hold the dictionary of the LZ encoder, and propose an algorithm to search over it. We show that the resulting encoder attains roughly the same compression ratios as those based on suffix trees. However, the amount of memory required by the suffix array is fixed, and much lower than the variable amount of memory used by encoders based on suffix trees (which depends on the text to encode). We conclude that suffix arrays, when compared to suffix trees in terms of the trade-off among time, memory, and compression ratio, may be preferable in scenarios (e.g., embedded systems) where memory is at a premium and high speed is not critical.
Resumo:
In the Sparse Point Representation (SPR) method the principle is to retain the function data indicated by significant interpolatory wavelet coefficients, which are defined as interpolation errors by means of an interpolating subdivision scheme. Typically, a SPR grid is coarse in smooth regions, and refined close to irregularities. Furthermore, the computation of partial derivatives of a function from the information of its SPR content is performed in two steps. The first one is a refinement procedure to extend the SPR by the inclusion of new interpolated point values in a security zone. Then, for points in the refined grid, such derivatives are approximated by uniform finite differences, using a step size proportional to each point local scale. If required neighboring stencils are not present in the grid, the corresponding missing point values are approximated from coarser scales using the interpolating subdivision scheme. Using the cubic interpolation subdivision scheme, we demonstrate that such adaptive finite differences can be formulated in terms of a collocation scheme based on the wavelet expansion associated to the SPR. For this purpose, we prove some results concerning the local behavior of such wavelet reconstruction operators, which stand for SPR grids having appropriate structures. This statement implies that the adaptive finite difference scheme and the one using the step size of the finest level produce the same result at SPR grid points. Consequently, in addition to the refinement strategy, our analysis indicates that some care must be taken concerning the grid structure, in order to keep the truncation error under a certain accuracy limit. Illustrating results are presented for 2D Maxwell's equation numerical solutions.
Resumo:
A necessidade de poder computacional é crescente nas diversas áreas de actuação humana, tanto na indústria, como em ambientes académicos. Grid Computing permite a ligação de recursos computacionais dispersos de maneira a permitir a sua utilização mais eficaz, fornecendo aos utilizadores um acesso simplificado ao poder computacional de diversos sistemas. Os primeiros projectos de Grid Computing implicavam a ligação de máquinas paralelas ou aglomerados de alto desempenho e alto custo, disponíveis apenas em algumas instituições. Contrastando com o elevado custo dos super-computadores, os computadores pessoais e a Internet sofreram uma evolução significativa nos últimos anos. O uso de computadores dispersos em uma WAN pode representar um ambiente muito interessante para processamento de alto desempenho. Os sistemas em Grid fornecem a possibilidade de se utilizar um conjunto de computadores pessoais de modo a fornecer uma computação que utiliza recursos que de outra maneira estariam omissos. Este trabalho consiste no estudo de Grid Computing a nível de conceito e de arquitectura e numa análise ao seu estado actual hoje em dia. Como complemento foi desenvolvido um componente que permite o desenvolvimento de serviços para Grids (Grid Services) mais eficaz do que o modelo de suporte a serviços actualmente utilizado. Este componente é disponibilizado sob a forma um plug-in para a plataforma Eclipse IDE.
Resumo:
5th. European Congress on Computational Methods in Applied Sciences and Engineering (ECCOMAS 2008) 8th. World Congress on Computational Mechanics (WCCM8)
Resumo:
Object-oriented programming languages presently are the dominant paradigm of application development (e. g., Java,. NET). Lately, increasingly more Java applications have long (or very long) execution times and manipulate large amounts of data/information, gaining relevance in fields related with e-Science (with Grid and Cloud computing). Significant examples include Chemistry, Computational Biology and Bio-informatics, with many available Java-based APIs (e. g., Neobio). Often, when the execution of such an application is terminated abruptly because of a failure (regardless of the cause being a hardware of software fault, lack of available resources, etc.), all of its work already performed is simply lost, and when the application is later re-initiated, it has to restart all its work from scratch, wasting resources and time, while also being prone to another failure and may delay its completion with no deadline guarantees. Our proposed solution to address these issues is through incorporating mechanisms for checkpointing and migration in a JVM. These make applications more robust and flexible by being able to move to other nodes, without any intervention from the programmer. This article provides a solution to Java applications with long execution times, by extending a JVM (Jikes research virtual machine) with such mechanisms. Copyright (C) 2011 John Wiley & Sons, Ltd.
Resumo:
This paper proposes the use of a Modular Marx Multilevel Converter, as a solution for energy integration between an offshore Wind Farm and the power grid network. The Marx modular multilevel converter is based on the Marx generator, and solves two typical problems in this type of multilevel topologies: modularity and dc capacitor voltage balancing. This paper details the strategy for dc capacitor voltage equalization. The dynamic models of the converter and power grid are presented in order to design the converter ac output voltages and the dc capacitor voltage controller. The average current control is presented and used for power flow control, harmonics and reactive power compensation. Simulation results are presented in order to show the effectiveness of the proposed (MC)-C-3 topology.
Resumo:
We study the implications of the searches based on H -> tau(+)tau-by the ATLAS and CMS collaborations on the parameter space of the two-Higgs-doublet model (2HDM). In the 2HDM, the scalars can decay into a tau pair with a branching ratio larger than the SM one, leading to constraints on the 2HDM parameter space. We show that in model II, values of tan beta > 1.8 are definitively excluded if the pseudoscalar is in the mass range 110 GeV < m(A) < 145 GeV. We have also discussed the implications for the 2HDM of the recent dimuon search by the ATLAS collaboration for a CP-odd scalar in the mass range 4-12 GeV.
Resumo:
The increasing number of players that operate in power systems leads to a more complex management. In this paper a new multi-agent platform is proposed, which simulates the real operation of power system players. MASGriP – A Multi-Agent Smart Grid Simulation Platform is presented. Several consumer and producer agents are implemented and simulated, considering real characteristics and different goals and actuation strategies. Aggregator entities, such as Virtual Power Players and Curtailment Service Providers are also included. The integration of MASGriP agents in MASCEM (Multi-Agent System for Competitive Electricity Markets) simulator allows the simulation of technical and economical activities of several players. An energy resources management architecture used in microgrids is also explained.
Resumo:
This paper presents an algorithm to efficiently generate the state-space of systems specified using the IOPT Petri-net modeling formalism. IOPT nets are a non-autonomous Petri-net class, based on Place-Transition nets with an extended set of features designed to allow the rapid prototyping and synthesis of system controllers through an existing hardware-software co-design framework. To obtain coherent and deterministic operation, IOPT nets use a maximal-step execution semantics where, in a single execution step, all enabled transitions will fire simultaneously. This fact increases the resulting state-space complexity and can cause an arc "explosion" effect. Real-world applications, with several million states, will reach a higher order of magnitude number of arcs, leading to the need for high performance state-space generator algorithms. The proposed algorithm applies a compilation approach to read a PNML file containing one IOPT model and automatically generate an optimized C program to calculate the corresponding state-space.
Resumo:
The present research paper presents five different clustering methods to identify typical load profiles of medium voltage (MV) electricity consumers. These methods are intended to be used in a smart grid environment to extract useful knowledge about customer’s behaviour. The obtained knowledge can be used to support a decision tool, not only for utilities but also for consumers. Load profiles can be used by the utilities to identify the aspects that cause system load peaks and enable the development of specific contracts with their customers. The framework presented throughout the paper consists in several steps, namely the pre-processing data phase, clustering algorithms application and the evaluation of the quality of the partition, which is supported by cluster validity indices. The process ends with the analysis of the discovered knowledge. To validate the proposed framework, a case study with a real database of 208 MV consumers is used.