37 resultados para modeling and visualization
Resumo:
Acquired brain injury (ABI) 1-2 refers to any brain damage occurring after birth. It usually causes certain damage to portions of the brain. ABI may result in a significant impairment of an individuals physical, cognitive and/or psychosocial functioning. The main causes are traumatic brain injury (TBI), cerebrovascular accident (CVA) and brain tumors. The main consequence of ABI is a dramatic change in the individuals daily life. This change involves a disruption of the family, a loss of future income capacity and an increase of lifetime cost. One of the main challenges in neurorehabilitation is to obtain a dysfunctional profile of each patient in order to personalize the treatment. This paper proposes a system to generate a patient s dysfunctional profile by integrating theoretical, structural and neuropsychological information on a 3D brain imaging-based model. The main goal of this dysfunctional profile is to help therapists design the most suitable treatment for each patient. At the same time, the results obtained are a source of clinical evidence to improve the accuracy and quality of our rehabilitation system. Figure 1 shows the diagram of the system. This system is composed of four main modules: image-based extraction of parameters, theoretical modeling, classification and co-registration and visualization module.
Resumo:
To study the fluid motion-vehicle dynamics interaction, a model of four, liquid filled two-axle container freight wagons was set up. The railway vehicle has been modelled as a multi-body system (MBS). To include fluid sloshing, an equivalent mechanical model has been developed and incorporated. The influence of several factors has been studied in computer simulations, such as track defects, curve negotiation, train velocity, wheel wear, liquid and solid wagonload, and container baffles. SIMPACK has been used for MBS analysis, and ANSYS for liquid sloshing modelling and equivalent mechanical systems validation. Acceleration and braking manoeuvres of the freight train set the liquid cargo into motion. This longitudinal sloshing motion of the fluid cargo inside the tanks initiated a swinging motion of some components of the coupling gear. The coupling gear consists of UIC standard traction hooks and coupling screws that are located between buffers. One of the coupling screws is placed in the traction hook of the opposite wagon thus joining the two wagons, whereas the unused coupling screw rests on a hanger. Simulation results showed that, for certain combinations of type of liquid, filling level and container dimensions, the liquid cargo could provoke an undesirable, although not hazardous, release of the unused coupling screw from its hanger. The coupling screw's release was especially obtained when a period of acceleration was followed by an abrupt braking manoeuvre at 1 m/s2. It was shown that a resonance effect between the liquid's oscillation and the coupling screw's rotary motion could be the reason for the coupling screw's undesired release. Possible solutions to avoid the phenomenon are given.Acceleration and braking manoeuvres of the freight train set the liquid cargo into motion. This longitudinal sloshing motion of the fluid cargo inside the tanks initiated a swinging motion of some components of the coupling gear. The coupling gear consists of UIC standard traction hooks and coupling screws that are located between buffers. One of the coupling screws is placed in the traction hook of the opposite wagon thus joining the two wagons, whereas the unused coupling screw rests on a hanger. This paper reports on a study of the fluid motion-train vehicle dynamics interaction. In the study, a model of four, liquid-filled two-axle container freight wagons was developed. The railway vehicle has been modeled as a multi-body system (MBS). To include fluid sloshing, an equivalent mechanical model has been developed and incorporated. The influence of several factors has been studied in computer simulations, such as track defects, curve negotiation, train velocity, wheel wear, liquid and solid wagonload, and container baffles. A simulation program was used for MBS analysis, and a finite element analysis program was used for liquid sloshing modeling and equivalent mechanical systems validation. Acceleration and braking maneuvers of the freight train set the liquid cargo into motion. This longitudinal sloshing motion of the fluid cargo inside the tanks initiated a swinging motion of some components of the coupling gear. Simulation results showed that, for certain combinations of type of liquid, filling level and container dimensions, the liquid cargo could provoke an undesirable, although not hazardous, release of an unused coupling screw from its hanger. It was shown that a resonance effect between the liquid's oscillation and the coupling screw's rotary motion could be the reason for the coupling screw's undesired release. Solutions are suggested to avoid the resonance problem, and directions for future research are given.
Resumo:
The control part of the execution of a constraint logic program can be conceptually shown as a search-tree, where nodes correspond to calis, and whose branches represent conjunctions and disjunctions. This tree represents the search space traversed by the program, and has also a direct relationship with the amount of work performed by the program. The nodes of the tree can be used to display information regarding the state and origin of instantiation of the variables involved in each cali. This depiction can also be used for the enumeration process. These are the features implemented in APT, a tool which runs constraint logic programs while depicting a (modified) search-tree, keeping at the same time information about the state of the variables at every moment in the execution. This information can be used to replay the execution at will, both forwards and backwards in time. These views can be abstracted when the size of the execution requires it. The search-tree view is used as a framework onto which constraint-level visualizations (such as those presented in the following chapter) can be attached.
Resumo:
We present a generic preprocessor for combined static/dynamic validation and debugging of constraint logic programs. Passing programs through the preprocessor prior to execution allows detecting many bugs automatically. This is achieved by performing a repertoire of tests which range from simple syntactic checks to much more advanced checks based on static analysis of the program. Together with the program, the user may provide a series of assertions which trigger further automatic checking of the program. Such assertions are written using the assertion language presented in Chapter 2, which allows expressing a wide variety of properties. These properties extend beyond the predefined set which may be understandable by the available static analyzers and include properties defined by means of user programs. In addition to user-provided assertions, in each particular CLP system assertions may be available for predefined system predicates. Checking of both user-provided assertions and assertions for system predicates is attempted first at compile-time by comparing them with the results of static analysis. This may allow statically proving that the assertions hold (Le., they are validated) or that they are violated (and thus bugs detected). User-provided assertions (or parts of assertions) which cannot be statically proved ñor disproved are optionally translated into run-time tests. The implementation of the preprocessor is generic in that it can be easily customized to different CLP systems and dialects and in that it is designed to allow the integration of additional analyses in a simple way. We also report on two tools which are instances of the generic preprocessor: CiaoPP (for the Ciao Prolog system) and CHIPRE (for the CHIP CLP(FL>) system). The currently existing analyses include types, modes, non-failure, determinacy, and computational cost, and can treat modules separately, performing incremental analysis.
Resumo:
This introduction gives a general perspective of the debugging methodology and the tools developed in the ESPRIT IV project DiSCiPl Debugging Systems for Constraint Programming. It has been prepared by the editors of this volume by substantial rewriting of the DiSCiPl deliverable CP Debugging Tools [1]. This introduction is organised as follows. Section 1 outlines the DiSCiPl view of debugging, its associated debugging methodology, and motivates the kinds of tools proposed: the assertion based tools, the declarative diagnoser and the visualisation tools. Sections 2 through 4 provide a short presentation of the tools of each kind. Finally, Section 5 presents a summary of the tools developed in the project. This introduction gives only a general view of the DiSCiPl debugging methodology and tools. For details and for specific bibliographic referenees the reader is referred to the subsequent chapters.
Resumo:
Visualization of program executions has been used in applications which include education and debugging. However, traditional visualization techniques often fall short of expectations or are altogether inadequate for new programming paradigms, such as Constraint Logic Programming (CLP), whose declarative and operational semantics differ in some crucial ways from those of other paradigms. In particular, traditional ideas regarding the behavior of data often cannot be lifted in a straightforward way to (C)LP from other families of programming languages. In this chapter we discuss techniques for visualizing data evolution in CLP. We briefly review some previously proposed visualization paradigms, and also propose a number of (to our knowledge) novel ones. The graphical representations have been chosen based on the perceived needs of a programmer trying to analyze the behavior and characteristics of an execution. In particular, we concéntrate on the representation of the run-time valúes of the variables, and the constraints among them. Given our interest in visualizing large executions, we also pay attention to abstraction techniques, i.e., techniques which are intended to help in reducing the complexity of the visual information.
Resumo:
Modeling and prediction of the overall elastic–plastic response and local damage mechanisms in heterogeneous materials, in particular particle reinforced composites, is a very complex problem. Microstructural complexities such as the inhomogeneous spatial distribution of particles, irregular morphology of the particles, and anisotropy in particle orientation after secondary processing, such as extrusion, significantly affect deformation behavior. We have studied the effect of particle/matrix interface debonding in SiC particle reinforced Al alloy matrix composites with (a) actual microstructure consisting of angular SiC particles and (b) idealized ellipsoidal SiC particles. Tensile deformation in SiC particle reinforced Al matrix composites was modeled using actual microstructures reconstructed from serial sectioning approach. Interfacial debonding was modeled using user-defined cohesive zone elements. Modeling with the actual microstructure (versus idealized ellipsoids) has a significant influence on: (a) localized stresses and strains in particle and matrix, and (b) far-field strain at which localized debonding takes place. The angular particles exhibited higher degree of load transfer and are more sensitive to interfacial debonding. Larger decreases in stress are observed in the angular particles, because of the flat surfaces, normal to the loading axis, which bear load. Furthermore, simplification of particle morphology may lead to erroneous results.
Resumo:
Probabilistic modeling is the de�ning characteristic of estimation of distribution algorithms (EDAs) which determines their behavior and performance in optimization. Regularization is a well-known statistical technique used for obtaining an improved model by reducing the generalization error of estimation, especially in high-dimensional problems. `1-regularization is a type of this technique with the appealing variable selection property which results in sparse model estimations. In this thesis, we study the use of regularization techniques for model learning in EDAs. Several methods for regularized model estimation in continuous domains based on a Gaussian distribution assumption are presented, and analyzed from di�erent aspects when used for optimization in a high-dimensional setting, where the population size of EDA has a logarithmic scale with respect to the number of variables. The optimization results obtained for a number of continuous problems with an increasing number of variables show that the proposed EDA based on regularized model estimation performs a more robust optimization, and is able to achieve signi�cantly better results for larger dimensions than other Gaussian-based EDAs. We also propose a method for learning a marginally factorized Gaussian Markov random �eld model using regularization techniques and a clustering algorithm. The experimental results show notable optimization performance on continuous additively decomposable problems when using this model estimation method. Our study also covers multi-objective optimization and we propose joint probabilistic modeling of variables and objectives in EDAs based on Bayesian networks, speci�cally models inspired from multi-dimensional Bayesian network classi�ers. It is shown that with this approach to modeling, two new types of relationships are encoded in the estimated models in addition to the variable relationships captured in other EDAs: objectivevariable and objective-objective relationships. An extensive experimental study shows the e�ectiveness of this approach for multi- and many-objective optimization. With the proposed joint variable-objective modeling, in addition to the Pareto set approximation, the algorithm is also able to obtain an estimation of the multi-objective problem structure. Finally, the study of multi-objective optimization based on joint probabilistic modeling is extended to noisy domains, where the noise in objective values is represented by intervals. A new version of the Pareto dominance relation for ordering the solutions in these problems, namely �-degree Pareto dominance, is introduced and its properties are analyzed. We show that the ranking methods based on this dominance relation can result in competitive performance of EDAs with respect to the quality of the approximated Pareto sets. This dominance relation is then used together with a method for joint probabilistic modeling based on `1-regularization for multi-objective feature subset selection in classi�cation, where six di�erent measures of accuracy are considered as objectives with interval values. The individual assessment of the proposed joint probabilistic modeling and solution ranking methods on datasets with small-medium dimensionality, when using two di�erent Bayesian classi�ers, shows that comparable or better Pareto sets of feature subsets are approximated in comparison to standard methods.
Resumo:
This paper describes a novel approach to phonotactic LID, where instead of using soft-counts based on phoneme lattices, we use posteriogram to obtain n-gram counts. The high-dimensional vectors of counts are reduced to low-dimensional units for which we adapted the commonly used term i-vectors. The reduction is based on multinomial subspace modeling and is designed to work in the total-variability space. The proposed technique was tested on the NIST 2009 LRE set with better results to a system based on using soft-counts (Cavg on 30s: 3.15% vs 3.43%), and with very good results when fused with an acoustic i-vector LID system (Cavg on 30s acoustic 2.4% vs 1.25%). The proposed technique is also compared with another low dimensional projection system based on PCA. In comparison with the original soft-counts, the proposed technique provides better results, reduces the problems due to sparse counts, and avoids the process of using pruning techniques when creating the lattices.
Resumo:
Desde la aparición del turborreactor, el motor aeróbico con turbomaquinaria ha demostrado unas prestaciones excepcionales en los regímenes subsónico y supersónico bajo. No obstante, la operación a velocidades superiores requiere sistemas más complejos y pesados, lo cual ha imposibilitado la ejecución de estos conceptos. Los recientes avances tecnológicos, especialmente en materiales ligeros, han restablecido el interés por los motores de ciclo combinado. La simulación numérica de estos nuevos conceptos es esencial para estimar las prestaciones de la planta propulsiva, así como para abordar las dificultades de integración entre célula y motor durante las primeras etapas de diseño. Al mismo tiempo, la evaluación de estos extraordinarios motores requiere una metodología de análisis distinta. La tesis doctoral versa sobre el diseño y el análisis de los mencionados conceptos propulsivos mediante el modelado numérico y la simulación dinámica con herramientas de vanguardia. Las distintas arquitecturas presentadas por los ciclos combinados basados en sendos turborreactor y motor cohete, así como los diversos sistemas comprendidos en cada uno de ellos, hacen necesario establecer una referencia común para su evaluación. Es más, la tendencia actual hacia aeronaves "más eléctricas" requiere una nueva métrica para juzgar la aptitud de un proceso de generación de empuje en el que coexisten diversas formas de energía. A este respecto, la combinación del Primer y Segundo Principios define, en un marco de referencia absoluto, la calidad de la trasferencia de energía entre los diferentes sistemas. Esta idea, que se ha estado empleando desde hace mucho tiempo en el análisis de plantas de potencia terrestres, ha sido extendida para relacionar la misión de la aeronave con la ineficiencia de cada proceso involucrado en la generación de empuje. La metodología se ilustra mediante el estudio del motor de ciclo combinado variable de una aeronave para el crucero a Mach 5. El diseño de un acelerador de ciclo combinado basado en el turborreactor sirve para subrayar la importancia de la integración del motor y la célula. El diseño está limitado por la trayectoria ascensional y el espacio disponible en la aeronave de crucero supersónico. Posteriormente se calculan las prestaciones instaladas de la planta propulsiva en función de la velocidad y la altitud de vuelo y los parámetros de control del motor: relación de compresión, relación aire/combustible y área de garganta. ABSTRACT Since the advent of the turbojet, the air-breathing engine with rotating machinery has demonstrated exceptional performance in the subsonic and low supersonic regimes. However, the operation at higher speeds requires further system complexity and weight, which so far has impeded the realization of these concepts. Recent technology developments, especially in lightweight materials, have restored the interest towards combined-cycle engines. The numerical simulation of these new concepts is essential at the early design stages to compute a first estimate of the engine performance in addition to addressing airframe-engine integration issues. In parallel, a different analysis methodology is required to evaluate these unconventional engines. The doctoral thesis concerns the design and analysis of the aforementioned engine concepts by means of numerical modeling and dynamic simulation with state-of-the-art tools. A common reference is needed to evaluate the different architectures of the turbine and the rocket-based combined-cycle engines as well as the various systems within each one of them. Furthermore, the actual trend towards more electric aircraft necessitates a common metric to judge the suitability of a thrust generation process where different forms of energy coexist. In line with this, the combination of the First and the Second Laws yields the quality of the energy being transferred between the systems on an absolute reference frame. This idea, which has been since long applied to the analysis of on-ground power plants, was extended here to relate the aircraft mission with the inefficiency of every process related to the thrust generation. The methodology is illustrated with the study of a variable- combined-cycle engine for a Mach 5 cruise aircraft. The design of a turbine-based combined-cycle booster serves to highlight the importance of the engine-airframe integration. The design is constrained by the ascent trajectory and the allocated space in the supersonic cruise aircraft. The installed performance of the propulsive plant is then computed as a function of the flight speed and altitude and the engine control parameters: pressure ratio, air-to-fuel ratio and throat area.
Resumo:
Nowadays, Software Product Line (SPL) engineering [1] has been widely-adopted in software development due to the significant improvements that has provided, such as reducing cost and time-to-market and providing flexibility to respond to planned changes [2]. SPL takes advantage of common features among the products of a family through the systematic reuse of the core-assets and the effective management of variabilities across the products. SPL features are realized at the architectural level in product-line architecture (PLA) models. Therefore, suitable modeling and specification techniques are required to model variability. In fact, architectural variability modeling has become a challenge for SPLE due to the fact that PLA modeling requires not only modeling variability at the level of the external architecture configuration (see [3,4] literature reviews), but also at the level of internal specification of components [5]. In addition, PLA modeling requires preserving the traceability between features and PLAs. Finally, it is important to take into account that PLA modeling should guide architects in modeling the PLA core assets and variability, and in deriving the customized products. To deal with these needs, we present in this demonstration the FPLA Modeling Framework.
Resumo:
Virtual reality (VR) techniques to understand and obtain conclusions of data in an easy way are being used by the scientific community. However, these techniques are not used frequently for analyzing large amounts of data in life sciences, particularly in genomics, due to the high complexity of data (curse of dimensionality). Nevertheless, new approaches that allow to bring out the real important data characteristics, arise the possibility of constructing VR spaces to visually understand the intrinsic nature of data. It is well known the benefits of representing high dimensional data in tridimensional spaces by means of dimensionality reduction and transformation techniques, complemented with a strong component of interaction methods. Thus, a novel framework, designed for helping to visualize and interact with data about diseases, is presented. In this paper, the framework is applied to the Van't Veer breast cancer dataset is used, while oncologists from La Paz Hospital (Madrid) are interacting with the obtained results. That is to say a first attempt to generate a visually tangible model of breast cancer disease in order to support the experience of oncologists is presented.
Resumo:
We recover and develop some robotic systems concepts (on the light of present systems tools) that were originated for an intended Mars Rover in the sixties of the last century at the Instrumentation Laboratory of MIT, where one of the authors was involved. The basic concepts came from the specifications for a type of generalized robot inspired in the structure of the vertebrate nervous systems, where the decision system was based in the structure and function of the Reticular Formation (RF). The vertebrate RF is supposed to commit the whole organism to one among various modes of behavior, so taking the decisions about the present overall task. That is, it is a kind of control and command system. In this concepts updating, the basic idea is that the RF comprises a set of computing units such that each computing module receives information only from a reduced part of the overall, little processed sensory inputs. Each computing unit is capable of both general diagnostics about overall input situations and of specialized diagnostics according to the values of a concrete subset of the input lines. Slave systems to this command and control computer, there are the sensors, the representations of external environment, structures for modeling and planning and finally, the effectors acting in the external world.
Resumo:
Mechanical degradation of tungsten alloys at extreme temperatures in vacuum and oxidation atmospheres.
Resumo:
This special issue gathers together a number of recent papers on fractal geometry and its applications to the modeling of flow and transport in porous media. The aim is to provide a systematic approach for analyzing the statics and dynamics of fluids in fractal porous media by means of theory, modeling and experimentation. The topics covered include lacunarity analyses of multifractal and natural grayscale patterns, random packing's of self-similar pore/particle size distributions, Darcian and non-Darcian hydraulic flows, diffusion within fractals, models for the permeability and thermal conductivity of fractal porous media and hydrophobicity and surface erosion properties of fractal structures.