379 resultados para Toolkit


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Incluye Bibliografía

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)

Relevância:

10.00% 10.00%

Publicador:

Resumo:

El presente Boletín FAL se basa en el documento "First United Nations Global Road Safety Week. A toolkit for organizers of events" desarrollado por la Organización Mundial de la Salud, 2006, que puede ser accedido a través de http://www.paho.org/Spanish/DD/PUB/seguridadVial-ToolKit.pdf en idioma español. La versión completa en idioma inglés se encuentra en: http://www.who.int/roadsafety/week/toolkit/en/index.html

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The huge demand for procedures involving ionizing radiation promotes the need for safe methods of experimentation considering the danger of their biological e ects with consequent risk to humans. Brazilian's legislation prohibits experiments involving this type of radiation in humans through Decree 453 of Ministry of Health with determines that such procedures comply with the principles of justi cation, optimization and dose limitation. In this line, concurrently with the advancement of available computer processing power, computing simulations have become relevant in those situations where experimental procedures are too cost or impractical. The Monte Carlo method, created along the Manhattan Project duringWorldWar II, is a powerful strategy to simulations in computational physics. In medical physics, this technique has been extensively used with applications in diagnostics and cancer treatment. The objective of this work is to simulate the production and detection of X-rays for the energy range of diagnostic radiology, for molybdenum target, using the Geant4 toolkit. X-ray tubes with this kind of target material are used in diagnostic radiology, speci cally in mammography, one of the most used techniques for screening of breast cancer in women. During the simulations, we used di erent models for bremsstrahlung available in physical models for low energy, in situations already covered by the literature in earlier versions of Geant4. Our results show that although the physical situations seems qualitatively adequate, quantitative comparisons to available analytical data shows aws in the code of Geant4 Low Energy source

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The mobile application proposed in this undergraduate project, classified as a Location-Based Service and named Traveller, was developed to support users of mobile phones equipped with GPS in unknown locations by providing information about weather, location of users and stores in urban areas. The mobile devices whose this project is intended are those equipped with Android. The programming language Java was selected and the Eclipse development environment was used along with the toolkit for developing Android (ADT - Android Development Tools)

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This work presents a study about the use of standards and directions on parallel programming in distributed systems, using the MPI standard and PETSc toolkit, performing an analysis of their performances over certain mathematic operations involving matrices. The concepts are used to develop applications to solve problems involving Principal Components Analysis (PCA), which are executed in a Beowulf cluster. The results are compared to the ones of an analogous application with sequencial execution, and then it is analized if there was any performance boost on the parallel application

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The use of stones to crack open encapsulated fruit is widespread among wild bearded capuchin monkeys (Cebus libidinosus) inhabiting savanna-like environments. Some populations in Serra da Capivara National Park (Piaui, Brazil), though, exhibit a seemingly broader toolkit, using wooden sticks as probes, and employing stone tools for a variety of purposes. Over the course of 701.5 hr of visual contact of two wild capuchin groups we recorded 677 tool use episodes. Five hundred and seventeen of these involved the use of stones, and 160 involved the use of sticks (or other plant parts) as probes to access water, arthropods, or the contents of insects` nests. Stones were mostly used as ""hammers""-not only to open fruit or seeds, or smash other food items, but also to break dead wood, conglomerate rock, or cement in search of arthropods, to dislodge bigger stones, and to pulverize embedded quartz pebbles (licking, sniffing, or rubbing the body with the powder produced). Stones also were used in a ""hammer-like"" fashion to loosen the soil for digging out roots and arthropods, and sometimes as ""hoes"" to pull the loosened soil. In a few cases, we observed the re-utilization of stone tools for different purposes (N = 3), or the combined use of two tools-stones and sticks (N = 4) or two stones (N = 5), as sequential or associative tools. On three occasions, the monkeys used smaller stones to loosen bigger quartz pebbles embedded in conglomerate rock, which were subsequently used as tools. These could be considered the first reports of secondary tool use by wild capuchin monkeys. Am. J. Primatol. 71:242-251, 2009. (c) 2008 Wiley-Liss, Inc.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

[EN] The General Curvilinear Environmental Model is a high-resolution system composed of the General Curvilinear Coastal Ocean Model (GCCOM) and the General Curvilinear Atmospheric Model (GCAM). Both modules are capable of reading a general curvilinear grid, orthogonal as well as non-orthogonal in all three directions. These two modules are weakly coupled using the distributed coupling toolkit (DCT). The model can also be nested within larger models and users are able to interact with the model and run it using a web based computational environment.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

[ES] En este proyecto se trata el proceso de análisis y desarrollo llevado a cabo con el objetivo de construir un prototipo funcional de simulador virtual de endoscopia rígida monocanal orientado a la histeroscopia. Para el desarrollo del prototipo se toma como base el entorno ESQUI, un entorno de simulación virtual médica de código libre. Este entorno provee una librería, basada a su vez en la conocida librería gráfica VTK(Visual ToolKit), cuyo propósito es poner a disposición del programador toda la algoritmia necesaria para construir una simulación médica virtual. En este proyecto, esta librería se depuró y amplió para mejorar el soporte a las técnicas de endoscopia rígida que se persiguen simular. Por otro lado se emplea el Simball 4D, un dispositivo de interfaz humana de la empresa G-coder Systems, para capturar la interacción del usuario emulando la morfología y dinámica de un endoscopio rígido. Todos estos elementos se conectan con una interfaz gráfica sencilla, intuitiva y práctica soportada por wxWidgets y utilizando Python como lenguaje de scripting. Finalmente, se analiza el prototipo resultante y se proponen una serie de líneas futuras de cara a la aplicación didáctica del mismo, tanto en relación a los objetivos conceptuales del prototipo como a los aspectos específicos del entorno ESQUI.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Coordinating activities in a distributed system is an open research topic. Several models have been proposed to achieve this purpose such as message passing, publish/subscribe, workflows or tuple spaces. We have focused on the latter model, trying to overcome some of its disadvantages. In particular we have applied spatial database techniques to tuple spaces in order to increase their performance when handling a large number of tuples. Moreover, we have studied how structured peer to peer approaches can be applied to better distribute tuples on large networks. Using some of these result, we have developed a tuple space implementation for the Globus Toolkit that can be used by Grid applications as a coordination service. The development of such a service has been quite challenging due to the limitations imposed by XML serialization that have heavily influenced its design. Nevertheless, we were able to complete its implementation and use it to implement two different types of test applications: a completely parallelizable one and a plasma simulation that is not completely parallelizable. Using this last application we have compared the performance of our service against MPI. Finally, we have developed and tested a simple workflow in order to show the versatility of our service.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this thesis we describe in detail the Monte Carlo simulation (LVDG4) built to interpret the experimental data collected by LVD and to measure the muon-induced neutron yield in iron and liquid scintillator. A full Monte Carlo simulation, based on the Geant4 (v 9.3) toolkit, has been developed and validation tests have been performed. We used the LVDG4 to determine the active vetoing and the shielding power of LVD. The idea was to evaluate the feasibility to host a dark matter detector in the most internal part, called Core Facility (LVD-CF). The first conclusion is that LVD is a good moderator, but the iron supporting structure produce a great number of neutrons near the core. The second conclusions is that if LVD is used as an active veto for muons, the neutron flux in the LVD-CF is reduced by a factor 50, of the same order of magnitude of the neutron flux in the deepest laboratory of the world, Sudbury. Finally, the muon-induced neutron yield has been measured. In liquid scintillator we found $(3.2 \pm 0.2) \times 10^{-4}$ n/g/cm$^2$, in agreement with previous measurements performed at different depths and with the general trend predicted by theoretical calculations and Monte Carlo simulations. Moreover we present the first measurement, in our knowledge, of the neutron yield in iron: $(1.9 \pm 0.1) \times 10^{-3}$ n/g/cm$^2$. That measurement provides an important check for the MC of neutron production in heavy materials that are often used as shield in low background experiments.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

To assist rational compound design of organic semiconductors, two problems need to be addressed. First, the material morphology has to be known at an atomistic level. Second, with the morphology at hand, an appropriate charge transport model needs to be developed in order to link charge carrier mobility to structure.rnrnThe former can be addressed by generating atomistic morphologies using molecular dynamics simulations. However, the accessible range of time- and length-scales is limited. To overcome these limitations, systematic coarse-graining methods can be used. In the first part of the thesis, the Versatile Object-oriented Toolkit for Coarse-graining Applications is introduced, which provides a platform for the implementation of coarse-graining methods. Tools to perform Boltzmann inversion, iterative Boltzmann inversion, inverse Monte Carlo, and force-matching are available and have been tested on a set of model systems (water, methanol, propane and a single hexane chain). Advantages and problems of each specific method are discussed.rnrnIn partially disordered systems, the second issue is closely connected to constructing appropriate diabatic states between which charge transfer occurs. In the second part of the thesis, the description initially used for small conjugated molecules is extended to conjugated polymers. Here, charge transport is modeled by introducing conjugated segments on which charge carriers are localized. Inter-chain transport is then treated within a high temperature non-adiabatic Marcus theory while an adiabatic rate expression is used for intra-chain transport. The charge dynamics is simulated using the kinetic Monte Carlo method.rnrnThe entire framework is finally employed to establish a relation between the morphology and the charge mobility of the neutral and doped states of polypyrrole, a conjugated polymer. It is shown that for short oligomers, charge carrier mobility is insensitive to the orientational molecular ordering and is determined by the threshold transfer integral which connects percolating clusters of molecules that form interconnected networks. The value of this transfer integral can be related to the radial distribution function. Hence, charge mobility is mainly determined by the local molecular packing and is independent of the global morphology, at least in such a non-crystalline state of a polymer.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In der vorliegenden Arbeit werden verschiedene Wassermodelle in sogenannten Multiskalen-Computersimulationen mit zwei Auflösungen untersucht, in atomistischer Auflösung und in einer vergröberten Auflösung, die als "coarse-grained" bezeichnet wird. In der atomistischen Auflösung wird ein Wassermolekül, entsprechend seiner chemischen Struktur, durch drei Atome beschrieben, im Gegensatz dazu wird ein Molekül in der coarse-grained Auflösung durch eine Kugel dargestellt.rnrnDie coarse-grained Modelle, die in dieser Arbeit vorgestellt werden, werden mit verschiedenen coarse-graining Methoden entwickelt. Hierbei kommen hauptsächlich die "iterative Boltzmann Inversion" und die "iterative Monte Carlo Inversion" zum Einsatz. Beides sind struktur-basierte Ansätze, die darauf abzielen bestimmte strukturelle Eigenschaften, wie etwa die Paarverteilungsfunktionen, des zugrundeliegenden atomistischen Systems zu reproduzieren. Zur automatisierten Anwendung dieser Methoden wurde das Softwarepaket "Versatile Object-oriented Toolkit for Coarse-Graining Applications" (VOTCA) entwickelt.rnrnEs wird untersucht, in welchem Maße coarse-grained Modelle mehrere Eigenschaftenrndes zugrundeliegenden atomistischen Modells gleichzeitig reproduzieren können, z.B. thermodynamische Eigenschaften wie Druck und Kompressibilität oder strukturelle Eigenschaften, die nicht zur Modellbildung verwendet wurden, z.B. das tetraedrische Packungsverhalten, welches für viele spezielle Eigenschaft von Wasser verantwortlich ist.rnrnMit Hilfe des "Adaptive Resolution Schemes" werden beide Auflösungen in einer Simulation kombiniert. Dabei profitiert man von den Vorteilen beider Modelle:rnVon der detaillierten Darstellung eines räumlich kleinen Bereichs in atomistischer Auflösung und von der rechnerischen Effizienz des coarse-grained Modells, die den Bereich simulierbarer Zeit- und Längenskalen vergrössert.rnrnIn diesen Simulationen kann der Einfluss des Wasserstoffbrückenbindungsnetzwerks auf die Hydration von Fullerenen untersucht werden. Es zeigt sich, dass die Struktur der Wassermoleküle an der Oberfläche hauptsächlich von der Art der Wechselwirkung zwischen dem Fulleren und Wasser und weniger von dem Wasserstoffbrückenbindungsnetzwerk dominiert wird.rn

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Despite the scientific achievement of the last decades in the astrophysical and cosmological fields, the majority of the Universe energy content is still unknown. A potential solution to the “missing mass problem” is the existence of dark matter in the form of WIMPs. Due to the very small cross section for WIMP-nuleon interactions, the number of expected events is very limited (about 1 ev/tonne/year), thus requiring detectors with large target mass and low background level. The aim of the XENON1T experiment, the first tonne-scale LXe based detector, is to be sensitive to WIMP-nucleon cross section as low as 10^-47 cm^2. To investigate the possibility of such a detector to reach its goal, Monte Carlo simulations are mandatory to estimate the background. To this aim, the GEANT4 toolkit has been used to implement the detector geometry and to simulate the decays from the various background sources: electromagnetic and nuclear. From the analysis of the simulations, the level of background has been found totally acceptable for the experiment purposes: about 1 background event in a 2 tonne-years exposure. Indeed, using the Maximum Gap method, the XENON1T sensitivity has been evaluated and the minimum for the WIMP-nucleon cross sections has been found at 1.87 x 10^-47 cm^2, at 90% CL, for a WIMP mass of 45 GeV/c^2. The results have been independently cross checked by using the Likelihood Ratio method that confirmed such results with an agreement within less than a factor two. Such a result is completely acceptable considering the intrinsic differences between the two statistical methods. Thus, in the PhD thesis it has been proven that the XENON1T detector will be able to reach the designed sensitivity, thus lowering the limits on the WIMP-nucleon cross section by about 2 orders of magnitude with respect to the current experiments.