18 resultados para Micro parallel kinematic manipulator

em Repositório Científico do Instituto Politécnico de Lisboa - Portugal


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Purpose: The most recent Varian® micro multileaf collimator(MLC), the High Definition (HD120) MLC, was modeled using the BEAMNRCMonte Carlo code. This model was incorporated into a Varian medical linear accelerator, for a 6 MV beam, in static and dynamic mode. The model was validated by comparing simulated profiles with measurements. Methods: The Varian® Trilogy® (2300C/D) accelerator model was accurately implemented using the state-of-the-art Monte Carlo simulation program BEAMNRC and validated against off-axis and depth dose profiles measured using ionization chambers, by adjusting the energy and the full width at half maximum (FWHM) of the initial electron beam. The HD120 MLC was modeled by developing a new BEAMNRC component module (CM), designated HDMLC, adapting the available DYNVMLC CM and incorporating the specific characteristics of this new micro MLC. The leaf dimensions were provided by the manufacturer. The geometry was visualized by tracing particles through the CM and recording their position when a leaf boundary is crossed. The leaf material density and abutting air gap between leaves were adjusted in order to obtain a good agreement between the simulated leakage profiles and EBT2 film measurements performed in a solid water phantom. To validate the HDMLC implementation, additional MLC static patterns were also simulated and compared to additional measurements. Furthermore, the ability to simulate dynamic MLC fields was implemented in the HDMLC CM. The simulation results of these fields were compared with EBT2 film measurements performed in a solid water phantom. Results: Overall, the discrepancies, with and without MLC, between the opened field simulations and the measurements using ionization chambers in a water phantom, for the off-axis profiles are below 2% and in depth-dose profiles are below 2% after the maximum dose depth and below 4% in the build-up region. On the conditions of these simulations, this tungsten-based MLC has a density of 18.7 g cm− 3 and an overall leakage of about 1.1 ± 0.03%. The discrepancies between the film measured and simulated closed and blocked fields are below 2% and 8%, respectively. Other measurements were performed for alternated leaf patterns and the agreement is satisfactory (to within 4%). The dynamic mode for this MLC was implemented and the discrepancies between film measurements and simulations are within 4%. Conclusions: The Varian® Trilogy® (2300 C/D) linear accelerator including the HD120 MLC was successfully modeled and simulated using the Monte CarloBEAMNRC code by developing an independent CM, the HDMLC CM, either in static and dynamic modes.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

International Conference with Peer Review 2012 IEEE International Conference in Geoscience and Remote Sensing Symposium (IGARSS), 22-27 July 2012, Munich, Germany

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Helically twisted fibers can be produced by electrospinning liquid-crystalline cellulose solutions. Fiber topographies are studied by atomic force microscopy, scanning electron microscopy (see figure) and polarized optical microscopy. The fibers have a nearly universal pitch-to-diameter ratio and comprise both right- and left-handed helices.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Micro- and nano-patterned materials are of great importance for the design of new nanoscale electronic, optical and mechanical devices, ranging from sensors to displays. A prospective system that can support a designed functionality is elastomeric polyurethane thin films with nano- or micromodulated surface structures ("wrinkles"). These wrinkles can be induced on different lengthscales by mechanically stretching the films, without the need for any sophisticated lithographic techniques. In the present article we focus on the experimental control of the wrinkling process. A simple model for wrinkle formation is also discussed, and some preliminary results reported. Hierarchical assembly of these tunable structures paves the way for the development of a new class of materials with a wide range of applications, from electronics to biomedicine.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A celulose é o polímero renovável mais abundante do mundo. É conhecido pela sua excelente biocompatibilidade, propriedades térmicas e mecânicas. A celulose assim como os polipéptideos e o ADN, pertence a uma família de moléculas orgânicas que dão origem à formação de fases líquidas cristalinas (LCs) colestéricas. A Passiflora Edulis, tal como outras plantas trepadeiras, possui longas e flexíveis gavinhas que permitem à planta encontrar um suporte para se fixar. As gavinhas podem assumir a forma de espirais ou de hélices consoante sejam sustentadas por apenas uma ou por ambas as extremidades. As hélices apresentam muitas vezes duas porções helicoidais, uma esquerda e outra direita, separadas por um segmento recto denominado perversão. Este comportamento é consequência da curvatura intrínseca das gavinhas produzidas pela planta trepadeira. O mesmo comportamento pode ser observado em micro e nanofibras celulósicas fabricadas a partir de soluções líquido-cristalinas, numa escala três a quatro ordens de grandeza inferior à das gavinhas. Este facto sugere que o modelo físico utilizado tenha invariância de escala. Neste trabalho é feito o estudo de fibras e jactos que imitam as estruturas helicoidais apresentadas pelas gavinhas das plantas trepadeiras. As fibras e jactos são produzidos a partir de soluções líquidas cristalinas celulósicas. De modo a determinar as características morfológicas e estruturais, que contribuem para a curvatura das fibras, foram utilizadas técnicas de imagem por ressonância magnética (MRI), microscopia óptica com luz polarisada (MOP), microscopia electrónica de varrimento (SEM) e microscopia de força atómica (AFM) . A variação da forma das estruturas helicoidais com a temperatura parece ser relevante para o fabrico de membranas não tecidas para aplicação em sensores termo-mecânicos.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Micro-generation is the small scale production of heat and/or electricity from a low carbon source and can be a powerful driver for carbon reduction, behavior change, security of supply and economic value. The energy conversion technologies can include photovoltaic panels, micro combined heat and power, micro wind, heat pumps, solar thermal systems, fuel cells and micro hydro schemes. In this paper, a small research of the availability of the conversion apparatus and the prices for the micro wind turbines and photovoltaic systems is made and a comparison between these two technologies is performed in terms of the availability of the resource and costs. An analysis of the new legal framework published in Portugal is done to realize if the incentives to individualspsila investment in sustainable and local energy production is worth for their point of view. An economic evaluation for these alternatives, accounting with the governmentpsilas incentives should lead, in most cases, into attractive return rates for the investment. Apart from the attractiveness of the investment there are though other aspects that should be taken into account and those are the benefits that these choices have to us all. The idea is that micro-generation will not only make a significant direct contribution to carbon reduction targets, it will also trigger a multiplier effect in behavior change by engaging hearts and minds, and providing more efficient use of energy by householders. The diversified profile of power generation by micro-generators, both in terms of location and timing, should reduce the impact of intermittency or plant failures with significant gains for security of supply.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Mestrado em Fiscalidade

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper presents a micro power light energy harvesting system for indoor environments. Light energy is collected by amorphous silicon photovoltaic (a-Si:H PV) cells, processed by a switched capacitor (SC) voltage doubler circuit with maximum power point tracking (MPPT), and finally stored in a large capacitor. The MPPT fractional open circuit voltage (V-OC) technique is implemented by an asynchronous state machine (ASM) that creates and dynamically adjusts the clock frequency of the step-up SC circuit, matching the input impedance of the SC circuit to the maximum power point condition of the PV cells. The ASM has a separate local power supply to make it robust against load variations. In order to reduce the area occupied by the SC circuit, while maintaining an acceptable efficiency value, the SC circuit uses MOSFET capacitors with a charge sharing scheme for the bottom plate parasitic capacitors. The circuit occupies an area of 0.31 mm(2) in a 130 nm CMOS technology. The system was designed in order to work under realistic indoor light intensities. Experimental results show that the proposed system, using PV cells with an area of 14 cm(2), is capable of starting-up from a 0 V condition, with an irradiance of only 0.32 W/m(2). After starting-up, the system requires an irradiance of only 0.18 W/m(2) (18 mu W/cm(2)) to remain operating. The ASM circuit can operate correctly using a local power supply voltage of 453 mV, dissipating only 0.085 mu W. These values are, to the best of the authors' knowledge, the lowest reported in the literature. The maximum efficiency of the SC converter is 70.3 % for an input power of 48 mu W, which is comparable with reported values from circuits operating at similar power levels.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This letter presents a new parallel method for hyperspectral unmixing composed by the efficient combination of two popular methods: vertex component analysis (VCA) and sparse unmixing by variable splitting and augmented Lagrangian (SUNSAL). First, VCA extracts the endmember signatures, and then, SUNSAL is used to estimate the abundance fractions. Both techniques are highly parallelizable, which significantly reduces the computing time. A design for the commodity graphics processing units of the two methods is presented and evaluated. Experimental results obtained for simulated and real hyperspectral data sets reveal speedups up to 100 times, which grants real-time response required by many remotely sensed hyperspectral applications.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This work demonstrates the feasibility of using polymeric micro- and nanofiber-composed films and liquid crystals as electrically switchable scattering light shutters. We present a concept of electro-optic device based on an innovative combination of two mature technologies: optics of nematic liquid crystals and electrospinning of nanofibers. These devices have electric and optical characteristics far superior to other comparable methods. The simulation presented shows results that are highly consistent with those of experiments and that explain the working mechanism of the devices.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Single processor architectures are unable to provide the required performance of high performance embedded systems. Parallel processing based on general-purpose processors can achieve these performances with a considerable increase of required resources. However, in many cases, simplified optimized parallel cores can be used instead of general-purpose processors achieving better performance at lower resource utilization. In this paper, we propose a configurable many-core architecture to serve as a co-processor for high-performance embedded computing on Field-Programmable Gate Arrays. The architecture consists of an array of configurable simple cores with support for floating-point operations interconnected with a configurable interconnection network. For each core it is possible to configure the size of the internal memory, the supported operations and number of interfacing ports. The architecture was tested in a ZYNQ-7020 FPGA in the execution of several parallel algorithms. The results show that the proposed many-core architecture achieves better performance than that achieved with a parallel generalpurpose processor and that up to 32 floating-point cores can be implemented in a ZYNQ-7020 SoC FPGA.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Hyperspectral imaging has become one of the main topics in remote sensing applications, which comprise hundreds of spectral bands at different (almost contiguous) wavelength channels over the same area generating large data volumes comprising several GBs per flight. This high spectral resolution can be used for object detection and for discriminate between different objects based on their spectral characteristics. One of the main problems involved in hyperspectral analysis is the presence of mixed pixels, which arise when the spacial resolution of the sensor is not able to separate spectrally distinct materials. Spectral unmixing is one of the most important task for hyperspectral data exploitation. However, the unmixing algorithms can be computationally very expensive, and even high power consuming, which compromises the use in applications under on-board constraints. In recent years, graphics processing units (GPUs) have evolved into highly parallel and programmable systems. Specifically, several hyperspectral imaging algorithms have shown to be able to benefit from this hardware taking advantage of the extremely high floating-point processing performance, compact size, huge memory bandwidth, and relatively low cost of these units, which make them appealing for onboard data processing. In this paper, we propose a parallel implementation of an augmented Lagragian based method for unsupervised hyperspectral linear unmixing on GPUs using CUDA. The method called simplex identification via split augmented Lagrangian (SISAL) aims to identify the endmembers of a scene, i.e., is able to unmix hyperspectral data sets in which the pure pixel assumption is violated. The efficient implementation of SISAL method presented in this work exploits the GPU architecture at low level, using shared memory and coalesced accesses to memory.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Remote hyperspectral sensors collect large amounts of data per flight usually with low spatial resolution. It is known that the bandwidth connection between the satellite/airborne platform and the ground station is reduced, thus a compression onboard method is desirable to reduce the amount of data to be transmitted. This paper presents a parallel implementation of an compressive sensing method, called parallel hyperspectral coded aperture (P-HYCA), for graphics processing units (GPU) using the compute unified device architecture (CUDA). This method takes into account two main properties of hyperspectral dataset, namely the high correlation existing among the spectral bands and the generally low number of endmembers needed to explain the data, which largely reduces the number of measurements necessary to correctly reconstruct the original data. Experimental results conducted using synthetic and real hyperspectral datasets on two different GPU architectures by NVIDIA: GeForce GTX 590 and GeForce GTX TITAN, reveal that the use of GPUs can provide real-time compressive sensing performance. The achieved speedup is up to 20 times when compared with the processing time of HYCA running on one core of the Intel i7-2600 CPU (3.4GHz), with 16 Gbyte memory.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The application of compressive sensing (CS) to hyperspectral images is an active area of research over the past few years, both in terms of the hardware and the signal processing algorithms. However, CS algorithms can be computationally very expensive due to the extremely large volumes of data collected by imaging spectrometers, a fact that compromises their use in applications under real-time constraints. This paper proposes four efficient implementations of hyperspectral coded aperture (HYCA) for CS, two of them termed P-HYCA and P-HYCA-FAST and two additional implementations for its constrained version (CHYCA), termed P-CHYCA and P-CHYCA-FAST on commodity graphics processing units (GPUs). HYCA algorithm exploits the high correlation existing among the spectral bands of the hyperspectral data sets and the generally low number of endmembers needed to explain the data, which largely reduces the number of measurements necessary to correctly reconstruct the original data. The proposed P-HYCA and P-CHYCA implementations have been developed using the compute unified device architecture (CUDA) and the cuFFT library. Moreover, this library has been replaced by a fast iterative method in the P-HYCA-FAST and P-CHYCA-FAST implementations that leads to very significant speedup factors in order to achieve real-time requirements. The proposed algorithms are evaluated not only in terms of reconstruction error for different compressions ratios but also in terms of computational performance using two different GPU architectures by NVIDIA: 1) GeForce GTX 590; and 2) GeForce GTX TITAN. Experiments are conducted using both simulated and real data revealing considerable acceleration factors and obtaining good results in the task of compressing remotely sensed hyperspectral data sets.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

One of the main problems of hyperspectral data analysis is the presence of mixed pixels due to the low spatial resolution of such images. Linear spectral unmixing aims at inferring pure spectral signatures and their fractions at each pixel of the scene. The huge data volumes acquired by hyperspectral sensors put stringent requirements on processing and unmixing methods. This letter proposes an efficient implementation of the method called simplex identification via split augmented Lagrangian (SISAL) which exploits the graphics processing unit (GPU) architecture at low level using Compute Unified Device Architecture. SISAL aims to identify the endmembers of a scene, i.e., is able to unmix hyperspectral data sets in which the pure pixel assumption is violated. The proposed implementation is performed in a pixel-by-pixel fashion using coalesced accesses to memory and exploiting shared memory to store temporary data. Furthermore, the kernels have been optimized to minimize the threads divergence, therefore achieving high GPU occupancy. The experimental results obtained for the simulated and real hyperspectral data sets reveal speedups up to 49 times, which demonstrates that the GPU implementation can significantly accelerate the method's execution over big data sets while maintaining the methods accuracy.