981 resultados para Loosely packed array


Relevância:

20.00% 20.00%

Publicador:

Resumo:

3D dose reconstruction is a verification of the delivered absorbed dose. Our aim was to describe and evaluate a 3D dose reconstruction method applied to phantoms in the context of narrow beams. A solid water phantom and a phantom containing a bone-equivalent material were irradiated on a 6 MV linac. The transmitted dose was measured by using one array of a 2D ion chamber detector. The dose reconstruction was obtained by an iterative algorithm. A phantom set-up error and organ interfraction motion were simulated to test the algorithm sensitivity. In all configurations convergence was obtained within three iterations. A local reconstructed dose agreement of at least 3% / 3mm with respect to the planned dose was obtained, except in a few points of the penumbra. The reconstructed primary fluences were consistent with the planned ones, which validates the whole reconstruction process. The results validate our method in a simple geometry and for narrow beams. The method is sensitive to a set-up error of a heterogeneous phantom and interfraction heterogeneous organ motion.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this investigation, high-resolution, 1x1x1-mm(3) functional magnetic resonance imaging (fMRI) at 7 T is performed using a multichannel array head coil and a surface coil approach. Scan geometry was optimized for each coil separately to exploit the strengths of both coils. Acquisitions with the surface coil focused on partial brain coverage, while whole-brain coverage fMRI experiments were performed with the array head coil. BOLD sensitivity in the occipital lobe was found to be higher with the surface coil than with the head array, suggesting that restriction of signal detection to the area of interest may be beneficial for localized activation studies. Performing independent component analysis (ICA) decomposition of the fMRI data, we consistently detected BOLD signal changes and resting state networks. In the surface coil data, a small negative BOLD response could be detected in these resting state network areas. Also in the data acquired with the surface coil, two distinct components of the positive BOLD signal were consistently observed. These two components were tentatively assigned to tissue and venous signal changes.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Methods for the extraction of features from physiological datasets are growing needs as clinical investigations of Alzheimer’s disease (AD) in large and heterogeneous population increase. General tools allowing diagnostic regardless of recording sites, such as different hospitals, are essential and if combined to inexpensive non-invasive methods could critically improve mass screening of subjects with AD. In this study, we applied three state of the art multiway array decomposition (MAD) methods to extract features from electroencephalograms (EEGs) of AD patients obtained from multiple sites. In comparison to MAD, spectral-spatial average filter (SSFs) of control and AD subjects were used as well as a common blind source separation method, algorithm for multiple unknown signal extraction (AMUSE). We trained a feed-forward multilayer perceptron (MLP) to validate and optimize AD classification from two independent databases. Using a third EEG dataset, we demonstrated that features extracted from MAD outperformed features obtained from SSFs AMUSE in terms of root mean squared error (RMSE) and reaching up to 100% of accuracy in test condition. We propose that MAD maybe a useful tool to extract features for AD diagnosis offering great generalization across multi-site databases and opening doors to the discovery of new characterization of the disease.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Rapid amplification of cDNA ends (RACE) is a widely used approach for transcript identification. Random clone selection from the RACE mixture, however, is an ineffective sampling strategy if the dynamic range of transcript abundances is large. To improve sampling efficiency of human transcripts, we hybridized the products of the RACE reaction onto tiling arrays and used the detected exons to delineate a series of reverse-transcriptase (RT)-PCRs, through which the original RACE transcript population was segregated into simpler transcript populations. We independently cloned the products and sequenced randomly selected clones. This approach, RACEarray, is superior to direct cloning and sequencing of RACE products because it specifically targets new transcripts and often results in overall normalization of transcript abundance. We show theoretically and experimentally that this strategy leads indeed to efficient sampling of new transcripts, and we investigated multiplexing the strategy by pooling RACE reactions from multiple interrogated loci before hybridization.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We report on two patients with de novo subtelomeric terminal deletion of chromosome 6p. Patient 1 is an 8-month-old female born with normal growth parameters, typical facial features of 6pter deletion, bilateral corectopia, and protruding tongue. She has severe developmental delay, profound bilateral neurosensory deafness, poor visual contact, and hypsarrhythmia since the age of 6 months. Patient 2 is a 5-year-old male born with normal growth parameters and unilateral hip dysplasia; he has a characteristic facial phenotype, bilateral embryotoxon, and moderate mental retardation. Further characterization of the deletion, using high-resolution array comparative genomic hybridization (array-CGH; Agilent Human Genome kit 244 K), revealed that Patient 1 has a 8.1 Mb 6pter-6p24.3 deletion associated with a contiguous 5.8 Mb 6p24.3-6p24.1 duplication and Patient 2 a 5.7 Mb 6pter-6p25.1 deletion partially overlapping with that of Patient 1. Complementary FISH and array analysis showed that the inv del dup(6) in Patient 1 originated de novo. Our results demonstrate that simple rearrangements are often more complex than defined by standard techniques. We also discuss genotype-phenotype correlations including previously reported cases of deletion 6p.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Process variations are a major bottleneck for digital CMOS integrated circuits manufacturability and yield. That iswhy regular techniques with different degrees of regularity are emerging as possible solutions. Our proposal is a new regular layout design technique called Via-Configurable Transistors Array (VCTA) that pushes to the limit circuit layout regularity for devices and interconnects in order to maximize regularity benefits. VCTA is predicted to perform worse than the Standard Cell approach designs for a certain technology node but it will allow the use of a future technology on an earlier time. Ourobjective is to optimize VCTA for it to be comparable to the Standard Cell design in an older technology. Simulations for the first unoptimized version of our VCTA of delay and energy consumption for a Full Adder circuit in the 90 nm technology node are presented and also the extrapolation for Carry-RippleAdders from 4 bits to 64 bits.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The well-known structure of an array combiner along with a maximum likelihood sequence estimator (MLSE) receiveris the basis for the derivation of a space-time processor presentinggood properties in terms of co-channel and intersymbol interferencerejection. The use of spatial diversity at the receiver front-endtogether with a scalar MLSE implies a joint design of the spatialcombiner and the impulse response for the sequence detector. Thisis faced using the MMSE criterion under the constraint that thedesired user signal power is not cancelled, yielding an impulse responsefor the sequence detector that is matched to the channel andcombiner response. The procedure maximizes the signal-to-noiseratio at the input of the detector and exhibits excellent performancein realistic multipath channels.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Combustion of wood is increasing because of the needs of decreasing the emissions of carbon dioxide and the amount of waste going to landfills. Wood based fuels are often scattered on a large area. The transport distances should be short enough to prevent too high costs, and so the size of heating and power plants using wood fuels is often rather small. Combustion technologies of small-size units have to be developed to reach efficient and environmentally friendly energy production. Furnaces that use different packed bed combustion or gasification techniques areoften most economic in small-scale energy production. Ignition front propagation rate affects the stability, heat release rate and emissions of packed bed combustion. Ignition front propagation against airflow in packed beds of wood fuels has been studied. The research has been carried out mainly experimentally. Theoretical aspects have been considered to draw conclusions about the experimental results. The effects of airflow rate, moisture content of the fuel, size, shape and density of particles, and porosity of the bed on the propagation rate of the ignition front have been studied. The experiments were carried out in a pot furnace. The fuels used in the experiments were mainly real wood fuels that are often burned in the production of energy. The fuel types were thin wood chips, saw dust, shavings, wood chips, and pellets with different sizes. Also a few mixturesof the above were tested. Increase in the moisture content of the fuel decreases the propagation rates of the ignition front and makes the range of possible airflow rates narrower because of the energy needed for the evaporation of water and the dilution of volatile gases due to evaporated steam. Increase in the airflow rate increases the ignition rate until a maximum rate of propagation is reached after which it decreases. The maximum flame propagation rate is not always reached in stoichiometric combustion conditions. Increase in particle size and density transfers the optimum airflow rate towards fuel lean conditions. Mixing of small and large particles is often advantageous, because small particles make itpossible to reach the maximum ignition rate in fuel rich conditions, and large particles widen the range of possible airflow rates. A correlation was found forthe maximum rate of ignition front propagation in different wood fuels. According to the correlation, the maximum ignition mass flux is increased when the sphericity of the particles and the porosity of the bed are increased and the moisture content of the fuel is decreased. Another fit was found between sphericity and porosity. Increase in sphericity decreases the porosity of the bed. The reasons of the observed results are discussed.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Value chain collaboration has been a prevailing topic for research, and there is a constantly growing interest in developing collaborative models for improved efficiency in logistics. One area of collaboration is demand information management, which enables improved visibility and decrease of inventories in the value chain. Outsourcing of non-core competencies has changed the nature of collaboration from intra-enterprise to cross-enterprise activity, and this together with increasing competition in the globalizing markets have created a need for methods and tools for collaborative work. The retailer part in the value chain of consumer packaged goods (CPG) has been studied relatively widely, proven models have been defined, and there exist several best practice collaboration cases. The information and communications technology has developed rapidly, offering efficient solutions and applications to exchange information between value chain partners. However, the majority of CPG industry still works with traditional business models and practices. This concerns especially companies operating in the upstream of the CPG value chain. Demand information for consumer packaged goods originates at retailers' counters, based on consumers' buying decisions. As this information does not get transferred along the value chain towards the upstream parties, each player needs to optimize their part, causing safety margins for inventories and speculation in purchasing decisions. The safety margins increase with each player, resulting in a phenomenon known as the bullwhip effect. The further the company is from the original demand information source, the more distorted the information is. This thesis concentrates on the upstream parts of the value chain of consumer packaged goods, and more precisely the packaging value chain. Packaging is becoming a part of the product with informative and interactive features, and therefore is not just a cost item needed to protect the product. The upstream part of the CPG value chain is distinctive, as the product changes after each involved party, and therefore the original demand information from the retailers cannot be utilized as such – even if it were transferred seamlessly. The objective of this thesis is to examine the main drivers for collaboration, and barriers causing the moderate adaptation level of collaborative models. Another objective is to define a collaborative demand information management model and test it in a pilot business situation in order to see if the barriers can be eliminated. The empirical part of this thesis contains three parts, all related to the research objective, but involving different target groups, viewpoints and research approaches. The study shows evidence that the main barriers for collaboration are very similar to the barriers in the lower part of the same value chain; lack of trust, lack of business case and lack of senior management commitment. Eliminating one of them – the lack of business case – is not enough to eliminate the two other barriers, as the operational model in this thesis shows. The uncertainty of the future, fear of losing an independent position in purchasing decision making and lack of commitment remain strong enough barriers to prevent the implementation of the proposed collaborative business model. The study proposes a new way of defining the value chain processes: it divides the contracting and planning process into two processes, one managing the commercial parts and the other managing the quantity and specification related issues. This model can reduce the resistance to collaboration, as the commercial part of the contracting process would remain the same as in the traditional model. The quantity/specification-related issues would be managed by the parties with the best capabilities and resources, as well as access to the original demand information. The parties in between would be involved in the planning process as well, as their impact for the next party upstream is significant. The study also highlights the future challenges for companies operating in the CPG value chain. The markets are becoming global, with toughening competition. Also, the technology development will most likely continue with a speed exceeding the adaptation capabilities of the industry. Value chains are also becoming increasingly dynamic, which means shorter and more agile business relationships, and at the same time the predictability of consumer demand is getting more difficult due to shorter product life cycles and trends. These changes will certainly have an effect on companies' operational models, but it is very difficult to estimate when and how the proven methods will gain wide enough adaptation to become standards.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Ground-based gamma-ray astronomy has had a major breakthrough with the impressive results obtained using systems of imaging atmospheric Cherenkov telescopes. Ground-based gamma-ray astronomy has a huge potential in astrophysics, particle physics and cosmology. CTA is an international initiative to build the next generation instrument, with a factor of 5-10 improvement in sensitivity in the 100 GeV-10 TeV range and the extension to energies well below 100 GeV and above 100 TeV. CTA will consist of two arrays (one in the north, one in the south) for full sky coverage and will be operated as open observatory. The design of CTA is based on currently available technology. This document reports on the status and presents the major design concepts of CTA.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Ground-based gamma-ray astronomy has had a major breakthrough with the impressive results obtained using systems of imaging atmospheric Cherenkov telescopes. Ground-based gamma-ray astronomy has a huge potential in astrophysics, particle physics and cosmology. CTA is an international initiative to build the next generation instrument, with a factor of 5-10 improvement in sensitivity in the 100 GeV-10 TeV range and the extension to energies well below 100 GeV and above 100 TeV. CTA will consist of two arrays (one in the north, one in the south) for full sky coverage and will be operated as open observatory. The design of CTA is based on currently available technology. This document reports on the status and presents the major design concepts of CTA.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Ground-based gamma-ray astronomy has had a major breakthrough with the impressive results obtained using systems of imaging atmospheric Cherenkov telescopes. Ground-based gamma-ray astronomy has a huge potential in astrophysics, particle physics and cosmology. CTA is an international initiative to build the next generation instrument, with a factor of 5-10 improvement in sensitivity in the 100 GeV-10 TeV range and the extension to energies well below 100 GeV and above 100 TeV. CTA will consist of two arrays (one in the north, one in the south) for full sky coverage and will be operated as open observatory. The design of CTA is based on currently available technology. This document reports on the status and presents the major design concepts of CTA.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A simple and sensitive on-line flow injection system for determination of zinc with FAAS has been described. The method is based on the separation and preconcentration of zinc on a microcolumn of immobilized Alizarin Red S on alumina. The adsorbed analyte is then eluted with 250 µL of nitric acid (1 mol L-1) and is transported to flame atomic absorption spectrometer for quantification. The effect of pH, sample and eluent flow rates and presence of various cations and anions on the retention of zinc was investigated. The sorption of zinc was quantitative in the pH range of 5.5-8.5. For a sample volume of 25 mL an enrichment factor of 144 and a detection limit (3S) of 0.2 µg L-1 was obtained. The precision (RSD, n=7) was 3.0% at the 20 µg L-1 level. The developed system was successfully applied to the determination of zinc in water samples, hair, urine and saliva.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Separations using supercritical fluid chromatography (SFC) with packed columns have been re-discovered and explored in recent years. SFC enables fast and efficient separations and, in some cases, gives better results than high performance liquid chromatography (HPLC). This paper provides an overview of recent advances in SFC separations using packed columns for both achiral and chiral separations. The most important types of stationary phases used in SFC are discussed as well as the most critical parameters involved in the separations and some recent applications.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The main goal of the present Master’s Thesis project was to create a field-programmable gate array (FPGA) based system for the control of single-electron transistors or other cryoelectronic devices. The FPGA and similar technologies are studied in the present work. The fixed and programmable logic are compared with each other. The main features and limitations of the hardware used in the project are investigated. The hardware and software connections of the device to the computer are shown in detail. The software development techniques for FPGA-based design are described. The steps of design for programmable logic are considered. Furthermore, the results of filters implemented in the software are illustrated.