934 resultados para Parallel design patterns
Resumo:
Nine ruminally cannulated cows fed different energy sources were used to evaluate an avian-derived polyclonal antibody preparation (PAP-MV) against the specific ruminal bacteria Streptococcus bovis, Fusobacterium necrophorum, Clostridium aminophilum, Peptostreptococcus anaerobius, and Clostridium stick-landii and monensin (MON) on ruminal fermentation patterns and in vivo digestibility. The experimental design was three 3 x 3 Latin squares distinguished by the main energy source in the diet [dry-ground corn grain (CG), high-moisture corn silage (HMCS), or citrus pulp (CiPu)]. Inside each Latin square, animals received one of the feed additives per period [none (CON), MON, or PAP-MV]. Dry matter intake and ruminal fermentation variables such as pH, total short-chain fatty acids (tSCFA), which included acetate, propionate, and butyrate, as well as lactic acid and NH(3)-N concentration were analyzed in this trial. Total tract DM apparent digestibility and its fractions were estimated using chromic oxide as an external marker. Each experimental period lasted 21 d. Ruminal fluid sampling was carried out on the last day of the period at 0, 2, 4, 6, 8, 10, and 12 h after the morning meal. Ruminal pH was higher (P = 0.006) 4 h postfeeding in MON and PAP-MV groups when compared with CON. Acetate: propionate ratio was greater in PAP-MV compared with MON across sampling times. Polyclonal antibodies did not alter (P > 0.05) tSCFA, molar proportion of acetate and butyrate, or lactic acid and NH(3)-N concentration. Ruminal pH was higher (P = 0.01), 4 h postfeeding in CiPu diets compared with CG and HMCS. There was no interaction between feed additive and energy source (P > 0.05) for any of the digestibility coefficients analyzed. Starch digestibility was less (P = 0.008) in PAP-MV when compared with CON and MON. In relation to energy sources, NDF digestibility was greater (P = 0.007) in CG and CiPu vs. the HMCS diet. The digestibility of ADF was greater (P = 0.002) in CiPu diets followed by CG and HMCS. Feeding PAP-MV or monensin altered ruminal fermentation patterns and digestive function in cows; however, those changes were independent of the main energy source of the diet.
Resumo:
Outcome after traumatic brain injury (TBI) is characterized by a high degree of variability which has often been difficult to capture in traditional outcome studies. The purpose of this study was to describe patterns of community integration 2-5 years after TBI. Participants were 208 patients admitted to a Brain Injury Rehabilitation Unit between 1991-1995 in Brisbane, Australia. The design comprised retrospective data collection and questionnaire follow-up by mail. Mean follow-up was 3.5 years. Demographic, injury severity and functional status variables were retrieved from hospital records. Community integration was assessed using the Community Integration Questionnaire (CIQ), and vocational status measured by a self administered questionnaire. Data was analysed using cluster analysis which divided the data into meaningful subsets. Based on the CIQ subscale scores of home, social and productive integration, a three cluster solution was selected, with groups labelled as working (n = 78), balanced (n = 46) and poorly integrated (n = 84). Although 38% of the sample returned to a high level of productive activity and 22% achieved a balanced lifestyle, overall community integration was poor for the remainder. This poorly integrated group had more severe injury characterized by longer periods of acute care and post-traumatic amnesia (PTA) and greater functional disability on discharge. These findings have implications for service delivery prior to and during the process of reintegration after brain injury.
Resumo:
The design and development of two X-band amplifying reflectarrays is presented. The arrays use dual-polarized aperture coupled patch antennas with FET transistors and phasing circuits to amplify a microwave signal and to radiate it in a chosen direction. Two cases are considered, one when a reflectarray converts a spherical wave due to a feed horn into a plane wave radiated into a boresight direction, and two, when the reflectarray converts a spherical wave due to a dual-polarized four-element feed array into a co-focal spherical wave. This amplified signal is received in an orthogonal port of the feed array so that the entire structure acts as a spatial power combiner. The two amplifying arrays are tested in the near-field zone for phase distribution over their apertures to achieve the required beam formation. Alternatively, their radiation patterns or gains are investigated.
Resumo:
A method is presented for calculating the currents and winding patterns required to design independent zonal and tesseral shim coils for magnetic resonance imaging. Both actively shielded and unshielded configurations are considered, and the region of interest can be located asymmetrically with respect to the coil's length. Streamline, target-field and Fourier-series methods are utilized. The desired target-field is specified at two cylindrical radii, on and inside a circular conducting cylinder of length 2L and radius a. The specification is over some asymmetric portion pL < z < qL of the coil's length (-1 < p < q < 1). Arbitrary functions are used in the outer sections, -L < z < pL and qL < z < L, to ensure continuity of the magnetic field across the entire length of the coil. The entire field is then periodically extended as a half-range cosine Fourier series about either end of the coil. The resultant Fourier coefficients are then substituted into the Fourier-series expressions for the internal and external magnetic fields, and current densities and stream functions on both the primary coil and shield. A contour plot of the stream function directly gives the required coil winding patterns. Spherical harmonic analysis and shielding analysis on field calculations from a ZX shim coil indicate that example designs and theory are well matched.
Resumo:
The integration and composition of software systems requires a good architectural design phase to speed up communications between (remote) components. However, during implementation phase, the code to coordinate such components often ends up mixed in the main business code. This leads to maintenance problems, raising the need for, on the one hand, separating the coordination code from the business code, and on the other hand, providing mechanisms for analysis and comprehension of the architectural decisions once made. In this context our aim is at developing a domain-specific language, CoordL, to describe typical coordination patterns. From our point of view, coordination patterns are abstractions, in a graph form, over the composition of coordination statements from the system code. These patterns would allow us to identify, by means of pattern-based graph search strategies, the code responsible for the coordination of the several components in a system. The recovering and separation of the architectural decisions for a better comprehension of the software is the main purpose of this pattern language
Resumo:
OBJECTIVE: To extend an existing computer programme for the evaluation and design of shift schedules (BASS 3) by integrating workload as well as economic aspects. METHODS: The redesigned prototype BASS 4 includes a new module with a suitable and easily applicable screening method (EBA) for the assessment of the intensity of physical, emotional and cognitive workload components and their temporal patterns. Specified criterion functions based on these ratings allow for an adjustment of shift and rest duration according to the intensity of physical and mental workload. Furthermore, with regard to interactive effects both workload and temporal conditions, e.g. time of day, are taken into account. In a second new module, important economic aspects and criteria have been implemented. Different ergonomic solutions for scheduling problems can now also be evaluated with regard to their economic costs. RESULTS: The new version of the computer programme (BASS 4) can now simultaneously take into account numerous ergonomic, legal, agreed and economic criteria for the design and evaluation of working hours. CONCLUSIONS: BASS 4 can now be used as an instrument for the design and the evaluation of working hours with regard to legal, ergonomic and economic aspects at the shop floor as well as in administrative (e.g. health and safety inspection) and research problems.
Resumo:
Embedded real-time applications increasingly present high computation requirements, which need to be completed within specific deadlines, but that present highly variable patterns, depending on the set of data available in a determined instant. The current trend to provide parallel processing in the embedded domain allows providing higher processing power; however, it does not address the variability in the processing pattern. Dimensioning each device for its worst-case scenario implies lower average utilization, and increased available, but unusable, processing in the overall system. A solution for this problem is to extend the parallel execution of the applications, allowing networked nodes to distribute the workload, on peak situations, to neighbour nodes. In this context, this report proposes a framework to develop parallel and distributed real-time embedded applications, transparently using OpenMP and Message Passing Interface (MPI), within a programming model based on OpenMP. The technical report also devises an integrated timing model, which enables the structured reasoning on the timing behaviour of these hybrid architectures.
Resumo:
Multicore platforms have transformed parallelism into a main concern. Parallel programming models are being put forward to provide a better approach for application programmers to expose the opportunities for parallelism by pointing out potentially parallel regions within tasks, leaving the actual and dynamic scheduling of these regions onto processors to be performed at runtime, exploiting the maximum amount of parallelism. It is in this context that this paper proposes a scheduling approach that combines the constant-bandwidth server abstraction with a priority-aware work-stealing load balancing scheme which, while ensuring isolation among tasks, enables parallel tasks to be executed on more than one processor at a given time instant.
Resumo:
An approach for the analysis of uncertainty propagation in reliability-based design optimization of composite laminate structures is presented. Using the Uniform Design Method (UDM), a set of design points is generated over a domain centered on the mean reference values of the random variables. A methodology based on inverse optimal design of composite structures to achieve a specified reliability level is proposed, and the corresponding maximum load is outlined as a function of ply angle. Using the generated UDM design points as input/output patterns, an Artificial Neural Network (ANN) is developed based on an evolutionary learning process. Then, a Monte Carlo simulation using ANN development is performed to simulate the behavior of the critical Tsai number, structural reliability index, and their relative sensitivities as a function of the ply angle of laminates. The results are generated for uniformly distributed random variables on a domain centered on mean values. The statistical analysis of the results enables the study of the variability of the reliability index and its sensitivity relative to the ply angle. Numerical examples showing the utility of the approach for robust design of angle-ply laminates are presented.
Resumo:
This letter presents a new parallel method for hyperspectral unmixing composed by the efficient combination of two popular methods: vertex component analysis (VCA) and sparse unmixing by variable splitting and augmented Lagrangian (SUNSAL). First, VCA extracts the endmember signatures, and then, SUNSAL is used to estimate the abundance fractions. Both techniques are highly parallelizable, which significantly reduces the computing time. A design for the commodity graphics processing units of the two methods is presented and evaluated. Experimental results obtained for simulated and real hyperspectral data sets reveal speedups up to 100 times, which grants real-time response required by many remotely sensed hyperspectral applications.
Resumo:
Dissertação apresentada para a obtenção do Grau de Doutor em Informática pela Universidade Nova de Lisboa, Faculdade de Ciências e Tecnologia.
Resumo:
Swarm Intelligence (SI) is the property of a system whereby the collective behaviors of (unsophisticated) agents interacting locally with their environment cause coherent functional global patterns to emerge. Particle swarm optimization (PSO) is a form of SI, and a population-based search algorithm that is initialized with a population of random solutions, called particles. These particles are flying through hyperspace and have two essential reasoning capabilities: their memory of their own best position and knowledge of the swarm's best position. In a PSO scheme each particle flies through the search space with a velocity that is adjusted dynamically according with its historical behavior. Therefore, the particles have a tendency to fly towards the best search area along the search process. This work proposes a PSO based algorithm for logic circuit synthesis. The results show the statistical characteristics of this algorithm with respect to number of generations required to achieve the solutions. It is also presented a comparison with other two Evolutionary Algorithms, namely Genetic and Memetic Algorithms.
Resumo:
IEEE International Symposium on Circuits and Systems, pp. 724 – 727, Seattle, EUA
Resumo:
Recent integrated circuit technologies have opened the possibility to design parallel architectures with hundreds of cores on a single chip. The design space of these parallel architectures is huge with many architectural options. Exploring the design space gets even more difficult if, beyond performance and area, we also consider extra metrics like performance and area efficiency, where the designer tries to design the architecture with the best performance per chip area and the best sustainable performance. In this paper we present an algorithm-oriented approach to design a many-core architecture. Instead of doing the design space exploration of the many core architecture based on the experimental execution results of a particular benchmark of algorithms, our approach is to make a formal analysis of the algorithms considering the main architectural aspects and to determine how each particular architectural aspect is related to the performance of the architecture when running an algorithm or set of algorithms. The architectural aspects considered include the number of cores, the local memory available in each core, the communication bandwidth between the many-core architecture and the external memory and the memory hierarchy. To exemplify the approach we did a theoretical analysis of a dense matrix multiplication algorithm and determined an equation that relates the number of execution cycles with the architectural parameters. Based on this equation a many-core architecture has been designed. The results obtained indicate that a 100 mm(2) integrated circuit design of the proposed architecture, using a 65 nm technology, is able to achieve 464 GFLOPs (double precision floating-point) for a memory bandwidth of 16 GB/s. This corresponds to a performance efficiency of 71 %. Considering a 45 nm technology, a 100 mm(2) chip attains 833 GFLOPs which corresponds to 84 % of peak performance These figures are better than those obtained by previous many-core architectures, except for the area efficiency which is limited by the lower memory bandwidth considered. The results achieved are also better than those of previous state-of-the-art many-cores architectures designed specifically to achieve high performance for matrix multiplication.
Resumo:
Single processor architectures are unable to provide the required performance of high performance embedded systems. Parallel processing based on general-purpose processors can achieve these performances with a considerable increase of required resources. However, in many cases, simplified optimized parallel cores can be used instead of general-purpose processors achieving better performance at lower resource utilization. In this paper, we propose a configurable many-core architecture to serve as a co-processor for high-performance embedded computing on Field-Programmable Gate Arrays. The architecture consists of an array of configurable simple cores with support for floating-point operations interconnected with a configurable interconnection network. For each core it is possible to configure the size of the internal memory, the supported operations and number of interfacing ports. The architecture was tested in a ZYNQ-7020 FPGA in the execution of several parallel algorithms. The results show that the proposed many-core architecture achieves better performance than that achieved with a parallel generalpurpose processor and that up to 32 floating-point cores can be implemented in a ZYNQ-7020 SoC FPGA.