951 resultados para Worst-case execution-time
Resumo:
Coupled-cluster theory in its single-reference formulation represents one of the most successful approaches in quantum chemistry for the description of atoms and molecules. To extend the applicability of single-reference coupled-cluster theory to systems with degenerate or near-degenerate electronic configurations, multireference coupled-cluster methods have been suggested. One of the most promising formulations of multireference coupled cluster theory is the state-specific variant suggested by Mukherjee and co-workers (Mk-MRCC). Unlike other multireference coupled-cluster approaches, Mk-MRCC is a size-extensive theory and results obtained so far indicate that it has the potential to develop to a standard tool for high-accuracy quantum-chemical treatments. This work deals with developments to overcome the limitations in the applicability of the Mk-MRCC method. Therefore, an efficient Mk-MRCC algorithm has been implemented in the CFOUR program package to perform energy calculations within the singles and doubles (Mk-MRCCSD) and singles, doubles, and triples (Mk-MRCCSDT) approximations. This implementation exploits the special structure of the Mk-MRCC working equations that allows to adapt existing efficient single-reference coupled-cluster codes. The algorithm has the correct computational scaling of d*N^6 for Mk-MRCCSD and d*N^8 for Mk-MRCCSDT, where N denotes the system size and d the number of reference determinants. For the determination of molecular properties as the equilibrium geometry, the theory of analytic first derivatives of the energy for the Mk-MRCC method has been developed using a Lagrange formalism. The Mk-MRCC gradients within the CCSD and CCSDT approximation have been implemented and their applicability has been demonstrated for various compounds such as 2,6-pyridyne, the 2,6-pyridyne cation, m-benzyne, ozone and cyclobutadiene. The development of analytic gradients for Mk-MRCC offers the possibility of routinely locating minima and transition states on the potential energy surface. It can be considered as a key step towards routine investigation of multireference systems and calculation of their properties. As the full inclusion of triple excitations in Mk-MRCC energy calculations is computational demanding, a parallel implementation is presented in order to circumvent limitations due to the required execution time. The proposed scheme is based on the adaption of a highly efficient serial Mk-MRCCSDT code by parallelizing the time-determining steps. A first application to 2,6-pyridyne is presented to demonstrate the efficiency of the current implementation.
Resumo:
Modern software systems, in particular distributed ones, are everywhere around us and are at the basis of our everyday activities. Hence, guaranteeing their cor- rectness, consistency and safety is of paramount importance. Their complexity makes the verification of such properties a very challenging task. It is natural to expect that these systems are reliable and above all usable. i) In order to be reliable, compositional models of software systems need to account for consistent dynamic reconfiguration, i.e., changing at runtime the communication patterns of a program. ii) In order to be useful, compositional models of software systems need to account for interaction, which can be seen as communication patterns among components which collaborate together to achieve a common task. The aim of the Ph.D. was to develop powerful techniques based on formal methods for the verification of correctness, consistency and safety properties related to dynamic reconfiguration and communication in complex distributed systems. In particular, static analysis techniques based on types and type systems appeared to be an adequate methodology, considering their success in guaranteeing not only basic safety properties, but also more sophisticated ones like, deadlock or livelock freedom in a concurrent setting. The main contributions of this dissertation are twofold. i) On the components side: we design types and a type system for a concurrent object-oriented calculus to statically ensure consistency of dynamic reconfigurations related to modifications of communication patterns in a program during execution time. ii) On the communication side: we study advanced safety properties related to communication in complex distributed systems like deadlock-freedom, livelock- freedom and progress. Most importantly, we exploit an encoding of types and terms of a typical distributed language, session π-calculus, into the standard typed π- calculus, in order to understand their expressive power.
Resumo:
This thesis deals with heterogeneous architectures in standard workstations. Heterogeneous architectures represent an appealing alternative to traditional supercomputers because they are based on commodity components fabricated in large quantities. Hence their price-performance ratio is unparalleled in the world of high performance computing (HPC). In particular, different aspects related to the performance and consumption of heterogeneous architectures have been explored. The thesis initially focuses on an efficient implementation of a parallel application, where the execution time is dominated by an high number of floating point instructions. Then the thesis touches the central problem of efficient management of power peaks in heterogeneous computing systems. Finally it discusses a memory-bounded problem, where the execution time is dominated by the memory latency. Specifically, the following main contributions have been carried out: A novel framework for the design and analysis of solar field for Central Receiver Systems (CRS) has been developed. The implementation based on desktop workstation equipped with multiple Graphics Processing Units (GPUs) is motivated by the need to have an accurate and fast simulation environment for studying mirror imperfection and non-planar geometries. Secondly, a power-aware scheduling algorithm on heterogeneous CPU-GPU architectures, based on an efficient distribution of the computing workload to the resources, has been realized. The scheduler manages the resources of several computing nodes with a view to reducing the peak power. The two main contributions of this work follow: the approach reduces the supply cost due to high peak power whilst having negligible impact on the parallelism of computational nodes. from another point of view the developed model allows designer to increase the number of cores without increasing the capacity of the power supply unit. Finally, an implementation for efficient graph exploration on reconfigurable architectures is presented. The purpose is to accelerate graph exploration, reducing the number of random memory accesses.
Resumo:
This thesis, after presenting recent advances obtained for the two-dimensional bin packing problem, focuses on the case where guillotine restrictions are imposed. A mathematical characterization of non-guillotine patterns is provided and the relation between the solution value of the two-dimensional problem with guillotine restrictions and the two-dimensional problem unrestricted is being studied from a worst-case perspective. Finally it presents a new heuristic algorithm, for the two-dimensional problem with guillotine restrictions, based on partial enumeration, and computationally evaluates its performance on a large set of instances from the literature. Computational experiments show that the algorithm is able to produce proven optimal solutions for a large number of problems, and gives a tight approximation of the optimum in the remaining cases.
Resumo:
Accurate placement of lesions is crucial for the effectiveness and safety of a retinal laser photocoagulation treatment. Computer assistance provides the capability for improvements to treatment accuracy and execution time. The idea is to use video frames acquired from a scanning digital ophthalmoscope (SDO) to compensate for retinal motion during laser treatment. This paper presents a method for the multimodal registration of the initial frame from an SDO retinal video sequence to a retinal composite image, which may contain a treatment plan. The retinal registration procedure comprises the following steps: 1) detection of vessel centerline points and identification of the optic disc; 2) prealignment of the video frame and the composite image based on optic disc parameters; and 3) iterative matching of the detected vessel centerline points in expanding matching regions. This registration algorithm was designed for the initialization of a real-time registration procedure that registers the subsequent video frames to the composite image. The algorithm demonstrated its capability to register various pairs of SDO video frames and composite images acquired from patients.
Resumo:
Extension of 3-D atmospheric data products back into the past is desirable for a wide range of applications. Historical upper-air data are important in this endeavour, particularly in the maritime regions of the tropics and the southern hemisphere, where observations are extremely sparse. Here we present newly digitized and re-evaluated early ship-based upper-air data from two cruises: (1) kite and registering balloon profiles from onboard the ship SMS Planet on a cruise from Europe around South Africa and across the Indian Ocean to the western Pacific in 1906/1907, and (2) ship-based radiosonde data from onboard the MS Schwabenland on a cruise from Europe across the Atlantic to Antarctica and back in 1938/1939. We describe the data and provide estimations of the errors. We compare the data with a recent reanalysis (the Twentieth Century Reanalysis Project, 20CR, Compo et al., 2011) that provides global 3-D data back to the 19th century based on an assimilation of surface pressure data only (plus monthly mean sea-surface temperatures). In cruise (1), the agreement is generally good, but large temperature differences appear during a period with a strong inversion. In cruise (2), after a subset of the data are corrected, close agreement between observations and 20CR is found for geopotential height (GPH) and temperature notwithstanding a likely cold bias of 20CR at the tropopause level. Results are considerably worse for relative humidity, which was reportedly inaccurately measured. Note that comparing 20CR, which has limited skill in the tropical regions, with measurements from ships in remote regions made under sometimes difficult conditions can be considered a worst case assessment. In view of that fact, the anomaly correlations for temperature of 0.3–0.6 in the lower troposphere in cruise (1) and of 0.5–0.7 for tropospheric temperature and GPH in cruise (2) are considered as promising results. Moreover, they are consistent with the error estimations. The results suggest room for further improvement of data products in remote regions.
Resumo:
The purpose of this study was to evaluate the effect of continuously released BDNF on peripheral nerve regeneration in a rat model. Initial in vitro evaluation of calcium alginate prolonged-release-capsules (PRC) proved a consistent release of BDNF for a minimum of 8 weeks. In vivo, a worst case scenario was created by surgical removal of a 20-mm section of the sciatic nerve of the rat. Twenty-four autologous fascia tubes were filled with calcium alginate spheres and sutured to the epineurium of both nerve ends. The animals were divided into 3 groups. In group 1, the fascial tube contained plain calcium alginate spheres. In groups 2 and 3, the fascial tube contained calcium alginate spheres with BDNF alone or BDNF stabilized with bovine serum albumin, respectively. The autocannibalization of the operated extremity was clinically assessed and documented in 12 additional rats. The regeneration was evaluated histologically at 4 weeks and 10 weeks in a blinded manner. The length of nerve fibers and the numbers of axons formed in the tube was measured. Over a 10-week period, axons have grown significantly faster in groups 2 and 3 with continuously released BDNF compared to the control. The rats treated with BDNF (groups 2 and 3) demonstrated significantly less autocannibalization than the control group (group 1). These results suggest that BDNF may not only stimulate faster peripheral nerve regeneration provided there is an ideal, biodegradable continuous delivery system but that it significantly reduces the neuropathic pain in the rat model.
Resumo:
The goal of this study was to propose a general numerical analysis methodology to evaluate the magnetic resonance imaging (MRI)-safety of active implants. Numerical models based on the finite element (FE) technique were used to estimate if the normal operation of an active device was altered during MRI imaging. An active implanted pump was chosen to illustrate the method. A set of controlled experiments were proposed and performed to validate the numerical model. The calculated induced voltages in the important electronic components of the device showed dependence with the MRI field strength. For the MRI radiofrequency fields, significant induced voltages of up to 20 V were calculated for a 0.3T field-strength MRI. For the 1.5 and 3.0T MRIs, the calculated voltages were insignificant. On the other hand, induced voltages up to 11 V were calculated in the critical electronic components for the 3.0T MRI due to the gradient fields. Values obtained in this work reflect to the worst case situation which is virtually impossible to achieve in normal scanning situations. Since the calculated voltages may be removed by appropriate protection circuits, no critical problems affecting the normal operation of the pump were identified. This study showed that the proposed methodology helps the identification of the possible incompatibilities between active implants and MR imaging, and can be used to aid the design of critical electronic systems to ensure MRI-safety
Resumo:
This paper describes a method for DRR generation as well as for volume gradients projection using hardware accelerated 2D texture mapping and accumulation buffering and demonstrates its application in 2D-3D registration of X-ray fluoroscopy to CT images. The robustness of the present registration scheme are guaranteed by taking advantage of a coarse-to-fine processing of the volume/image pyramids based on cubic B-splines. A human cadaveric spine specimen together with its ground truth was used to compare the present scheme with a purely software-based scheme in three aspects: accuracy, speed, and capture ranges. Our experiments revealed an equivalent accuracy and capture ranges but with much shorter registration time with the present scheme. More specifically, the results showed 0.8 mm average target registration error, 55 second average execution time per registration, and 10 mm and 10° capture ranges for the present scheme when tested on a 3.0 GHz Pentium 4 computer.
Resumo:
As the performance gap between microprocessors and memory continues to increase, main memory accesses result in long latencies which become a factor limiting system performance. Previous studies show that main memory access streams contain significant localities and SDRAM devices provide parallelism through multiple banks and channels. These locality and parallelism have not been exploited thoroughly by conventional memory controllers. In this thesis, SDRAM address mapping techniques and memory access reordering mechanisms are studied and applied to memory controller design with the goal of reducing observed main memory access latency. The proposed bit-reversal address mapping attempts to distribute main memory accesses evenly in the SDRAM address space to enable bank parallelism. As memory accesses to unique banks are interleaved, the access latencies are partially hidden and therefore reduced. With the consideration of cache conflict misses, bit-reversal address mapping is able to direct potential row conflicts to different banks, further improving the performance. The proposed burst scheduling is a novel access reordering mechanism, which creates bursts by clustering accesses directed to the same rows of the same banks. Subjected to a threshold, reads are allowed to preempt writes and qualified writes are piggybacked at the end of the bursts. A sophisticated access scheduler selects accesses based on priorities and interleaves accesses to maximize the SDRAM data bus utilization. Consequentially burst scheduling reduces row conflict rate, increasing and exploiting the available row locality. Using a revised SimpleScalar and M5 simulator, both techniques are evaluated and compared with existing academic and industrial solutions. With SPEC CPU2000 benchmarks, bit-reversal reduces the execution time by 14% on average over traditional page interleaving address mapping. Burst scheduling also achieves a 15% reduction in execution time over conventional bank in order scheduling. Working constructively together, bit-reversal and burst scheduling successfully achieve a 19% speedup across simulated benchmarks.
Resumo:
Planning in realistic domains typically involves reasoning under uncertainty, operating under time and resource constraints, and finding the optimal subset of goals to work on. Creating optimal plans that consider all of these features is a computationally complex, challenging problem. This dissertation develops an AO* search based planner named CPOAO* (Concurrent, Probabilistic, Over-subscription AO*) which incorporates durative actions, time and resource constraints, concurrent execution, over-subscribed goals, and probabilistic actions. To handle concurrent actions, action combinations rather than individual actions are taken as plan steps. Plan optimization is explored by adding two novel aspects to plans. First, parallel steps that serve the same goal are used to increase the plan’s probability of success. Traditionally, only parallel steps that serve different goals are used to reduce plan execution time. Second, actions that are executing but are no longer useful can be terminated to save resources and time. Conventional planners assume that all actions that were started will be carried out to completion. To reduce the size of the search space, several domain independent heuristic functions and pruning techniques were developed. The key ideas are to exploit dominance relations for candidate action sets and to develop relaxed planning graphs to estimate the expected rewards of states. This thesis contributes (1) an AO* based planner to generate parallel plans, (2) domain independent heuristics to increase planner efficiency, and (3) the ability to execute redundant actions and to terminate useless actions to increase plan efficiency.
Resumo:
A basic prerequisite for in vivo X-ray imaging of the lung is the exact determination of radiation dose. Achieving resolutions of the order of micrometres may become particularly challenging owing to increased dose, which in the worst case can be lethal for the imaged animal model. A framework for linking image quality to radiation dose in order to optimize experimental parameters with respect to dose reduction is presented. The approach may find application for current and future in vivo studies to facilitate proper experiment planning and radiation risk assessment on the one hand and exploit imaging capabilities on the other.
Resumo:
This paper examines the impact of disastrous and ‘ordinary’ floods on human societies in what is now Austria. The focus is on urban areas and their neighbourhoods. Examining institutional sources such as accounts of the bridge masters, charters, statutes and official petitions, it can be shown that city communities were well acquainted with this permanent risk: in fact, an office was established for the restoration of bridges and the maintenance of water defences and large depots for timber and water pipes ensured that the reconstruction of bridges and the system of water supply could start immediately after the floods had subsided. Carpenters and similar groups gained 10 to 20 per cent of their income from the repair of bridges and other flood damage. The construction of houses in endangered zones was adapted in order to survive the worst case experiences. Thus, we may describe those communities living along the central European rivers as ‘cultures of flood management’. This special knowledge vanished, however, from the mid-nineteenth century onwards, when river regulations gave the people a false feeling of security.
Resumo:
STUDY DESIGN Technical note and case series. OBJECTIVE To introduce an innovative minimal-invasive surgical procedure reducing surgery time and blood loss in management of U-shaped sacrum fractures. SUMMARY OF BACKGROUND Despite their seldom appearance, U-shaped fractures can cause severe neurological deficits and surgical management difficulties. According to the nature of the injury normally occurring in multi-injured patients after a fall from height, a jump, or road traffic accident, U-shaped fractures create a spinopelvic dissociation and hence are highly unstable. In the past, time-consuming open procedures like large posterior constructs or shortening osteotomies with or without decompression were the method of choice, sacrificing spinal mobility. Insufficient restoration of sacrococcygeal angle and pelvic incidence with conventional techniques may have adverse long-term effects in these patients. METHODS In a consecutive series of 3 patients, percutaneous reduction of the fracture with Schanz pins inserted in either the pedicles of L5 or the S1 body and the posterior superior iliac crest was achieved. The Schanz pins act as lever, allowing a good manipulation of the fracture. The reduction is secured by a temporary external fixator to permit optimal restoration of pelvic incidence and sacral kyphosis. Insertion of 2 transsacral screws allow fixation of the restored spinopelvic alignment. RESULTS Anatomic alignment of the sacrum was possible in each case. Surgery time ranged from 90 to 155 minutes and the blood loss was <50 mL in all 3 cases. Two patients had very good results in the long term regarding maintenance of pelvic incidence and sacrococcygeal angle. One patient with previous cauda equina decompression had loss of correction after 6 months. CONCLUSIONS Percutaneous reduction and transsacral screw fixation offers a less invasive method for treating U-shaped fractures. This can be advantageous in treatment of patients with multiple injuries.
Resumo:
BACKGROUND Microvascular anastomosis is the cornerstone of free tissue transfers. Irrespective of the microsurgical technique that one seeks to integrate or improve, the time commitment in the laboratory is significant. After extensive previous training on several animal models, we sought to identify an animal model that circumvents the following issues: ethical rules, cost, time-consuming and expensive anesthesia, and surgical preparation of tissues required to access vessels before performing the microsurgical training, not to mention that laboratories are closed on weekends. METHODS Between January 2012 and April 2012, a total of 91 earthworms were used for 150 microsurgical training exercises to simulate vascular end-to-side microanastomosis. The training sessions were divided into ten periods of 7 days. Each training session included 15 simulations of end-to-side vascular microanastomoses: larger than 1.5 mm (n=5), between 1.0 and 1.5 mm (n=5), and smaller than 1.0 mm (n=5). A linear model with the main variables being the number of weeks (as a numerical covariate) and the size of the animal (as a factor) was used to determine the trend in time of anastomosis over subsequent weeks as well as the differences between the different size groups. RESULTS The linear model shows a significant trend (p<0.001) in time of anastomosis in the course of the training, as well as significant differences (p<0.001) between the groups of animals of different sizes. For microanastomoses larger than 1.5 mm, the mean anastomosis time decreased from 19.3±1.0 to 11.1±0.4 min between the first and last week of training (decrease of 42.5%). For training with smaller diameters, the results showed a decrease in execution time of 43.2% (diameter between 1.0 and 1.5 mm) and 40.9% (diameter<1.0 mm) between the first and last periods. The study demonstrates an improvement in the dexterity and speed of nodes execution. CONCLUSION The earthworm appears to be a reliable experimental model for microsurgical training of end-to-side microanastomoses. Its numerous advantages are discussed here and we predict training on earthworms will significantly grow and develop in the near future. LEVEL OF EVIDENCE III This journal requires that authors assign a level of evidence to each article. For a full description of these Evidence-Based Medicine ratings, please refer to the Table of Contents or the online Instructions to Authors www.springer.com/00266 .