991 resultados para Compiler tuning automatization process
Resumo:
A rápida evolução do hardware demanda uma evolução contínua dos compiladores. Um processo de ajuste deve ser realizado pelos projetistas de compiladores para garantir que o código gerado pelo compilador mantenha uma determinada qualidade, seja em termos de tempo de processamento ou outra característica pré-definida. Este trabalho visou automatizar o processo de ajuste de compiladores por meio de técnicas de aprendizado de máquina. Como resultado os planos de compilação obtidos usando aprendizado de máquina com as características propostas produziram código para programas cujos valores para os tempos de execução se aproximaram daqueles seguindo o plano padrão utilizado pela LLVM.
Resumo:
Modern compilers present a great and ever increasing number of options which can modify the features and behavior of a compiled program. Many of these options are often wasted due to the required comprehensive knowledge about both the underlying architecture and the internal processes of the compiler. In this context, it is usual, not having a single design goal but a more complex set of objectives. In addition, the dependencies between different goals are difficult to be a priori inferred. This paper proposes a strategy for tuning the compilation of any given application. This is accomplished by using an automatic variation of the compilation options by means of multi-objective optimization and evolutionary computation commanded by the NSGA-II algorithm. This allows finding compilation options that simultaneously optimize different objectives. The advantages of our proposal are illustrated by means of a case study based on the well-known Apache web server. Our strategy has demonstrated an ability to find improvements up to 7.5% and up to 27% in context switches and L2 cache misses, respectively, and also discovers the most important bottlenecks involved in the application performance.
Resumo:
Aims. In a recent measurement, Meléndez & Barbuy (2009, A&A, 497, 611) report accurate log gf values for 142 important astrophysical lines with wavelengths in the range 4000 Å to 8000 Å. Their results include both solar and laboratory measurements. In this paper, we describe a theoretical study of these lines. Methods. The CIV3 structure codes, combined with our "fine-tuning" extrapolation process, are used to undertake a large-scale CI calculation involving the lowest 262 fine-structure levels belonging to the 3d4s, 3d, 3d4s, 3d4p, and 3d4s4p configurations. Results. We find that many of the 142 transitions are very weak intercombination lines. Other transitions are weak because the dominant configurations in the two levels differ by two orbitals. Conclusions. The comparison between our log gf values and the experimental values generally shows good agreement for most of these transitions, with our theoretical values agreeing slightly more closely with the solar than with the laboratory measurements. A detailed analysis of the small number of transitions for which the agreement between theory and experiment is not as good shows that such disagreements largely arise from severe cancellation due to CI mixing.
Resumo:
Este artículo describe una estrategia de selección de frases para hacer el ajuste de un sistema de traducción estadístico basado en el decodificador Moses que traduce del español al inglés. En este trabajo proponemos dos posibilidades para realizar esta selección de las frases del corpus de validación que más se parecen a las frases que queremos traducir (frases de test en lengua origen). Con esta selección podemos obtener unos mejores pesos de los modelos para emplearlos después en el proceso de traducción y, por tanto, mejorar los resultados. Concretamente, con el método de selección basado en la medida de similitud propuesta en este artículo, mejoramos la medida BLEU del 27,17% con el corpus de validación completo al 27,27% seleccionando las frases para el ajuste. Estos resultados se acercan a los del experimento ORACLE: se utilizan las mismas frases de test para hacer el ajuste de los pesos. En este caso, el BLEU obtenido es de 27,51%.
Resumo:
Tuning compilations is the process of adjusting the values of a compiler options to improve some features of the final application. In this paper, a strategy based on the use of a genetic algorithm and a multi-objective scheme is proposed to deal with this task. Unlike previous works, we try to take advantage of the knowledge of this domain to provide a problem-specific genetic operation that improves both the speed of convergence and the quality of the results. The evaluation of the strategy is carried out by means of a case of study aimed to improve the performance of the well-known web server Apache. Experimental results show that a 7.5% of overall improvement can be achieved. Furthermore, the adaptive approach has shown an ability to markedly speed-up the convergence of the original strategy.
Resumo:
Thesis (M.A.)--University of Illinois at Urbana-Champaign.
Resumo:
Mode of access: Internet.
Resumo:
This study explored how Korean men married to migrant women construct meaning around married life. Data were collected through in-depth interviews with 10 men who had had been married to migrant women for ≥ 2 years. Data collection and analysis were performed concurrently using a grounded theory approach. The core category generated was the process of sustaining a family unit. The men came to understand the importance of a distribution of power within the family in sustaining the family unit. Constituting this process were four stages: recognizing an imbalance of power, relinquishing power, empowering, and fine-tuning the balance of power. This study provides important insight into the dynamics of marital power from men's point of view by demonstrating a link between the way people adjust to married life and the process by which married couples adjust through the distribution and redistribution of power.
Resumo:
A simple, uniquely plasma-enabled and environment-friendly process to reduce the thickness of vertically standing graphenes to only 4–5 graphene layers and arranging them in dense, ultra-large surface area, ultra-open-edge-length, self-organized and interconnected networks is demonstrated. The approach for the ultimate thickness reduction to 1–2 graphene layers is also proposed. The vertical graphene networks are optically transparent and show tunable electric properties from semiconducting to semi-metallic and metallic at room and near-room temperature, thus recovering semi-metallic properties of a single-layer graphene.
Resumo:
Nowadays, integration of small-scale electricity generators, known as Distributed Generation (DG), into distribution networks has become increasingly popular. This tendency together with the falling price of DG units has a great potential in giving the DG a better chance to participate in voltage regulation process, in parallel with other regulating devices already available in the distribution systems. The voltage control issue turns out to be a very challenging problem for distribution engineers, since existing control coordination schemes need to be reconsidered to take into account the DG operation. In this paper, a control coordination approach is proposed, which is able to utilize the ability of the DG as a voltage regulator, and at the same time minimize the interaction of DG with another DG or other active devices, such as On-load Tap Changing Transformer (OLTC). The proposed technique has been developed based on the concepts of protection principles (magnitude grading and time grading) for response coordination of DG and other regulating devices and uses Advanced Line Drop Compensators (ALDCs) for implementation. A distribution feeder with tap changing transformer and DG units has been extracted from a practical system to test the proposed control technique. The results show that the proposed method provides an effective solution for coordination of DG with another DG or voltage regulating devices and the integration of protection principles has considerably reduced the control interaction to achieve the desired voltage correction.
Resumo:
Owing to widespread applications, synthesis and characterization of silver nanoparticles is recently attracting considerable attention. Increasing environmental concerns over chemical synthesis routes have resulted in attempts to develop biomimetic approaches. One of them is synthesis using plant parts, which eliminates the elaborate process of maintaining the microbial culture and often found to be kinetically favourable than other bioprocesses. The present study deals with investigating the effect of process variables like reductant concentrations, reaction pH, mixing ratio of the reactants and interaction time on the morphology and size of silver nanoparticles synthesized using aqueous extract of Azadirachta indica (Neem) leaves. The formation of crystalline silver nanoparticles was confirmed using X-ray diffraction analysis. By means of UV spectroscopy, Scanning and Transmission Electron Microscopy techniques, it was observed that the morphology and size of the nanoparticles were strongly dependent on the process parameters. Within 4 h interaction period, nanoparticles below 20-nm-size with nearly spherical shape were produced. On increasing interaction time (ageing) to 66 days, both aggregation and shape anisotropy (ellipsoidal, polyhedral and capsular) of the particles increased. In alkaline pH range, the stability of cluster distribution increased with a declined tendency for aggregation of the particles. It can be inferred from the study that fine tuning the bioprocess parameters will enhance possibilities of desired nano-product tailor made for particular applications.
Resumo:
We describe a compiler for the Flat Concurrent Prolog language on a message passing multiprocessor architecture. This compiler permits symbolic and declarative programming in the syntax of Guarded Horn Rules, The implementation has been verified and tested on the 64-node PARAM parallel computer developed by C-DAC (Centre for the Development of Advanced Computing, India), Flat Concurrent Prolog (FCP) is a logic programming language designed for concurrent programming and parallel execution, It is a process oriented language, which embodies dataflow synchronization and guarded-command as its basic control mechanisms. An identical algorithm is executed on every processor in the network, We assume regular network topologies like mesh, ring, etc, Each node has a local memory, The algorithm comprises of two important parts: reduction and communication, The most difficult task is to integrate the solutions of problems that arise in the implementation in a coherent and efficient manner. We have tested the efficacy of the compiler on various benchmark problems of the ICOT project that have been reported in the recent book by Evan Tick, These problems include Quicksort, 8-queens, and Prime Number Generation, The results of the preliminary tests are favourable, We are currently examining issues like indexing and load balancing to further optimize our compiler.
Resumo:
A series of deoxycholic and cholic acid-derived oligomers were synthesized and their ability to extract hydrophilic dye molecules of different structure, size, and functional groups into nonpolar media was studied. The structure of the dye and dendritic effect in the extraction process was examined using absorption spectroscopy and dynamic light scattering (DLS). The efficiency of structurally preorganized oligomers in the aggregation process was evaluated by 1-anilinonaphthalene-8-sulfonic acid (ANS) fluorescence studies. The possible formation of globular structures for higher-generation molecules was investigated by molecular modeling studies and the results were correlated with the anomaly observed in the extraction process with this molecule. The ability of these molecules for selective extraction of specific dyes from blended colors is also reported.
Resumo:
Epitaxial La0.5Sr0.5CoO3 (LSCO) thin films are grown on LaAlO3 (100) and SrTiO3 (100) substrates by pulsed laser ablation. By tuning the growth parameters, we are able to enhance the ferromagnetic transition temperature (T-C) up to 262 K as evident from ac susceptibility, dc magnetization, and resistivity measurements. The magnitude of T-C is the same as that for the bulk stoichiometric LSCO illustrating the high quality of the grown films. Detailed structural analysis clearly reveals that the induced strain in the LSCO film has no role behind this enhancement; in fact, the determining factor is the oxygen stoichiometry. The films also exhibit ageing effect as the T-C decreases with time. This is considered in terms of gradual change in the oxygen stoichiometry through diffusion process as the time progresses. (C) 2013 AIP Publishing LLC.
Resumo:
This paper presents the design and implementation of PolyMage, a domain-specific language and compiler for image processing pipelines. An image processing pipeline can be viewed as a graph of interconnected stages which process images successively. Each stage typically performs one of point-wise, stencil, reduction or data-dependent operations on image pixels. Individual stages in a pipeline typically exhibit abundant data parallelism that can be exploited with relative ease. However, the stages also require high memory bandwidth preventing effective utilization of parallelism available on modern architectures. For applications that demand high performance, the traditional options are to use optimized libraries like OpenCV or to optimize manually. While using libraries precludes optimization across library routines, manual optimization accounting for both parallelism and locality is very tedious. The focus of our system, PolyMage, is on automatically generating high-performance implementations of image processing pipelines expressed in a high-level declarative language. Our optimization approach primarily relies on the transformation and code generation capabilities of the polyhedral compiler framework. To the best of our knowledge, this is the first model-driven compiler for image processing pipelines that performs complex fusion, tiling, and storage optimization automatically. Experimental results on a modern multicore system show that the performance achieved by our automatic approach is up to 1.81x better than that achieved through manual tuning in Halide, a state-of-the-art language and compiler for image processing pipelines. For a camera raw image processing pipeline, our performance is comparable to that of a hand-tuned implementation.