925 resultados para citation
Resumo:
Traditional logic programming languages, such as Prolog, use a fixed left-to-right atom scheduling rule. Recent logic programming languages, however, usually provide more flexible scheduling in which computation generally proceeds leftto- right but in which some calis are dynamically "delayed" until their arguments are sufRciently instantiated to allow the cali to run efficiently. Such dynamic scheduling has a significant cost. We give a framework for the global analysis of logic programming languages with dynamic scheduling and show that program analysis based on this framework supports optimizations which remove much of the overhead of dynamic scheduling.
Resumo:
While logic programming languages offer a great deal of scope for parallelism, there is usually some overhead associated with the execution of goals in parallel because of the work involved in task creation and scheduling. In practice, therefore, the "granularity" of a goal, i.e. an estimate of the work available under it, should be taken into account when deciding whether or not to execute a goal concurrently as a sepárate task. This paper describes a method for estimating the granularity of a goal at compile time. The runtime overhead associated with our approach is usually quite small, and the performance improvements resulting from the incorporation of grainsize control can be quite good. This is shown by means of experimental results.
Resumo:
This report addresses speculative parallelism (the assignment of spare processing resources to tasks which are not known to be strictly required for the successful completion of a computation) at the user and application level. At this level, the execution of a program is seen as a (dynamic) tree —a graph, in general. A solution for a problem is a traversal of this graph from the initial state to a node known to be the answer. Speculative parallelism then represents the assignment of resources to múltiple branches of this graph even if they are not positively known to be on the path to a solution. In highly non-deterministic programs the branching factor can be very high and a naive assignment will very soon use up all the resources. This report presents work assignment strategies other than the usual depth-first and breadth-first. Instead, best-first strategies are used. Since their definition is application-dependent, the application language contains primitives that allow the user (or application programmer) to a) indícate when intelligent OR-parallelism should be used; b) provide the functions that define "best," and c) indícate when to use them. An abstract architecture enables those primitives to perform the search in a "speculative" way, using several processors, synchronizing them, killing the siblings of the path leading to the answer, etc. The user is freed from worrying about these interactions. Several search strategies are proposed and their implementation issues are addressed. "Armageddon," a global pruning method, is introduced, together with both a software and a hardware implementation for it. The concepts exposed are applicable to áreas of Artificial Intelligence such as extensive expert systems, planning, game playing, and in general to large search problems. The proposed strategies, although showing promise, have not been evaluated by simulation or experimentation.
Resumo:
Abstract is not available
Resumo:
Proof-Carrying Code (PCC) is a general approach to mobile code safety in which the code supplier augments the program with a certifícate (or proof). The intended benefit is that the program consumer can locally validate the certifícate w.r.t. the "untrusted" program by means of a certifícate checker—a process which should be much simpler, eíñcient, and automatic than generating the original proof. Abstraction Carrying Code (ACC) is an enabling technology for PCC in which an abstract model of the program plays the role of certifícate. The generation of the certifícate, Le., the abstraction, is automatically carried out by an abstract interpretation-based analysis engine, which is parametric w.r.t. different abstract domains. While the analyzer on the producer side typically has to compute a semantic fixpoint in a complex, iterative process, on the receiver it is only necessary to check that the certifícate is indeed a fixpoint of the abstract semantics equations representing the program. This is done in a single pass in a much more efficient process. ACC addresses the fundamental issues in PCC and opens the door to the applicability of the large body of frameworks and domains based on abstract interpretation as enabling technology for PCC. We present an overview of ACC and we describe in a tutorial fashion an application to the problem of resource-aware security in mobile code. Essentially the information computed by a cost analyzer is used to genérate cost certificates which attest a safe and efficient use of a mobile code. A receiving side can then reject code which brings cost certificates (which it cannot validate or) which have too large cost requirements in terms of computing resources (in time and/or space) and accept mobile code which meets the established requirements.
Resumo:
The aim of program specialization is to optimize programs by exploiting certain knowledge about the context in which the program will execute. There exist many program manipulation techniques which allow specializing the program in different ways. Among them, one of the best known techniques is partial evaluation, often referred to simply as program specialization, which optimizes programs by specializing them for (partially) known input data. In this work we describe abstract specialization, a technique whose main features are: (1) specialization is performed with respect to "abstract" valúes rather than "concrete" ones, and (2) abstract interpretation rather than standard interpretation of the program is used in order to propágate information about execution states. The concept of abstract specialization is at the heart of the specialization system in CiaoPP, the Ciao system preprocessor. In this paper we present a unifying view of the different specialization techniques used in CiaoPP and discuss their potential applications by means of examples. The applications discussed include program parallelization, optimization of dynamic scheduling (concurreney), and integration of partial evaluation techniques.
Resumo:
An abstract is not available.
Resumo:
A Kuhnian approach to research assessment requires us to consider that the important scientific breakthroughs that drive scientific progress are infrequent and that the progress of science does not depend on normal research. Consequently, indicators of research performance based on the total number of papers do not accurately measure scientific progress. Similarly, those universities with the best reputations in terms of scientific progress differ widely from other universities in terms of the scale of investments made in research and in the higher concentrations of outstanding scientists present, but less so in terms of the total number of papers or citations. This study argues that indicators for the 1% high-citation tail of the citation distribution reveal the contribution of universities to the progress of science and provide quantifiable justification for the large investments in research made by elite research universities. In this tail, which follows a power low, the number of the less frequent and highly cited important breakthroughs can be predicted from the frequencies of papers in the upper part of the tail. This study quantifies the false impression of excellence produced by multinational papers, and by other types of papers that do not contribute to the progress of science. Many of these papers are concentrated in and dominate lists of highly cited papers, especially in lower-ranked universities. The h-index obscures the differences between higher- and lower-ranked universities because the proportion of h-core papers in the 1% high-citation tail is not proportional to the value of the h-index.
Resumo:
The use of barometric altimetry is to some extent a limiting factor on safety, predictability and efficiency of aircraft operations, and reduces the potential of the trajectory based operations capabilities. However, geometric altimetry could be used to improve all of these aspects. Nowadays aircraft altitude is estimated by applying the International Standard Atmosphere which differs from real altitude. At different temperatures for an assigned barometric altitude, aerodynamic forces are different and this has a direct relationship with time, fuel consumption and range of the flight. The study explores the feasibility of using sensors providing geometric reference altitude, in particular, to supply capabilities for the optimization of vertical profiles and also, their impact on the vertical Air Traffic Management separation assurance processes. One of the aims of the thesis is to assess if geometric altitude fulfils the aeronautical requirements through existing sensors. Also the thesis will elaborate on the advantages of geometric altitude over the barometric altitude in terms of efficiency for vertical navigation. The evidence that geometric altitude is the best choice to improve the efficiency in vertical profile and aircraft capacity by reducing vertical uncertainties will also be shown. In this paper, an atmospheric study is presented, as well as the impact of temperature deviation from International Standard Atmosphere model is analyzed in order to obtain relationship between geometric and barometric altitude. Furthermore, an aircraft model to study aircraft vertical profile is provided to analyse trajectories based on geometric altitudes.
Resumo:
This paper analyzes the relationship among research collaboration, number of documents and number of citations of computer science research activity. It analyzes the number of documents and citations and how they vary by number of authors. They are also analyzed (according to author set cardinality) under different circumstances, that is, when documents are written in different types of collaboration, when documents are published in different document types, when documents are published in different computer science subdisciplines, and, finally, when documents are published by journals with different impact factor quartiles. To investigate the above relationships, this paper analyzes the publications listed in the Web of Science and produced by active Spanish university professors between 2000 and 2009, working in the computer science field. Analyzing all documents, we show that the highest percentage of documents are published by three authors, whereas single-authored documents account for the lowest percentage. By number of citations, there is no positive association between the author cardinality and citation impact. Statistical tests show that documents written by two authors receive more citations per document and year than documents published by more authors. In contrast, results do not show statistically significant differences between documents published by two authors and one author. The research findings suggest that international collaboration results on average in publications with higher citation rates than national and institutional collaborations. We also find differences regarding citation rates between journals and conferences, across different computer science subdisciplines and journal quartiles as expected. Finally, our impression is that the collaborative level (number of authors per document) will increase in the coming years, and documents published by three or four authors will be the trend in computer science literature.
Resumo:
Wireless sensor networks (WSNs) appeal to a wide range of applications that involve the monitoring of various physical phenomena. However, WSNs are subject to many threats. In particular, lack of pervasive tamper-resistant hardware results in sensors being easy targets for compromise. Having compromised a sensor, the adversary learns all the sensor secrets, allowing it to later encrypt/decrypt or authenticate messages on behalf of that sensor. This threat is particularly relevant in the novel unattended wireless sensor networks (UWSNs) scenario. UWSNs operate without constant supervision by a trusted sink. UWSN?s unattended nature and increased exposure to attacks prompts the need for special techniques geared towards regaining security after being compromised. In this article, we investigate cooperative self-healing in UWSNs and propose various techniques to allow unattended sensors to recover security after compromise. Our techniques provide seamless healing rates even against a very agile and powerful adversary. The effectiveness and viability of our proposed techniques are assessed by thorough analysis and supported by simulation results. Finally, we introduce some real-world issues affecting UWSN deployment and provide some solutions for them as well as a few open problems calling for further investigation.
Resumo:
Aiming to address requirements concerning integration of services in the context of ?big data?, this paper presents an innovative approach that (i) ensures a flexible, adaptable and scalable information and computation infrastructure, and (ii) exploits the competences of stakeholders and information workers to meaningfully confront information management issues such as information characterization, classification and interpretation, thus incorporating the underlying collective intelligence. Our approach pays much attention to the issues of usability and ease-of-use, not requiring any particular programming expertise from the end users. We report on a series of technical issues concerning the desired flexibility of the proposed integration framework and we provide related recommendations to developers of such solutions. Evaluation results are also discussed.
Resumo:
In this paper we investigate the effect of biasing the axonal connection delay values in the number of polychronous groups produced for a spiking neuron network model. We use an estimation of distribution algorithm (EDA) that learns tree models to search for optimal delay configurations. Our results indicate that the introduced approach can be used to considerably increase the number of such groups.
Resumo:
This paper describes the main goals and outcomes of the EU-funded Framework 7 project entitled Semantic Evaluation at Large Scale (SEALS). The growth and success of the Semantic Web is built upon a wide range of Semantic technologies from ontology engineering tools through to semantic web service discovery and semantic search. The evaluation of such technologies ? and, indeed, assessments of their mutual compatibility ? is critical for their sustained improvement and adoption. The SEALS project is creating an open and sustainable platform on which all aspects of an evaluation can be hosted and executed and has been designed to accommodate most technology types. It is envisaged that the platform will become the de facto repository of test datasets and will allow anyone to organise, execute and store the results of technology evaluations free of charge and without corporate bias. The demonstration will show how individual tools can be prepared for evaluation, uploaded to the platform, evaluated according to some criteria and the subsequent results viewed. In addition, the demonstration will show the flexibility and power of the SEALS Platform for evaluation organisers by highlighting some of the key technologies used.
Resumo:
En esta tesis se investiga de forma experimental el transporte pasivo de magnitudes físicas en micro-sistemas con carácter de inmediata aplicación industrial, usando métodos innovadores para mejorar la eficiencia de los mismos optimizando parámetros críticos del diseño o encontrar nuevos destinos de posible aplicación. Parte de los resultados obtenidos en estos experimentos han sido publicados en revistas con un índice de impacto tal que pertenecen al primer cuarto del JCR. Primero de todo se ha analizado el efecto que produce en un intercambiador de calor basado en micro-canales el hecho de dejar un espacio entre canales y tapa superior para la interconexión de los mismos. Esto genera efectos tridimensionales que mejoran la exracción de calor del intercambiador y reducen la caída de presión que aparece por el transcurso del fluido a través de los micro-canales, lo que tiene un gran impacto en la potencia que ha de suministrar la bomba de refrigerante. Se ha analizado también la mejora producida en términos de calor disipado de un micro-procesador refrigerado con un ampliamente usado plato de aletas al implementar en éste una cámara de vapor que almacena un fluido bifásico. Se ha desarrollado de forma paralela un modelo numérico para optimizar las nuevas dimensiones del plato de aletas modificado compatibles con una serie de requerimientos de diseño en el que tanto las dimensiones como el peso juegan un papel esencial. Por otro lado, se han estudiado los fenomenos fluido-dinámicos que aparecen aguas abajo de un cuerpo romo en el seno de un fluido fluyendo por un canal con una alta relación de bloqueo. Los resultados de este estudio confirman, de forma experimental, la existencia de un régimen intermedio, caracterizado por el desarrollo de una burbuja de recirculación oscilante entre los regímenes, bien diferenciados, de burbuja de recirculación estacionaria y calle de torbellinos de Karman, como función del número de Reynolds del flujo incidente. Para la obtención, análisis y post-proceso de los datos, se ha contado con la ayuda de un sistema de Velocimetría por Imágenes de Partículas (PIV). Finalmente y como adición a este último punto, se ha estudiado las vibraciones de un cuerpo romo producidas por el desprendimiento de torbellinos en un canal de alta relación de bloqueo con la base obtenida del estudio anterior. El prisma se mueve con un movimiento armónico simple para un intervalo de números de Reynolds y este movimiento se transforma en vibración alrededor de su eje a partir de un ciero número de Reynolds. En relación al fluido, el régimen de desprendimiento de torbellinos se alcanza a menores números de Reynolds que en el caso de tener el cuerpo romo fijo. Uniendo estos dos registros de movimientos y variando la relación de masas entre prisma y fluido se obtiene un mapa con diferentes estados globales del sistema. Esto no solo tiene aplicación como método para promover el mezclado sino también como método para obtener energía a partir del movimiento del cuerpo en el seno del fluido. Abstract In this thesis, experimental research focused on passive scalar transport is performed in micro-systems with marked sense of industrial application, using innovative methods in order to obtain better performances optimizing critical design parameters or finding new utilities. Part of the results obtained in these experiments have been published into high impact factor journals belonged to the first quarter of the Journal Citation Reports (JCR). First of all the effect of tip clearance in a micro-channel based heat sink is analyzed. Leaving a gap between channels and top cover, letting the channels communicate each other causes three-dimensional effects which improve the heat transfer between fluid and heat sink and also reducing the pressure drop caused by the fluid passing through the micro-channels which has a great impact on the total cooling pumping power needed. It is also analyzed the enhancement produced in terms of dissipated heat in a micro-processor cooling system by improving the predominantly used fin plate with a vapour chamber based heat spreader which contains a two-phase fluid inside. It has also been developed at the same time a numerical model to optimize the new fin plate dimensions compatible with a series of design requirements in which both size and wight plays a very restrictive role. On the other hand, fluid-dynamics phenomena that appears downstream of a bluff body in the bosom of a fluid flow with high blockage ratio has been studied. This research experimentally confirms the existence of an intermediate regime characterized by an oscillating closed recirculation bubble intermediate regime between the steady closed recirculation bubble regime and the vortex shedding regime (Karman street like regime) as a function of the incoming flow Reynolds number. A particle image velocimetry technique (PIV) has been used in order to obtain, analyze and post-process the fluid-dynamic data. Finally and as an addition to the last point, a study on the vortexinduced vibrations (VIV) of a bluff body inside a high blockage ratio channel has been carried out taking advantage of the results obtained with the fixed square prism. The prism moves with simple harmonic motion for a Reynolds number interval and this movement becomes vibrational around its axial axis after overcoming at definite Reynolds number. Regarding the fluid, vortex shedding regime is reached at Reynolds numbers lower than the previous critical ones. Merging both movement spectra and varying the square prism to fluid mass ratio, a map with different global states is reached. This is not only applicable as a mixing enhancement technique but as an energy harvesting method.