743 resultados para correctional programs


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Abstract interpretation-based data-flow analysis of logic programs is at this point relatively well understood from the point of view of general frameworks and abstract domains. On the other hand, comparatively little attention has been given to the problems which arise when analysis of a full, practical dialect of the Prolog language is attempted, and only few solutions to these problems have been proposed to date. Such problems relate to dealing correctly with all builtins, including meta-logical and extra-logical predicates, with dynamic predicates (where the program is modified during execution), and with the absence of certain program text during compilation. Existing proposals for dealing with such issues generally restrict in one way or another the classes of programs which can be analyzed if the information from analysis is to be used for program optimization. This paper attempts to fill this gap by considering a full dialect of Prolog, essentially following the recently proposed ISO standard, pointing out the problems that may arise in the analysis of such a dialect, and proposing a combination of known and novel solutions that together allow the correct analysis of arbitrary programs using the full power of the language.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Global analyzers traditionally read and analyze the entire program at once, in a non-incremental way. However, there are many situations which are not well suited to this simple model and which instead require reanalysis of certain parts of a program which has already been analyzed. In these cases, it appears inefficient to perform the analysis of the program again from scratch, as needs to be done with current systems. We describe how the fixpoint algorithms in current generic analysis engines can be extended to support incremental analysis. The possible changes to a program are classified into three types: addition, deletion, and arbitrary change. For each one of these, we provide one or more algorithms for identifying the parts of the analysis that must be recomputed and for performing the actual recomputation. The potential benefits and drawbacks of these algorithms are discussed. Finally, we present some experimental results obtained with an implementation of the algorithms in the PLAI generic abstract interpretation framework. The results show significant benefits when using the proposed incremental analysis algorithms.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper, we examine the issue of memory management in the parallel execution of logic programs. We concentrate on non-deterministic and-parallel schemes which we believe present a relatively general set of problems to be solved, including most of those encountered in the memory management of or-parallel systems. We present a distributed stack memory management model which allows flexible scheduling of goals. Previously proposed models (based on the "Marker model") are lacking in that they impose restrictions on the selection of goals to be executed or they may require consume a large amount of virtual memory. This paper first presents results which imply that the above mentioned shortcomings can have significant performance impacts. An extension of the Marker Model is then proposed which allows flexible scheduling of goals while keeping (virtual) memory consumption down. Measurements are presented which show the advantage of this solution. Methods for handling forward and backward execution, cut and roll back are discussed in the context of the proposed scheme. In addition, the paper shows how the same mechanism for flexible scheduling can be applied to allow the efficient handling of the very general form of suspension that can occur in systems which combine several types of and-parallelism and more sophisticated methods of executing logic programs. We believe that the results are applicable to many and- and or-parallel systems.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Knowing the size of the terms to which program variables are bound at run-time in logic programs is required in a class of applications related to program optimization such as, for example, recursion elimination and granularity analysis. Such size is difficult to even approximate at compile time and is thus generally computed at run-time by using (possibly predefined) predicates which traverse the terms involved. We propose a technique based on program transformation which has the potential of performing this computation much more efficiently. The technique is based on finding program procedures which are called before those in which knowledge regarding term sizes is needed and which traverse the terms whose size is to be determined, and transforming such procedures so that they compute term sizes "on the fly". We present a systematic way of determining whether a given program can be transformed in order to compute a given term size at a given program point without additional term traversal. Also, if several such transformations are possible our approach allows finding minimal transformations under certain criteria. We also discuss the advantages and present some applications of our technique.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Bruynooghe described a framework for the top-down abstract interpretation of logic programs. In this framework, abstract interpretation is carried out by constructing an abstract and-or tree in a top-down fashion for a given query and program. Such an abstract interpreter requires fixpoint computation for programs which contain recursive predicates. This paper presents in detail a fixpoint algorithm that has been developed for this purpose and the motivation behind it. We start off by describing a simple-minded algorithm. After pointing out its shortcomings, we present a series of refinements to this algorithm, until we reach the final version. The aim is to give an intuitive grasp and provide justification for the relative complexity of the final algorithm. We also present an informal proof of correctness of the algorithm and some results obtained from an implementation.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Este trabajo aborda la metodología seguida para llevar a cabo el proyecto de investigación PRONAF (Clinical Trials Gov.: number NCT01116856.) Background: At present, scientific consensus exists on the multifactorial etiopatogenia of obesity. Both professionals and researchers agree that treatment must also have a multifactorial approach, including diet, physical activity, pharmacology and/or surgical treatment. These two last ones should be reserved for those cases of morbid obesities or in case of failure of the previous ones. The aim of the PRONAF study is to determine what type of exercise combined with caloric restriction is the most appropriate to be included in overweigth and obesity intervention programs, and the aim of this paper is to describe the design and the evaluation methods used to carry out the PRONAF study. Methods/design: One-hundred nineteen overweight (46 males) and 120 obese (61 males) subjects aged 18–50 years were randomly assigned to a strength training group, an endurance training group, a combined strength + endurance training group or a diet and physical activity recommendations group. The intervention period was 22 weeks (in all cases 3 times/wk of training for 22 weeks and 2 weeks for pre and post evaluation). All subjects followed a hypocaloric diet (25-30% less energy intake than the daily energy expenditure estimated by accelerometry). 29–34% of the total energy intake came from fat, 14–20% from protein, and 50–55% from carbohydrates. The mayor outcome variables assesed were, biochemical and inflamatory markers, body composition, energy balance, physical fitness, nutritional habits, genetic profile and quality of life. 180 (75.3%) subjects finished the study, with a dropout rate of 24.7%. Dropout reasons included: personal reasons 17 (28.8%), low adherence to exercise 3 (5.1%), low adherence to diet 6 (10.2%), job change 6 (10.2%), and lost interest 27 (45.8%). Discussion: Feasibility of the study has been proven, with a low dropout rate which corresponds to the estimated sample size. Transfer of knowledge is foreseen as a spin-off, in order that overweight and obese subjects can benefit from the results. The aim is to transfer it to sports centres. Effectiveness on individual health-related parameter in order to determine the most effective training programme will be analysed in forthcoming publications.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Background and aim: Many exercise studies, although generally showing the beneficial effects of supervised aerobic, resistance or combined exercise on blood lipids, have sometimes reached equivocal conclusions. The aim of this study is to evaluate the impact of different programs that combined exercise and dietary restriction on blood lipids versus a clinical practice intervention for weight loss, in overweight adults. Methods: For this study 66 subjects participated in a supervised 22 weeks training program, composed of three sessions per week and they were randomized in three groups: strength training (S; n = 19), endurance training (E; n = 25), a combination of E and S (SE; n = 22). Eighteen subjects served as physical activity group (PA) that followed a clinical intervention consisted of physical activity recommendations. All groups followed the same dietary treatment, and blood samples were obtained for lipids measurements, at the beginning and end of the study. Results: Lipid profile improved in all groups. No significant differences for baseline and post-training values were observed between groups. In general, SE and PA decreased low-density lipoprotein cholesterol (LDL-C) values (p menor que 0.01). S decreased triglyceride levels (p menor que 0.01) and E, SE, and PA decreased total cholesterol levels (p menor que 0.05, p menor que 0.01 and p menor que 0.01, respectively). Conclusions: These results suggest that an intervention program of supervised exercise combined with diet restriction did not achieved further improvements in blood lipid profile than diet restriction and physical activity recommendations, in overweight adults. (Clinical Trials gov number: NCT01116856).

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Correlación entre adherencia a la dieta y el ejercicio en programas de perdida de peso en función del sexo

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The characteristics of CC and CLP systems are in principle very dierent However a recent trend towards convergence in the implementation techniques for these systems can be observed While CLP and Prolog systems have been incorporating capabilities to deal with userdened suspension and coroutining CC compilers have been trying to coalesce negrained tasks into coarsergrained sequential threads This convergence of techniques opens up the possibility of having a general purpose kernel language and abstract machine to serve as a compilation target for a variety of userlevel languages We propose a transformation technique directed towards such an objective In particular we report on techniques to support the Andorra computational model essentially emulating the AndorraI system via program transformation into a sequential language with delay primitives The system is automatic comprising an optional program analyzer and a basic transformer to the kernel language It turns out that a simple parallel CLP or Prolog system with dynamic scheduling is sucient as a kernel language for this purpose The preliminary results are quite encouraging performance of the resulting system is comparable to the current AndorraI implementation.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

La seguridad verificada es una metodología para demostrar propiedades de seguridad de los sistemas informáticos que se destaca por las altas garantías de corrección que provee. Los sistemas informáticos se modelan como programas probabilísticos y para probar que verifican una determinada propiedad de seguridad se utilizan técnicas rigurosas basadas en modelos matemáticos de los programas. En particular, la seguridad verificada promueve el uso de demostradores de teoremas interactivos o automáticos para construir demostraciones completamente formales cuya corrección es certificada mecánicamente (por ordenador). La seguridad verificada demostró ser una técnica muy efectiva para razonar sobre diversas nociones de seguridad en el área de criptografía. Sin embargo, no ha podido cubrir un importante conjunto de nociones de seguridad “aproximada”. La característica distintiva de estas nociones de seguridad es que se expresan como una condición de “similitud” entre las distribuciones de salida de dos programas probabilísticos y esta similitud se cuantifica usando alguna noción de distancia entre distribuciones de probabilidad. Este conjunto incluye destacadas nociones de seguridad de diversas áreas como la minería de datos privados, el análisis de flujo de información y la criptografía. Ejemplos representativos de estas nociones de seguridad son la indiferenciabilidad, que permite reemplazar un componente idealizado de un sistema por una implementación concreta (sin alterar significativamente sus propiedades de seguridad), o la privacidad diferencial, una noción de privacidad que ha recibido mucha atención en los últimos años y tiene como objetivo evitar la publicación datos confidenciales en la minería de datos. La falta de técnicas rigurosas que permitan verificar formalmente este tipo de propiedades constituye un notable problema abierto que tiene que ser abordado. En esta tesis introducimos varias lógicas de programa quantitativas para razonar sobre esta clase de propiedades de seguridad. Nuestra principal contribución teórica es una versión quantitativa de una lógica de Hoare relacional para programas probabilísticos. Las pruebas de correción de estas lógicas son completamente formalizadas en el asistente de pruebas Coq. Desarrollamos, además, una herramienta para razonar sobre propiedades de programas a través de estas lógicas extendiendo CertiCrypt, un framework para verificar pruebas de criptografía en Coq. Confirmamos la efectividad y aplicabilidad de nuestra metodología construyendo pruebas certificadas por ordendor de varios sistemas cuyo análisis estaba fuera del alcance de la seguridad verificada. Esto incluye, entre otros, una meta-construcción para diseñar funciones de hash “seguras” sobre curvas elípticas y algoritmos diferencialmente privados para varios problemas de optimización combinatoria de la literatura reciente. ABSTRACT The verified security methodology is an emerging approach to build high assurance proofs about security properties of computer systems. Computer systems are modeled as probabilistic programs and one relies on rigorous program semantics techniques to prove that they comply with a given security goal. In particular, it advocates the use of interactive theorem provers or automated provers to build fully formal machine-checked versions of these security proofs. The verified security methodology has proved successful in modeling and reasoning about several standard security notions in the area of cryptography. However, it has fallen short of covering an important class of approximate, quantitative security notions. The distinguishing characteristic of this class of security notions is that they are stated as a “similarity” condition between the output distributions of two probabilistic programs, and this similarity is quantified using some notion of distance between probability distributions. This class comprises prominent security notions from multiple areas such as private data analysis, information flow analysis and cryptography. These include, for instance, indifferentiability, which enables securely replacing an idealized component of system with a concrete implementation, and differential privacy, a notion of privacy-preserving data mining that has received a great deal of attention in the last few years. The lack of rigorous techniques for verifying these properties is thus an important problem that needs to be addressed. In this dissertation we introduce several quantitative program logics to reason about this class of security notions. Our main theoretical contribution is, in particular, a quantitative variant of a full-fledged relational Hoare logic for probabilistic programs. The soundness of these logics is fully formalized in the Coq proof-assistant and tool support is also available through an extension of CertiCrypt, a framework to verify cryptographic proofs in Coq. We validate the applicability of our approach by building fully machine-checked proofs for several systems that were out of the reach of the verified security methodology. These comprise, among others, a construction to build “safe” hash functions into elliptic curves and differentially private algorithms for several combinatorial optimization problems from the recent literature.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Evaluation of three solar and daylighting control systems based on Calumen II, Ecotect and Radiance simulation programs to obtain an energy efficient and healthy interior in the experimental building prototype SDE10

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We present a novel general resource analysis for logic programs based on sized types.Sized types are representations that incorporate structural (shape) information and allow expressing both lower and upper bounds on the size of a set of terms and their subterms at any position and depth. They also allow relating the sizes of terms and subterms occurring at different argument positions in logic predicates. Using these sized types, the resource analysis can infer both lower and upper bounds on the resources used by all the procedures in a program as functions on input term (and subterm) sizes, overcoming limitations of existing analyses and enhancing their precision. Our new resource analysis has been developed within the abstract interpretation framework, as an extension of the sized types abstract domain, and has been integrated into the Ciao preprocessor, CiaoPP. The abstract domain operations are integrated with the setting up and solving of recurrence equations for both, inferring size and resource usage functions. We show that the analysis is an improvement over the previous resource analysis present in CiaoPP and compares well in power to state of the art systems.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We present a novel analysis for relating the sizes of terms and subterms occurring at diferent argument positions in logic predicates. We extend and enrich the concept of sized type as a representation that incorporates structural (shape) information and allows expressing both lower and upper bounds on the size of a set of terms and their subterms at any position and depth. For example, expressing bounds on the length of lists of numbers, together with bounds on the values of all of their elements. The analysis is developed using abstract interpretation and the novel abstract operations are based on setting up and solving recurrence relations between sized types. It has been integrated, together with novel resource usage and cardinality analyses, in the abstract interpretation framework in the Ciao preprocessor, CiaoPP, in order to assess both the accuracy of the new size analysis and its usefulness in the resource usage estimation application. We show that the proposed sized types are a substantial improvement over the previous size analyses present in CiaoPP, and also benefit the resource analysis considerably, allowing the inference of equal or better bounds than comparable state of the art systems.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

European Universities are involved in series of great changes regarding teaching and education organization during the last few years. The origin of these changes is the creation of the so-called European Higher Education Area (EHEA), which main target is to harmonize the different University studies throughout Europe. As a consequence, most of the programs of studies in all degrees are suffering changes in order to converge to common structures. Taking advantage of the actual process, some European universities are moving from traditional Agricultural Engineering programs to a more wide discipline named recently as Biosystems Engineering, which is a science- based engineering discipline that integrates engineering science and design with applied biological, environmental and agricultural sciences, broadening in this way the area of application of Engineering sciences not strictly to agricultural sciences, but to the biologic al sciences in general, including the agricultural sciences. This paper presents a comparative study of different Bachelor of Science degrees offered by American and European Universities in the field of Agricultural/Biosystems Engineering. To carry out the analysis 40 programs accredited by ABET in American Universities and 50 European programs. Among other questions, the total number of credits, the number of semesters, the kind of modules and the distribution of subjects in groups (Basic Sciences, Engineering Fundamentals, Agricultural/Biological Sciences, Humanities & Economic Sciences, Applied Agricultural/Biological Engineering and electives) are discussed in the paper. The information provided can be an useful starting point in future definitions of new or renewed degrees with the aim of advancing in internationalization of the programs and helping student’s mobility.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We present a novel general resource analysis for logic programs based on sized types. Sized types are representations that incorporate structural (shape) information and allow expressing both lower and upper bounds on the size of a set of terms and their subterms at any position and depth. They also allow relating the sizes of terms and subterms occurring at different argument positions in logic predicates. Using these sized types, the resource analysis can infer both lower and upper bounds on the resources used by all the procedures in a program as functions on input term (and subterm) sizes, overcoming limitations of existing resource analyses and enhancing their precision. Our new resource analysis has been developed within the abstract interpretation framework, as an extension of the sized types abstract domain, and has been integrated into the Ciao preprocessor, CiaoPP. The abstract domain operations are integrated with the setting up and solving of recurrence equations for inferring both size and resource usage functions. We show that the analysis is an improvement over the previous resource analysis present in CiaoPP and compares well in power to state of the art systems.