947 resultados para 005 Computer programming, programs


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Community-based participatory research necessitates that community members act as partners in decision making and mutual learning and discovery. In the same light, for programs/issues involving youth, youth should be partners in knowledge sharing and evaluation (Checkoway & Richards-Schuster, 2004). This study is a youth-focused empowerment evaluation for the Successful Youth program. Successful Youth is a multi-component youth development after-school program for Latino middle school youth, created with the goal of reducing teen pregnancy. An empowerment evaluation is collaborative and participatory (Balcazar and Harper 2003). The three steps of an empowerment evaluation are: (1) defining mission, (2) taking stock, and (3) planning for the future (Fetterman 2001).^ In a program where youth are developing leadership skills, making choices, and learning how to self reflect and evaluate, the empowerment evaluation could not be more aligned with promoting and enhancing these skills. In addition, an empowerment evaluation is designed to "foster improvement and self-determination" and "build capacity" (Fetterman 2001). Four empowerment groups were conducted with approximately 6-9 Latino 7th grade students per group. All participants were enrolled in the Successful Youth program. Results indicate points where students' perceptions of the program were aligned with the program's mission and where gaps were identified. Students offered recommendations for program improvements. Additionally, students enjoyed expressing their feelings about the program and appreciated that their opinions were valued. Youth recommendations will be brought to program staff; and, where possible, gaps will be addressed. Empowerment evaluations with youth will continue during the duration of the program so that youth involvement and input remains integral in the evaluation and to ascertain whether the program's goals are being met. ^

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background. This study was designed to evaluate the effects of the Young Leaders for Healthy Change program, an internet-delivered program in the school setting that emphasized health advocacy skills-development, on nutrition and physical activity behaviors among older adolescents (1318 years). The program consisted of online curricular modules, training modules, social media, peer and parental support, and a community service project. Module content was developed based on Social Cognitive Theory and known determinants of behavior for older adolescents. ^ Methods. Of the 283 students who participated in the fall 2011 YL program, 38 students participated in at least ten of the 12 weeks and were eligible for this study. This study used a single group-only pretest/posttest evaluation design. Participants were 68% female, 58% white/Caucasian, 74% 10th or 11th graders, and 89% mostly A and/or B students. The primary behavioral outcomes for this analysis were participation in 60-minutes of physical activity per day, 20-minutes of vigorous- or moderate- intensity physical activity (MVPA) participation per day, television and computer time, fruit and vegetable (FV) intake, sugar-sweetened beverage intake, and consumption of breakfast, home-cooked meals, and fast food. Other outcomes included knowledge, beliefs, and attitudes related to healthy eating, physical activity, and advocacy skills. ^ Findings. Among the 38 participants, no significant changes in any variables were observed. However, among those who did not previously meet behavioral goals there was an 89% increase in students who participated in more than 20 minutes of MVPA per day and a 58% increase in students who ate home-cooked meals 57 days per week. The majority of participants met program goals related to knowledge, beliefs, and attitudes prior to the start of the program. Participants reported either maintaining or improving to the goal at posttest for all items except FV intake knowledge, taste and affordability of healthy foods, interest in teaching others about being healthy, and ease of finding ways to advocate in the community. ^ Conclusions. The results of this evaluation indicated that promoting healthy behaviors requires different strategies than maintaining healthy behaviors among high school students. In the school setting, programs need to target the promotion and maintenance of health behaviors to engage all students who participate in the program as part of a class or club activity. Tailoring the program using screening and modifying strategies to meet the needs of all students may increase the potential reach of the program. The Transtheoretical Model may provide information on how to develop a tailored program. Additional research on how to utilize the constructs of TTM effectively among high school students needs to be conducted. Further evaluation studies should employ a more expansive evaluation to assess the long-term effectiveness of health advocacy programming.^

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Introduction: Both a systems approach to change and a focus on multi-sector interventions ensures obesity prevention programming within the community is equitable, sustainable, and cost-effective. An authentic community engagement approach is required to implement interventions guided by best-evidence research and practice. Although there are examples illustrating the benefits of community engagement, there is no standardized method to implement it. The San Antonio Sports Foundation (SA Sports), a non-profit community-based organization, implements a variety of free events and programs promoting active life styles. One such program is the Fit Family Challenge which is a summer-long program implemented at the school level targeted at families. ^ Aims: This thesis was a culmination of the experience from the student collaborating with SA Sports as part of a practicum opportunity. Using secondary data collected by the Fit Family Challenge during the 2011 year, the goals of this thesis were: to assess individual changes; evaluate short-term impact; and describe the community engagement process. ^ Methods: SA Sports collected quantitative and qualitative data during the implementation and evaluation of the FFC program. SA Sports allowed the used of de-identified data to be analyzed to study the aims of this thesis. ^ Results: The program was able to provide families with the knowledge, information, and opportunity to exercise as a family and cook healthier meals. School district coordinators were generally satisfied and illustrated the benefits of a community partnership. An authentic community engagement was present highlighting the importance of communication, collaboration and the sustainability of such partnerships in the community. ^ Conclusion: The success of an obesity program should focus on triggers that initiate behavioral change rather than physiological changes. The evaluation was guided by a community engagement approach, which illustrated the development of new partnerships and the strengthening of other collaborations. Ultimately, the engagement approach empowered the community to identify their own problems and build collaboration, rather than tackling obesity prevention alone. ^

Relevância:

30.00% 30.00%

Publicador:

Resumo:

SIMBAA is a spatially explicit, individual-based simulation model. It was developed to analyse the response of populations of Antarctic benthic species and their diversity to iceberg scouring. This disturbance is causing a high local mortality providing potential space for new colonisation. Traits can be attributed to model species, e.g. in terms of reproduction, dispersal, and life span. Physical disturbances can be designed in space and time, e.g. in terms of size, shape, and frequency. Environmental heterogeneity can be considered by cell-specific capacities to host a certain number of individuals. When grid cells become empty (after a disturbance event or due to natural mortality of of an individual), a lottery decides which individual from which species stored in a pool of candidates (for this cell) will recruit in that cell. After a defined period the individuals become mature and their offspring are dispersed and stored in the pool of candidates. The biological parameters and disturbance regimes decide on how long an individual lives. Temporal development of single populations of species as well as Shannon diversity are depicted in the main window graphically and primary values are listed. Examples for simulations can be loaded and saved as sgf-files. The results are also shown in an additional window in a dimensionless area with 50 x 50 cells, which contain single individuals depicted as circles; their colour indicates the assignment to the self-designed model species and the size represents their age. Dominant species per cell and disturbed areas can also be depicted. Output of simulation runs can be saved as images, which can be assembled to video-clips by standard computer programs (see GIF-examples of which "Demo 1" represents the response of the Antarctic benthos to iceberg scouring and "Demo 2" represents a simulation of a deep-sea benthic habitat).

Relevância:

30.00% 30.00%

Publicador:

Resumo:

With over 700 million illiterate adults in the world, many governments have implemented adult literacy programs across the world, although typically with low rates of success partly because the quality of teaching is low. One solution may lie in the standardization of teaching provided by computer-aided instruction. We present the first rigorous evidence of the effectiveness of a computer-based adult literacy program. A randomized control trial study of TARA Akshar Plus, an Indian adult literacy program, was implemented in the state of Uttar Pradesh in India. We find large, significant impacts of this computer-aided program on literacy and numeracy outcomes. We compare the improvement in learning to that of other traditional adult literacy programs and conclude that TARA Akshar Plus is effective in increasing literacy and numeracy for illiterate adult women.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We show a method for parallelizing top down dynamic programs in a straightforward way by a careful choice of a lock-free shared hash table implementation and randomization of the order in which the dynamic program computes its subproblems. This generic approach is applied to dynamic programs for knapsack, shortest paths, and RNA structure alignment, as well as to a state-of-the-art solution for minimizing the mximum number of open stacks. Experimental results are provided on three different modern multicore architectures which show that this parallelization is effective and reasonably scalable. In particular, we obtain over 10 times speedup for 32 threads on the open stacks problem.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Since the early days of logic programming, researchers in the field realized the potential for exploitation of parallelism present in the execution of logic programs. Their high-level nature, the presence of nondeterminism, and their referential transparency, among other characteristics, make logic programs interesting candidates for obtaining speedups through parallel execution. At the same time, the fact that the typical applications of logic programming frequently involve irregular computations, make heavy use of dynamic data structures with logical variables, and involve search and speculation, makes the techniques used in the corresponding parallelizing compilers and run-time systems potentially interesting even outside the field. The objective of this article is to provide a comprehensive survey of the issues arising in parallel execution of logic programming languages along with the most relevant approaches explored to date in the field. Focus is mostly given to the challenges emerging from the parallel execution of Prolog programs. The article describes the major techniques used for shared memory implementation of Or-parallelism, And-parallelism, and combinations of the two. We also explore some related issues, such as memory management, compile-time analysis, and execution visualization.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

El clculo de relaciones binarias fue creado por De Morgan en 1860 para ser posteriormente desarrollado en gran medida por Peirce y Schrder. Tarski, Givant, Freyd y Scedrov demostraron que las lgebras relacionales son capaces de formalizar la lgica de primer orden, la lgica de orden superior as como la teora de conjuntos. A partir de los resultados matemticos de Tarski y Freyd, esta tesis desarrolla semnticas denotacionales y operacionales para la programacin lgica con restricciones usando el lgebra relacional como base. La idea principal es la utilizacin del concepto de semntica ejecutable, semnticas cuya caracterstica principal es el que la ejecucin es posible utilizando el razonamiento estndar del universo semntico, este caso, razonamiento ecuacional. En el caso de este trabajo, se muestra que las lgebras relacionales distributivas con un operador de punto fijo capturan toda la teora y metateora estndar de la programacin lgica con restricciones incluyendo los rboles utilizados en la bsqueda de demostraciones. La mayor parte de tcnicas de optimizacin de programas, evaluacin parcial e interpretacin abstracta pueden ser llevadas a cabo utilizando las semnticas aqu presentadas. La demostracin de la correccin de la implementacin resulta extremadamente sencilla. En la primera parte de la tesis, un programa lgico con restricciones es traducido a un conjunto de trminos relacionales. La interpretacin estndar en la teora de conjuntos de dichas relaciones coincide con la semntica estndar para CLP. Las consultas contra el programa traducido son llevadas a cabo mediante la reescritura de relaciones. Para concluir la primera parte, se demuestra la correccin y equivalencia operacional de esta nueva semntica, as como se define un algoritmo de unificacin mediante la reescritura de relaciones. La segunda parte de la tesis desarrolla una semntica para la programacin lgica con restricciones usando la teora de alegorasversin categrica del lgebra de relacionesde Freyd. Para ello, se definen dos nuevos conceptos de Categora Regular de Lawvere y _-Alegora, en las cuales es posible interpretar un programa lgico. La ventaja fundamental que el enfoque categrico aporta es la definicin de una mquina categrica que mejora e sistema de reescritura presentado en la primera parte. Gracias al uso de relaciones tabulares, la mquina modela la ejecucin eficiente sin salir de un marco estrictamente formal. Utilizando la reescritura de diagramas, se define un algoritmo para el clculo de pullbacks en Categoras Regulares de Lawvere. Los dominios de las tabulaciones aportan informacin sobre la utilizacin de memoria y variable libres, mientras que el estado compartido queda capturado por los diagramas. La especificacin de la mquina induce la derivacin formal de un juego de instrucciones eficiente. El marco categrico aporta otras importantes ventajas, como la posibilidad de incorporar tipos de datos algebraicos, funciones y otras extensiones a Prolog, a la vez que se conserva el carcter 100% declarativo de nuestra semntica. ABSTRACT The calculus of binary relations was introduced by De Morgan in 1860, to be greatly developed by Peirce and Schrder, as well as many others in the twentieth century. Using different formulations of relational structures, Tarski, Givant, Freyd, and Scedrov have shown how relation algebras can provide a variable-free way of formalizing first order logic, higher order logic and set theory, among other formal systems. Building on those mathematical results, we develop denotational and operational semantics for Constraint Logic Programming using relation algebra. The idea of executable semantics plays a fundamental role in this work, both as a philosophical and technical foundation. We call a semantics executable when program execution can be carried out using the regular theory and tools that define the semantic universe. Throughout this work, the use of pure algebraic reasoning is the basis of denotational and operational results, eliminating all the classical non-equational meta-theory associated to traditional semantics for Logic Programming. All algebraic reasoning, including execution, is performed in an algebraic way, to the point we could state that the denotational semantics of a CLP program is directly executable. Techniques like optimization, partial evaluation and abstract interpretation find a natural place in our algebraic models. Other properties, like correctness of the implementation or program transformation are easy to check, as they are carried out using instances of the general equational theory. In the first part of the work, we translate Constraint Logic Programs to binary relations in a modified version of the distributive relation algebras used by Tarski. Execution is carried out by a rewriting system. We prove adequacy and operational equivalence of the semantics. In the second part of the work, the relation algebraic approach is improved by using allegory theory, a categorical version of the algebra of relations developed by Freyd and Scedrov. The use of allegories lifts the semantics to typed relations, which capture the number of logical variables used by a predicate or program state in a declarative way. A logic program is interpreted in a _-allegory, which is in turn generated from a new notion of Regular Lawvere Category. As in the untyped case, program translation coincides with program interpretation. Thus, we develop a categorical machine directly from the semantics. The machine is based on relation composition, with a pullback calculation algorithm at its core. The algorithm is defined with the help of a notion of diagram rewriting. In this operational interpretation, types represent information about memory allocation and the execution mechanism is more efficient, thanks to the faithful representation of shared state by categorical projections. We finish the work by illustrating how the categorical semantics allows the incorporation into Prolog of constructs typical of Functional Programming, like abstract data types, and strict and lazy functions.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Irregular computations pose sorne of the most interesting and challenging problems in automatic parallelization. Irregularity appears in certain kinds of numerical problems and is pervasive in symbolic applications. Such computations often use dynamic data structures, which make heavy use of pointers. This complicates all the steps of a parallelizing compiler, from independence detection to task partitioning and placement. Starting in the mid 80s there has been significant progress in the development of parallelizing compilers for logic programming (and more recently, constraint programming) resulting in quite capable parallelizers. The typical applications of these paradigms frequently involve irregular computations, and make heavy use of dynamic data structures with pointers, since logical variables represent in practice a well-behaved form of pointers. This arguably makes the techniques used in these compilers potentially interesting. In this paper, we introduce in a tutoral way, sorne of the problems faced by parallelizing compilers for logic and constraint programs and provide pointers to sorne of the significant progress made in the area. In particular, this work has resulted in a series of achievements in the areas of inter-procedural pointer aliasing analysis for independence detection, cost models and cost analysis, cactus-stack memory management, techniques for managing speculative and irregular computations through task granularity control and dynamic task allocation such as work-stealing schedulers), etc.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Global analyzers traditionally read and analyze the entire program at once, in a nonincremental way. However, there are many situations which are not well suited to this simple model and which instead require reanalysis of certain parts of a program which has already been analyzed. In these cases, it appears inecient to perform the analysis of the program again from scratch, as needs to be done with current systems. We describe how the xed-point algorithms used in current generic analysis engines for (constraint) logic programming languages can be extended to support incremental analysis. The possible changes to a program are classied into three types: addition, deletion, and arbitrary change. For each one of these, we provide one or more algorithms for identifying the parts of the analysis that must be recomputed and for performing the actual recomputation. The potential benets and drawbacks of these algorithms are discussed. Finally, we present some experimental results obtained with an implementation of the algorithms in the PLAI generic abstract interpretation framework. The results show signicant benets when using the proposed incremental analysis algorithms.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We report on a detailed study of the application and effectiveness of program analysis based on abstract interpretation to automatic program parallelization. We study the case of parallelizing logic programs using the notion of strict independence. We first propose and prove correct a methodology for the application in the parallelization task of the information inferred by abstract interpretation, using a parametric domain. The methodology is generic in the sense of allowing the use of different analysis domains. A number of well-known approximation domains are then studied and the transformation into the parametric domain defined. The transformation directly illustrates the relevance and applicability of each abstract domain for the application. Both local and global analyzers are then built using these domains and embedded in a complete parallelizing compiler. Then, the performance of the domains in this context is assessed through a number of experiments. A comparatively wide range of aspects is studied, from the resources needed by the analyzers in terms of time and memory to the actual benefits obtained from the information inferred. Such benefits are evaluated both in terms of the characteristics of the parallelized code and of the actual speedups obtained from it. The results show that data flow analysis plays an important role in achieving efficient parallelizations, and that the cost of such analysis can be reasonable even for quite sophisticated abstract domains. Furthermore, the results also offer significant insight into the characteristics of the domains, the demands of the application, and the trade-offs involved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This article presents and illustrates a practical approach to the dataow analysis of constraint logic programming languages using abstract interpretation. It is rst argued that from the framework point of view it suces to propose relatively simple extensions of traditional analysis methods which have already been proved useful and practical and for exist. This is shown by proposing a simple extension of Bruynooghes traditional framework which allows it to analyze constraint logic programs. Then and using this generalized framework two abstract domains and their required abstract functions are presented the rst abstract domain approximates deniteness information and the second one freeness. Finally an approach for cobining those domains is proposed The two domains and their combination have been implemented and used in the analysis of CLP and Prolog III applications. Results from this implementation showing its performance and accuracy are also presented

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Este trabajo tiene como objetivos la monitorizacin en tiempo real de la actividad ssmica, tanto prxima como lejana, a partir de los datos ssmicos registrados por una estacin de banda ancha, y el desarrollo de un sistema de difusin interactiva de informacin actualizada de terremotos, destinado al pblico general. Ambas fuentes de informacin se mostrarn a travs de una Unidad de Visualizacin denominada Monitor Ssmico Interactivo. El registro de los datos ssmicos se realiza utilizando el sensor de tres componentes de la estacin ssmica GUD, perteneciente a la Red Digital de Banda Ancha y transmisin digital del Instituto Geogrfico Nacional, instalada en la Baslica del Valle de los Cados, en lalocalidad de Guadarrama (Madrid). En la E.T.S.I. Topografa, Geodesia y Cartografa se ha instalado un ordenador con conexin a Internet, para la recepcin y almacenamiento de los datos, y los programas Scream y Drumplot desarrollados por Guralp, necesarios para la monitorizacin de la seal ssmica en tiempo real. A partir de estos datos, mediante aplicaciones desarrolladas bajo programacin Linux y haciendo uso de las herramientas que ofrece el software SAC (Seismic Analysis Code), se genera adems un registro grfico y una pelcula animada de dicha segmentacin para cada evento. Se ha configurado un servidor de correo y una cuenta para la recepcin de dos tipos de mensajes de correo, enviados desde la sede central del Instituto Geogrfico Nacional, con la informacin de los eventos registrados por GUD una vez revisados: - Mensajes enviados diariamente, con un listado de eventos ocurridos en los 30 ltimos das. - Mensajes con la informacin en cuasi tiempo real de la ltima alerta ssmica. Se ha desarrollado el programa saco para la gestin del correo recibido que analiza la informacin ssmica, la almacena en ficheros y ejecuta sobre ellos las aplicaciones de dibujo. Estas aplicaciones han sido previamente desarrolladas bajo programacin Linux y software GMT (Generic Mapping Tools), y a partir de ellas se generan automticamente las distintas imgenes que se visualizan en el Monitor Ssmico: un mapa de sismicidad prxima en la Pennsula Ibrica, un mapa de sismicidad lejana en el mundo, un mapa de detalle para localizar y representar la ltima alerta generada, los listados con la informacin de los eventos representados en los mapas, los registros grficos y las pelculas animadas de dichos sismogramas. Monitor Ssmico Interactivo ha sido desarrollado para ofrecer adems la posibilidad de interactuar con la Unidad de Visualizacin: se ha creado una base de datos para uso cientfico donde se almacenan todos los eventos registrados por GUD. As el usuario puede realizar una peticin, a travs del envo de un mensaje de correo, que le permite visualizar de forma instantnea las imgenes que muestran la informacin de cualquier terremoto de su inters. ABSTRACT This study is aimed at real-time monitoring of both near and distant seismic activityfrom the seismic data recorded by a broadband seismic station, and the development of an interactive broadcast system of updated information of earthquakes, for the general public. Bothsources of information are displayed through a display unit called "Interactive Seismic Monitor". The seismic data recording is carried out by using the three-component sensor of the GUD seismic station, which belongs to the Digital Network Broadband and digital broadcast of the National Geographic Institute, housed in the Basilica of The Valley of the Fallen, in the town of Guadarrama (Madrid). A computer with Internet connection has been installed in E.T.S.I. Surveying, Geodesy and Cartography for receiving and storing data, together with Scream and Drumplot programs, developed by Guralp, which are necessary for monitoring the real time seismic signal. Based on the data collected, through programming applications developed under Linux system and using the software tools provided by the SAC (Seismic Analysis Code), a chart recorder and an animated gif image of the segmentation for each event are also generated. A mail server and a mail account have been configured for the receipt of two types of email messages, sent from the National Geographic Institute head office, with the information of the events recorded by GUD after being reviewed: - Messages sent daily, providing a list of events in the past 30 days. - Messages containing information on near real-time seismic of the last seismic alert. A program called "saco" has also been developed for handling mail received that analyzes the seismic data, which stores it in files and runs drawing applications on them. These applications have been previously developed under Linux system and software programming GMT (Generic Mapping Tools), and from them different images that are displayed on the Seismic Monitor are automatically generated: a near seismicity Iberian peninsula map, a distant seismicity world map, a detailed map to locate and represent the last seismic alert generated, the lists with the information of the events depicted in the maps,together with the charts and the animated gif image of such seismograms. Interactive Seismic Monitor has been developed to offer any user the possibility of interacting with the display unit: a database has been created for scientific use which stores all the events recorded by GUD. Thus, any user could make a request, by sending an e-mail that allows them to view instantly all the images showing the information of any earthquake of interest on the display unit.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A framework for the automatic parallelization of (constraint) logic programs is proposed and proved correct. Intuitively, the parallelization process replaces conjunctions of literals with parallel expressions. Such expressions trigger at run-time the exploitation of restricted, goal-level, independent and-parallelism. The parallelization process performs two steps. The first one builds a conditional dependency graph (which can be implified using compile-time analysis information), while the second transforms the resulting graph into linear conditional expressions, the parallel expressions of the &-Prolog language. Several heuristic algorithms for the latter ("annotation") process are proposed and proved correct. Algorithms are also given which determine if there is any loss of parallelism in the linearization process with respect to a proposed notion of maximal parallelism. Finally, a system is presented which implements the proposed approach. The performance of the different annotation algorithms is compared experimentally in this system by studying the time spent in parallelization and the effectiveness of the results in terms of speedups.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper illustrates the use of a top-down framework to obtain goal independent analyses of logic programs, a task which is usually associated with the bottom-up approach. While it is well known that the bottomup approach can be used, through the magic set transformation, for goal dependent analysis, it is less known that the top-down approach can be used for goal independent analysis. The paper describes two ways of doing the latter. We show how the results of a goal independent analysis can be used to speed up subsequent goal dependent analyses. However this speed-up may result in a loss of precisin. The influence of domain characteristics on this precisin is discussed and an experimental evaluation using a generic top-down analyzer is described.