12 resultados para Administrative support
em Boston University Digital Common
Resumo:
Reproduction of copy held by Special Collections, Bridewell Library, Perkins School of Theology, Southern Methodist University. Includes both DjVu and PDF files for download. Mode of access: World Wide Web.
Resumo:
Background Chronic illness and premature mortality from malaria, water-borne diseases, and respiratory illnesses have long been known to diminish the welfare of individuals and households in developing countries. Previous research has also shown that chronic diseases among farming populations suppress labor productivity and agricultural output. As the illness and death toll from HIV/AIDS continues to climb in most of sub-Saharan Africa, concern has arisen that the loss of household labor it causes will reduce crop yields, impoverish farming households, intensify malnutrition, and suppress growth in the agricultural sector. If chronic morbidity and premature mortality among individuals in farming households have substantial impacts on household production, and if a large number of households are affected, it is possible that an increase in morbidity and mortality from HIV/AIDS or other diseases could affect national aggregate output and exports. If, on the other hand, the impact at the household farm level is modest, or if relatively few households are affected, there is likely to be little effect on aggregate production across an entire country. Which of these outcomes is more likely in West Africa is unknown. Little rigorous, quantitative research has been published on the impacts of AIDS on smallholder farm production, particularly in West Africa. The handful of studies that have been conducted have looked mainly at small populations in areas of very high HIV prevalence in southern and eastern Africa. Conclusions about how HIV/AIDS, and other causes of chronic morbidity and mortality, are affecting agriculture across the continent cannot be drawn from these studies. In view of the importance of agriculture, and particularly smallholder agriculture, in the economies of most African countries and the scarcity of resources for health interventions, it is valuable to identify, describe, and quantify the impact of chronic morbidity and mortality on smallholder production of important crops in West Africa. One such crop is cocoa. In Ghana, cocoa is a crop of national importance that is produced almost exclusively by smallholder households. In 2003, Ghana was the world’s second-largest producer of cocoa. Cocoa accounted for a quarter of Ghana’s export revenues that year and generated 15 percent of employment. The success and growth of the cocoa industry is thus vital to the country’s overall social and economic development. Study Objectives and Methods In February and March 2005, the Center for International Health and Development of Boston University (CIHD) and the Department of Agricultural Economics and Agribusiness (DAEA) of the University of Ghana, with financial support from the Africa Bureau of the U.S. Agency for International Development and from Mars, Inc., which is a major purchaser of West African cocoa, conducted a survey of a random sample of cocoa farming households in the Western Region of Ghana. The survey documented the extent of chronic morbidity and mortality in cocoa growing households in the Western Region of Ghana, the country’s largest cocoa growing region, and analyzed the impact of morbidity and mortality on cocoa production. It aimed to answer three specific research questions. (1) What is the baseline status of the study population in terms of household size and composition, acute and chronic morbidity, recent mortality, and cocoa production? (2) What is the relationship between household size and cocoa production, and how can this relationship be used to understand the impact of adult mortality and chronic morbidity on the production of cocoa at the household level? The study population was the approximately 42,000 cocoa farming households in the southern part of Ghana’s Western Region. A random sample of households was selected from a roster of eligible households developed from existing administrative information. Under the supervision of the University of Ghana field team, enumerators were graduate students of the Department of Agricultural Economics and Agribusiness or employees of the Cocoa Services Division. A total of 632 eligible farmers participated in the survey. Of these, 610 provided complete responses to all questions needed to complete the multivariate statistical analysis reported here.
Resumo:
A working paper written for Boston University Libraries to foster discussion about how to provide better support for BU faculty authors.
Resumo:
Predictability -- the ability to foretell that an implementation will not violate a set of specified reliability and timeliness requirements -- is a crucial, highly desirable property of responsive embedded systems. This paper overviews a development methodology for responsive systems, which enhances predictability by eliminating potential hazards resulting from physically-unsound specifications. The backbone of our methodology is the Time-constrained Reactive Automaton (TRA) formalism, which adopts a fundamental notion of space and time that restricts expressiveness in a way that allows the specification of only reactive, spontaneous, and causal computation. Using the TRA model, unrealistic systems – possessing properties such as clairvoyance, caprice, infinite capacity, or perfect timing -- cannot even be specified. We argue that this "ounce of prevention" at the specification level is likely to spare a lot of time and energy in the development cycle of responsive systems -- not to mention the elimination of potential hazards that would have gone, otherwise, unnoticed. The TRA model is presented to system developers through the Cleopatra programming language. Cleopatra features a C-like imperative syntax for the description of computation, which makes it easier to incorporate in applications already using C. It is event-driven, and thus appropriate for embedded process control applications. It is object-oriented and compositional, thus advocating modularity and reusability. Cleopatra is semantically sound; its objects can be transformed, mechanically and unambiguously, into formal TRA automata for verification purposes, which can be pursued using model-checking or theorem proving techniques. Since 1989, an ancestor of Cleopatra has been in use as a specification and simulation language for embedded time-critical robotic processes.
Resumo:
This paper proposes the use of in-network caches (which we call Angels) to reduce the Minimum Distribution Time (MDT) of a file from a seeder – a node that possesses the file – to a set of leechers – nodes who are interested in downloading the file. An Angel is not a leecher in the sense that it is not interested in receiving the entire file, but rather it is interested in minimizing the MDT to all leechers, and as such uses its storage and up/down-link capacity to cache and forward parts of the file to other peers. We extend the analytical results by Kumar and Ross [1] to account for the presence of angels by deriving a new lower bound for the MDT. We show that this newly derived lower bound is tight by proposing a distribution strategy under assumptions of a fluid model. We present a GroupTree heuristic that addresses the impracticalities of the fluid model. We evaluate our designs through simulations that show that our Group-Tree heuristic outperforms other heuristics, that it scales well with the increase of the number of leechers, and that it closely approaches the optimal theoretical bounds.
Resumo:
This thesis proposes the use of in-network caches (which we call Angels) to reduce the Minimum Distribution Time (MDT) of a file from a seeder – a node that possesses the file – to a set of leechers – nodes who are interested in downloading the file. An Angel is not a leecher in the sense that it is not interested in receiving the entire file, but rather it is interested in minimizing the MDT to all leechers, and as such uses its storage and up/down-link capacity to cache and forward parts of the file to other peers. We extend the analytical results by Kumar and Ross (Kumar and Ross, 2006) to account for the presence of angels by deriving a new lower bound for the MDT. We show that this newly derived lower bound is tight by proposing a distribution strategy under assumptions of a fluid model. We present a GroupTree heuristic that addresses the impracticalities of the fluid model. We evaluate our designs through simulations that show that our GroupTree heuristic outperforms other heuristics, that it scales well with the increase of the number of leechers, and that it closely approaches the optimal theoretical bounds.
Resumo:
In work that involves mathematical rigor, there are numerous benefits to adopting a representation of models and arguments that can be supplied to a formal reasoning or verification system: reusability, automatic evaluation of examples, and verification of consistency and correctness. However, accessibility has not been a priority in the design of formal verification tools that can provide these benefits. In earlier work [Lap09a], we attempt to address this broad problem by proposing several specific design criteria organized around the notion of a natural context: the sphere of awareness a working human user maintains of the relevant constructs, arguments, experiences, and background materials necessary to accomplish the task at hand. This work expands one aspect of the earlier work by considering more extensively an essential capability for any formal reasoning system whose design is oriented around simulating the natural context: native support for a collection of mathematical relations that deal with common constructs in arithmetic and set theory. We provide a formal definition for a context of relations that can be used to both validate and assist formal reasoning activities. We provide a proof that any algorithm that implements this formal structure faithfully will necessary converge. Finally, we consider the efficiency of an implementation of this formal structure that leverages modular implementations of well-known data structures: balanced search trees and transitive closures of hypergraphs.
Resumo:
It is useful in systems that must support multiple applications with various temporal requirements to allow application-specific policies to manage resources accordingly. However, there is a tension between this goal and the desire to control and police possibly malicious programs. The Java-based Sensor Execution Environment (SXE) in snBench presents a situation where such considerations add value to the system. Multiple applications can be run by multiple users with varied temporal requirements, some Real-Time and others best effort. This paper outlines and documents an implementation of a hierarchical and configurable scheduling system with which different applications can be executed using application-specific scheduling policies. Concurrently the system administrator can define fairness policies between applications that are imposed upon the system. Additionally, to ensure forward progress of system execution in the face of malicious or malformed user programs, an infrastructure for execution using multiple threads is described.
Resumo:
The SNBENCH is a general-purpose programming environment and run-time system targeted towards a variety of Sensor applications such as environmental sensing, location sensing, video sensing, etc. In its current structure, the run-time engine of the SNBENCH namely, the Sensorium Execution Environment (SXE) processes the entities of execution in a single thread of operation. In order to effectively support applications that are time-sensitive and need priority, it is imperative to process the tasks discretely so that specific policies can be applied at a much granular level. The goal of this project was to modify the SXE to enable efficient use of system resources by way of multi-tasking the individual components. Additionally, the transformed SXE offers the ability to classify and employ different schemes of processing to the individual tasks.
Resumo:
The purpose of this project is the creation of a graphical "programming" interface for a sensor network tasking language called STEP. The graphical interface allows the user to specify a program execution graphically from an extensible pallet of functionalities and save the results as a properly formatted STEP file. Moreover, the software is able to load a file in STEP format and convert it into the corresponding graphical representation. During both phases a type-checker is running on the background to ensure that both the graphical representation and the STEP file are syntactically correct. This project has been motivated by the Sensorium project at Boston University. In this technical report we present the basic features of the software, the process that has been followed during the design and implementation. Finally, we describe the approach used to test and validate our software.
Resumo:
In this project we design and implement a centralized hashing table in the snBench sensor network environment. We discuss the feasibility of this approach and compare and contrast with the distributed hashing architecture, with particular discussion regarding the conditions under which a centralized architecture makes sense. There are numerous computational tasks that require persistence of data in a sensor network environment. To help motivate the need for data storage in snBench we demonstrate a practical application of the technology whereby a video camera can monitor a room to detect the presence of a person and send an alert to the appropriate authorities.
Resumo:
We present a thorough characterization of the access patterns in blogspace -- a fast-growing constituent of the content available through the Internet -- which comprises a rich interconnected web of blog postings and comments by an increasingly prominent user community that collectively define what has become known as the blogosphere. Our characterization of over 35 million read, write, and administrative requests spanning a 28-day period is done from three different blogosphere perspectives. The server view characterizes the aggregate access patterns of all users to all blogs; the user view characterizes how individual users interact with blogosphere objects (blogs); the object view characterizes how individual blogs are accessed. Our findings support two important conclusions. First, we show that the nature of interactions between users and objects is fundamentally different in blogspace than that observed in traditional web content. Access to objects in blogspace could be conceived as part of an interaction between an author and its readership. As we show in our work, such interactions range from one-to-many "broadcast-type" and many-to-one "registration-type" communication between an author and its readers, to multi-way, iterative "parlor-type" dialogues among members of an interest group. This more-interactive nature of the blogosphere leads to interesting traffic and communication patterns, which are different from those observed in traditional web content. Second, we identify and characterize novel features of the blogosphere workload, and we investigate the similarities and differences between typical web server workloads and blogosphere server workloads. Given the increasing share of blogspace traffic, understanding such differences is important for capacity planning and traffic engineering purposes, for example.