962 resultados para Epidemiological data
Resumo:
Managing the physical and compute infrastructure of a large data center is an embodiment of a Cyber-Physical System (CPS). The physical parameters of the data center (such as power, temperature, pressure, humidity) are tightly coupled with computations, even more so in upcoming data centers, where the location of workloads can vary substantially due, for example, to workloads being moved in a cloud infrastructure hosted in the data center. In this paper, we describe a data collection and distribution architecture that enables gathering physical parameters of a large data center at a very high temporal and spatial resolutionof the sensor measurements. We think this is an important characteristic to enable more accurate heat-flow models of the data center andwith them, _and opportunities to optimize energy consumption. Havinga high resolution picture of the data center conditions, also enables minimizing local hotspots, perform more accurate predictive maintenance (pending failures in cooling and other infrastructure equipment can be more promptly detected) and more accurate billing. We detail this architecture and define the structure of the underlying messaging system that is used to collect and distribute the data. Finally, we show the results of a preliminary study of a typical data center radio environment.
Resumo:
Consider the problem of designing an algorithm for acquiring sensor readings. Consider specifically the problem of obtaining an approximate representation of sensor readings where (i) sensor readings originate from different sensor nodes, (ii) the number of sensor nodes is very large, (iii) all sensor nodes are deployed in a small area (dense network) and (iv) all sensor nodes communicate over a communication medium where at most one node can transmit at a time (a single broadcast domain). We present an efficient algorithm for this problem, and our novel algorithm has two desired properties: (i) it obtains an interpolation based on all sensor readings and (ii) it is scalable, that is, its time-complexity is independent of the number of sensor nodes. Achieving these two properties is possible thanks to the close interlinking of the information processing algorithm, the communication system and a model of the physical world.
Resumo:
Mestrado em Contabilidade e Gestão das Instituições Financeiras
Resumo:
OBJECTIVE To analyze spatial changes in the risk of AIDS and the relationship between AIDS incidence and socioeconomic variables in the state of Rondonia, Amazon region. METHODS A spatial, population case-control study in Rondonia, Brazil, based on 1,780 cases reported to the Epidemiological Surveillance System and controls based on demographic data from 1987 to 2006. The cases were grouped into five consecutive four-year periods. A generalized additive model was adjusted to the data; the dependent variable was the status of the individuals (case or control), and the independent variables were a bi-dimensional spline of the geographic coordinates and some municipality-level socioeconomic variables. The observed values of the Moran’s I test were compared to a reference distribution of values generated under conditions of spatial randomness. RESULTS AIDS risk shows a marked spatial and temporal pattern. The disease incidence is related to socioeconomic variables at the municipal level in Rondônia, such as urbanization and human capital. The highest incidence rates of AIDS are in municipalities along the BR-364 highway and calculations of the Moran’s I test show positive spatial correlation associated with proximity of the municipality to the highway in the third and fourth periods (p = 0.05). CONCLUSIONS Incidence of the disease is higher in municipalities of greater economic wealth and urbanization, and in those municipalities bisected by Rondônia’s main roads. The rapid development associated with the opening up of once remote regions may be accompanied by an increase in these risks to health.
Resumo:
Network control systems (NCSs) are spatially distributed systems in which the communication between sensors, actuators and controllers occurs through a shared band-limited digital communication network. However, the use of a shared communication network, in contrast to using several dedicated independent connections, introduces new challenges which are even more acute in large scale and dense networked control systems. In this paper we investigate a recently introduced technique of gathering information from a dense sensor network to be used in networked control applications. Obtaining efficiently an approximate interpolation of the sensed data is exploited as offering a good tradeoff between accuracy in the measurement of the input signals and the delay to the actuation. These are important aspects to take into account for the quality of control. We introduce a variation to the state-of-the-art algorithms which we prove to perform relatively better because it takes into account the changes over time of the input signal within the process of obtaining an approximate interpolation.
Resumo:
Cluster scheduling and collision avoidance are crucial issues in large-scale cluster-tree Wireless Sensor Networks (WSNs). The paper presents a methodology that provides a Time Division Cluster Scheduling (TDCS) mechanism based on the cyclic extension of RCPS/TC (Resource Constrained Project Scheduling with Temporal Constraints) problem for a cluster-tree WSN, assuming bounded communication errors. The objective is to meet all end-to-end deadlines of a predefined set of time-bounded data flows while minimizing the energy consumption of the nodes by setting the TDCS period as long as possible. Sinceeach cluster is active only once during the period, the end-to-end delay of a given flow may span over several periods when there are the flows with opposite direction. The scheduling tool enables system designers to efficiently configure all required parameters of the IEEE 802.15.4/ZigBee beaconenabled cluster-tree WSNs in the network design time. The performance evaluation of thescheduling tool shows that the problems with dozens of nodes can be solved while using optimal solvers.
Resumo:
The simulation analysis is important approach to developing and evaluating the systems in terms of development time and cost. This paper demonstrates the application of Time Division Cluster Scheduling (TDCS) tool for the configuration of IEEE 802.15.4/ZigBee beaconenabled cluster-tree WSNs using the simulation analysis, as an illustrative example that confirms the practical applicability of the tool. The simulation study analyses how the number of retransmissions impacts the reliability of data transmission, the energy consumption of the nodes and the end-to-end communication delay, based on the simulation model that was implemented in the Opnet Modeler. The configuration parameters of the network are obtained directly from the TDCS tool. The simulation results show that the number of retransmissions impacts the reliability, the energy consumption and the end-to-end delay, in a way that improving the one may degrade the others.
Resumo:
Cooperating objects (COs) is a recently coined term used to signify the convergence of classical embedded computer systems, wireless sensor networks and robotics and control. We present essential elements of a reference architecture for scalable data processing for the CO paradigm.
Resumo:
Doctoral Thesis in Information Systems and Technologies Area of Engineering and Manag ement Information Systems
Resumo:
Environment monitoring has an important role in occupational exposure assessment. However, due to several factors is done with insufficient frequency and normally don´t give the necessary information to choose the most adequate safety measures to avoid or control exposure. Identifying all the tasks developed in each workplace and conducting a task-based exposure assessment help to refine the exposure characterization and reduce assessment errors. A task-based assessment can provide also a better evaluation of exposure variability, instead of assessing personal exposures using continuous 8-hour time weighted average measurements. Health effects related with exposure to particles have mainly been investigated with mass-measuring instruments or gravimetric analysis. However, more recently, there are some studies that support that size distribution and particle number concentration may have advantages over particle mass concentration for assessing the health effects of airborne particles. Several exposure assessments were performed in different occupational settings (bakery, grill house, cork industry and horse stable) and were applied these two resources: task-based exposure assessment and particle number concentration by size. The results showed interesting results: task-based approach applied permitted to identify the tasks with higher exposure to the smaller particles (0.3 μm) in the different occupational settings. The data obtained allow more concrete and effective risk assessment and the identification of priorities for safety investments.
Resumo:
OBJECTIVE Develop an index to evaluate the maternal and neonatal hospital care of the Brazilian Unified Health System.METHODS This descriptive cross-sectional study of national scope was based on the structure-process-outcome framework proposed by Donabedian and on comprehensive health care. Data from the Hospital Information System and the National Registry of Health Establishments were used. The maternal and neonatal network of Brazilian Unified Health System consisted of 3,400 hospitals that performed at least 12 deliveries in 2009 or whose number of deliveries represented 10.0% or more of the total admissions in 2009. Relevance and reliability were defined as criteria for the selection of variables. Simple and composite indicators and the index of completeness were constructed and evaluated, and the distribution of maternal and neonatal hospital care was assessed in different regions of the country.RESULTS A total of 40 variables were selected, from which 27 single indicators, five composite indicators, and the index of completeness of care were built. Composite indicators were constructed by grouping simple indicators and included the following variables: hospital size, level of complexity, delivery care practice, recommended hospital practice, and epidemiological practice. The index of completeness of care grouped the five variables and classified them in ascending order, thereby yielding five levels of completeness of maternal and neonatal hospital care: very low, low, intermediate, high, and very high. The hospital network was predominantly of small size and low complexity, with inadequate child delivery care and poor development of recommended and epidemiological practices. The index showed that more than 80.0% hospitals had a low index of completeness of care and that most qualified heath care services were concentrated in the more developed regions of the country.CONCLUSIONS The index of completeness proved to be of great value for monitoring the maternal and neonatal hospital care of Brazilian Unified Health System and indicated that the quality of health care was unsatisfactory. However, its application does not replace specific evaluations.
Resumo:
Epidemiological studies have shown the effect of diet on the incidence of chronic diseases; however, proper planning, designing, and statistical modeling are necessary to obtain precise and accurate food consumption data. Evaluation methods used for short-term assessment of food consumption of a population, such as tracking of food intake over 24h or food diaries, can be affected by random errors or biases inherent to the method. Statistical modeling is used to handle random errors, whereas proper designing and sampling are essential for controlling biases. The present study aimed to analyze potential biases and random errors and determine how they affect the results. We also aimed to identify ways to prevent them and/or to use statistical approaches in epidemiological studies involving dietary assessments.
Resumo:
OBJECTIVE To assess the validity of dengue fever reports and how they relate to the definition of case and severity. METHODS Diagnostic test assessment was conducted using cross-sectional sampling from a universe of 13,873 patients treated during the fifth epidemiological period in health institutions from 11 Colombian departments in 2013. The test under analyses was the reporting to the National Public Health Surveillance System, and the reference standard was the review of histories identified by active institutional search. We reviewed all histories of patients diagnosed with dengue fever, as well as a random sample of patients with febrile syndromes. The specificity and sensitivity of reports were estimated for this purpose, considering the inverse of the probability of being selected for weighting. The concordance between reporting and the findings of the active institutional search was calculated using Kappa statistics. RESULTS We included 4,359 febrile patients, and 31.7% were classified as compatible with dengue fever (17 with severe dengue fever; 461 with dengue fever and warning signs; 904 with dengue fever and no warning signs). The global sensitivity of reports was 13.2% (95%CI 10.9;15.4) and specificity was 98.4% (95%CI 97.9;98.9). Sensitivity varied according to severity: 12.1% (95%CI 9.3;14.8) for patients presenting dengue fever with no warning signs; 14.5% (95%CI 10.6;18.4) for those presenting dengue fever with warning signs, and 40.0% (95%CI 9.6;70.4) for those with severe dengue fever. Concordance between reporting and the findings of the active institutional search resulted in a Kappa of 10.1%. CONCLUSIONS Low concordance was observed between reporting and the review of clinical histories, which was associated with the low reporting of dengue fever compatible cases, especially milder cases.
Resumo:
XVIII Jornadas de Paleontología, 2002
Resumo:
OBJECTIVE To analyze the cases of tuberculosis and the impact of direct follow-up on the assessment of treatment outcomes.METHODS This open prospective cohort study evaluated 504 cases of tuberculosis reported in the Sistema de Informação de Agravos de Notificação (SINAN – Notifiable Diseases Information System) in Juiz de Fora, MG, Southeastern Brazil, between 2008 and 2009. The incidence of treatment outcomes was compared between a group of patients diagnosed with tuberculosis and directly followed up by monthly consultations during return visits (287) and a patient group for which the information was indirectly collected (217) through the city’s surveillance system. The Chi-square test was used to compare the percentages, with a significance level of 0.05. The relative risk (RR) was used to evaluate the differences in the incidence rate of each type of treatment outcome between the two groups.RESULTS Of the outcomes directly and indirectly evaluated, 18.5% and 3.2% corresponded to treatment default and 3.8% and 0.5% corresponded to treatment failure, respectively. The incidence of treatment default and failure was higher in the group with direct follow-up (p < 0.05) (RR = 5.72, 95%CI 2.65;12.34, and RR = 8.31, 95%CI 1.08;63.92, respectively).CONCLUSIONS A higher incidence of treatment default and failure was observed in the directly followed up group, and most of these cases were neglected by the disease reporting system. Therefore, effective measures are needed to improve the control of tuberculosis and data quality.