541 resultados para violations


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Decades of research show that environmental exposure to the chemical benzene is associated with severe carcinogenic, hematoxic and genotoxic effects on the human body. As such, the Environmental Protection Agency (EPA) has designated the chemical as a Hazardous Air Pollutant and prescribed benzene air concentration guidelines that provide cities with an ideal ambient level to protect human health. However, in Houston, Texas, a city home to the top industrial benzene emitters in the US who undoubtedly contribute greatly to the potentially unsafe levels of ambient benzene, regulations beyond the EPA’s unenforceable guidelines are critical to protecting public health. Despite this, the EPA has failed to establish National Ambient Air Quality Standards (NAAQS) for benzene. States are thus left to regulate air benzene levels on their own; in the case of Texas, the Texas Commission on Environmental Quality (TCEQ) and state legislature have failed to proactively develop legally enforceable policies to reduce major source benzene emissions. This inaction continues to exacerbate a public health problem, which may only be solved through a legal framework that restricts preventable benzene emissions to protect human health and holds industrial companies accountable for violations of such regulations and standards. This analysis explores legal barriers that the City of Houston and other relevant agencies currently face in their attempt to demand and bring about such change. ^

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This dissertation develops and explores the methodology for the use of cubic spline functions in assessing time-by-covariate interactions in Cox proportional hazards regression models. These interactions indicate violations of the proportional hazards assumption of the Cox model. Use of cubic spline functions allows for the investigation of the shape of a possible covariate time-dependence without having to specify a particular functional form. Cubic spline functions yield both a graphical method and a formal test for the proportional hazards assumption as well as a test of the nonlinearity of the time-by-covariate interaction. Five existing methods for assessing violations of the proportional hazards assumption are reviewed and applied along with cubic splines to three well known two-sample datasets. An additional dataset with three covariates is used to explore the use of cubic spline functions in a more general setting. ^

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The research presented in this article aims to identify and to analyze the cases of environmental conflicts in the microrregião de Viçosa, Minas Gerais, Brasil. For this were performed data surveys were conducted in the archives of the Ministério Público regarding the municipalities of microrregião in study and a workshop with various social movements of the mesorregião da Zona da Mata. From this survey on we check the conflict established between environmental legislation and farm workers, and also the concentration of the penalty on small cases of violations of environmental legislation. Thus, we conclude that to think about environmental conflicts is require the recognition of inequalities of power and the different types of knowledge and rationalities involved in the appropriation of natural resources by society.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The research presented in this article aims to identify and to analyze the cases of environmental conflicts in the microrregião de Viçosa, Minas Gerais, Brasil. For this were performed data surveys were conducted in the archives of the Ministério Público regarding the municipalities of microrregião in study and a workshop with various social movements of the mesorregião da Zona da Mata. From this survey on we check the conflict established between environmental legislation and farm workers, and also the concentration of the penalty on small cases of violations of environmental legislation. Thus, we conclude that to think about environmental conflicts is require the recognition of inequalities of power and the different types of knowledge and rationalities involved in the appropriation of natural resources by society.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The research presented in this article aims to identify and to analyze the cases of environmental conflicts in the microrregião de Viçosa, Minas Gerais, Brasil. For this were performed data surveys were conducted in the archives of the Ministério Público regarding the municipalities of microrregião in study and a workshop with various social movements of the mesorregião da Zona da Mata. From this survey on we check the conflict established between environmental legislation and farm workers, and also the concentration of the penalty on small cases of violations of environmental legislation. Thus, we conclude that to think about environmental conflicts is require the recognition of inequalities of power and the different types of knowledge and rationalities involved in the appropriation of natural resources by society.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The BSRN Toolbox is a software package supplied by the WRMC and is freely available to all station scientists and data users. The main features of the package include a download manager for Station- to-Archive files, a tool to convert files into human readable TAB-separated ASCII-tables (similar to those output by the PANGAEA database), and a tool to check data sets for violations of the "BSRN Global Network recommended QC tests, V2.0" quality criteria. The latter tool creates quality codes, one per measured value, indicating if the data are "physically possible," "extremely rare," or if "intercomparison limits are exceeded." In addition, auxiliary data such as solar zenith angle or global calculated from diffuse and direct can be output. All output from the QC tool can be visualized using PanPlot (doi:10.1594/PANGAEA.816201).

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Since the year 2000 when the Protocol to Prevent, Suppress and Punish Trafficking in Persons, Especially Women and Children, human trafficking has been regarded as one of the egregious violations of human rights, and global efforts have been made to eradicate it. The anti-trafficking framework has multiple dimensions, and the way the anti-trafficking framework is constructed influences its impact on the victims and non-trafficked migrants. This paper will analyze the impact of the anti-trafficking framework on the experiences of Burmese victims and non-trafficked migrants in Thailand. I will question the conventional framework of anti-trafficking, and seek to construct a framework more appropriate for addressing victims' actual needs. In conclusion, the anti-trafficking framework should serve the best interest of the victim; still, it should not be one which might adversely affect the interest of the would-be victim who is not identified as a victim according to the law.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper explains how the Armington-Krugman-Melitz supermodel developed by Dixon and Rimmer can be parameterized, and demonstrates that only two kinds of additional information are required in order to extend a standard trade model to include Melitz-type monopolistic competition and heterogeneous firms. Further, it is shown how specifying too much additional information leads to violations of the model constraints, necessitating adjustment and reconciliation of the data. Once a Melitz-type model is parameterized, a Krugman-type model can also be parameterized using the calibrated values in the Melitz-type model without any additional data. Sample code for the General Algebraic Modeling System (GAMS) has also been prepared to promote the innovative supermodel in the AGE community.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We have designed and implemented a framework that unifies unit testing and run-time verification (as well as static verification and static debugging). A key contribution of our approach is that a unified assertion language is used for all of these tasks. We first propose methods for compiling runtime checks for (parts of) assertions which cannot be verified at compile-time via program transformation. This transformation allows checking preconditions and postconditions, including conditional postconditions, properties at arbitrary program points, and certain computational properties. The implemented transformation includes several optimizations to reduce run-time overhead. We also propose a minimal addition to the assertion language which allows defining unit tests to be run in order to detect possible violations of the (partial) specifications expressed by the assertions. This language can express for example the input data for performing the unit tests or the number of times that the unit tests should be repeated. We have implemented the framework within the Ciao/CiaoPP system and effectively applied it to the verification of ISO-prolog compliance and to the detection of different types of bugs in the Ciao system source code. Several experimental results are presented that ¡Ilústrate different trade-offs among program size, running time, or levéis of verbosity of the messages shown to the user.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We propose a general framework for assertion-based debugging of constraint logic programs. Assertions are linguistic constructions for expressing properties of programs. We define several assertion schemas for writing (partial) specifications for constraint logic programs using quite general properties, including user-defined programs. The framework is aimed at detecting deviations of the program behavior (symptoms) with respect to the given assertions, either at compile-time (i.e., statically) or run-time (i.e., dynamically). We provide techniques for using information from global analysis both to detect at compile-time assertions which do not hold in at least one of the possible executions (i.e., static symptoms) and assertions which hold for all possible executions (i.e., statically proved assertions). We also provide program transformations which introduce tests in the program for checking at run-time those assertions whose status cannot be determined at compile-time. Both the static and the dynamic checking are provably safe in the sense that all errors flagged are definite violations of the pecifications. Finally, we report briefly on the currently implemented instances of the generic framework.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We propose a general framework for assertion-based debugging of constraint logic programs. Assertions are linguistic constructions which allow expressing properties of programs. We define assertion schemas which allow writing (partial) specifications for constraint logic programs using quite general properties, including user-defined programs. The framework is aimed at detecting deviations of the program behavior (symptoms) with respect to the given assertions, either at compile-time or run-time. We provide techniques for using information from global analysis both to detect at compile-time assertions which do not hold in at least one of the possible executions (i.e., static symptoms) and assertions which hold for all possible executions (i.e., statically proved assertions). We also provide program transformations which introduce tests in the program for checking at run-time those assertions whose status cannot be determined at compile-time. Both the static and the dynamic checking are provably safe in the sense that all errors flagged are definite violations of the specifications. Finally, we report on an implemented instance of the assertion language and framework.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

As the number of data sources publishing their data on the Web of Data is growing, we are experiencing an immense growth of the Linked Open Data cloud. The lack of control on the published sources, which could be untrustworthy or unreliable, along with their dynamic nature that often invalidates links and causes conflicts or other discrepancies, could lead to poor quality data. In order to judge data quality, a number of quality indicators have been proposed, coupled with quality metrics that quantify the “quality level” of a dataset. In addition to the above, some approaches address how to improve the quality of the datasets through a repair process that focuses on how to correct invalidities caused by constraint violations by either removing or adding triples. In this paper we argue that provenance is a critical factor that should be taken into account during repairs to ensure that the most reliable data is kept. Based on this idea, we propose quality metrics that take into account provenance and evaluate their applicability as repair guidelines in a particular data fusion setting.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Ciao is a logic-based, multi-paradigm programming system. One of its most distinguishing features is that it supports a large number of semantic and syntactic language features which can be selectively activated or deactivated for each program module. As a result, a module can be written in, for example, ISO-Prolog plus constraints and higher order, while another can be a puré logic module with a different control rule such as iterative deepening and/or tabling, and perhaps using constructive negation. A powerful and modular extensión mechanism allows user-level design and implementation of such features and sub-languages. Another distinguishing feature of Ciao is its powerful assertion language, which allows expressing many kinds of program properties (ranging from, e.g., moded types to resource consumption), as well as tests and documentation. The compiler is capable of statically ñnding violations of these properties or verifying that programs comply with them, and issuing certiñcates of this compliance. The compiler also performs many types of optimizations, including automatic parallelization. It offers very competitive performance, while retaining the flexibility and interactive development of a dynamic language. We will present a hands-on overview of the system, through small examples which emphasize the novel aspects and the motivations which lie behind Ciao's design and implementation.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We have designed and implemented a framework that unifies unit testing and run-time verification (as well as static verification and static debugging). A key contribution of our approach is that a unified assertion language is used for all of these tasks. We first propose methods for compiling runtime checks for (parts of) assertions which cannot be verified at compile-time via program transformation. This transformation allows checking preconditions and postconditions, including conditional postconditions, properties at arbitrary program points, and certain computational properties. The implemented transformation includes several optimizations to reduce run-time overhead. We also propose a minimal addition to the assertion language which allows defining unit tests to be run in order to detect possible violations of the (partial) specifications expressed by the assertions. This language can express for example the input data for performing the unit tests or the number of times that the unit tests should be repeated. We have implemented the framework within the Ciao/CiaoPP system and effectively applied it to the verification of ISO-prolog compliance and to the detection of different types of bugs in the Ciao system source code. Several experimental results are presented that illustrate different trade-offs among program size, running time, or levels of verbosity of the messages shown to the user.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

La computación basada en servicios (Service-Oriented Computing, SOC) se estableció como un paradigma ampliamente aceptado para el desarollo de sistemas de software flexibles, distribuidos y adaptables, donde las composiciones de los servicios realizan las tareas más complejas o de nivel más alto, frecuentemente tareas inter-organizativas usando los servicios atómicos u otras composiciones de servicios. En tales sistemas, las propriedades de la calidad de servicio (Quality of Service, QoS), como la rapídez de procesamiento, coste, disponibilidad o seguridad, son críticas para la usabilidad de los servicios o sus composiciones en cualquier aplicación concreta. El análisis de estas propriedades se puede realizarse de una forma más precisa y rica en información si se utilizan las técnicas de análisis de programas, como el análisis de complejidad o de compartición de datos, que son capables de analizar simultáneamente tanto las estructuras de control como las de datos, dependencias y operaciones en una composición. El análisis de coste computacional para la composicion de servicios puede ayudar a una monitorización predictiva así como a una adaptación proactiva a través de una inferencia automática de coste computacional, usando los limites altos y bajos como funciones del valor o del tamaño de los mensajes de entrada. Tales funciones de coste se pueden usar para adaptación en la forma de selección de los candidatos entre los servicios que minimizan el coste total de la composición, basado en los datos reales que se pasan al servicio. Las funciones de coste también pueden ser combinadas con los parámetros extraídos empíricamente desde la infraestructura, para producir las funciones de los límites de QoS sobre los datos de entrada, cuales se pueden usar para previsar, en el momento de invocación, las violaciones de los compromisos al nivel de servicios (Service Level Agreements, SLA) potenciales or inminentes. En las composiciones críticas, una previsión continua de QoS bastante eficaz y precisa se puede basar en el modelado con restricciones de QoS desde la estructura de la composition, datos empiricos en tiempo de ejecución y (cuando estén disponibles) los resultados del análisis de complejidad. Este enfoque se puede aplicar a las orquestaciones de servicios con un control centralizado del flujo, así como a las coreografías con participantes multiples, siguiendo unas interacciones complejas que modifican su estado. El análisis del compartición de datos puede servir de apoyo para acciones de adaptación, como la paralelización, fragmentación y selección de los componentes, las cuales son basadas en dependencias funcionales y en el contenido de información en los mensajes, datos internos y las actividades de la composición, cuando se usan construcciones de control complejas, como bucles, bifurcaciones y flujos anidados. Tanto las dependencias funcionales como el contenido de información (descrito a través de algunos atributos definidos por el usuario) se pueden expresar usando una representación basada en la lógica de primer orden (claúsulas de Horn), y los resultados del análisis se pueden interpretar como modelos conceptuales basados en retículos. ABSTRACT Service-Oriented Computing (SOC) is a widely accepted paradigm for development of flexible, distributed and adaptable software systems, in which service compositions perform more complex, higher-level, often cross-organizational tasks using atomic services or other service compositions. In such systems, Quality of Service (QoS) properties, such as the performance, cost, availability or security, are critical for the usability of services and their compositions in concrete applications. Analysis of these properties can become more precise and richer in information, if it employs program analysis techniques, such as the complexity and sharing analyses, which are able to simultaneously take into account both the control and the data structures, dependencies, and operations in a composition. Computation cost analysis for service composition can support predictive monitoring and proactive adaptation by automatically inferring computation cost using the upper and lower bound functions of value or size of input messages. These cost functions can be used for adaptation by selecting service candidates that minimize total cost of the composition, based on the actual data that is passed to them. The cost functions can also be combined with the empirically collected infrastructural parameters to produce QoS bounds functions of input data that can be used to predict potential or imminent Service Level Agreement (SLA) violations at the moment of invocation. In mission-critical applications, an effective and accurate continuous QoS prediction, based on continuations, can be achieved by constraint modeling of composition QoS based on its structure, known data at runtime, and (when available) the results of complexity analysis. This approach can be applied to service orchestrations with centralized flow control, and choreographies with multiple participants with complex stateful interactions. Sharing analysis can support adaptation actions, such as parallelization, fragmentation, and component selection, which are based on functional dependencies and information content of the composition messages, internal data, and activities, in presence of complex control constructs, such as loops, branches, and sub-workflows. Both the functional dependencies and the information content (described using user-defined attributes) can be expressed using a first-order logic (Horn clause) representation, and the analysis results can be interpreted as a lattice-based conceptual models.