35 resultados para Checks
em Universidad Politécnica de Madrid
Resumo:
El objetivo de este trabajo es generar un modelo Edafogeomorfológico útil en la identificación de necesidades de manejo de suelos, se realizó un estudio en el área de la cuenca Cañada La Gorda Machiques-Colón, estado Zulia, Venezuela, caracterizada por un clima tropical de condición subhúmeda, con duración del periodo de crecimiento (DPC) de 230 días, régimen de humedad Ustic y de temperatura Isohipertémico. Se empleó el enfoque de la ecuación factorial de formación de suelos para el análisis y descripción biofísica de los factores a lo largo de una carena. El relieve caracterizado a partir de fotografías aéreas, imágenes de satélites y de chequeos sistemático mediante transectos en el sentido del flujo del escurrimiento; la vegetación a través del uso de la tierra, la cobertura vegetal, la identificación de las especies dominantes a partir de sus nombres vernáculos y la definición de indicadores de vegetales (Iv). Los suelos fueron descritos y clasificados según la Taxonomía de suelos y valorados mediante el modelo paramétricode Riquier et al. (1970) para determinar el índice de productividad (Ip). Se caracterizaron dos paisajes gemorfológicos: Colinar (C) y Valle (V), seis posiciones geomorfológicas entre ambos paisajes definidas por la sucesión de relieves en el sentido de la pendiente: Tope de colina-loma (TC), mesa conservada (MC), vertientes de mesa alta (VA), media (VM), baja (VB) y valle intracolinar (VI); e igual número de perfiles de suelos representativos, los cuales mostraron edafogénesis muy avanzada con Ip inferiores a 8% en todas las posiciones, exceptuando la VB, con una productividad de 13%. El uso de la tierra es a base de pastoreo semi-intensivo de plantas forrajeras introducidas. Las formaciones vegetales predominante fueron los matorrales y arbustales dispersos, acompañados con restos de una selva tropófila fuertemente afectada por la extracción forestal y la conversión en áreas de pastoreo. Se identificaron 8 Iv, asociados fuertemente con condiciones de físicas e hidrológicas del suelo. El alto impacto de las actividades humanas sobre el suelo y vegetación, expresado a través de los procesos de erosión activa, la ausencia de áreas boscosas y la baja productividad de los sistemas de ganadería reportada para la zona, señalan la necesidad de reorientación del uso actual de la tierra, para lo cual se plantean alternativas como la incorporación de bosques protectores y sistemas agrosilvopastoriles In order to generate an Edaphogeomorphological model to be used for the identification of management requirements of soils, a study was carried out in the area of the Cañada La Gorda watershed, Machiques Colon, Zulia State with a tropical climate, subsumid conditions with a growing period of 230 days, an Ustic soil moisture and Isohypertermic regimes. The soils factorial equation approach was used for the analisis and description of the factors of soil formation throughout a soil catena. Relief was characterized through aerial photographs, satellite images and systematic checks of transects drawn in the sense of surface runoff and also taking into account geomorphological features. Vegetation cover and land use were described and vegetation components were indified by its local names to defined vegetations indicators (VI) for the local conditions. Soils were described and classified according to soil taxonomy and valued by means of a parametric model proposed by Riquier et al, (1970) for determining the productivity index (PI). Two geomorphological landscapes were defined: Hilly and Valley with six positions within the landscapes: hilltop (round or elongated), preserved tableland summit, slopes of high, medium and low tableland and valleys between hills. Representative soils of each position were studied showing a highly advanced degree of edaphogenesis with PI values below 8% in all positions except the valleys with a PI of 13%. Land use type is based on semi intensive pasturing of introduced forage species, with a vegetation of brushwood and scattered shrubs, with some trees relicts of woods affected by timber extraction and turn to grassland Eight VI were identified, highly associated to local physical and hidrological soil conditions. The enormous impact of human activity on soils and vegetation as shown by active erosion processes and absence of wooded areas and the low productivity of livestock systems reported for the area, indicates the necessity of a reorientation of the present land use introducing alternatives like the incorporation of protective woods and agrosilvopastoral management systems.
Resumo:
The progressive depletion of fossil fuels and their high contribution to the energy supply in this modern society forces that will be soon replaced by renewable fuels. But the dispersion and alternation of renewable energy production also undertake to reduce their costs to use as energy storage and hydrogen carrier. It is necessary to develop technologies for hydrogen production from all renewable energy storage technologies and the development of energy production from hydrogen fuel cells and cogeneration and tri generation systems. In order to propel this technological development discussed where the hydrogen plays a key role as energy storage and renewable energy, the National Centre of Hydrogen and Fuel Cell Technology Experimentation in Spain equipped with installations that enable scientific and technological design, develop, verify, certify, approve, test, measure and, more importantly, the facility ensures continuous operation for 24 hours a day, 365 days year. At the same time, the system is scalable so as to allow continuous adaptation of new technologies are developed and incorporated into the assembly to verify integration at the same time it checks the validity of their development. The transformation sector can be said to be the heart of the system, because without neglecting the other sectors, this should prove the validity of hydrogen as a carrier - energy storage are important efforts that have to do to demonstrate the suitability of fuel cells or internal combustion systems to realize the energy stored in hydrogen at prices competitive with conventional systems. The multiple roles to meet the fuel cells under different conditions of operation require to cover their operating conditions, many different sizes and applications. The fourth area focuses on integration is an essential complement within the installation. We must integrate not only the electricity produced, but also hydrogen is used and the heat generated in the process of using hydrogen energy. The energy management in its three forms: hydrogen chemical, electrical and thermal integration requires complicated and require a logic and artificial intelligence extremes to ensure maximum energy efficiency at the same time optimum utilization is achieved. Verification of the development and approval in the entire production system and, ultimately, as a demonstrator set to facilitate the simultaneous evolution of production technology, storage and distribution of hydrogen fuel cells has been assessed.
Resumo:
Large cross section structural timber have been used in many structures over long periods of time and still make up an important part of the market due to its mechanical properties. Furthermore, it is frequent its employment in new construction site. It involves the need for a visual grading standard for timber used in construction according to the quality assessment. The material has to satisfy the requirements according to the currently regulations. UNE 56544 is the Spanish visual grading standard for coniferous structural timber. The 2007 version defined a new visual grade in the standard for large section termed Structural Large Timber (MEG). This research checks the new visual grading and consists of 116 structural size specimens of sawn coniferous timber of Scotch pine (Pinus sylvestris L.) from Segovia, Spain. The pieces had a cross section of 150 by 200 mm. They were visually graded according to UNE 56544:2007. Also, mechanical properties have been obtained according to standard EN 408. The results show very low output with an excessive percentage of rejected pieces (33%). The main reasons for the rejection of pieces are fissures and twist
Resumo:
During the past years, the industry has shifted position and moved towards “the luxury universe” whose customers are demanding, treating individuals as unique and valued customer for the business, offering vehicles produced with the state of the art technologies and implementing the highest finishing standards. Due to the competitive level in the market, car makers enable processes which equalizes customer services to E.R. management, being dealt with the maximum urgency that allows the comparison between both, car workshops and emergency rooms, where workshop bays or ramps will be equal to emergency boxes and skilled technicians are equivalent to the health care specialist, who will carry out tests and checks prior to afford any final operation, keeping the “patient” under control before it is back to normal utilization. This paper establishes a valid model for the automotive industry to estimate customer service demand forecasting under variable demand conditions using analogies with patient demand models used for the medical ER.
Resumo:
During the past years, the industry has shifted position and moved towards “the luxury universe” whose customers are demanding, treating individuals as unique and valued customer for the business, offering vehicles produced with the state of the art technologies and implementing the highest finishing standards. Due to the competitive level in the market, motor makers enable processes which equalizes customer services to E.R. management, being dealt with the maximum urgency that allows the comparison between both, car workshops and emergency rooms, where workshop bays or ramps will be equal to emergency boxes and skilled technicians are equivalent to the health care specialist, who will carry out tests and checks prior to afford any final operation, keeping the “patient” under control before it is back to normal utilization. This paper ratify a valid model for the automotive industry to estimate customer service demand forecasting under variable demand conditions using analogies with patient demand models used for the medical ER
Resumo:
We present two concurrent semantics (i.e. semantics where concurrency is explicitely represented) for CC programs with atomic tells. One is based on simple partial orders of computation steps, while the other one is based on contextual nets and it is an extensión of a previous one for eventual CC programs. Both such semantics allow us to derive concurrency, dependency, and nondeterminism information for the considered languages. We prove some properties about the relation between the two semantics, and also about the relation between them and the operational semantics. Moreover, we discuss how to use the contextual net semantics in the context of CLP programs. More precisely, by interpreting concurrency as possible parallelism, our semantics can be useful for a safe parallelization of some CLP computation steps. Dually, the dependency information may also be interpreted as necessary sequentialization, thus possibly exploiting it for the task of scheduling CC programs. Moreover, our semantics is also suitable for CC programs with a new kind of atomic tell (called locally atomic tell), which checks for consistency only the constraints it depends on. Such a tell achieves a reasonable trade-off between efficiency and atomicity, since the checked constraints can be stored in a local memory and are thus easily accessible even in a distributed implementation.
Resumo:
This paper addresses the issue of the practicality of global flow analysis in logic program compilation, in terms of speed of the analysis, precisión, and usefulness of the information obtained. To this end, design and implementation aspects are discussed for two practical abstract interpretation-based flow analysis systems: MA , the MCC And-parallel Analyzer and Annotator; and Ms, an experimental mode inference system developed for SB-Prolog. The paper also provides performance data obtained (rom these implementations and, as an example of an application, a study of the usefulness of the mode information obtained in reducing run-time checks in independent and-parallelism.Based on the results obtained, it is concluded that the overhead of global flow analysis is not prohibitive, while the results of analysis can be quite precise and useful.
Resumo:
Traditional schemes for abstract interpretation-based global analysis of logic programs generally focus on obtaining procedure argument mode and type information. Variable sharing information is often given only the attention needed to preserve the correctness of the analysis. However, such sharing information can be very useful. In particular, it can be used for predicting runtime goal independence, which can eliminate costly run-time checks in and-parallel execution. In this paper, a new algorithm for doing abstract interpretation in logic programs is described which concentrates on inferring the dependencies of the terms bound to program variables with increased precisión and at all points in the execution of the program, rather than just at a procedure level. Algorithms are presented for computing abstract entry and success substitutions which extensively keep track of variable aliasing and term dependence information. In addition, a new, abstract domain independent ñxpoint algorithm is presented and described in detail. The algorithms are illustrated with examples. Finally, results from an implementation of the abstract interpreter are presented.
Resumo:
We have designed and implemented a framework that unifies unit testing and run-time verification (as well as static verification and static debugging). A key contribution of our approach is that a unified assertion language is used for all of these tasks. We first propose methods for compiling runtime checks for (parts of) assertions which cannot be verified at compile-time via program transformation. This transformation allows checking preconditions and postconditions, including conditional postconditions, properties at arbitrary program points, and certain computational properties. The implemented transformation includes several optimizations to reduce run-time overhead. We also propose a minimal addition to the assertion language which allows defining unit tests to be run in order to detect possible violations of the (partial) specifications expressed by the assertions. This language can express for example the input data for performing the unit tests or the number of times that the unit tests should be repeated. We have implemented the framework within the Ciao/CiaoPP system and effectively applied it to the verification of ISO-prolog compliance and to the detection of different types of bugs in the Ciao system source code. Several experimental results are presented that ¡Ilústrate different trade-offs among program size, running time, or levéis of verbosity of the messages shown to the user.
Resumo:
Proof carrying code (PCC) is a general is originally a roof in ñrst-order logic of certain vermethodology for certifying that the execution of an un- ification onditions and the checking process involves trusted mobile code is safe. The baste idea is that the ensuring that the certifícate is indeed a valid ñrst-order code supplier attaches a certifícate to the mobile code proof. which the consumer checks in order to ensure that the The main practical difñculty of PCC techniques is in code is indeed safe. The potential benefit is that the generating safety certiñeates which at the same time: i) consumer's task is reduced from the level of proving to allow expressing interesting safety properties, ii) can be the level of checking. Recently, the abstract interpre- generated automatically and, iii) are easy and efficient tation techniques developed, in logic programming have to check. In [1], the abstract interpretation techniques been proposed as a basis for PCC. This extended ab- [5] developed in logic programming1 are proposed as stract reports on experiments which illustrate several is- a basis for PCC. They offer a number of advantages sues involved in abstract interpretation-based certifica- for dealing with the aforementioned issues. In particution. First, we describe the implementation of our sys- lar, the xpressiveness of existing abstract domains will tem in the context of CiaoPP: the preprocessor of the be implicitly available in abstract interpretation-based Ciao multi-paradigm programming system. Then, by code certification to deñne a wide range of safety propermeans of some experiments, we show how code certifi- ties. Furthermore, the approach inherits the automation catión is aided in the implementation of the framework. and inference power of the abstract interpretation en- Finally, we discuss the application of our method within gines used in (Constraint) Logic Programming, (C)LP. the área, of pervasive systems
Resumo:
We present a concurrent semantics (i.e. a semantics where concurrency is explicitely represented) for CC programs with atomic tells. This allows to derive concurrency, dependency, and nondeterminism information for such languages. The ability to treat failure information puts CLP programs also in the range of applicability of our semantics: although such programs are not concurrent, the concurrency information derived in the semantics may be interpreted as possible parallelism, thus allowing to safely parallelize those computation steps which appear to be concurrent in the net. Dually, the dependency information may also be interpreted as necessary sequentialization, thus possibly exploiting it to schedule CC programs. The fact that the semantical structure contains dependency information suggests a new tell operation, which checks for consistency only the constraints it depends on, achieving a reasonable trade-off between efficiency and atomicity.
Resumo:
Logic programming systems which exploit and-parallelism among non-deterministic goals rely on notions of independence among those goals in order to ensure certain efficiency properties. "Non-strict" independence (NSI) is a more relaxed notion than the traditional notion of "strict" independence (SI) which still ensures the relevant efficiency properties and can allow considerable more parallelism than SI. However, all compilation technology developed to date has been based on SI, because of the intrinsic complexity of exploiting NSI. This is related to the fact that NSI cannot be determined "a priori" as SI. This paper filis this gap by developing a technique for compile-time detection and annotation of NSI. It also proposes algorithms for combined compiletime/ run-time detection, presenting novel run-time checks for this type of parallelism. Also, a transformation procedure to eliminate shared variables among parallel goals is presented, aimed at performing as much work as possible at compile-time. The approach is based on the knowledge of certain properties regarding the run-time instantiations of program variables —sharing and freeness— for which compile-time technology is available, with new approaches being currently proposed. Thus, the paper does not deal with the analysis itself, but rather with how the analysis results can be used to parallelize programs.
Resumo:
This paper presents a technique for achieving a class of optimizations related to the reduction of checks within cycles. The technique uses both Program Transformation and Abstract Interpretation. After a ñrst pass of an abstract interpreter which detects simple invariants, program transformation is used to build a hypothetical situation that simpliñes some predicates that should be executed within the cycle. This transformation implements the heuristic hypothesis that once conditional tests hold they may continué doing so recursively. Specialized versions of predicates are generated to detect and exploit those cases in which the invariance may hold. Abstract interpretation is then used again to verify the truth of such hypotheses and conñrm the proposed simpliñcation. This allows optimizations that go beyond those possible with only one pass of the abstract interpreter over the original program, as is normally the case. It also allows selective program specialization using a standard abstract interpreter not speciñcally designed for this purpose, thus simplifying the design of this already complex module of the compiler. In the paper, a class of programs amenable to such optimization is presented, along with some examples and an evaluation of the proposed techniques in some application áreas such as floundering detection and reducing run-time tests in automatic logic program parallelization. The analysis of the examples presented has been performed automatically by an implementation of the technique using existing abstract interpretation and program transformation tools.
Resumo:
Traditional schemes for abstract interpretation-based global analysis of logic programs generally focus on obtaining procedure argument mode and type information. Variable sharing information is often given only the attention needed to preserve the correctness of the analysis. However, such sharing information can be very useful. In particular, it can be used for predicting run-time goal independence, which can eliminate costly run-time checks in and-parallel execution. In this paper, a new algorithm for doing abstract interpretation in logic programs is described which infers the dependencies of the terms bound to program variables with increased precisión and at all points in the execution of the program, rather than just at a procedure level. Algorithms are presented for computing abstract entry and success substitutions which extensively keep track of variable aliasing and term dependence information. The algorithms are illustrated with examples.
Resumo:
Although the sequential execution speed of logic programs has been greatly improved by the concepts introduced in the Warren Abstract Machine (WAM), parallel execution represents the only way to increase this speed beyond the natural limits of sequential systems. However, most proposed parallel logic programming execution models lack the performance optimizations and storage efficiency of sequential systems. This paper presents a parallel abstract machine which is an extension of the WAM and is thus capable of supporting ANDParallelism without giving up the optimizations present in sequential implementations. A suitable instruction set, which can be used as a target by a variety of logic programming languages, is also included. Special instructions are provided to support a generalized version of "Restricted AND-Parallelism" (RAP), a technique which reduces the overhead traditionally associated with the run-time management of variable binding conflicts to a series of simple run-time checks, which select one out of a series of compiled execution graphs.