15 resultados para ergonomic approach to the activity

em Universidad Politécnica de Madrid


Relevância:

100.00% 100.00%

Publicador:

Resumo:

The European construction industry is supposed to consume the 40% of the natural European resources and to generate the 40% of the European solid waste. Conscious of the great damage being suffered by the environment because of construction activity, this work tries to provide the building actors with a new tool to improve the current situation. The tool proposed is a model for the comprehensive evaluation of construction products by determining their environmental level. In this research, the environmental level of a construction product has been defined as its quality of accomplishing the construction requirements needed by causing the minimum ecological impact in its surrounding environment. This information allows building actors to choose suitable materials for building needs and also for the environment, mainly in the project stage or on the building site, contributing to improve the relationship between buildings and environment. For the assessment of the environmental level of construction products, five indicators have been identified regarding their global environmental impact through the product life cycle: CO2 emissions provoked during their production, volume and toxicity of waste generated on the building site, durability and recycling capacity after their useful life. Therefore, the less environmental impact one construction product produces, the higher environmental level performs. The model has been tested in 30 construction products that include environmental criteria in their description. The results obtained will be discussed in this article. Furthermore, this model can lay down guidelines for the selection of ecoefficient construction products and the design of new eco-competitive and eco-committed ones

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The aim of this work is to solve a question raised for average sampling in shift-invariant spaces by using the well-known matrix pencil theory. In many common situations in sampling theory, the available data are samples of some convolution operator acting on the function itself: this leads to the problem of average sampling, also known as generalized sampling. In this paper we deal with the existence of a sampling formula involving these samples and having reconstruction functions with compact support. Thus, low computational complexity is involved and truncation errors are avoided. In practice, it is accomplished by means of a FIR filter bank. An answer is given in the light of the generalized sampling theory by using the oversampling technique: more samples than strictly necessary are used. The original problem reduces to finding a polynomial left inverse of a polynomial matrix intimately related to the sampling problem which, for a suitable choice of the sampling period, becomes a matrix pencil. This matrix pencil approach allows us to obtain a practical method for computing the compactly supported reconstruction functions for the important case where the oversampling rate is minimum. Moreover, the optimality of the obtained solution is established.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Static analyses of object-oriented programs usually rely on intermediate representations that respect the original semantics while having a more uniform and basic syntax. Most of the work involving object-oriented languages and abstract interpretation usually omits the description of that language or just refers to the Control Flow Graph(CFG) it represents. However, this lack of formalization on one hand results in an absence of assurances regarding the correctness of the transformation and on the other it typically strongly couples the analysis to the source language. In this work we present a framework for analysis of object-oriented languages in which in a first phase we transform the input program into a representation based on Horn clauses. This allows on one hand proving the transformation correct attending to a simple condition and on the other being able to apply an existing analyzer for (constraint) logic programming to automatically derive a safe approximation of the semantics of the original program. The approach is flexible in the sense that the first phase decouples the analyzer from most languagedependent features, and correct because the set of Horn clauses returned by the transformation phase safely approximates the standard semantics of the input program. The resulting analysis is also reasonably scalable due to the use of mature, modular (C)LP-based analyzers. The overall approach allows us to report results for medium-sized programs.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper presents and illustrates with an example a practical approach to the dataflow analysis of programs written in constraint logic programming (CLP) languages using abstract interpretation. It is first argued that, from the framework point of view, it sufnces to propose relatively simple extensions of traditional analysis methods which have already been proved useful and practical and for which efncient fixpoint algorithms have been developed. This is shown by proposing a simple but quite general extensión of Bruynooghe's traditional framework to the analysis of CLP programs. In this extensión constraints are viewed not as "suspended goals" but rather as new information in the store, following the traditional view of CLP. Using this approach, and as an example of its use, a complete, constraint system independent, abstract analysis is presented for approximating definiteness information. The analysis is in fact of quite general applicability. It has been implemented and used in the analysis of CLP(R) and Prolog-III applications. Results from the implementation of this analysis are also presented.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A boundary element approach for time harmonic axisymmetric problems using the complete space point load fundamental solution is presented. The fundamental solution is integrated numerically along the azimuthal co-ordinate of each axisymmetric element. To increase the accuracy of the numerical integration a simple co-ordinate transformation is proposed. The approach is applied to the computation of the dynamic stiffness functions of rigid circular foundations on layered viscoelastic soils. Three different sites are considered: a uniform half-space, a soil layer on a half-space, and a soil consisting of four horizontal layers and a compliant half-space. The numerical results obtained by the proposed approach for surface circular foundations are very close to corresponding published results obtained by different procedures.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The Kolmogorov approach to turbulence is applied to the Burgers turbulence in the stochastic adhesion model of large-scale structure formation. As the perturbative approach to this model is unreliable, here a new, non-perturbative approach, based on a suitable formulation of Kolmogorov's scaling laws, is proposed. This approach suggests that the power-law exponent of the matter density two-point correlation function is in the range 1–1.33, but it also suggests that the adhesion model neglects important aspects of the gravitational dynamics.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A linear method is developed for solving the nonlinear differential equations of a lumped-parameter thermal model of a spacecraft moving in a closed orbit. This method, based on perturbation theory, is compared with heuristic linearizations of the same equations. The essential feature of the linear approach is that it provides a decomposition in thermal modes, like the decomposition of mechanical vibrations in normal modes. The stationary periodic solution of the linear equations can be alternately expressed as an explicit integral or as a Fourier series. This method is applied to a minimal thermal model of a satellite with ten isothermal parts (nodes), and the method is compared with direct numerical integration of the nonlinear equations. The computational complexity of this method is briefly studied for general thermal models of orbiting spacecraft, and it is concluded that it is certainly useful for reduced models and conceptual design but it can also be more efficient than the direct integration of the equations for large models. The results of the Fourier series computations for the ten-node satellite model show that the periodic solution at the second perturbative order is sufficiently accurate.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

There is general agreement within the scientific community in considering Biology as the science with more potential to develop in the XXI century. This is due to several reasons, but probably the most important one is the state of development of the rest of experimental and technological sciences. In this context, there are a very rich variety of mathematical tools, physical techniques and computer resources that permit to do biological experiments that were unbelievable only a few years ago. Biology is nowadays taking advantage of all these newly developed technologies, which are been applied to life sciences opening new research fields and helping to give new insights in many biological problems. Consequently, biologists have improved a lot their knowledge in many key areas as human function and human diseases. However there is one human organ that is still barely understood compared with the rest: The human brain. The understanding of the human brain is one of the main challenges of the XXI century. In this regard, it is considered a strategic research field for the European Union and the USA. Thus, there is a big interest in applying new experimental techniques for the study of brain function. Magnetoencephalography (MEG) is one of these novel techniques that are currently applied for mapping the brain activity1. This technique has important advantages compared to the metabolic-based brain imagining techniques like Functional Magneto Resonance Imaging2 (fMRI). The main advantage is that MEG has a higher time resolution than fMRI. Another benefit of MEG is that it is a patient friendly clinical technique. The measure is performed with a wireless set up and the patient is not exposed to any radiation. Although MEG is widely applied in clinical studies, there are still open issues regarding data analysis. The present work deals with the solution of the inverse problem in MEG, which is the most controversial and uncertain part of the analysis process3. This question is addressed using several variations of a new solving algorithm based in a heuristic method. The performance of those methods is analyzed by applying them to several test cases with known solutions and comparing those solutions with the ones provided by our methods.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Bus rapid transit (BRT) systems are massive transport systems with medium/high capacity, high quality service and low infrastructure and operating costs. TransMilenio is Bogotá's most important mass transportation system and one of the biggest BRT systems in the world, although it only has completed its third construction phase out of a total of eight. In this paper we review the proposals in the literature to optimize BRT system operation, with a special emphasis on TransMilenio, and propose a mathematical model that adapts elements of the above proposals and incorporates novel elements accounting for the features of TransMilenio system.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A method for formulating and algorithmically solving the equations of finite element problems is presented. The method starts with a parametric partition of the domain in juxtaposed strips that permits sweeping the whole region by a sequential addition (or removal) of adjacent strips. The solution of the difference equations constructed over that grid proceeds along with the addition removal of strips in a manner resembling the transfer matrix approach, except that different rules of composition that lead to numerically stable algorithms are used for the stiffness matrices of the strips. Dynamic programming and invariant imbedding ideas underlie the construction of such rules of composition. Among other features of interest, the present methodology provides to some extent the analyst's control over the type and quantity of data to be computed. In particular, the one-sweep method presented in Section 9, with no apparent counterpart in standard methods, appears to be very efficient insofar as time and storage is concerned. The paper ends with the presentation of a numerical example

Relevância:

100.00% 100.00%

Publicador:

Resumo:

When applying computational mathematics in practical applications, even though one may be dealing with a problem that can be solved algorithmically, and even though one has good algorithms to approach the solution, it can happen, and often it is the case, that the problem has to be reformulated and analyzed from a different computational point of view. This is the case of the development of approximate algorithms. This paper frames in the research area of approximate algebraic geometry and commutative algebra and, more precisely, on the problem of the approximate parametrization.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Growing energy crops on marginal land has been promoted as a way of ensuring that biomass production involves an acceptable and sustainable use of land. Saline and saline-prone agricultural lands represent an opportunity for growing energy crops avoiding the displacement of food production and contributing to restoration of degraded land. Giant reed (Arundo donax L.) is a perennial grass that has been proposed as a promising energy crop for lignocellulosic biomass production while its tolerance to salinity has been proved. In this work, the identification of surplus saline lands that could be irrigated with saline waters for growing tolerant-energy crops (giant reed) in the mainland of Spain and the assessment of the agronomically attainable yield in these limiting growing conditions were undertaken. To this purpose, a GIS analysis was conducted using geodatabases related to saline areas, agro-climatic conditions, irrigation water requirements, agricultural land availability, restrictions regarding the range of electrical conductivity tolerated by the crop, competition with agro-food crops and irrigation water provisions. According to the approach developed, the irrigated and saline agricultural area available and suitable for biomass production from giant reed amounted up to 34 412 ha. The agronomically attainable yield in these limiting conditions was estimated at 12.7 – 22.2 t dm ha−1 yr−1 and the potential production of lignocellulosic biomass, 597 338 t dm yr−1. The methodology followed in this study can be applied to other target regions; it allows the identification of this type of marginal lands, where salinity-tolerant plant species could be grown for bioenergy purposes, avoiding competition with agro-food crops, and where soil restoration measurements should be undertaken.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Este trabajo aborda la metodología seguida para llevar a cabo el proyecto de investigación PRONAF (Clinical Trials Gov.: number NCT01116856.) Background: At present, scientific consensus exists on the multifactorial etiopatogenia of obesity. Both professionals and researchers agree that treatment must also have a multifactorial approach, including diet, physical activity, pharmacology and/or surgical treatment. These two last ones should be reserved for those cases of morbid obesities or in case of failure of the previous ones. The aim of the PRONAF study is to determine what type of exercise combined with caloric restriction is the most appropriate to be included in overweigth and obesity intervention programs, and the aim of this paper is to describe the design and the evaluation methods used to carry out the PRONAF study. Methods/design: One-hundred nineteen overweight (46 males) and 120 obese (61 males) subjects aged 18–50 years were randomly assigned to a strength training group, an endurance training group, a combined strength + endurance training group or a diet and physical activity recommendations group. The intervention period was 22 weeks (in all cases 3 times/wk of training for 22 weeks and 2 weeks for pre and post evaluation). All subjects followed a hypocaloric diet (25-30% less energy intake than the daily energy expenditure estimated by accelerometry). 29–34% of the total energy intake came from fat, 14–20% from protein, and 50–55% from carbohydrates. The mayor outcome variables assesed were, biochemical and inflamatory markers, body composition, energy balance, physical fitness, nutritional habits, genetic profile and quality of life. 180 (75.3%) subjects finished the study, with a dropout rate of 24.7%. Dropout reasons included: personal reasons 17 (28.8%), low adherence to exercise 3 (5.1%), low adherence to diet 6 (10.2%), job change 6 (10.2%), and lost interest 27 (45.8%). Discussion: Feasibility of the study has been proven, with a low dropout rate which corresponds to the estimated sample size. Transfer of knowledge is foreseen as a spin-off, in order that overweight and obese subjects can benefit from the results. The aim is to transfer it to sports centres. Effectiveness on individual health-related parameter in order to determine the most effective training programme will be analysed in forthcoming publications.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Choosing an appropriate accounting system for manufacturing has always been a challenge for managers. In this article we try to compare three accounting systems designed since 1980 to address problems of traditional accounting system. In the first place we are going to present a short overview on background and definition of three accounting systems: Activity Based costing, Time-Driven Activity Based Costing and Lean Accounting. Comparisons are made based on the three basic roles of information generated by accounting systems: financial reporting, decision making, and operational control and improvement. The analysis in this paper reveals how decisions are made over the value stream in the companies using Lean Accounting while decisions under the ABC Accounting system are taken at individual product level, and finally we will show how TD-ABC covers both product and process levels for decision making. In addition, this paper shows the importance of nonfinancial measures for operational control and improvement under the Lean Accounting and TD-ABC methods whereas ABC relies mostly on financial measures in this context.