34 resultados para Contracts of execution

em Universidad Politécnica de Madrid


Relevância:

100.00% 100.00%

Publicador:

Resumo:

In the year 1999 approves the Law of Construction Building (LOE, in Spanish) to regulate a sector such as construction, which contained some shortcomings from the legal point of view. Currently, the LOE has been in force 12 years, changing the spanish world of the construction, due to influenced by internationalization. Within the LOE, there regulating the different actors involved in the construction building, as the Projects design, the Director of Construction, the developer, The builder, Director of execution of the construction (actor only in Spain, similar as construcion engineer and abroad in), control entities and the users, but lacks figure Project manager will assume the delegation of the promoter helping and you organize, direct and management the process. This figure assumes that the market and contracts are not legally regulated in Spain, then should define and establish its regulation in the LOE. (Spain Construction Law) The translation in spanish of the words "Project Manager is owed to Professor Rafael de Heredia in his book Integrated Project Management, as agent acting on behalf of the organization and promoter assuming control of the project, ie Integraded Project Management . Already exist in Spain, AEDIP (Spanish Association Integrated of Project Construction management) which comprises the major companies in “Project Management” in Spain, and MeDIP (Master in Integrated Construction Project) the largest and most advanced studies at the Polytechnic University of Madrid, in "Construction Project Management" they teach which is also in Argentina. The Integrated Project ("Project Management") applied to the construction process is a methodological technique that helps to organize, control and manage the resources of the promoters in the building process. When resources are limited (which is usually most situations) to manage them efficiently becomes very important. Well, we find that in this situation, the resources are not only limited, but it is limited, so a comprehensive control and monitoring of them becomes not only important if not crucial. The alternative of starting from scratch with a team that specializes in developing these follow directly intervening to ensure that scarce resources are used in the best possible way requires the use of a specific methodology (Manual DIP, Matrix Foreign EDR breakdown structure EDP Project, Risk Management and Control, Design Management, et ..), that is the methodology used by "Projects managers" to ensure that the initial objectives of the promoters or investors are met and all actors in process, from design to construction company have the mind aim of the project will do, trying to get their interests do not prevail over the interests of the project. Among the agents listed in the building process, "Project Management" or DIPE (Director Comprehensive building process, a proposed name for possible incorporation into the LOE, ) currently not listed as such in the LOE (Act on Construction Planning ), one of the agents that exist within the building process is not regulated from the legal point of view, no obligations, ie, as is required by law to have a project, a builder, a construction management, etc. DIPE only one who wants to hire you as have been advanced knowledge of their services by the clients they have been hiring these agents, there being no legal obligation as mentioned above, then the market is dictating its ruling on this new figure, as if it were necessary, he was not hired and eventually disappeared from the building process. As the aim of this article is regular the process and implement the name of DIPE in the Spanish Law of buildings construction (LOE)

Relevância:

100.00% 100.00%

Publicador:

Resumo:

La reproducibilidad de estudios y resultados científicos es una meta a tener en cuenta por cualquier científico a la hora de publicar el producto de una investigación. El auge de la ciencia computacional, como una forma de llevar a cabo estudios empíricos haciendo uso de modelos matemáticos y simulaciones, ha derivado en una serie de nuevos retos con respecto a la reproducibilidad de dichos experimentos. La adopción de los flujos de trabajo como método para especificar el procedimiento científico de estos experimentos, así como las iniciativas orientadas a la conservación de los datos experimentales desarrolladas en las últimas décadas, han solucionado parcialmente este problema. Sin embargo, para afrontarlo de forma completa, la conservación y reproducibilidad del equipamiento computacional asociado a los flujos de trabajo científicos deben ser tenidas en cuenta. La amplia gama de recursos hardware y software necesarios para ejecutar un flujo de trabajo científico hace que sea necesario aportar una descripción completa detallando que recursos son necesarios y como estos deben de ser configurados. En esta tesis abordamos la reproducibilidad de los entornos de ejecución para flujos de trabajo científicos, mediante su documentación usando un modelo formal que puede ser usado para obtener un entorno equivalente. Para ello, se ha propuesto un conjunto de modelos para representar y relacionar los conceptos relevantes de dichos entornos, así como un conjunto de herramientas que hacen uso de dichos módulos para generar una descripción de la infraestructura, y un algoritmo capaz de generar una nueva especificación de entorno de ejecución a partir de dicha descripción, la cual puede ser usada para recrearlo usando técnicas de virtualización. Estas contribuciones han sido aplicadas a un conjunto representativo de experimentos científicos pertenecientes a diferentes dominios de la ciencia, exponiendo cada uno de ellos diferentes requisitos hardware y software. Los resultados obtenidos muestran la viabilidad de propuesta desarrollada, reproduciendo de forma satisfactoria los experimentos estudiados en diferentes entornos de virtualización. ABSTRACT Reproducibility of scientific studies and results is a goal that every scientist must pursuit when announcing research outcomes. The rise of computational science, as a way of conducting empirical studies by using mathematical models and simulations, have opened a new range of challenges in this context. The adoption of workflows as a way of detailing the scientific procedure of these experiments, along with the experimental data conservation initiatives that have been undertaken during last decades, have partially eased this problem. However, in order to fully address it, the conservation and reproducibility of the computational equipment related to them must be also considered. The wide range of software and hardware resources required to execute a scientific workflow implies that a comprehensive description detailing what those resources are and how they are arranged is necessary. In this thesis we address the issue of reproducibility of execution environments for scientific workflows, by documenting them in a formalized way, which can be later used to obtain and equivalent one. In order to do so, we propose a set of semantic models for representing and relating the relevant information of those environments, as well as a set of tools that uses these models for generating a description of the infrastructure, and an algorithmic process that consumes these descriptions for deriving a new execution environment specification, which can be enacted into a new equivalent one using virtualization solutions. We apply these three contributions to a set of representative scientific experiments, belonging to different scientific domains, and exposing different software and hardware requirements. The obtained results prove the feasibility of the proposed approach, by successfully reproducing the target experiments under different virtualization environments.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The price formation of the Iberian Energy Derivatives Market-the power futures market-starting in July 2006, is assessed until November 2011, through the evolution of the difference between forward and spot prices in the delivery period (“ex-post forward risk premium”) and the comparison with the forward generation costs from natural gas (“clean spark spread”). The premium tends to be positive in all existing mechanisms (futures, Over-the-Counter and auctions for catering part of the last resort supplies). Since year 2011, the values are smaller due to regulatorily recognized prices for coal power plants. The power futures are strongly correlated with European gas prices. The spreads built with prompt contracts tend also to be positive. The biggest ones are for the month contract, followed by the quarter contract and then by the year contract. Therefore, gas fired generation companies can maximize profits trading with contracts of shorter maturity.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

El mercado ibérico de futuros de energía eléctrica gestionado por OMIP (“Operador do Mercado Ibérico de Energia, Pólo Português”, con sede en Lisboa), también conocido como el mercado ibérico de derivados de energía, comenzó a funcionar el 3 de julio de 2006. Se analiza la eficiencia de este mercado organizado, por lo que se estudia la precisión con la que sus precios de futuros predicen el precio de contado. En dicho mercado coexisten dos modos de negociación: el mercado continuo (modo por defecto) y la contratación mediante subasta. En la negociación en continuo, las órdenes anónimas de compra y de venta interactúan de manera inmediata e individual con órdenes contrarias, dando lugar a operaciones con un número indeterminado de precios para cada contrato. En la negociación a través de subasta, un precio único de equilibrio maximiza el volumen negociado, liquidándose todas las operaciones a ese precio. Adicionalmente, los miembros negociadores de OMIP pueden liquidar operaciones “Over-The-Counter” (OTC) a través de la cámara de compensación de OMIP (OMIClear). Las cinco mayores empresas españolas de distribución de energía eléctrica tenían la obligación de comprar electricidad hasta julio de 2009 en subastas en OMIP, para cubrir parte de sus suministros regulados. De igual manera, el suministrador de último recurso portugués mantuvo tal obligación hasta julio de 2010. Los precios de equilibrio de esas subastas no han resultado óptimos a efectos retributivos de tales suministros regulados dado que dichos precios tienden a situarse ligeramente sesgados al alza. La prima de riesgo ex-post, definida como la diferencia entre los precios a plazo y de contado en el periodo de entrega, se emplea para medir su eficiencia de precio. El mercado de contado, gestionado por OMIE (“Operador de Mercado Ibérico de la Energía”, conocido tradicionalmente como “OMEL”), tiene su sede en Madrid. Durante los dos primeros años del mercado de futuros, la prima de riesgo media tiende a resultar positiva, al igual que en otros mercados europeos de energía eléctrica y gas natural. En ese periodo, la prima de riesgo ex-post tiende a ser negativa en los mercados de petróleo y carbón. Los mercados de energía tienden a mostrar niveles limitados de eficiencia de mercado. La eficiencia de precio del mercado de futuros aumenta con el desarrollo de otros mecanismos coexistentes dentro del mercado ibérico de electricidad (conocido como “MIBEL”) –es decir, el mercado dominante OTC, las subastas de centrales virtuales de generación conocidas en España como Emisiones Primarias de Energía, y las subastas para cubrir parte de los suministros de último recurso conocidas en España como subastas CESUR– y con una mayor integración de los mercados regionales europeos de energía eléctrica. Se construye un modelo de regresión para analizar la evolución de los volúmenes negociados en el mercado continuo durante sus cuatro primeros años como una función de doce indicadores potenciales de liquidez. Los únicos indicadores significativos son los volúmenes negociados en las subastas obligatorias gestionadas por OMIP, los volúmenes negociados en el mercado OTC y los volúmenes OTC compensados por OMIClear. El número de creadores de mercado, la incorporación de agentes financieros y compañías de generación pertenecientes a grupos integrados con suministradores de último recurso, y los volúmenes OTC compensados por OMIClear muestran una fuerte correlación con los volúmenes negociados en el mercado continuo. La liquidez de OMIP está aún lejos de los niveles alcanzados por los mercados europeos más maduros (localizados en los países nórdicos (Nasdaq OMX Commodities) y Alemania (EEX)). El operador de mercado y su cámara de compensación podrían desarrollar acciones eficientes de marketing para atraer nuevos agentes activos en el mercado de contado (p.ej. industrias consumidoras intensivas de energía, suministradores, pequeños productores, compañías energéticas internacionales y empresas de energías renovables) y agentes financieros, captar volúmenes del opaco OTC, y mejorar el funcionamiento de los productos existentes aún no líquidos. Resultaría de gran utilidad para tales acciones un diálogo activo con todos los agentes (participantes en el mercado, operador de mercado de contado, y autoridades supervisoras). Durante sus primeros cinco años y medio, el mercado continuo presenta un crecimento de liquidez estable. Se mide el desempeño de sus funciones de cobertura mediante la ratio de posición neta obtenida al dividir la posición abierta final de un contrato de derivados mensual entre su volumen acumulado en la cámara de compensación. Los futuros carga base muestran la ratio más baja debido a su buena liquidez. Los futuros carga punta muestran una mayor ratio al producirse su menor liquidez a través de contadas subastas fijadas por regulación portuguesa. Las permutas carga base liquidadas en la cámara de compensación ubicada en Madrid –MEFF Power, activa desde el 21 de marzo de 2011– muestran inicialmente valores altos debido a bajos volúmenes registrados, dado que esta cámara se emplea principalmente para vencimientos pequeños (diario y semanal). Dicha ratio puede ser una poderosa herramienta de supervisión para los reguladores energéticos cuando accedan a todas las transacciones de derivados en virtud del Reglamento Europeo sobre Integridad y Transparencia de los Mercados de Energía (“REMIT”), en vigor desde el 28 de diciembre de 2011. La prima de riesgo ex-post tiende a ser positiva en todos los mecanismos (futuros en OMIP, mercado OTC y subastas CESUR) y disminuye debido a la curvas de aprendizaje y al efecto, desde el año 2011, del precio fijo para la retribución de la generación con carbón autóctono. Se realiza una comparativa con los costes a plazo de generación con gas natural (diferencial “clean spark spread”) obtenido como la diferencia entre el precio del futuro eléctrico y el coste a plazo de generación con ciclo combinado internalizando los costes de emisión de CO2. Los futuros eléctricos tienen una elevada correlación con los precios de gas europeos. Los diferenciales de contratos con vencimiento inmediato tienden a ser positivos. Los mayores diferenciales se dan para los contratos mensuales, seguidos de los trimestrales y anuales. Los generadores eléctricos con gas pueden maximizar beneficios con contratos de menor vencimiento. Los informes de monitorización por el operador de mercado que proporcionan transparencia post-operacional, el acceso a datos OTC por el regulador energético, y la valoración del riesgo regulatorio pueden contribuir a ganancias de eficiencia. Estas recomendaciones son también válidas para un potencial mercado ibérico de futuros de gas, una vez que el hub ibérico de gas –actualmente en fase de diseño, con reuniones mensuales de los agentes desde enero de 2013 en el grupo de trabajo liderado por el regulador energético español– esté operativo. El hub ibérico de gas proporcionará transparencia al atraer más agentes y mejorar la competencia, incrementando su eficiencia, dado que en el mercado OTC actual no se revela precio alguno de gas. ABSTRACT The Iberian Power Futures Market, managed by OMIP (“Operador do Mercado Ibérico de Energia, Pólo Português”, located in Lisbon), also known as the Iberian Energy Derivatives Market, started operations on 3 July 2006. The market efficiency, regarding how well the future price predicts the spot price, is analysed for this energy derivatives exchange. There are two trading modes coexisting within OMIP: the continuous market (default mode) and the call auction. In the continuous trading, anonymous buy and sell orders interact immediately and individually with opposite side orders, generating trades with an undetermined number of prices for each contract. In the call auction trading, a single price auction maximizes the traded volume, being all trades settled at the same price (equilibrium price). Additionally, OMIP trading members may settle Over-the-Counter (OTC) trades through OMIP clearing house (OMIClear). The five largest Spanish distribution companies have been obliged to purchase in auctions managed by OMIP until July 2009, in order to partly cover their portfolios of end users’ regulated supplies. Likewise, the Portuguese last resort supplier kept that obligation until July 2010. The auction equilibrium prices are not optimal for remuneration purposes of regulated supplies as such prices seem to be slightly upward biased. The ex-post forward risk premium, defined as the difference between the forward and spot prices in the delivery period, is used to measure its price efficiency. The spot market, managed by OMIE (Market Operator of the Iberian Energy Market, Spanish Pool, known traditionally as “OMEL”), is located in Madrid. During the first two years of the futures market, the average forward risk premium tends to be positive, as it occurs with other European power and natural gas markets. In that period, the ex-post forward risk premium tends to be negative in oil and coal markets. Energy markets tend to show limited levels of market efficiency. The price efficiency of the Iberian Power Futures Market improves with the market development of all the coexistent forward contracting mechanisms within the Iberian Electricity Market (known as “MIBEL”) – namely, the dominant OTC market, the Virtual Power Plant Auctions known in Spain as Energy Primary Emissions, and the auctions catering for part of the last resort supplies known in Spain as CESUR auctions – and with further integration of European Regional Electricity Markets. A regression model tracking the evolution of the traded volumes in the continuous market during its first four years is built as a function of twelve potential liquidity drivers. The only significant drivers are the traded volumes in OMIP compulsory auctions, the traded volumes in the OTC market, and the OTC cleared volumes by OMIClear. The amount of market makers, the enrolment of financial members and generation companies belonging to the integrated group of last resort suppliers, and the OTC cleared volume by OMIClear show strong correlation with the traded volumes in the continuous market. OMIP liquidity is still far from the levels reached by the most mature European markets (located in the Nordic countries (Nasdaq OMX Commodities) and Germany (EEX)). The market operator and its clearing house could develop efficient marketing actions to attract new entrants active in the spot market (e.g. energy intensive industries, suppliers, small producers, international energy companies and renewable generation companies) and financial agents as well as volumes from the opaque OTC market, and to improve the performance of existing illiquid products. An active dialogue with all the stakeholders (market participants, spot market operator, and supervisory authorities) will help to implement such actions. During its firs five and a half years, the continuous market shows steady liquidity growth. The hedging performance is measured through a net position ratio obtained from the final open interest of a month derivatives contract divided by its accumulated cleared volume. The base load futures in the Iberian energy derivatives exchange show the lowest ratios due to good liquidity. The peak futures show bigger ratios as their reduced liquidity is produced by auctions fixed by Portuguese regulation. The base load swaps settled in the clearing house located in Spain – MEFF Power, operating since 21 March 2011, with a new denomination (BME Clearing) since 9 September 2013 – show initially large values due to low registered volumes, as this clearing house is mainly used for short maturity (daily and weekly swaps). The net position ratio can be a powerful oversight tool for energy regulators when accessing to all the derivatives transactions as envisaged by European regulation on Energy Market Integrity and Transparency (“REMIT”), in force since 28 December 2011. The ex-post forward risk premium tends to be positive in all existing mechanisms (OMIP futures, OTC market and CESUR auctions) and diminishes due to the learning curve and the effect – since year 2011 – of the fixed price retributing the indigenous coal fired generation. Comparison with the forward generation costs from natural gas (“clean spark spread”) – obtained as the difference between the power futures price and the forward generation cost with a gas fired combined cycle plant taking into account the CO2 emission rates – is also performed. The power futures are strongly correlated with European gas prices. The clean spark spreads built with prompt contracts tend to be positive. The biggest clean spark spreads are for the month contract, followed by the quarter contract and then by the year contract. Therefore, gas fired generation companies can maximize profits trading with contracts of shorter maturity. Market monitoring reports by the market operator providing post-trade transparency, OTC data access by the energy regulator, and assessment of the regulatory risk can contribute to efficiency gains. The same recommendations are also valid for a potential Iberian gas futures market, once an Iberian gas hub – currently in a design phase, with monthly meetings amongst the stakeholders in a Working Group led by the Spanish energy regulatory authority since January 2013 – is operating. The Iberian gas hub would bring transparency attracting more shippers and improving competition and thus its efficiency, as no gas price is currently disclosed in the existing OTC market.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

One of the most promising areas in which probabilistic graphical models have shown an incipient activity is the field of heuristic optimization and, in particular, in Estimation of Distribution Algorithms. Due to their inherent parallelism, different research lines have been studied trying to improve Estimation of Distribution Algorithms from the point of view of execution time and/or accuracy. Among these proposals, we focus on the so-called distributed or island-based models. This approach defines several islands (algorithms instances) running independently and exchanging information with a given frequency. The information sent by the islands can be either a set of individuals or a probabilistic model. This paper presents a comparative study for a distributed univariate Estimation of Distribution Algorithm and a multivariate version, paying special attention to the comparison of two alternative methods for exchanging information, over a wide set of parameters and problems ? the standard benchmark developed for the IEEE Workshop on Evolutionary Algorithms and other Metaheuristics for Continuous Optimization Problems of the ISDA 2009 Conference. Several analyses from different points of view have been conducted to analyze both the influence of the parameters and the relationships between them including a characterization of the configurations according to their behavior on the proposed benchmark.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

El presente trabajo se basa en la filosofía de la Construcción sin Pérdidas (“Lean Construction”), analizando la situación de esta filosofía en el sector de la edificación en el contexto internacional y español, respondiendo las siguientes preguntas: 1. ¿Cómo surge el “Lean Construction”? 2. ¿Cuáles son sus actividades, funciones y cometidos? 3. ¿Existe regulación del ¨Lean Construction” en otros países? 4. ¿Existe demanda del ¨Lean Construction” en España? 5. ¿Existe regulación del ¨Lean Construction” en España? 6. ¿Cómo debería ser la regulación ¨Lean Construction” en España? 7. ¿Cuál es la relación del “Lean Construction” con el “Project & Construction Management”? 8. ¿Cómo debería ser la regulación de “Lean Construction” en España considerando su relación con el “Project & Construction Management”? Las preguntas indicadas las hemos respondido detalladamente en el presente trabajo, a continuación se resume las respuestas a dichas preguntas: 1. El “Lean Construction” surge en agosto de 1992, cuando el investigador finlandés Lauri Koskela publicó en la Universidad de Stanford el reporte TECHNICAL REPORT N° 72 titulado “Application of the New Production Philosophy to Construction”. Un año más tarde el Dr. Koskela invitó a un grupo de especialistas en construcción al primer workshop de esta materia en Finlandia, dando origen al International Group for Lean Construction (IGLC) lo que ha permitido extender la filosofía a EEUU, Europa, América, Asia, Oceanía y África. “Lean Construction” es un sistema basado en el enfoque “Lean Production” desarrollado en Japón por Toyota Motors a partir de los años cincuenta, sistema que permitió a sus fábricas producir unidades con mayor eficiencia que las industrias americanas, con menores recursos, en menor tiempo, y con un número menor de errores de fabricación. 2. El sistema “Lean Construction” busca maximizar el valor y disminuir las pérdidas de los proyectos generando una coordinación eficiente entre los involucrados, manejando un proyecto como un sistema de producción, estrechando la colaboración entre los participantes de los proyectos, capacitándoles y empoderándoles, fomentando una cultura de cambio. Su propósito es desarrollar un proceso de construcción en el que no hayan accidentes, ni daños a equipos, instalaciones, entorno y comunidad, que se realice en conformidad con los requerimientos contractuales, sin defectos, en el plazo requerido, respetando los costes presupuestados y con un claro enfoque en la eliminación o reducción de las pérdidas, es decir, las actividades que no generen beneficios. El “Last Planner System”, o “Sistema del Último Planificador”, es un sistema del “Lean Construction” que por su propia naturaleza protege a la planificación y, por ende, ayuda a maximizar el valor y minimizar las pérdidas, optimizando de manera sustancial los sistemas de seguridad y salud. El “Lean Construction” se inició como un concepto enfocado a la ejecución de las obras, posteriormente se aplicó la filosofía a todas las etapas del proyecto. Actualmente considera el desarrollo total de un proyecto, desde que nace la idea hasta la culminación de la obra y puesta en marcha, considerando el ciclo de vida completo del proyecto. Es una filosofía de gestión, metodologías de trabajo y una cultura empresarial orientada a la eficiencia de los procesos y flujos. La filosofía “Lean Construction” se está expandiendo en todo el mundo, además está creciendo en su alcance, influyendo en la gestión contractual de los proyectos. Su primera evolución consistió en la creación del sistema “Lean Project Delivery System”, que es el concepto global de desarrollo de proyectos. Posteriormente, se proponen el “Target Value Design”, que consiste en diseñar de forma colaborativa para alcanzar los costes y el valor requerido, y el “Integrated Project Delivery”, en relación con sistemas de contratos relacionales (colaborativos) integrados, distintos a los contratos convencionales. 3. Se verificó que no existe regulación específica del ¨Lean Construction” en otros países, en otras palabras, no existe el agente con el nombre específico de “Especialista en Lean Construction” o similar, en consecuencia, es un agente adicional en el proyecto de la edificación, cuyas funciones y cometidos se pueden solapar con los del “Project Manager”, “Construction Manager”, “Contract Manager”, “Safety Manager”, entre otros. Sin embargo, se comprobó la existencia de formatos privados de contratos colaborativos de Integrated Project Delivery, los cuales podrían ser tomados como unas primeras referencias para futuras regulaciones. 4. Se verificó que sí existe demanda del ¨Lean Construction” en el desarrollo del presente trabajo, aunque aún su uso es incipiente, cada día existe más interesados en el tema. 5. No existe regulación del ¨Lean Construction” en España. 6. Uno de los objetivos fundamentales de esta tesis es el de regular esta figura cuando actúe en un proyecto, definir y realizar una estructura de Agente de la Edificación, según la Ley de Ordenación de la Edificación (LOE), y de esta manera poder introducirla dentro de la Legislación Española, protegiéndola de eventuales responsabilidades civiles. En España existe jurisprudencia (sentencias de los tribunales de justicia españoles) con jurisdicción civil basada en la LOE para absolver o condenar a agentes de la edificación que son definidos en los tribunales como “gestores constructivos” o similares. Por este motivo, en un futuro los tribunales podrían dictaminar responsabilidades solidarias entre el especialista “Lean Construction” y otros agentes del proyecto, dependiendo de sus actuaciones, y según se implemente el “Lean Project Delivery System”, el “Target Value Design” y el “Integrated Project Delivery”. Por otro lado, es posible que el nivel de actuación del especialista “Lean Construcción” pueda abarcar la gestión del diseño, la gestión de la ejecución material (construcción), la gestión de contratos, o la gestión integral de todo el proyecto de edificación, esto último, en concordancia con la última Norma ISO 21500:2012 o UNE-ISO 21500:2013 Directrices para la dirección y gestión de proyectos. En consecuencia, se debería incorporar adecuadamente a uno o más agentes de la edificación en la LOE de acuerdo a sus funciones y responsabilidades según los niveles de actuación del “Especialista en Lean Construction”. Se propone la creación de los siguientes agentes: Gestor del Diseño, Gestor Constructivo y Gestor de Contratos, cuyas definiciones están desarrolladas en este trabajo. Estas figuras son definidas de manera general, puesto que cualquier “Project Manager” o “DIPE”, gestor BIM (Building Information Modeling), o similar, puede actuar como uno o varios de ellos. También se propone la creación del agente “Gestor de la Construcción sin Pérdidas”, como aquel agente que asume las actuaciones del “gestor de diseño”, “gestor constructivo” y “gestor de contratos” con un enfoque en los principios del Lean Production. 7. En la tesis se demuestra, por medio del uso de la ISO 21500, que ambos sistemas son complementarios, de manera que los proyectos pueden tener ambos enfoques y ser compatibilizados. Un proyecto que use el “Project & Construction Management” puede perfectamente apoyarse en las herramientas y técnicas del “Lean Construction” para asegurar la eliminación o reducción de las pérdidas, es decir, las actividades que no generen valor, diseñando el sistema de producción, el sistema de diseño o el sistema de contratos. 8. Se debería incorporar adecuadamente al agente de la edificación “Especialista en Lean Construction” o similar y al agente ¨Especialista en Project & Construction Management” o DIPE en la Ley de Ordenación de la Edificación (LOE) de acuerdo a sus funciones y responsabilidades, puesto que la jurisprudencia se ha basado para absolver o condenar en la referida Ley. Uno de los objetivos fundamentales de esta tesis es el de regular la figura del “Especialista en Lean Construction” cuando actúa simultáneamente con el DIPE, y realizar una estructura de Agente de la Edificación según la LOE, y de esta manera protegerlo de eventuales responsabilidades solidarias. Esta investigación comprueba que la propuesta de definición del agente de edificación DIPE, según la LOE, presentada en la tesis doctoral del Doctor Manuel Soler Severino es compatible con las nuevas definiciones propuestas. El agente DIPE puede asumir los roles de los diferentes gestores propuestos en esta tesis si es que se especializa en dichas materias, o, si lo estima pertinente, recomendar sus contrataciones. ABSTRACT This work is based on the Lean Construction philosophy; an analysis is made herein with regard to the situation of this philosophy in the building sector within the international and Spanish context, replying to the following questions: 1. How did the concept of Lean Construction emerge? 2. Which are the activities, functions and objectives of Lean Construction? 3. Are there regulations on Lean Construction in other countries? 4. Is there a demand for Lean Construction in Spain? 5. Are there regulations on Lean Construction in Spain? 6. How should regulations on Lean Construction be developed in Spain? 7. What is the relationship between Lean Construction and the Project & Construction Management? 8. How should regulations on Lean Construction be developed in Spain considering its relationship with the Project & Construction Management? We have answered these questions in detail here and the replies are summarized as follows: 1. The concept of Lean Construction emerged in august of 1992, when Finnish researcher Lauri Koskela published in Stanford University TECHNICAL REPORT N° 72 entitled “Application of the New Production Philosophy to Construction”. A year later, Professor Koskela invited a group of construction specialists to Finland to the first workshop conducted on this matter; thus, the International Group for Lean Construction (IGLC) was established, which has contributed to extending the philosophy to the United States, Europe, the Americas, Asia, Oceania, and Africa. Lean Construction is a system based on the Lean Production approach, which was developed in Japan by Toyota Motors in the 1950s. Thanks to this system, the Toyota plants were able to produce more units, with greater efficiency than the American industry, less resources, in less time, and with fewer manufacturing errors. 2. The Lean Construction system aims at maximizing the value of projects while reducing waste, producing an effective coordination among those involved; it manages projects as a production system, enhancing collaboration between the parties that participate in the projects while building their capacities, empowering them, and promoting a culture of change. Its purpose is to develop a construction process free of accidents, without damages to the equipment, facilities, environment and community, flawless, in accordance with contractual requirements, within the terms established, respecting budgeted costs, and with a clear approach to eliminating or reducing waste, that is, activities that do not generate benefits. The Last Planner System is a Lean Construction system, which by its own nature protects planning and, therefore, helps to maximize the value and minimize waste, optimizing substantially the safety and health systems. Lean Construction started as a concept focused on the execution of works, and subsequently the philosophy was applied to all the stages of the project. At present it considers the project’s total development, since the time ideas are born until the completion and start-up of the work, taking into account the entire life cycle of the project. It is a philosophy of management, work methodologies, and entrepreneurial culture aimed at the effectiveness of processes and flows. The Lean Construction philosophy is extending all over the world and its scope is becoming broader, having greater influence on the contractual management of projects. It evolved initially through the creation of the Lean Project Delivery System, a global project development concept. Later on, the Target Value Design was developed, based on collaborative design to achieve the costs and value required, as well as the Integrated Project Delivery, in connection with integrated relational (collaborative) contract systems, as opposed to conventional contracts. 3. It was verified that no specific regulations on Lean Construction exist in other countries, in other words, there are no agents with the specific name of “Lean Construction Specialist” or other similar names; therefore, it is an additional agent in building projects, which functions and objectives can overlap those of the Project Manager, Construction Manager, Contract Manager, or Safety Manager, among others. However, the existence of private collaborative contracts of Integrated Project Delivery was confirmed, which could be considered as first references for future regulations. 4. There is a demand for Lean Construction in the development of this work; even though it is still emerging, there is a growing interest in this topic. 5. There are no regulations on Lean Construction in Spain. 6. One of the main objectives of this thesis is to regulate this role when acting in a project, and to define and develop a Building Agent structure, according to the Building Standards Law (LOE by its acronym in Spanish), in order to be able to incorporate it into the Spanish law, protecting it from civil liabilities. In Spain there is jurisprudence in civil jurisdiction based on the LOE to acquit or convict building agents, which are defined in the courts as “construction managers” or similar. For this reason, courts could establish in the future joint and several liabilities between the Lean Construction Specialist and other agents of the project, depending on their actions and based on the implementation of the Lean Project Delivery System, the Target Value Design, and the Integrated Project Delivery. On the other hand, it is possible that the level of action of the Lean Construction Specialist may comprise design management, construction management and contract management, or the integral management of the entire building project in accordance with the last ISO 21500:2012 or UNE-ISO 21500:2013, guidelines for the management of projects. Accordingly, one or more building agents should be appropriately incorporated into the LOE according to their functions and responsibilities and based on the levels of action of the Lean Construction Specialist. The creation of the following agents is proposed: Design Manager, Construction Manager, and Contract Manager, which definitions are developed in this work. These agents are defined in general, since any Project Manager or DIPE, Building Information Modeling (BIM) Manager or similar, may act as one or as many of them. The creation of the Lean Construction Manager is also proposed, as the agent that takes on the role of the Design Manager, Construction Manager and Contract Manager with a focus on the Lean Production principles. 7. In the thesis it is demonstrated that through the implementation of the ISO 21500, both systems are supplementary, so projects may have both approaches and be compatible. A project that applies the Project & Construction Management may perfectly have the support of the tools, techniques and practices of Lean Construction to ensure the elimination or reduction of losses, that is, those activities that do not generate value, thus designing the production system, the design system, or the contract system. 8. The Lean Construction Specialist or similar and the Specialist in Project & Construction Management should be incorporated appropriately into the LOE according to their functions and responsibilities, since jurisprudence has been based on such Law to acquit or convict. One of the main objectives of this thesis is the regulate the role of the Lean Construction Specialist when acting simultaneously with the DIPE, and to develop a structure of the building agent, according to the LOE, and in this way protect such agent from joint and several liabilities. This research proves that the proposal to define the DIPE building agent, according to the LOE, and presented in the doctoral dissertation of Manuel Soler Severino, Ph.D. is compatible with the new definitions proposed. The DIPE agent may assume the roles of the different managers proposed in this thesis if he specializes in those topics or, if deemed pertinent, recommends that they be engaged.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Compilation techniques such as those portrayed by the Warren Abstract Machine(WAM) have greatly improved the speed of execution of logic programs. The research presented herein is geared towards providing additional performance to logic programs through the use of parallelism, while preserving the conventional semantics of logic languages. Two áreas to which special attention is given are the preservation of sequential performance and storage efficiency, and the use of low overhead mechanisms for controlling parallel execution. Accordingly, the techniques used for supporting parallelism are efficient extensions of those which have brought high inferencing speeds to sequential implementations. At a lower level, special attention is also given to design and simulation detail and to the architectural implications of the execution model behavior. This paper offers an overview of the basic concepts and techniques used in the parallel design, simulation tools used, and some of the results obtained to date.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper presents and develops a generalized concept of Non-Strict Independent And Parallelism (NSIAP). NSIAP extends the applicability of Independent And- Parallelism (IAP) by enlarging the class of goals which are eligible for parallel execution. At the same time it maintains IAP's ability to run non-deterministic goals in parallel and to preserve the computational complexity expected in the execution of the program by the programmer. First, a parallel execution framework is defined and some fundamental correctness results, in the sense of equivalence of solutions with the sequential model, are discussed for this framework. The issue of efficiency is then considered. Two new definitions of NSI are given for the cases of puré and impure goals respectively and efficiency results are provided for programs parallelized under these definitions which include treatment of the case of goal failure: not only is reduction of execution time guaranteed (modulo run-time overheads) in the absence of failure but it is also shown that in the worst case of failure no speed-down will occur. In addition to applying to NSI, these results carry over and complete previous results shown in the context of IAP which did not deal with the case of goal failure. Finally, some practical examples of the application of the NSIAP concept to the parallelization of a set of programs are presented and performance results, showing the advantage of using NSI, are given.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Performance studies of actual parallel systems usually tend to concéntrate on the effectiveness of a given implementation. This is often done in the absolute, without quantitave reference to the potential parallelism contained in the programs from the point of view of the execution paradigm. We feel that studying the parallelism inherent to the programs is interesting, as it gives information about the best possible behavior of any implementation and thus allows contrasting the results obtained. We propose a method for obtaining ideal speedups for programs through a combination of sequential or parallel execution and simulation, and the algorithms that allow implementing the method. Our approach is novel and, we argüe, more accurate than previously proposed methods, in that a crucial part of the data - the execution times of tasks - is obtained from actual executions, while speedup is computed by simulation. This allows obtaining speedup (and other) data under controlled and ideal assumptions regarding issues such as number of processor, scheduling algorithm and overheads, etc. The results obtained can be used for example to evalúate the ideal parallelism that a program contains for a given model of execution and to compare such "perfect" parallelism to that obtained by a given implementation of that model. We also present a tool, IDRA, which implements the proposed method, and results obtained with IDRA for benchmark programs, which are then compared with those obtained in actual executions on real parallel systems.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Los problemas de programación de tareas son muy importantes en el mundo actual. Se puede decir que se presentan en todos los fundamentos de la industria moderna, de ahí la importancia de que estos sean óptimos, de forma que se puedan ahorrar recursos que estén asociados al problema. La programación adecuada de trabajos en procesos de manufactura, constituye un importante problema que se plantea dentro de la producción en muchas empresas. El orden en que estos son procesados, no resulta indiferente, sino que determinará algún parámetro de interés, cuyos valores convendrá optimizar en la medida de lo posible. Así podrá verse afectado el coste total de ejecución de los trabajos, el tiempo necesario para concluirlos o el stock de productos en curso que será generado. Esto conduce de forma directa al problema de determinar cuál será el orden más adecuado para llevar a cabo los trabajos con vista a optimizar algunos de los anteriores parámetros u otros similares. Debido a las limitaciones de las técnicas de optimización convencionales, en la presente tesis se presenta una metaheurística basada en un Algoritmo Genético Simple (Simple Genetic Algorithm, SGA), para resolver problemas de programación de tipo flujo general (Job Shop Scheduling, JSS) y flujo regular (Flow Shop Scheduling, FSS), que están presentes en un taller con tecnología de mecanizado con el objetivo de optimizar varias medidas de desempeño en un plan de trabajo. La aportación principal de esta tesis, es un modelo matemático para medir el consumo de energía, como criterio para la optimización, de las máquinas que intervienen en la ejecución de un plan de trabajo. Se propone además, un método para mejorar el rendimiento en la búsqueda de las soluciones encontradas, por parte del Algoritmo Genético Simple, basado en el aprovechamiento del tiempo ocioso. ABSTRACT The scheduling problems are very important in today's world. It can be said to be present in all the basics of modern industry, hence the importance that these are optimal, so that they can save resources that are associated with the problem. The appropriate programming jobs in manufacturing processes is an important problem that arises in production in many companies. The order in which they are processed, it is immaterial, but shall determine a parameter of interest, whose values agree optimize the possible. This may be affected the total cost of execution of work, the time needed to complete them or the stock of work in progress that will be generated. This leads directly to the problem of determining what the most appropriate order to carry out the work in order to maximize some of the above parameters or other similar. Due to the limitations of conventional optimization techniques, in this work present a metaheuristic based on a Simple Genetic Algorithm (Simple Genetic Algorithm, SGA) to solve programming problems overall flow rate (Job Shop Scheduling, JSS) and regular flow (Flow Shop Scheduling, FSS), which are present in a workshop with machining technology in order to optimize various performance measures in a plan. The main contribution of this thesis is a mathematical model to measure the energy consumption as a criterion for the optimization of the machines involved in the implementation of a work plan. It also proposes a method to improve performance in finding the solutions, by the simple genetic algorithm, based on the use of idle time.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Ciao is a public domain, next generation multi-paradigm programming environment with a unique set of features: Ciao offers a complete Prolog system, supporting ISO-Prolog, but its novel modular design allows both restricting and extending the language. As a result, it allows working with fully declarative subsets of Prolog and also to extend these subsets (or ISO-Prolog) both syntactically and semantically. Most importantly, these restrictions and extensions can be activated separately on each program module so that several extensions can coexist in the same application for different modules. Ciao also supports (through such extensions) programming with functions, higher-order (with predicate abstractions), constraints, and objects, as well as feature terms (records), persistence, several control rules (breadth-first search, iterative deepening, ...), concurrency (threads/engines), a good base for distributed execution (agents), and parallel execution. Libraries also support WWW programming, sockets, external interfaces (C, Java, TclTk, relational databases, etc.), etc. Ciao offers support for programming in the large with a robust module/object system, module-based separate/incremental compilation (automatically -no need for makefiles), an assertion language for declaring (optional) program properties (including types and modes, but also determinacy, non-failure, cost, etc.), automatic static inference and static/dynamic checking of such assertions, etc. Ciao also offers support for programming in the small producing small executables (including only those builtins used by the program) and support for writing scripts in Prolog. The Ciao programming environment includes a classical top-level and a rich emacs interface with an embeddable source-level debugger and a number of execution visualization tools. The Ciao compiler (which can be run outside the top level shell) generates several forms of architecture-independent and stand-alone executables, which run with speed, efficiency and executable size which are very competive with other commercial and academic Prolog/CLP systems. Library modules can be compiled into compact bytecode or C source files, and linked statically, dynamically, or autoloaded. The novel modular design of Ciao enables, in addition to modular program development, effective global program analysis and static debugging and optimization via source to source program transformation. These tasks are performed by the Ciao preprocessor ( ciaopp, distributed separately). The Ciao programming environment also includes lpdoc, an automatic documentation generator for LP/CLP programs. It processes Prolog files adorned with (Ciao) assertions and machine-readable comments and generates manuals in many formats including postscript, pdf, texinfo, info, HTML, man, etc. , as well as on-line help, ascii README files, entries for indices of manuals (info, WWW, ...), and maintains WWW distribution sites.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

En la coyuntura actual, en la que existe por un lado, exceso en la oferta de vivienda (de alto precio o de segunda residencia), y aparece por otro demanda de vivienda (de bajo precio y/o social), el mercado inmobiliario se encuentra paradójicamente bloqueado. Así, surge esta investigación como fruto de este momento histórico, en el cual se somete a debate económico el producto vivienda, no solo como consecuencia de la profunda crisis económica, sino también para la correcta gestión de los recursos desde el punto de vista de lo eficiente y sostenible. Se parte de la hipótesis de que es necesario determinar un estimador de costes de construcción de vivienda autopromovida como una de las soluciones a la habitación en el medio rural de Extremadura, para lo cual se ha tomado como modelo de análisis concretamente la Vivienda Autopromovida subvencionada por la Junta de Extremadura en el marco de la provincia de Cáceres. Con esta investigación se pretende establecer una herramienta matemática precisa que permita determinar la inversión a los promotores, el posible margen de beneficios a los contratistas y el valor real de la garantía en el préstamo a las entidades financieras. Pero el objetivo de mayor proyección social de esta investigación consiste en facilitar una herramienta sencilla a la Junta de Extremadura para que pueda establecer las ayudas de una manera proporcional. De este modo se ayuda a optimizar los recursos, lo cual en época de crisis resulta aun más acuciante, ya que conociendo previamente y con bastante exactitud el importe de las obras se pueden dirigir las ayudas de forma proporcional a las necesidades reales de la ejecución. De hecho, ciertas características difíciles de cuantificar para determinar las ayudas en materia de vivienda, como la influencia del número de miembros familiares o la atención a la discapacidad, se verían contempladas de forma indirecta en el coste estimado con el método aquí propuesto, ya que suponen siempre un aumento de las superficies construidas y útiles, de los huecos de fachadas o del tamaño de locales húmedos y por tanto se contemplan en la ecuación del modelo determinado. Por último, contar con un estimador de costes potencia la forma de asentamiento de la construcción mediante autopromocion de viviendas ya que ayuda a la toma de decisiones al particular, subvencionado o no. En efecto, la herramienta es valida en cierta medida para cualquier autopromocion, constituye un sistema de construcción con las menores posibilidades especulativas y lo más sostenible, es abundante en toda Extremadura, y consigue que el sector de la construcción sea un sistema más eficiente al optimizar su proceso económico de producción. SUMMARY Under the present circumstances, in which there is, on one hand, an excess in the supply of housing (high-price or second-home), and on the other hand a demand for housing (low cost and/or social), paradoxically the property market is at a standstill. This research has come about as a result of this moment in time, in which the product: housing, is undergoing economic debate, not only on account of this serious economic crisis, but for the proper management of resources from the point of view of efficiency and sustainability. A building-costs estimator for owner-developed housing is deemed necessary as one of the solutions for the rural environment that is Extremadura. To this end, it is the Owner-Developed House which has been taken as analysis model. It is subsidized by the Extremadura Regional Government in Caceres Province. This research establishes an accurate mathematical tool to work out the developers’ investment, the builder’s potential profit margin and the reality of the loan for the Financial Institution. But the result of most social relevance in this research is to provide the Extremadura Regional Government with a simple tool, so that it can draw up the Subventions proportionally. Thus, the resources are optimized, an even more vital matter in times of economic slump, due to the fact that if the cost of the building works is known with some accuracy beforehand, the subventions can be allocated in a way that is proportional to the real needs of execution. In fact certain elements related to housing subventions which are hard to quantify, such as the influence of number of family members or disability support, would be covered indirectly in cost estimate with the proposed method, since they inevitably involve an increase in built area, exterior wall openings and the size of plumbed rooms. As such they are covered in the determined model equation. Lastly, the availability of a cost-estimator reinforces the ownerdeveloped building model, since it assists decision-making by the individual, whether subsidized or not. This is because the tool is valid to some extent in any owner-development, and this building scheme, which is common in Extremadura, is the most sustainable, and the least liable to speculation. It makes the building sector more efficient by optimizing the economic production process.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Las uniones estructurales mecánicas y adhesivas requieren la combinación de un número importante de parámetros para la obtención de la continuidad estructural que exigen las condiciones de diseño. Las características de las uniones presentan importantes variaciones, ligadas a las condiciones de ejecución, tanto en uniones mecánicas como especialmente en uniones adhesivas y mixtas (unión mecánica y adhesiva, también conocidas como uniones híbridas). Las propiedades mecánicas de las uniones adhesivas dependen de la naturaleza y propiedades de los adhesivos y también de muchos otros parámetros que influyen directamente en el comportamiento de estas uniones. Algunos de los parámetros más significativos son: el acabado superficial de los materiales, área y espesor de la capa adhesiva, diseño adecuado, secuencia de aplicación, propiedades químicas de la superficie y preparación de los sustratos antes de aplicar el adhesivo. Los mecanismos de adhesión son complejos. En general, cada unión adhesiva solo puede explicarse considerando la actuación conjunta de varios mecanismos de adhesión. No existen adhesivos universales para un determinado material o aplicación, por lo que cada pareja sustrato-adhesivo requiere un particular estudio y el comportamiento obtenido puede variar, significativamente, de uno a otro caso. El fallo de una junta adhesiva depende del mecanismo cohesión-adhesión, ligado a la secuencia y modo de ejecución de los parámetros operacionales utilizados en la unión. En aplicaciones estructurales existen un número muy elevado de sistemas de unión y de posibles sustratos. En este trabajo se han seleccionado cuatro adhesivos diferentes (cianoacrilato, epoxi, poliuretano y silano modificado) y dos procesos de unión mecánica (remachado y clinchado). Estas uniones se han aplicado sobre chapas de acero al carbono en diferentes estados superficiales (chapa blanca, galvanizada y prepintada). Los parámetros operacionales analizados han sido: preparación superficial, espesor del adhesivo, secuencia de aplicación y aplicación de presión durante el curado. Se han analizado tanto las uniones individuales como las uniones híbridas (unión adhesiva y unión mecánica). La combinación de procesos de unión, sustratos y parámetros operacionales ha dado lugar a la preparación y ensayo de más de mil muestras. Pues, debido a la dispersión de resultados característica de las uniones adhesivas, para cada condición analizada se han ensayado seis probetas. Los resultados obtenidos han sido: El espesor de adhesivo utilizado es una variable muy importante en los adhesivos flexibles, donde cuanto menor es el espesor del adhesivo mayor es la resistencia mecánica a cortadura de la unión. Sin embargo en los adhesivos rígidos su influencia es mucho menor. La naturaleza de la superficie es fundamental para una buena adherencia del adhesivo al substrato, que repercute en la resistencia mecánica de la unión. La superficie que mejor adherencia presenta es la prepintada, especialmente cuando existe una alta compatibilidad entre la pintura y el adhesivo. La superficie que peor adherencia tiene es la galvanizada. La secuencia de aplicación ha sido un parámetro significativo en las uniones híbridas, donde los mejores resultados se han obtenido cuando se aplicaba primero el adhesivo y la unión mecánica se realizaba antes del curado del adhesivo. La aplicación de presión durante el curado se ha mostrado un parámetro significativo en los adhesivos con poca capacidad para el relleno de la junta. En los otros casos su influencia ha sido poco relevante. El comportamiento de las uniones estructurales mecánicas y adhesivas en cuanto a la resistencia mecánica de la unión puede variar mucho en función del diseño de dicha unión. La resistencia mecánica puede ser tan grande que falle antes el substrato que la unión. Las mejores resistencias se consiguen diseñando las uniones con adhesivo cianoacrilato, eligiendo adecuadamente las condiciones superficiales y operacionales, por ejemplo chapa blanca aplicando una presión durante el curado de la unión. La utilización de uniones mixtas aumenta muy poco o nada la resistencia mecánica, pero a cambio proporciona una baja dispersión de resultados, siendo destacable para la superficie galvanizada, que es la que presenta peor reproducibilidad cuando se realizan uniones sólo con adhesivo. Las uniones mixtas conducen a un aumento de la deformación antes de la rotura. Los adhesivos dan rotura frágil y las uniones mecánicas rotura dúctil. La unión mixta proporciona ductilidad a la unión. Las uniones mixtas también pueden dar rotura frágil, esto sucede cuando la resistencia del adhesivo es tres veces superior a la resistencia de la unión mecánica. Las uniones híbridas mejoran la rigidez de la junta, sobre todo se aprecia un aumento importante en las uniones mixtas realizadas con adhesivos flexibles, pudiendo decirse que para todos los adhesivos la rigidez de la unión híbrida es superior. ABSTRACT The mechanical and adhesive structural joints require the combination of a large number of parameters to obtain the structural continuity required for the design conditions. The characteristics of the junctions have important variations, linked to performance conditions, in mechanical joints as particular in mixed adhesive joints (mechanical and adhesive joints, also known as hybrid joints). The mechanical properties of the adhesive joints depend of the nature and properties of adhesives and also of many other parameters that directly influence in the behavior of these joints. Some of the most significant parameters are: the surface finished of the material, area and thickness of the adhesive layer, suitable design, and application sequence, chemical properties of the surface and preparation of the substrate before applying the adhesive. Adhesion mechanisms are complex. In general, each adhesive joint can only be explained by considering the combined action of several adhesions mechanisms. There aren’t universal adhesives for a given material or application, so that each pair substrate-adhesive requires a particular study and the behavior obtained can vary significantly from one to another case. The failure of an adhesive joint depends on the cohesion-adhesion mechanism, linked to the sequence and manner of execution of the operational parameters used in the joint. In the structural applications, there are a very high number of joining systems and possible substrates. In this work we have selected four different adhesives (cyanoacrylate, epoxy, polyurethane and silano modified) and two mechanical joining processes (riveting and clinching). These joints were applied on carbon steel with different types of surfaces (white sheet, galvanized and pre-painted). The operational parameters analyzed were: surface preparation, thickness of adhesive, application sequence and application of pressure during curing. We have analyzed individual joints both as hybrid joints (adhesive joint and mechanical joint). The combination of joining processes, substrates and operational parameters has resulted in the preparation and testing of over a thousand specimens. Then, due to the spread of results characteristic of adhesive joints, for each condition analyzed we have tested six samples. The results have been: The thickness of adhesive used is an important variable in the flexible adhesives, where the lower the adhesive thickness greater the shear strength of the joint. However in rigid adhesives is lower influence. The nature of the surface is essential for good adherence of the adhesive to the substrate, which affects the shear strength of the joint. The surface has better adherence is preprinted, especially when there is a high compatibility between the paint and the adhesive. The surface which has poor adherence is the galvanized. The sequence of application has been a significant parameter in the hybrid junctions, where the best results are obtained when applying first the adhesive and the mechanical joint is performed before cured of the adhesive. The application of pressure during curing has shown a significant parameter in the adhesives with little capacity for filler the joint. In other cases their influence has been less relevant. The behavior of structural mechanical and adhesive joints in the shear strength of the joint can vary greatly depending on the design of such a joint. The shear strength may be so large that the substrate fails before the joint. The best shear strengths are achieved by designing the junctions with cyanoacrylate adhesive, by selecting appropriately the surface and operating conditions, for example by white sheet applying a pressure during curing of the joint. The use of hybrid joints no increase shear strength, but instead provides a low dispersion of results, being remarkable for the galvanized surface, which is the having worst reproducibility when performed bonded joints. The hybrid joints leading to increased deformation before rupture. The joints witch adhesives give brittle fracture and the mechanics joints give ductile fracture. Hybrid joint provides ductility at the joint. Hybrid joint can also give brittle fracture, this happens when the shear strength of the adhesive is three times the shear strength of the mechanical joint. The hybrid joints improve stiffness of joint, especially seen a significant increase in hybrid joints bonding with flexible adhesives, can be said that for all the adhesives, the hybrid junction stiffness is higher.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Data-related properties of the activities involved in a service composition can be used to facilitate several design-time and run-time adaptation tasks, such as service evolution, distributed enactment, and instance-level adaptation. A number of these properties can be expressed using a notion of sharing. We present an approach for automated inference of data properties based on sharing analysis, which is able to handle service compositions with complex control structures, involving loops and sub-workflows. The properties inferred can include data dependencies, information content, domain-defined attributes, privacy or confidentiality levels, among others. The analysis produces characterizations of the data and the activities in the composition in terms of minimal and maximal sharing, which can then be used to verify compliance of potential adaptation actions, or as supporting information in their generation. This sharing analysis approach can be used both at design time and at run time. In the latter case, the results of analysis can be refined using the composition traces (execution logs) at the point of execution, in order to support run-time adaptation.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper presents a Focused Crawler in order to Get Semantic Web Resources (CSR). Structured data web are available in formats such as Extensible Markup Language (XML), Resource Description Framework (RDF) and Ontology Web Language (OWL) that can be used for processing. One of the main challenges for performing a manual search and download semantic web resources is that this task consumes a lot of time. Our research work propose a focused crawler which allow to download these resources automatically and store them on disk in order to have a collection that will be used for data processing. CRS consists of three layers: (a) The User Interface Layer, (b) The Focus Crawler Layer and (c) The Base Crawler Layer. CSR uses as a selection policie the Shark-Search method. CSR was conducted with two experiments. The first one starts on December 15 2012 at 7:11 am and ends on December 16 2012 at 4:01 were obtained 448,123,537 bytes of data. The CSR ends by itself after to analyze 80,4375 seeds with an unlimited depth. CSR got 16,576 semantic resources files where the 89 % was RDF, the 10 % was XML and the 1% was OWL. The second one was based on the Web Data Commons work of the Research Group Data and Web Science at the University of Mannheim and the Institute AIFB at the Karlsruhe Institute of Technology. This began at 4:46 am of June 2 2013 and 1:37 am June 9 2013. After 162.51 hours of execution the result was 285,279 semantic resources where predominated the XML resources with 99 % and OWL and RDF with 1 % each one.