38 resultados para lean implementation time

em Universidad Politécnica de Madrid


Relevância:

80.00% 80.00%

Publicador:

Resumo:

En esta tesis se presenta una nueva aproximación para la realización de mapas de calidad del aire, con objeto de que esta variable del medio físico pueda ser tenida en cuenta en los procesos de planificación física o territorial. La calidad del aire no se considera normalmente en estos procesos debido a su composición y a la complejidad de su comportamiento, así como a la dificultad de contar con información fiable y contrastada. Además, la variabilidad espacial y temporal de las medidas de calidad del aire hace que sea difícil su consideración territorial y exige la georeferenciación de la información. Ello implica la predicción de medidas para lugares del territorio donde no existen datos. Esta tesis desarrolla un modelo geoestadístico para la predicción de valores de calidad del aire en un territorio. El modelo propuesto se basa en la interpolación de las medidas de concentración de contaminantes registradas en las estaciones de monitorización, mediante kriging ordinario, previa homogeneización de estos datos para eliminar su carácter local. Con el proceso de eliminación del carácter local, desaparecen las tendencias de las series muestrales de datos debidas a las variaciones temporales y espaciales de la calidad del aire. La transformación de los valores de calidad del aire en cantidades independientes del lugar de muestreo, se realiza a través de parámetros de uso del suelo y de otras variables características de la escala local. Como resultado, se obtienen unos datos de entrada espacialmente homogéneos, que es un requisito fundamental para la utilización de cualquier algoritmo de interpolación, en concreto, del kriging ordinario. Después de la interpolación, se aplica una retransformación de los datos para devolver el carácter local al mapa final. Para el desarrollo del modelo, se ha elegido como área de estudio la Comunidad de Madrid, por la disponibilidad de datos reales. Estos datos, valores de calidad del aire y variables territoriales, se utilizan en dos momentos. Un momento inicial, donde se optimiza la selección de los parámetros más adecuados para la eliminación del carácter local de las medidas y se desarrolla cada una de las etapas del modelo. Y un segundo momento, en el que se aplica en su totalidad el modelo desarrollado y se contrasta su eficacia predictiva. El modelo se aplica para la estimación de los valores medios y máximos de NO2 del territorio de estudio. Con la implementación del modelo propuesto se acomete la territorialización de los datos de calidad del aire con la reducción de tres factores clave para su efectiva integración en la planificación territorial o en el proceso de toma de decisiones asociado: incertidumbre, tiempo empleado para generar la predicción y recursos (datos y costes) asociados. El modelo permite obtener una predicción de valores del contaminante objeto de análisis en unas horas, frente a los periodos de modelización o análisis requeridos por otras metodologías. Los recursos necesarios son mínimos, únicamente contar con los datos de las estaciones de monitorización del territorio que, normalmente, están disponibles en las páginas web viii institucionales de los organismos gestores de las redes de medida de la calidad del aire. Por lo que respecta a las incertidumbres de la predicción, puede decirse que los resultados del modelo propuesto en esta tesis son estadísticamente muy correctos y que los errores medios son, en general, similares o menores que los encontrados con la aplicación de las metodologías existentes. ABSTRACT This thesis presents a new approach for mapping air quality, so that this variable of physical environment can be taken into account in physical or territorial planning. Ambient air quality is not normally considered in territorial planning mainly due to the complexity of its composition and behavior and the difficulty of counting with reliable and contrasted information. In addition, the wide spatial and temporal variability of the measurements of air quality makes his territorial consideration difficult and requires georeferenced information. This involves predicting measurements in the places of the territory where there are no data. This thesis develops a geostatistical model for predicting air quality values in a territory. The proposed model is based on the interpolation of measurements of pollutants from the monitoring stations, using ordinary kriging, after a detrending or removal of the local character of sampling values process. With the detrending process, the local character of the time series of sampling data, due to temporal and spatial variations of air quality, is removed. The transformation of the air quality values into site-independent quantities is performed using land use parameters and other characteristic parameters of local scale. This detrending of the monitoring data process results in a spatial homogeneous input set which is a prerequisite for a correct use of any interpolation algorithm, particularly, ordinary kriging. After the interpolation step, a retrending or retransformation is applied in order to incorporate the local character in the final map at places where no monitoring data is available. For the development of this model, the Community of Madrid is chosen as study area, because of the availability of actual data. These data, air quality values and local parameters, are used in two moments. A starting point, to optimize the selection of the most suitable indicators for the detrending process and to develop each one of the model stages. And a second moment, to fully implement the developed model and to evaluate its predictive power. The model is applied to estimate the average and maximum values of NO2 in the study territory. With the implementation of the proposed model, the territorialization of air quality data is undertaken with the reduction in three key factors for the effective integration of this parameter in territorial planning or in the associated decision making process: uncertainty, time taken to generate the prediction and associated resources (data and costs). This model allows the prediction of pollutant values in hours, compared to the implementation time periods required for other modeling or analysis methodologies. The required resources are also minimal, only having data from monitoring stations in the territory, that are normally available on institutional websites of the authorities responsible for air quality networks control and management. With regard to the prediction uncertainties, it can be concluded that the results of the proposed model are statistically very accurate and the mean errors are generally similar to or lower than those found with the application of existing methodologies.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This paper presents an automatic modulation classifier for electronic warfare applications. It is a pattern recognition modulation classifier based on statistical features of the phase and instantaneous frequency. This classifier runs in a real time operation mode with sampling rates in excess of 1 Gsample/s. The hardware platform for this application is a Field Programmable Gate Array (FPGA). This AMC is subsidiary of a digital channelised receiver also implemented in the same platform.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Choosing an appropriate accounting system for manufacturing has always been a challenge for managers. In this article we try to compare three accounting systems designed since 1980 to address problems of traditional accounting system. In the first place we are going to present a short overview on background and definition of three accounting systems: Activity Based costing, Time-Driven Activity Based Costing and Lean Accounting. Comparisons are made based on the three basic roles of information generated by accounting systems: financial reporting, decision making, and operational control and improvement. The analysis in this paper reveals how decisions are made over the value stream in the companies using Lean Accounting while decisions under the ABC Accounting system are taken at individual product level, and finally we will show how TD-ABC covers both product and process levels for decision making. In addition, this paper shows the importance of nonfinancial measures for operational control and improvement under the Lean Accounting and TD-ABC methods whereas ABC relies mostly on financial measures in this context.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Disruptions in tokamaks devices are unavoidable, and they can have a significant impact on machine integrity. So it is very important have mechanisms to predict this phenomenon. Disruption prediction is a very complex task, not only because it is a multi-dimensional problem, but also because in order to be effective, it has to detect well in advance the actual disruptive event, in order to be able to use successful mitigation strategies. With these constraints in mind a real-time disruption predictor has been developed to be used in JET tokamak. The predictor has been designed to run in the Multithreaded Application Real-Time executor (MARTe) framework. The predictor ?Advanced Predictor Of DISruptions? (APODIS) is based on Support Vector Machine (SVM).

Relevância:

40.00% 40.00%

Publicador:

Resumo:

El presente trabajo se basa en la filosofía de la Construcción sin Pérdidas (“Lean Construction”), analizando la situación de esta filosofía en el sector de la edificación en el contexto internacional y español, respondiendo las siguientes preguntas: 1. ¿Cómo surge el “Lean Construction”? 2. ¿Cuáles son sus actividades, funciones y cometidos? 3. ¿Existe regulación del ¨Lean Construction” en otros países? 4. ¿Existe demanda del ¨Lean Construction” en España? 5. ¿Existe regulación del ¨Lean Construction” en España? 6. ¿Cómo debería ser la regulación ¨Lean Construction” en España? 7. ¿Cuál es la relación del “Lean Construction” con el “Project & Construction Management”? 8. ¿Cómo debería ser la regulación de “Lean Construction” en España considerando su relación con el “Project & Construction Management”? Las preguntas indicadas las hemos respondido detalladamente en el presente trabajo, a continuación se resume las respuestas a dichas preguntas: 1. El “Lean Construction” surge en agosto de 1992, cuando el investigador finlandés Lauri Koskela publicó en la Universidad de Stanford el reporte TECHNICAL REPORT N° 72 titulado “Application of the New Production Philosophy to Construction”. Un año más tarde el Dr. Koskela invitó a un grupo de especialistas en construcción al primer workshop de esta materia en Finlandia, dando origen al International Group for Lean Construction (IGLC) lo que ha permitido extender la filosofía a EEUU, Europa, América, Asia, Oceanía y África. “Lean Construction” es un sistema basado en el enfoque “Lean Production” desarrollado en Japón por Toyota Motors a partir de los años cincuenta, sistema que permitió a sus fábricas producir unidades con mayor eficiencia que las industrias americanas, con menores recursos, en menor tiempo, y con un número menor de errores de fabricación. 2. El sistema “Lean Construction” busca maximizar el valor y disminuir las pérdidas de los proyectos generando una coordinación eficiente entre los involucrados, manejando un proyecto como un sistema de producción, estrechando la colaboración entre los participantes de los proyectos, capacitándoles y empoderándoles, fomentando una cultura de cambio. Su propósito es desarrollar un proceso de construcción en el que no hayan accidentes, ni daños a equipos, instalaciones, entorno y comunidad, que se realice en conformidad con los requerimientos contractuales, sin defectos, en el plazo requerido, respetando los costes presupuestados y con un claro enfoque en la eliminación o reducción de las pérdidas, es decir, las actividades que no generen beneficios. El “Last Planner System”, o “Sistema del Último Planificador”, es un sistema del “Lean Construction” que por su propia naturaleza protege a la planificación y, por ende, ayuda a maximizar el valor y minimizar las pérdidas, optimizando de manera sustancial los sistemas de seguridad y salud. El “Lean Construction” se inició como un concepto enfocado a la ejecución de las obras, posteriormente se aplicó la filosofía a todas las etapas del proyecto. Actualmente considera el desarrollo total de un proyecto, desde que nace la idea hasta la culminación de la obra y puesta en marcha, considerando el ciclo de vida completo del proyecto. Es una filosofía de gestión, metodologías de trabajo y una cultura empresarial orientada a la eficiencia de los procesos y flujos. La filosofía “Lean Construction” se está expandiendo en todo el mundo, además está creciendo en su alcance, influyendo en la gestión contractual de los proyectos. Su primera evolución consistió en la creación del sistema “Lean Project Delivery System”, que es el concepto global de desarrollo de proyectos. Posteriormente, se proponen el “Target Value Design”, que consiste en diseñar de forma colaborativa para alcanzar los costes y el valor requerido, y el “Integrated Project Delivery”, en relación con sistemas de contratos relacionales (colaborativos) integrados, distintos a los contratos convencionales. 3. Se verificó que no existe regulación específica del ¨Lean Construction” en otros países, en otras palabras, no existe el agente con el nombre específico de “Especialista en Lean Construction” o similar, en consecuencia, es un agente adicional en el proyecto de la edificación, cuyas funciones y cometidos se pueden solapar con los del “Project Manager”, “Construction Manager”, “Contract Manager”, “Safety Manager”, entre otros. Sin embargo, se comprobó la existencia de formatos privados de contratos colaborativos de Integrated Project Delivery, los cuales podrían ser tomados como unas primeras referencias para futuras regulaciones. 4. Se verificó que sí existe demanda del ¨Lean Construction” en el desarrollo del presente trabajo, aunque aún su uso es incipiente, cada día existe más interesados en el tema. 5. No existe regulación del ¨Lean Construction” en España. 6. Uno de los objetivos fundamentales de esta tesis es el de regular esta figura cuando actúe en un proyecto, definir y realizar una estructura de Agente de la Edificación, según la Ley de Ordenación de la Edificación (LOE), y de esta manera poder introducirla dentro de la Legislación Española, protegiéndola de eventuales responsabilidades civiles. En España existe jurisprudencia (sentencias de los tribunales de justicia españoles) con jurisdicción civil basada en la LOE para absolver o condenar a agentes de la edificación que son definidos en los tribunales como “gestores constructivos” o similares. Por este motivo, en un futuro los tribunales podrían dictaminar responsabilidades solidarias entre el especialista “Lean Construction” y otros agentes del proyecto, dependiendo de sus actuaciones, y según se implemente el “Lean Project Delivery System”, el “Target Value Design” y el “Integrated Project Delivery”. Por otro lado, es posible que el nivel de actuación del especialista “Lean Construcción” pueda abarcar la gestión del diseño, la gestión de la ejecución material (construcción), la gestión de contratos, o la gestión integral de todo el proyecto de edificación, esto último, en concordancia con la última Norma ISO 21500:2012 o UNE-ISO 21500:2013 Directrices para la dirección y gestión de proyectos. En consecuencia, se debería incorporar adecuadamente a uno o más agentes de la edificación en la LOE de acuerdo a sus funciones y responsabilidades según los niveles de actuación del “Especialista en Lean Construction”. Se propone la creación de los siguientes agentes: Gestor del Diseño, Gestor Constructivo y Gestor de Contratos, cuyas definiciones están desarrolladas en este trabajo. Estas figuras son definidas de manera general, puesto que cualquier “Project Manager” o “DIPE”, gestor BIM (Building Information Modeling), o similar, puede actuar como uno o varios de ellos. También se propone la creación del agente “Gestor de la Construcción sin Pérdidas”, como aquel agente que asume las actuaciones del “gestor de diseño”, “gestor constructivo” y “gestor de contratos” con un enfoque en los principios del Lean Production. 7. En la tesis se demuestra, por medio del uso de la ISO 21500, que ambos sistemas son complementarios, de manera que los proyectos pueden tener ambos enfoques y ser compatibilizados. Un proyecto que use el “Project & Construction Management” puede perfectamente apoyarse en las herramientas y técnicas del “Lean Construction” para asegurar la eliminación o reducción de las pérdidas, es decir, las actividades que no generen valor, diseñando el sistema de producción, el sistema de diseño o el sistema de contratos. 8. Se debería incorporar adecuadamente al agente de la edificación “Especialista en Lean Construction” o similar y al agente ¨Especialista en Project & Construction Management” o DIPE en la Ley de Ordenación de la Edificación (LOE) de acuerdo a sus funciones y responsabilidades, puesto que la jurisprudencia se ha basado para absolver o condenar en la referida Ley. Uno de los objetivos fundamentales de esta tesis es el de regular la figura del “Especialista en Lean Construction” cuando actúa simultáneamente con el DIPE, y realizar una estructura de Agente de la Edificación según la LOE, y de esta manera protegerlo de eventuales responsabilidades solidarias. Esta investigación comprueba que la propuesta de definición del agente de edificación DIPE, según la LOE, presentada en la tesis doctoral del Doctor Manuel Soler Severino es compatible con las nuevas definiciones propuestas. El agente DIPE puede asumir los roles de los diferentes gestores propuestos en esta tesis si es que se especializa en dichas materias, o, si lo estima pertinente, recomendar sus contrataciones. ABSTRACT This work is based on the Lean Construction philosophy; an analysis is made herein with regard to the situation of this philosophy in the building sector within the international and Spanish context, replying to the following questions: 1. How did the concept of Lean Construction emerge? 2. Which are the activities, functions and objectives of Lean Construction? 3. Are there regulations on Lean Construction in other countries? 4. Is there a demand for Lean Construction in Spain? 5. Are there regulations on Lean Construction in Spain? 6. How should regulations on Lean Construction be developed in Spain? 7. What is the relationship between Lean Construction and the Project & Construction Management? 8. How should regulations on Lean Construction be developed in Spain considering its relationship with the Project & Construction Management? We have answered these questions in detail here and the replies are summarized as follows: 1. The concept of Lean Construction emerged in august of 1992, when Finnish researcher Lauri Koskela published in Stanford University TECHNICAL REPORT N° 72 entitled “Application of the New Production Philosophy to Construction”. A year later, Professor Koskela invited a group of construction specialists to Finland to the first workshop conducted on this matter; thus, the International Group for Lean Construction (IGLC) was established, which has contributed to extending the philosophy to the United States, Europe, the Americas, Asia, Oceania, and Africa. Lean Construction is a system based on the Lean Production approach, which was developed in Japan by Toyota Motors in the 1950s. Thanks to this system, the Toyota plants were able to produce more units, with greater efficiency than the American industry, less resources, in less time, and with fewer manufacturing errors. 2. The Lean Construction system aims at maximizing the value of projects while reducing waste, producing an effective coordination among those involved; it manages projects as a production system, enhancing collaboration between the parties that participate in the projects while building their capacities, empowering them, and promoting a culture of change. Its purpose is to develop a construction process free of accidents, without damages to the equipment, facilities, environment and community, flawless, in accordance with contractual requirements, within the terms established, respecting budgeted costs, and with a clear approach to eliminating or reducing waste, that is, activities that do not generate benefits. The Last Planner System is a Lean Construction system, which by its own nature protects planning and, therefore, helps to maximize the value and minimize waste, optimizing substantially the safety and health systems. Lean Construction started as a concept focused on the execution of works, and subsequently the philosophy was applied to all the stages of the project. At present it considers the project’s total development, since the time ideas are born until the completion and start-up of the work, taking into account the entire life cycle of the project. It is a philosophy of management, work methodologies, and entrepreneurial culture aimed at the effectiveness of processes and flows. The Lean Construction philosophy is extending all over the world and its scope is becoming broader, having greater influence on the contractual management of projects. It evolved initially through the creation of the Lean Project Delivery System, a global project development concept. Later on, the Target Value Design was developed, based on collaborative design to achieve the costs and value required, as well as the Integrated Project Delivery, in connection with integrated relational (collaborative) contract systems, as opposed to conventional contracts. 3. It was verified that no specific regulations on Lean Construction exist in other countries, in other words, there are no agents with the specific name of “Lean Construction Specialist” or other similar names; therefore, it is an additional agent in building projects, which functions and objectives can overlap those of the Project Manager, Construction Manager, Contract Manager, or Safety Manager, among others. However, the existence of private collaborative contracts of Integrated Project Delivery was confirmed, which could be considered as first references for future regulations. 4. There is a demand for Lean Construction in the development of this work; even though it is still emerging, there is a growing interest in this topic. 5. There are no regulations on Lean Construction in Spain. 6. One of the main objectives of this thesis is to regulate this role when acting in a project, and to define and develop a Building Agent structure, according to the Building Standards Law (LOE by its acronym in Spanish), in order to be able to incorporate it into the Spanish law, protecting it from civil liabilities. In Spain there is jurisprudence in civil jurisdiction based on the LOE to acquit or convict building agents, which are defined in the courts as “construction managers” or similar. For this reason, courts could establish in the future joint and several liabilities between the Lean Construction Specialist and other agents of the project, depending on their actions and based on the implementation of the Lean Project Delivery System, the Target Value Design, and the Integrated Project Delivery. On the other hand, it is possible that the level of action of the Lean Construction Specialist may comprise design management, construction management and contract management, or the integral management of the entire building project in accordance with the last ISO 21500:2012 or UNE-ISO 21500:2013, guidelines for the management of projects. Accordingly, one or more building agents should be appropriately incorporated into the LOE according to their functions and responsibilities and based on the levels of action of the Lean Construction Specialist. The creation of the following agents is proposed: Design Manager, Construction Manager, and Contract Manager, which definitions are developed in this work. These agents are defined in general, since any Project Manager or DIPE, Building Information Modeling (BIM) Manager or similar, may act as one or as many of them. The creation of the Lean Construction Manager is also proposed, as the agent that takes on the role of the Design Manager, Construction Manager and Contract Manager with a focus on the Lean Production principles. 7. In the thesis it is demonstrated that through the implementation of the ISO 21500, both systems are supplementary, so projects may have both approaches and be compatible. A project that applies the Project & Construction Management may perfectly have the support of the tools, techniques and practices of Lean Construction to ensure the elimination or reduction of losses, that is, those activities that do not generate value, thus designing the production system, the design system, or the contract system. 8. The Lean Construction Specialist or similar and the Specialist in Project & Construction Management should be incorporated appropriately into the LOE according to their functions and responsibilities, since jurisprudence has been based on such Law to acquit or convict. One of the main objectives of this thesis is the regulate the role of the Lean Construction Specialist when acting simultaneously with the DIPE, and to develop a structure of the building agent, according to the LOE, and in this way protect such agent from joint and several liabilities. This research proves that the proposal to define the DIPE building agent, according to the LOE, and presented in the doctoral dissertation of Manuel Soler Severino, Ph.D. is compatible with the new definitions proposed. The DIPE agent may assume the roles of the different managers proposed in this thesis if he specializes in those topics or, if deemed pertinent, recommends that they be engaged.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Distributed real-time embedded systems are becoming increasingly important to society. More demands will be made on them and greater reliance will be placed on the delivery of their services. A relevant subset of them is high-integrity or hard real-time systems, where failure can cause loss of life, environmental harm, or significant financial loss. Additionally, the evolution of communication networks and paradigms as well as the necessity of demanding processing power and fault tolerance, motivated the interconnection between electronic devices; many of the communications have the possibility of transferring data at a high speed. The concept of distributed systems emerged as systems where different parts are executed on several nodes that interact with each other via a communication network. Java’s popularity, facilities and platform independence have made it an interesting language for the real-time and embedded community. This was the motivation for the development of RTSJ (Real-Time Specification for Java), which is a language extension intended to allow the development of real-time systems. The use of Java in the development of high-integrity systems requires strict development and testing techniques. However, RTJS includes a number of language features that are forbidden in such systems. In the context of the HIJA project, the HRTJ (Hard Real-Time Java) profile was developed to define a robust subset of the language that is amenable to static analysis for high-integrity system certification. Currently, a specification under the Java community process (JSR- 302) is being developed. Its purpose is to define those capabilities needed to create safety critical applications with Java technology called Safety Critical Java (SCJ). However, neither RTSJ nor its profiles provide facilities to develop distributed realtime applications. This is an important issue, as most of the current and future systems will be distributed. The Distributed RTSJ (DRTSJ) Expert Group was created under the Java community process (JSR-50) in order to define appropriate abstractions to overcome this problem. Currently there is no formal specification. The aim of this thesis is to develop a communication middleware that is suitable for the development of distributed hard real-time systems in Java, based on the integration between the RMI (Remote Method Invocation) model and the HRTJ profile. It has been designed and implemented keeping in mind the main requirements such as the predictability and reliability in the timing behavior and the resource usage. iThe design starts with the definition of a computational model which identifies among other things: the communication model, most appropriate underlying network protocols, the analysis model, and a subset of Java for hard real-time systems. In the design, the remote references are the basic means for building distributed applications which are associated with all non-functional parameters and resources needed to implement synchronous or asynchronous remote invocations with real-time attributes. The proposed middleware separates the resource allocation from the execution itself by defining two phases and a specific threading mechanism that guarantees a suitable timing behavior. It also includes mechanisms to monitor the functional and the timing behavior. It provides independence from network protocol defining a network interface and modules. The JRMP protocol was modified to include two phases, non-functional parameters, and message size optimizations. Although serialization is one of the fundamental operations to ensure proper data transmission, current implementations are not suitable for hard real-time systems and there are no alternatives. This thesis proposes a predictable serialization that introduces a new compiler to generate optimized code according to the computational model. The proposed solution has the advantage of allowing us to schedule the communications and to adjust the memory usage at compilation time. In order to validate the design and the implementation a demanding validation process was carried out with emphasis in the functional behavior, the memory usage, the processor usage (the end-to-end response time and the response time in each functional block) and the network usage (real consumption according to the calculated consumption). The results obtained in an industrial application developed by Thales Avionics (a Flight Management System) and in exhaustive tests show that the design and the prototype are reliable for industrial applications with strict timing requirements. Los sistemas empotrados y distribuidos de tiempo real son cada vez más importantes para la sociedad. Su demanda aumenta y cada vez más dependemos de los servicios que proporcionan. Los sistemas de alta integridad constituyen un subconjunto de gran importancia. Se caracterizan por que un fallo en su funcionamiento puede causar pérdida de vidas humanas, daños en el medio ambiente o cuantiosas pérdidas económicas. La necesidad de satisfacer requisitos temporales estrictos, hace más complejo su desarrollo. Mientras que los sistemas empotrados se sigan expandiendo en nuestra sociedad, es necesario garantizar un coste de desarrollo ajustado mediante el uso técnicas adecuadas en su diseño, mantenimiento y certificación. En concreto, se requiere una tecnología flexible e independiente del hardware. La evolución de las redes y paradigmas de comunicación, así como la necesidad de mayor potencia de cómputo y de tolerancia a fallos, ha motivado la interconexión de dispositivos electrónicos. Los mecanismos de comunicación permiten la transferencia de datos con alta velocidad de transmisión. En este contexto, el concepto de sistema distribuido ha emergido como sistemas donde sus componentes se ejecutan en varios nodos en paralelo y que interactúan entre ellos mediante redes de comunicaciones. Un concepto interesante son los sistemas de tiempo real neutrales respecto a la plataforma de ejecución. Se caracterizan por la falta de conocimiento de esta plataforma durante su diseño. Esta propiedad es relevante, por que conviene que se ejecuten en la mayor variedad de arquitecturas, tienen una vida media mayor de diez anos y el lugar ˜ donde se ejecutan puede variar. El lenguaje de programación Java es una buena base para el desarrollo de este tipo de sistemas. Por este motivo se ha creado RTSJ (Real-Time Specification for Java), que es una extensión del lenguaje para permitir el desarrollo de sistemas de tiempo real. Sin embargo, RTSJ no proporciona facilidades para el desarrollo de aplicaciones distribuidas de tiempo real. Es una limitación importante dado que la mayoría de los actuales y futuros sistemas serán distribuidos. El grupo DRTSJ (DistributedRTSJ) fue creado bajo el proceso de la comunidad de Java (JSR-50) con el fin de definir las abstracciones que aborden dicha limitación, pero en la actualidad aun no existe una especificacion formal. El objetivo de esta tesis es desarrollar un middleware de comunicaciones para el desarrollo de sistemas distribuidos de tiempo real en Java, basado en la integración entre el modelo de RMI (Remote Method Invocation) y el perfil HRTJ. Ha sido diseñado e implementado teniendo en cuenta los requisitos principales, como la predecibilidad y la confiabilidad del comportamiento temporal y el uso de recursos. El diseño parte de la definición de un modelo computacional el cual identifica entre otras cosas: el modelo de comunicaciones, los protocolos de red subyacentes más adecuados, el modelo de análisis, y un subconjunto de Java para sistemas de tiempo real crítico. En el diseño, las referencias remotas son el medio básico para construcción de aplicaciones distribuidas las cuales son asociadas a todos los parámetros no funcionales y los recursos necesarios para la ejecución de invocaciones remotas síncronas o asíncronas con atributos de tiempo real. El middleware propuesto separa la asignación de recursos de la propia ejecución definiendo dos fases y un mecanismo de hebras especifico que garantiza un comportamiento temporal adecuado. Además se ha incluido mecanismos para supervisar el comportamiento funcional y temporal. Se ha buscado independencia del protocolo de red definiendo una interfaz de red y módulos específicos. También se ha modificado el protocolo JRMP para incluir diferentes fases, parámetros no funcionales y optimizaciones de los tamaños de los mensajes. Aunque la serialización es una de las operaciones fundamentales para asegurar la adecuada transmisión de datos, las actuales implementaciones no son adecuadas para sistemas críticos y no hay alternativas. Este trabajo propone una serialización predecible que ha implicado el desarrollo de un nuevo compilador para la generación de código optimizado acorde al modelo computacional. La solución propuesta tiene la ventaja que en tiempo de compilación nos permite planificar las comunicaciones y ajustar el uso de memoria. Con el objetivo de validar el diseño e implementación se ha llevado a cabo un exigente proceso de validación con énfasis en: el comportamiento funcional, el uso de memoria, el uso del procesador (tiempo de respuesta de extremo a extremo y en cada uno de los bloques funcionales) y el uso de la red (consumo real conforme al estimado). Los buenos resultados obtenidos en una aplicación industrial desarrollada por Thales Avionics (un sistema de gestión de vuelo) y en las pruebas exhaustivas han demostrado que el diseño y el prototipo son fiables para aplicaciones industriales con estrictos requisitos temporales.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Membrane systems are computational equivalent to Turing machines. However, their distributed and massively parallel nature obtains polynomial solutions opposite to traditional non-polynomial ones. At this point, it is very important to develop dedicated hardware and software implementations exploiting those two membrane systems features. Dealing with distributed implementations of P systems, the bottleneck communication problem has arisen. When the number of membranes grows up, the network gets congested. The purpose of distributed architectures is to reach a compromise between the massively parallel character of the system and the needed evolution step time to transit from one configuration of the system to the next one, solving the bottleneck communication problem. The goal of this paper is twofold. Firstly, to survey in a systematic and uniform way the main results regarding the way membranes can be placed on processors in order to get a software/hardware simulation of P-Systems in a distributed environment. Secondly, we improve some results about the membrane dissolution problem, prove that it is connected, and discuss the possibility of simulating this property in the distributed model. All this yields an improvement in the system parallelism implementation since it gets an increment of the parallelism of the external communication among processors. Proposed ideas improve previous architectures to tackle the communication bottleneck problem, such as reduction of the total time of an evolution step, increase of the number of membranes that could run on a processor and reduction of the number of processors.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Linear regression is a technique widely used in digital signal processing. It consists on finding the linear function that better fits a given set of samples. This paper proposes different hardware architectures for the implementation of the linear regression method on FPGAs, specially targeting area restrictive systems. It saves area at the cost of constraining the lengths of the input signal to some fixed values. We have implemented the proposed scheme in an Automatic Modulation Classifier, meeting the hard real-time constraints this kind of systems have.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Although several profiling techniques for identifying performance bottlenecks in logic programs have been developed, they are generally not automatic and in most cases they do not provide enough information for identifying the root causes of such bottlenecks. This complicates using their results for guiding performance improvement. We present a profiling method and tool that provides such explanations. Our profiler associates cost centers to certain program elements and can measure different types of resource-related properties that affect performance, preserving the precedence of cost centers in the cali graph. It includes an automatic method for detecting procedures that are performance bottlenecks. The profiling tool has been integrated in a previously developed run-time checking framework to allow verification of certain properties when they cannot be verified statically. The approach allows checking global computational properties which require complex instrumentation tracking information about previous execution states, such as, e.g., that the execution time accumulated by a given procedure is not greater than a given bound. We have built a prototype implementation, integrated it in the Ciao/CiaoPP system and successfully applied it to performance improvement, automatic optimization (e.g., resource-aware specialization of programs), run-time checking, and debugging of global computational properties (e.g., resource usage) in Prolog programs.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Traditional schemes for abstract interpretation-based global analysis of logic programs generally focus on obtaining procedure argument mode and type information. Variable sharing information is often given only the attention needed to preserve the correctness of the analysis. However, such sharing information can be very useful. In particular, it can be used for predicting runtime goal independence, which can eliminate costly run-time checks in and-parallel execution. In this paper, a new algorithm for doing abstract interpretation in logic programs is described which concentrates on inferring the dependencies of the terms bound to program variables with increased precisión and at all points in the execution of the program, rather than just at a procedure level. Algorithms are presented for computing abstract entry and success substitutions which extensively keep track of variable aliasing and term dependence information. In addition, a new, abstract domain independent ñxpoint algorithm is presented and described in detail. The algorithms are illustrated with examples. Finally, results from an implementation of the abstract interpreter are presented.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Abstract machines provide a certain separation between platformdependent and platform-independent concerns in compilation. Many of the differences between architectures are encapsulated in the speciflc abstract machine implementation and the bytecode is left largely architecture independent. Taking advantage of this fact, we present a framework for estimating upper and lower bounds on the execution times of logic programs running on a bytecode-based abstract machine. Our approach includes a one-time, programindependent proflling stage which calculates constants or functions bounding the execution time of each abstract machine instruction. Then, a compile-time cost estimation phase, using the instruction timing information, infers expressions giving platform-dependent upper and lower bounds on actual execution time as functions of input data sizes for each program. Working at the abstract machine level makes it possible to take into account low-level issues in new architectures and platforms by just reexecuting the calibration stage instead of having to tailor the analysis for each architecture and platform. Applications of such predicted execution times include debugging/veriflcation of time properties, certiflcation of time properties in mobile code, granularity control in parallel/distributed computing, and resource-oriented specialization.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

CIAO is an advanced programming environment supporting Logic and Constraint programming. It offers a simple concurrent kernel on top of which declarative and non-declarative extensions are added via librarles. Librarles are available for supporting the ISOProlog standard, several constraint domains, functional and higher order programming, concurrent and distributed programming, internet programming, and others. The source language allows declaring properties of predicates via assertions, including types and modes. Such properties are checked at compile-time or at run-time. The compiler and system architecture are designed to natively support modular global analysis, with the two objectives of proving properties in assertions and performing program optimizations, including transparently exploiting parallelism in programs. The purpose of this paper is to report on recent progress made in the context of the CIAO system, with special emphasis on the capabilities of the compiler, the techniques used for supporting such capabilities, and the results in the áreas of program analysis and transformation already obtained with the system.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Although several profiling techniques for identifying performance bottlenecks in logic programs have been developed, they are generally not automatic and in most cases they do not provide enough information for identifying the root causes of such bottlenecks. This complicates using their results for guiding performance improvement. We present a profiling method and tool that provides such explanations. Our profiler associates cost centers to certain program elements and can measure different types of resource-related properties that affect performance, preserving the precedence of cost centers in the call graph. It includes an automatic method for detecting procedures that are performance bottlenecks. The profiling tool has been integrated in a previously developed run-time checking framework to allow verification of certain properties when they cannot be verified statically. The approach allows checking global computational properties which require complex instrumentation tracking information about previous execution states, such as, e.g., that the execution time accumulated by a given procedure is not greater than a given bound. We have built a prototype implementation, integrated it in the Ciao/CiaoPP system and successfully applied it to performance improvement, automatic optimization (e.g., resource-aware specialization of programs), run-time checking, and debugging of global computational properties (e.g., resource usage) in Prolog programs.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Las técnicas de cirugía de mínima invasión (CMI) se están consolidando hoy en día como alternativa a la cirugía tradicional, debido a sus numerosos beneficios para los pacientes. Este cambio de paradigma implica que los cirujanos deben aprender una serie de habilidades distintas de aquellas requeridas en cirugía abierta. El entrenamiento y evaluación de estas habilidades se ha convertido en una de las mayores preocupaciones en los programas de formación de cirujanos, debido en gran parte a la presión de una sociedad que exige cirujanos bien preparados y una reducción en el número de errores médicos. Por tanto, se está prestando especial atención a la definición de nuevos programas que permitan el entrenamiento y la evaluación de las habilidades psicomotoras en entornos seguros antes de que los nuevos cirujanos puedan operar sobre pacientes reales. Para tal fin, hospitales y centros de formación están gradualmente incorporando instalaciones de entrenamiento donde los residentes puedan practicar y aprender sin riesgos. Es cada vez más común que estos laboratorios dispongan de simuladores virtuales o simuladores físicos capaces de registrar los movimientos del instrumental de cada residente. Estos simuladores ofrecen una gran variedad de tareas de entrenamiento y evaluación, así como la posibilidad de obtener información objetiva de los ejercicios. Los diferentes estudios de validación llevados a cabo dan muestra de su utilidad; pese a todo, los niveles de evidencia presentados son en muchas ocasiones insuficientes. Lo que es más importante, no existe un consenso claro a la hora de definir qué métricas son más útiles para caracterizar la pericia quirúrgica. El objetivo de esta tesis doctoral es diseñar y validar un marco de trabajo conceptual para la definición y validación de entornos para la evaluación de habilidades en CMI, en base a un modelo en tres fases: pedagógica (tareas y métricas a emplear), tecnológica (tecnologías de adquisición de métricas) y analítica (interpretación de la competencia en base a las métricas). Para tal fin, se describe la implementación práctica de un entorno basado en (1) un sistema de seguimiento de instrumental fundamentado en el análisis del vídeo laparoscópico; y (2) la determinación de la pericia en base a métricas de movimiento del instrumental. Para la fase pedagógica se diseñó e implementó un conjunto de tareas para la evaluación de habilidades psicomotoras básicas, así como una serie de métricas de movimiento. La validación de construcción llevada a cabo sobre ellas mostró buenos resultados para tiempo, camino recorrido, profundidad, velocidad media, aceleración media, economía de área y economía de volumen. Adicionalmente, los resultados obtenidos en la validación de apariencia fueron en general positivos en todos los grupos considerados (noveles, residentes, expertos). Para la fase tecnológica, se introdujo el EVA Tracking System, una solución para el seguimiento del instrumental quirúrgico basado en el análisis del vídeo endoscópico. La precisión del sistema se evaluó a 16,33ppRMS para el seguimiento 2D de la herramienta en la imagen; y a 13mmRMS para el seguimiento espacial de la misma. La validación de construcción con una de las tareas de evaluación mostró buenos resultados para tiempo, camino recorrido, profundidad, velocidad media, aceleración media, economía de área y economía de volumen. La validación concurrente con el TrEndo® Tracking System por su parte presentó valores altos de correlación para 8 de las 9 métricas analizadas. Finalmente, para la fase analítica se comparó el comportamiento de tres clasificadores supervisados a la hora de determinar automáticamente la pericia quirúrgica en base a la información de movimiento del instrumental, basados en aproximaciones lineales (análisis lineal discriminante, LDA), no lineales (máquinas de soporte vectorial, SVM) y difusas (sistemas adaptativos de inferencia neurodifusa, ANFIS). Los resultados muestran que en media SVM presenta un comportamiento ligeramente superior: 78,2% frente a los 71% y 71,7% obtenidos por ANFIS y LDA respectivamente. Sin embargo las diferencias estadísticas medidas entre los tres no fueron demostradas significativas. En general, esta tesis doctoral corrobora las hipótesis de investigación postuladas relativas a la definición de sistemas de evaluación de habilidades para cirugía de mínima invasión, a la utilidad del análisis de vídeo como fuente de información y a la importancia de la información de movimiento de instrumental a la hora de caracterizar la pericia quirúrgica. Basándose en estos cimientos, se han de abrir nuevos campos de investigación que contribuyan a la definición de programas de formación estructurados y objetivos, que puedan garantizar la acreditación de cirujanos sobradamente preparados y promocionen la seguridad del paciente en el quirófano. Abstract Minimally invasive surgery (MIS) techniques have become a standard in many surgical sub-specialties, due to their many benefits for patients. However, this shift in paradigm implies that surgeons must acquire a complete different set of skills than those normally attributed to open surgery. Training and assessment of these skills has become a major concern in surgical learning programmes, especially considering the social demand for better-prepared professionals and for the decrease of medical errors. Therefore, much effort is being put in the definition of structured MIS learning programmes, where practice with real patients in the operating room (OR) can be delayed until the resident can attest for a minimum level of psychomotor competence. To this end, skills’ laboratory settings are being introduced in hospitals and training centres where residents may practice and be assessed on their psychomotor skills. Technological advances in the field of tracking technologies and virtual reality (VR) have enabled the creation of new learning systems such as VR simulators or enhanced box trainers. These systems offer a wide range of tasks, as well as the capability of registering objective data on the trainees’ performance. Validation studies give proof of their usefulness; however, levels of evidence reported are in many cases low. More importantly, there is still no clear consensus on topics such as the optimal metrics that must be used to assess competence, the validity of VR simulation, the portability of tracking technologies into real surgeries (for advanced assessment) or the degree to which the skills measured and obtained in laboratory environments transfer to the OR. The purpose of this PhD is to design and validate a conceptual framework for the definition and validation of MIS assessment environments based on a three-pillared model defining three main stages: pedagogical (tasks and metrics to employ), technological (metric acquisition technologies) and analytical (interpretation of competence based on metrics). To this end, a practical implementation of the framework is presented, focused on (1) a video-based tracking system and (2) the determination of surgical competence based on the laparoscopic instruments’ motionrelated data. The pedagogical stage’s results led to the design and implementation of a set of basic tasks for MIS psychomotor skills’ assessment, as well as the definition of motion analysis parameters (MAPs) to measure performance on said tasks. Validation yielded good construct results for parameters such as time, path length, depth, average speed, average acceleration, economy of area and economy of volume. Additionally, face validation results showed positive acceptance on behalf of the experts, residents and novices. For the technological stage the EVA Tracking System is introduced. EVA provides a solution for tracking laparoscopic instruments from the analysis of the monoscopic video image. Accuracy tests for the system are presented, which yielded an average RMSE of 16.33pp for 2D tracking of the instrument on the image and of 13mm for 3D spatial tracking. A validation experiment was conducted using one of the tasks and the most relevant MAPs. Construct validation showed significant differences for time, path length, depth, average speed, average acceleration, economy of area and economy of volume; especially between novices and residents/experts. More importantly, concurrent validation with the TrEndo® Tracking System presented high correlation values (>0.7) for 8 of the 9 MAPs proposed. Finally, the analytical stage allowed comparing the performance of three different supervised classification strategies in the determination of surgical competence based on motion-related information. The three classifiers were based on linear (linear discriminant analysis, LDA), non-linear (support vector machines, SVM) and fuzzy (adaptive neuro fuzzy inference systems, ANFIS) approaches. Results for SVM show slightly better performance than the other two classifiers: on average, accuracy for LDA, SVM and ANFIS was of 71.7%, 78.2% and 71% respectively. However, when confronted, no statistical significance was found between any of the three. Overall, this PhD corroborates the investigated research hypotheses regarding the definition of MIS assessment systems, the use of endoscopic video analysis as the main source of information and the relevance of motion analysis in the determination of surgical competence. New research fields in the training and assessment of MIS surgeons can be proposed based on these foundations, in order to contribute to the definition of structured and objective learning programmes that guarantee the accreditation of well-prepared professionals and the promotion of patient safety in the OR.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The stepped and excessively slow execution of pseudo-dynamic tests has been found to be the source of some errors arising from strain-rate effect and stress relaxation. In order to control those errors, a new continuous test method which allows the selection of a more suitable time scale factor in the response is proposed in this work. By dimensional analysis, such scaled-time response is obtained theoretically by augmenting the inertial and damping properties of the structure, for which we propose the use of hydraulic pistons which are servo-controlled to produce active mass and damping, nevertheless using an equipment which is similar to that required in a pseudo-dynamic test. The results of the successful implementation of this technique for a simple specimen are shown here.