994 resultados para Fundamental operations


Relevância:

70.00% 70.00%

Publicador:

Resumo:

Operations Management is all around us and is integral to every industry. Using contemporary and engaging examples this brand new text book brings to life fundamental Operations Management principles and theories that are applicable to both manufacturing and service situations, reflecting the very latest developments in this dynamic field.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Rolling element bearings are the most critical components in the traction system of high speed trains. Monitoring their integrity is a fundamental operation in order to avoid catastrophic failures and to implement effective condition based maintenance strategies. Generally, diagnostics of rolling element bearings is usually performed by analyzing vibration signals measured by accelerometers placed in the proximity of the bearing under investigation. Several papers have been published on this subject in the last two decades, mainly devoted to the development and assessment of signal processing techniques for diagnostics. The experimental validation of such techniques has been traditionally performed by means of laboratory tests on artificially damaged bearings, while their actual effectiveness in specific industrial applications, particularly in rail industry, remains scarcely investigated. This paper is aimed at filling this knowledge gap, by addressing the diagnostics of bearings taken from the service after a long term operation on the traction system of a high speed train. Moreover, in order to test the effectiveness of the diagnostic procedures in the environmental conditions peculiar to the rail application, a specific test-rig has been built, consisting of a complete full-scale train traction system, able to reproduce the effects of wheeltrack interaction and bogie-wheelset dynamics. The results of the experimental campaign show that suitable signal processing techniques are able to diagnose bearing failures even in this harsh and noisy application. Moreover, the most suitable location of the sensors on the traction system is proposed, in order to limit their number.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The K-best detector is considered as a promising technique in the MIMO-OFDM detection because of its good performance and low complexity. In this paper, a new K-best VLSI architecture is presented. In the proposed architecture, the metric computation units (MCUs) expand each surviving path only to its partial branches, based on the novel expansion scheme, which can predetermine the branches' ascending order by their local distances. Then a distributed sorter sorts out the new K surviving paths from the expanded branches in pipelines. Compared to the conventional K-best scheme, the proposed architecture can approximately reduce fundamental operations by 50% and 75% for the 16-QAM and the 64-QAM cases, respectively, and, consequently, lower the demand on the hardware resource significantly. Simulation results prove that the proposed architecture can achieve a performance very similar to conventional K-best detectors. Hence, it is an efficient solution to the K-best detector's VLSI implementation for high-throughput MIMO-OFDM systems.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

本文在深刻理解空值语义的基础上,给出一种处理占位型空值的方法。讨论了空值环境下关系数据库的查询策略,定义了含三种查询操作的关系代数最小完备集中的关系代数运算,并对查询计算的有效性和完备性进行了分析。

Relevância:

60.00% 60.00%

Publicador:

Resumo:

本文以[1]中的扩展关系模型为基础在两种元组级的不完全信息──不确定及可能信息中引入属性级的不完全信息空值,使两种不同性质的不完全信息同时出现在同一关系中。为了能够查询到不同种类及不同确定程度的信息,文中制定了这种扩展关系模型上关系的查询策略,定义了能够体现这种策略的最小关系代数运算。

Relevância:

60.00% 60.00%

Publicador:

Resumo:

本文从空值语义及更新操作的关系出发,提出了一种新的扩展关系模型,用以组织更新操作下的含有空值的关系数据库中的信息.同时,定义了这种模型下的基本关系代数运算.为实现空值环境下关系数据库的数据更新奠定了基础。

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Transreal arithmetic is total, in the sense that the fundamental operations of addition, subtraction, multiplication and division can be applied to any transreal numbers with the result being a transreal number [1]. In particular division by zero is allowed. It is proved, in [3], that transreal arithmetic is consistent and contains real arithmetic. The entire set of transreal numbers is a total semantics that models all of the semantic values, that is truth values, commonly used in logics, such as the classical, dialetheaic, fuzzy and gap values [2]. By virtue of the totality of transreal arithmetic, these logics can be implemented using total, arithmetical functions, specifically operators, whose domain and counterdomain is the entire set of transreal numbers

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Pós-graduação em Educação Matemática - IGCE

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Distributed real-time embedded systems are becoming increasingly important to society. More demands will be made on them and greater reliance will be placed on the delivery of their services. A relevant subset of them is high-integrity or hard real-time systems, where failure can cause loss of life, environmental harm, or significant financial loss. Additionally, the evolution of communication networks and paradigms as well as the necessity of demanding processing power and fault tolerance, motivated the interconnection between electronic devices; many of the communications have the possibility of transferring data at a high speed. The concept of distributed systems emerged as systems where different parts are executed on several nodes that interact with each other via a communication network. Java’s popularity, facilities and platform independence have made it an interesting language for the real-time and embedded community. This was the motivation for the development of RTSJ (Real-Time Specification for Java), which is a language extension intended to allow the development of real-time systems. The use of Java in the development of high-integrity systems requires strict development and testing techniques. However, RTJS includes a number of language features that are forbidden in such systems. In the context of the HIJA project, the HRTJ (Hard Real-Time Java) profile was developed to define a robust subset of the language that is amenable to static analysis for high-integrity system certification. Currently, a specification under the Java community process (JSR- 302) is being developed. Its purpose is to define those capabilities needed to create safety critical applications with Java technology called Safety Critical Java (SCJ). However, neither RTSJ nor its profiles provide facilities to develop distributed realtime applications. This is an important issue, as most of the current and future systems will be distributed. The Distributed RTSJ (DRTSJ) Expert Group was created under the Java community process (JSR-50) in order to define appropriate abstractions to overcome this problem. Currently there is no formal specification. The aim of this thesis is to develop a communication middleware that is suitable for the development of distributed hard real-time systems in Java, based on the integration between the RMI (Remote Method Invocation) model and the HRTJ profile. It has been designed and implemented keeping in mind the main requirements such as the predictability and reliability in the timing behavior and the resource usage. iThe design starts with the definition of a computational model which identifies among other things: the communication model, most appropriate underlying network protocols, the analysis model, and a subset of Java for hard real-time systems. In the design, the remote references are the basic means for building distributed applications which are associated with all non-functional parameters and resources needed to implement synchronous or asynchronous remote invocations with real-time attributes. The proposed middleware separates the resource allocation from the execution itself by defining two phases and a specific threading mechanism that guarantees a suitable timing behavior. It also includes mechanisms to monitor the functional and the timing behavior. It provides independence from network protocol defining a network interface and modules. The JRMP protocol was modified to include two phases, non-functional parameters, and message size optimizations. Although serialization is one of the fundamental operations to ensure proper data transmission, current implementations are not suitable for hard real-time systems and there are no alternatives. This thesis proposes a predictable serialization that introduces a new compiler to generate optimized code according to the computational model. The proposed solution has the advantage of allowing us to schedule the communications and to adjust the memory usage at compilation time. In order to validate the design and the implementation a demanding validation process was carried out with emphasis in the functional behavior, the memory usage, the processor usage (the end-to-end response time and the response time in each functional block) and the network usage (real consumption according to the calculated consumption). The results obtained in an industrial application developed by Thales Avionics (a Flight Management System) and in exhaustive tests show that the design and the prototype are reliable for industrial applications with strict timing requirements. Los sistemas empotrados y distribuidos de tiempo real son cada vez más importantes para la sociedad. Su demanda aumenta y cada vez más dependemos de los servicios que proporcionan. Los sistemas de alta integridad constituyen un subconjunto de gran importancia. Se caracterizan por que un fallo en su funcionamiento puede causar pérdida de vidas humanas, daños en el medio ambiente o cuantiosas pérdidas económicas. La necesidad de satisfacer requisitos temporales estrictos, hace más complejo su desarrollo. Mientras que los sistemas empotrados se sigan expandiendo en nuestra sociedad, es necesario garantizar un coste de desarrollo ajustado mediante el uso técnicas adecuadas en su diseño, mantenimiento y certificación. En concreto, se requiere una tecnología flexible e independiente del hardware. La evolución de las redes y paradigmas de comunicación, así como la necesidad de mayor potencia de cómputo y de tolerancia a fallos, ha motivado la interconexión de dispositivos electrónicos. Los mecanismos de comunicación permiten la transferencia de datos con alta velocidad de transmisión. En este contexto, el concepto de sistema distribuido ha emergido como sistemas donde sus componentes se ejecutan en varios nodos en paralelo y que interactúan entre ellos mediante redes de comunicaciones. Un concepto interesante son los sistemas de tiempo real neutrales respecto a la plataforma de ejecución. Se caracterizan por la falta de conocimiento de esta plataforma durante su diseño. Esta propiedad es relevante, por que conviene que se ejecuten en la mayor variedad de arquitecturas, tienen una vida media mayor de diez anos y el lugar ˜ donde se ejecutan puede variar. El lenguaje de programación Java es una buena base para el desarrollo de este tipo de sistemas. Por este motivo se ha creado RTSJ (Real-Time Specification for Java), que es una extensión del lenguaje para permitir el desarrollo de sistemas de tiempo real. Sin embargo, RTSJ no proporciona facilidades para el desarrollo de aplicaciones distribuidas de tiempo real. Es una limitación importante dado que la mayoría de los actuales y futuros sistemas serán distribuidos. El grupo DRTSJ (DistributedRTSJ) fue creado bajo el proceso de la comunidad de Java (JSR-50) con el fin de definir las abstracciones que aborden dicha limitación, pero en la actualidad aun no existe una especificacion formal. El objetivo de esta tesis es desarrollar un middleware de comunicaciones para el desarrollo de sistemas distribuidos de tiempo real en Java, basado en la integración entre el modelo de RMI (Remote Method Invocation) y el perfil HRTJ. Ha sido diseñado e implementado teniendo en cuenta los requisitos principales, como la predecibilidad y la confiabilidad del comportamiento temporal y el uso de recursos. El diseño parte de la definición de un modelo computacional el cual identifica entre otras cosas: el modelo de comunicaciones, los protocolos de red subyacentes más adecuados, el modelo de análisis, y un subconjunto de Java para sistemas de tiempo real crítico. En el diseño, las referencias remotas son el medio básico para construcción de aplicaciones distribuidas las cuales son asociadas a todos los parámetros no funcionales y los recursos necesarios para la ejecución de invocaciones remotas síncronas o asíncronas con atributos de tiempo real. El middleware propuesto separa la asignación de recursos de la propia ejecución definiendo dos fases y un mecanismo de hebras especifico que garantiza un comportamiento temporal adecuado. Además se ha incluido mecanismos para supervisar el comportamiento funcional y temporal. Se ha buscado independencia del protocolo de red definiendo una interfaz de red y módulos específicos. También se ha modificado el protocolo JRMP para incluir diferentes fases, parámetros no funcionales y optimizaciones de los tamaños de los mensajes. Aunque la serialización es una de las operaciones fundamentales para asegurar la adecuada transmisión de datos, las actuales implementaciones no son adecuadas para sistemas críticos y no hay alternativas. Este trabajo propone una serialización predecible que ha implicado el desarrollo de un nuevo compilador para la generación de código optimizado acorde al modelo computacional. La solución propuesta tiene la ventaja que en tiempo de compilación nos permite planificar las comunicaciones y ajustar el uso de memoria. Con el objetivo de validar el diseño e implementación se ha llevado a cabo un exigente proceso de validación con énfasis en: el comportamiento funcional, el uso de memoria, el uso del procesador (tiempo de respuesta de extremo a extremo y en cada uno de los bloques funcionales) y el uso de la red (consumo real conforme al estimado). Los buenos resultados obtenidos en una aplicación industrial desarrollada por Thales Avionics (un sistema de gestión de vuelo) y en las pruebas exhaustivas han demostrado que el diseño y el prototipo son fiables para aplicaciones industriales con estrictos requisitos temporales.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Granulation is one of the fundamental operations in particulate processing and has a very ancient history and widespread use. Much fundamental particle science has occurred in the last two decades to help understand the underlying phenomena. Yet, until recently the development of granulation systems was mostly based on popular practice. The use of process systems approaches to the integrated understanding of these operations is providing improved insight into the complex nature of the processes. Improved mathematical representations, new solution techniques and the application of the models to industrial processes are yielding better designs, improved optimisation and tighter control of these systems. The parallel development of advanced instrumentation and the use of inferential approaches provide real-time access to system parameters necessary for improvements in operation. The use of advanced models to help develop real-time plant diagnostic systems provides further evidence of the utility of process system approaches to granulation processes. This paper highlights some of those aspects of granulation. (c) 2005 Elsevier Ltd. All rights reserved.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Large scale disasters, such as the one caused by the Typhoon Haiyan, which devastated portions of the Philippines in 2013, or the catastrophic 2010 Haiti earthquake, which caused major damage in Port-au-Prince and other settlements in the region, have massive and lasting effects on populations. Nowadays, disasters can be considered as a consequence of inappropriately managed risk. These risks are the product of hazards and vulnerability, which refers to the extent to which a community can be affected by the impact of a hazard. In this way, developing countries, due to their greater vulnerability, suffer the highest costs when a disaster occurs. Disaster relief is a challenge for politics, economies, and societies worldwide. Humanitarian organizations face multiple decision problems when responding to disasters. In particular, once a disaster strikes, the distribution of humanitarian aid to the population affected is one of the most fundamental operations in what is called humanitarian logistics. This term is defined as the process of planning, implementing and controlling the effcient, cost-effective ow and storage of goods and materials as well as related information, from the point of origin to the point of consumption, for the purpose of meeting the end bene- ciaries' requirements and alleviate the suffering of vulnerable people, [the Humanitarian Logistics Conference, 2004 (Fritz Institute)]. During the last decade there has been an increasing interest in the OR/MS community in studying this topic, pointing out the similarities and differences between humanitarian and business logistics, and developing models suited to handle the special characteristics of these problems. Several authors have pointed out that traditional logistic objectives, such as minimizing operation cost, are not the most relevant goals in humanitarian operations. Other factors, such as the time of operation, or the design of safe and equitable distribution plans, come to the front, and new models and algorithms are needed to cope with these special features. Up to six attributes related to the distribution plan are considered in our multi-criteria approach. Even though there are usually simple ways to measure the cost of an operation, the evaluation of some other attributes such as security or equity is not easy. As a result, several attribute measures are proposed and developed, focusing on different aspects of the solutions. Furthermore, when metaheuristic solution methods are used, considering non linear objective functions does not increase the complexity of the algorithms significantly, and thus more accurate measures can be utilized...