34 resultados para one-phase appliance supply

em Universidad Politécnica de Madrid


Relevância:

100.00% 100.00%

Publicador:

Resumo:

In October 2002, under the auspices of Spanish Cooperation, a pilot electrification project put into operation two centralised PV-diesel hybrid systems in two different Moroccan villages. These systems currently provide a full-time energy service and supply electricity to more than a hundred of families, six community buildings, street lighting and one running water system. The appearance of the electricity service is very similar to an urban one: one phase AC supply (230V/50Hz) distributed up to each dwelling using a low-voltage mini-grid, which has been designed to be fully compatible with a future arrival of the utility grid. The management of this electricity service is based on a “fee-for-service” scheme agreed between a local NGO, partner of the project, and electricity associations created in each village, which are in charge of, among other tasks, recording the daily energy production of systems and the monthly energy consumption of each house. This register of data allows a systematic evaluation of both the system performance and the energy consumption of users. Now, after four years of operation, this paper presents the experience of this pilot electrification project and draws lessons that can be useful for designing, managing and sizing this type of small village PV-hybrid system

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The generator differential protection is one of the most important electrical protections of synchronous generator stator windings. Its operation principle is based on the comparison of the input current and output current at each phase winding. Unwanted trip commands are usually caused by CT saturation, wrong CT selection, or the fact that they may come from different manufacturers. In generators grounded through high impedance, only phase-to-phase or three-phase faults can be detected by the differential protection. This kind of fault causes differential current to flow in, at least, two phases of the winding. Several cases of unwanted trip commands caused by the appearance of differential current in only one phase of the generator have been reported. In this paper multi-phase criterion is proposed for generator differential protection algorithm when applied to high impedance grounded generators.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this paper, a system that allows applying precision agriculture techniques is described. The application is based on the deployment of a team of unmanned aerial vehicles that are able to take georeferenced pictures in order to create a full map by applying mosaicking procedures for postprocessing. The main contribution of this work is practical experimentation with an integrated tool. Contributions in different fields are also reported. Among them is a new one-phase automatic task partitioning manager, which is based on negotiation among the aerial vehicles, considering their state and capabilities. Once the individual tasks are assigned, an optimal path planning algorithm is in charge of determining the best path for each vehicle to follow. Also, a robust flight control based on the use of a control law that improves the maneuverability of the quadrotors has been designed. A set of field tests was performed in order to analyze all the capabilities of the system, from task negotiations to final performance. These experiments also allowed testing control robustness under different weather conditions.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A través de los años las estructuras de hormigón armado han ido aumentando su cuota de mercado, sustituyendo a las estructuras de fábrica de piedra o ladrillo y restándole participación a las estructuras metálicas. Uno de los primeros problemas que surgieron al ejecutar las estructuras de hormigón armado, era cómo conectar una fase de una estructura de este tipo a una fase posterior o a una modificación posterior. Hasta los años 80-90 las conexiones de una fase de una estructura de hormigón armado, con otra posterior se hacían dejando en la primera fase placas de acero con garrotas embebidas en el hormigón fresco o barras grifadas recubiertas de poliestireno expandido. Una vez endurecido el hormigón se podían conectar nuevas barras, para la siguiente fase mediante soldadura a la placa de la superficie o enderezando las barras grifadas, para embeberlas en el hormigón fresco de la fase siguiente. Estos sistemas requerían conocer la existencia y alcance de la fase posterior antes de hormigonar la fase previa. Además requerían un replanteo muy exacto y complejo de los elementos de conexión. Otro problema existente en las estructuras de hormigón era la adherencia de un hormigón fresco a un hormigón endurecido previamente, ya que la superficie de contacto de ambos hormigones suponía un punto débil, con una adherencia baja. A partir de los años 80, la industria química de la construcción experimentó un gran avance en el desarrollo de productos capaces de generar una buena adherencia sobre el hormigón endurecido. Este avance tecnológico tenía aplicación tanto en la adherencia del hormigón fresco sobre el hormigón endurecido, como en la adherencia de barras post-instaladas en agujeros de hormigón endurecido. Este sistema se denominó “anclajes adherentes de barras de acero en hormigón endurecido”. La forma genérica de ejecutarlos es hacer una perforación cilíndrica en el soporte de hormigón, con una herramienta especifica como un taladro, limpiar la perforación, llenarla del material adherente y finalmente introducir la barra de acero. Los anclajes adherentes se dividen en anclajes cementosos y anclajes químicos, siendo estos últimos los más habituales, fiables, resistentes y fáciles de ejecutar. El uso del anclaje adherente de barras de acero en hormigón endurecido se ha extendido por todo el espectro productivo, siendo muy habitual tanto en construcción de obras de hormigón armado de obra civil y edificación, como en obras industriales, instalaciones o fijación de elementos. La ejecución de un anclaje de una barra de acero en hormigón endurecido depende de numerosas variables, que en su conjunto, o de forma aislada pueden afectar de forma notable a la resistencia del anclaje. Nos referimos a variables de los anclajes, que a menudo no se consideran tales como la dirección de la perforación, la máquina de perforación y el útil de perforación utilizado, la diferencia de diámetros entre el diámetro del taladro y la barra, el tipo de material de anclaje, la limpieza del taladro, la humedad del soporte, la altura del taladro, etc. La utilización en los últimos años de los hormigones Autocompactables, añade una variable adicional, que hasta ahora apenas ha sido estudiada. En línea con lo apuntado, la presente tesis doctoral tiene como objetivo principal el estudio de las condiciones de ejecución en la resistencia de los anclajes en hormigón convencional y autocompactable. Esta investigación se centra principalmente en la evaluación de la influencia de una serie de variables sobre la resistencia de los anclajes, tanto en hormigón convencional como en un hormigón autocompactable. Para este estudio ha sido necesaria la fabricación de dos soportes de hormigón sobre los cuales desarrollar los ensayos. Uno de los bloques se ha fabricado con hormigón convencional y el otro con hormigón autocompactable. En cada pieza de hormigón se han realizado 174 anclajes con barras de acero, variando los parámetros a estudiar, para obtener resultados de todas las variables consideradas. Los ensayos a realizar en ambos bloques son exactamente iguales, para poder comparar la diferencia entre un anclaje en un soporte de hormigón con vibrado convencional (HVC) y un hormigón autocompactante (HAC). De cada tipo de ensayo deseado se harán dos repeticiones en la misma pieza. El ensayo de arrancamiento de las barras se realizara con un gato hidráulico hueco, con un sistema de instrumentación de lectura y registro de datos en tiempo real. El análisis de los resultados, realizado con una potente herramienta estadística, ha permitido determinar y evaluar numéricamente la influencia de los variables consideradas en la resistencia de los anclajes realizados. Así mismo ha permitido diferenciar los resultados obtenidos en los hormigones convencionales y autocompactantes, tanto desde el punto de vista de la resistencia mecánica, como de las deformaciones sufridas en el arrancamiento. Se define la resistencia mecánica de un anclaje, como la fuerza desarrollada en la dirección de la barra, para hacer su arrancamiento del soporte. De la misma forma se considera desplazamiento, a la separación entre un punto fijo de la barra y otro del soporte, en la dirección de la barra. Dichos puntos se determinan cuando se ha terminado el anclaje, en la intersección de la superficie plana del soporte, con la barra. Las conclusiones obtenidas han permitido establecer qué variables afectan a la ejecución de los anclajes y en qué cuantía lo hacen, así como determinar la diferencia entre los anclajes en hormigón vibrado convencional y hormigón autocompactante, con resultados muy interesantes, que permiten valorar la influencia de dichas variables. Dentro de las conclusiones podemos destacar tres grupos, que denominaremos como de alta influencia, baja influencia y sin influencia. En todos los casos hay que hacer el estudio en términos de carga y de desplazamiento. Podemos considerar como de alta influencia, en términos de carga las variables de máquina de perforación y el material de anclaje. En términos de desplazamiento podemos considerar de alta influencia además de la máquina de perforación y el material de anclaje, el diámetro del taladro, así como la limpieza y humedad del soporte. Podemos considerar de baja influencia, en términos de carga las variables de tipo de hormigón, dirección de perforación, limpieza y humedad del soporte. En términos de desplazamiento podemos considerar de baja influencia el tipo de hormigón y la dirección de perforación. Podemos considerar en el apartado de “sin influencia”, en términos de carga las variables de diámetro de perforación y altura del taladro. En términos de desplazamiento podemos considerar como “sin influencia” la variable de altura del taladro. Podemos afirmar que las diferencias entre los valores de carga aumentan de forma muy importante en términos de desplazamiento. ABSTRACT Over the years the concrete structures have been increasing their market share, replacing the masonry structures of stone or brick and subtracting as well the participation of the metallic structures. One of the first problems encountered in the implementing of the reinforced concrete structures was connecting a phase structure of this type at a later stage or a subsequent amendment. Until the 80s and 90s the connections of one phase of a reinforced concrete structure with a subsequent first phase were done by leaving the steel plates embedded in the fresh concrete using hooks or bent bars coated with expanded polystyrene. Once the concrete had hardened new bars could be connected to the next stage by welding them to the surface plate or by straightening the bent bars to embed them in the fresh concrete of the next phase. These systems required a previous knowledge of the existence and scope of the subsequent phase before concreting the previous one. They also required a very precise and complex rethinking of the connecting elements. Another existing problem in the concrete structures was the adhesion of a fresh concrete to a previously hardened concrete, since the contact surface of both concretes leaded to a weak point with low adherence. Since the 80s, the chemicals construction industry experienced a breakthrough in the development of products that generate a good grip on the concrete. This technological advance had its application both in the grip on one hardened fresh concrete and in the adhesion of bar post-installed in holes of hardened concrete. This system was termed as adherent anchors of steel bars in hardened concrete. The generic way of executing this system is by firstly drilling a cylindrical hole in the concrete support using a specific tool such as a drill. Then, cleaning the bore and filling it with bonding material to lastly, introduce the steel bar. These adherent anchors are divided into cement and chemical anchors, the latter being the most common, reliable, durable and easy to run. The use of adhesive anchor of steel bars in hardened concrete has spread across the production spectrum turning itself into a very common solution in both construction of reinforced concrete civil engineering and construction, and industrial works, installations and fixing elements as well. The execution of an anchor of a steel bar in hardened concrete depends on numerous variables which together or as a single solution may significantly affect the strength of the anchor. We are referring to variables of anchors which are often not considered, such as the diameter difference between the rod and the bore, the drilling system, cleansing of the drill, type of anchor material, the moisture of the substrate, the direction of the drill, the drill’s height, etc. During recent years, the emergence of self-compacting concrete adds an additional variable which has hardly been studied so far. According to mentioned this thesis aims to study the main performance conditions in the resistance of conventional and self-compacting concrete anchors. This research is primarily focused on the evaluation of the influence of several variables on the strength of the anchoring, both in conventional concrete and self-compacting concrete. In order to complete this study it has been required the manufacture of two concrete supports on which to develop the tests. One of the blocks has been manufactured with conventional concrete and the other with self-compacting concrete. A total of 174 steel bar anchors have been made in each one of the concrete pieces varying the studied parameters in order to obtain results for all variables considered. The tests to be performed on both blocks are exactly the same in order to compare the difference between an anchor on a stand with vibrated concrete (HVC) and a self-compacting concrete (SCC). Each type of test required two repetitions in the same piece. The pulling test of the bars was made with a hollow jack and with an instrumentation system for reading and recording data in real time. The use of a powerful statistical tool in the analysis of the results allowed to numerically determine and evaluate the influence of the variables considered in the resistance of the anchors made. It has likewise enabled to differentiate the results obtained in the self-compacting and conventional concretes, from both the outlook of the mechanical strength and the deformations undergone by uprooting. The mechanical strength of an anchor is defined as the strength undergone in a direction of the bar to uproot it from the support. Likewise, the movement is defined as the separation between a fixed point of the bar and a fixed point from the support considering the direction of the bar. These points are only determined once the anchor is finished, with the bar, at the intersection in the flat surface of the support. The conclusions obtained have established which variables affect the execution of the anchors and in what quantity. They have also permitted to determine the difference between the anchors in vibrated concrete and selfcompacting concrete with very interesting results that also allow to assess the influence of these mentioned variables. Three groups are highlighted among the conclusions called high influence, low influence and no influence. In every case is necessary to perform the study in terms of loading and movement. In terms of loading, there are considered as high influence two variables: drilling machinery and anchorage material. In terms of movement, there are considered as high influence the drilling diameter and the cleaning and moisture of the support, besides the drilling machinery and the anchorage material. Variables such as type of concrete, drilling direction and cleaning and moisture of the support are considered of low influence in terms of load. In terms of movement, the type of concrete and the direction of the drilling are considered variables of low influence. Within the no influence section in terms of loading, there are included the diameter of the drilling and the height of the drill. In terms of loading, the height of the drill is considered as a no influence variable. We can affirm that the differences among the loading values increase significantly in terms of movement.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this paper, the applicability of the FRA technique is discussed as a method for detecting inter-turn faults in stator windings. Firstly, this method is tested in an individual medium-voltage stator coil with satisfactory results. Secondly, the tests are extended to a medium-voltage induction motor stator winding, in which inter-turn faults are performed in every coil end of one phase. Results of the frequency response in case of inter-turn faults are evaluated in both cases for different fault resistance values. The experimental setup is also described for each experiment. The results of the application of this technique to the detection of inter-turn faults justify further research in optimizing this technique for preventive maintenance.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Recientemente, el paradigma de la computación en la nube ha recibido mucho interés por parte tanto de la industria como del mundo académico. Las infraestructuras cloud públicas están posibilitando nuevos modelos de negocio y ayudando a reducir costes. Sin embargo, una compañía podría desear ubicar sus datos y servicios en sus propias instalaciones, o tener que atenerse a leyes de protección de datos. Estas circunstancias hacen a las infraestructuras cloud privadas ciertamente deseables, ya sea para complementar a las públicas o para sustituirlas por completo. Por desgracia, las carencias en materia de estándares han impedido que las soluciones para la gestión de infraestructuras privadas se hayan desarrollado adecuadamente. Además, la multitud de opciones disponibles ha creado en los clientes el miedo a depender de una tecnología concreta (technology lock-in). Una de las causas de este problema es la falta de alineación entre la investigación académica y los productos comerciales, ya que aquella está centrada en el estudio de escenarios idealizados sin correspondencia con el mundo real, mientras que éstos consisten en soluciones desarrolladas sin tener en cuenta cómo van a encajar con los estándares más comunes o sin preocuparse de hacer públicos sus resultados. Con objeto de resolver este problema, propongo un sistema de gestión modular para infraestructuras cloud privadas enfocado en tratar con las aplicaciones en lugar de centrarse únicamente en los recursos hardware. Este sistema de gestión sigue el paradigma de la computación autónoma y está diseñado en torno a un modelo de información sencillo, desarrollado para ser compatible con los estándares más comunes. Este modelo divide el entorno en dos vistas, que sirven para separar aquello que debe preocupar a cada actor involucrado del resto de información, pero al mismo tiempo permitiendo relacionar el entorno físico con las máquinas virtuales que se despliegan encima de él. En dicho modelo, las aplicaciones cloud están divididas en tres tipos genéricos (Servicios, Trabajos de Big Data y Reservas de Instancias), para que así el sistema de gestión pueda sacar partido de las características propias de cada tipo. El modelo de información está complementado por un conjunto de acciones de gestión atómicas, reversibles e independientes, que determinan las operaciones que se pueden llevar a cabo sobre el entorno y que es usado para hacer posible la escalabilidad en el entorno. También describo un motor de gestión encargado de, a partir del estado del entorno y usando el ya mencionado conjunto de acciones, la colocación de recursos. Está dividido en dos niveles: la capa de Gestores de Aplicación, encargada de tratar sólo con las aplicaciones; y la capa del Gestor de Infraestructura, responsable de los recursos físicos. Dicho motor de gestión obedece un ciclo de vida con dos fases, para así modelar mejor el comportamiento de una infraestructura real. El problema de la colocación de recursos es atacado durante una de las fases (la de consolidación) por un resolutor de programación entera, y durante la otra (la online) por un heurístico hecho ex-profeso. Varias pruebas han demostrado que este acercamiento combinado es superior a otras estrategias. Para terminar, el sistema de gestión está acoplado a arquitecturas de monitorización y de actuadores. Aquella estando encargada de recolectar información del entorno, y ésta siendo modular en su diseño y capaz de conectarse con varias tecnologías y ofrecer varios modos de acceso. ABSTRACT The cloud computing paradigm has raised in popularity within the industry and the academia. Public cloud infrastructures are enabling new business models and helping to reduce costs. However, the desire to host company’s data and services on premises, and the need to abide to data protection laws, make private cloud infrastructures desirable, either to complement or even fully substitute public oferings. Unfortunately, a lack of standardization has precluded private infrastructure management solutions to be developed to a certain level, and a myriad of diferent options have induced the fear of lock-in in customers. One of the causes of this problem is the misalignment between academic research and industry ofering, with the former focusing in studying idealized scenarios dissimilar from real-world situations, and the latter developing solutions without taking care about how they f t with common standards, or even not disseminating their results. With the aim to solve this problem I propose a modular management system for private cloud infrastructures that is focused on the applications instead of just the hardware resources. This management system follows the autonomic system paradigm, and is designed around a simple information model developed to be compatible with common standards. This model splits the environment in two views that serve to separate the concerns of the stakeholders while at the same time enabling the traceability between the physical environment and the virtual machines deployed onto it. In it, cloud applications are classifed in three broad types (Services, Big Data Jobs and Instance Reservations), in order for the management system to take advantage of each type’s features. The information model is paired with a set of atomic, reversible and independent management actions which determine the operations that can be performed over the environment and is used to realize the cloud environment’s scalability. From the environment’s state and using the aforementioned set of actions, I also describe a management engine tasked with the resource placement. It is divided in two tiers: the Application Managers layer, concerned just with applications; and the Infrastructure Manager layer, responsible of the actual physical resources. This management engine follows a lifecycle with two phases, to better model the behavior of a real infrastructure. The placement problem is tackled during one phase (consolidation) by using an integer programming solver, and during the other (online) with a custom heuristic. Tests have demonstrated that this combined approach is superior to other strategies. Finally, the management system is paired with monitoring and actuators architectures. The former able to collect the necessary information from the environment, and the later modular in design and capable of interfacing with several technologies and ofering several access interfaces.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

GaInP nucleation on Ge(100) often starts by annealing of the Ge(100) substrates under supply of phosphorus precursors. However, the influence on the Ge surface is not well understood. Here, we studied vicinal Ge(100) surfaces annealed under tertiarybutylphosphine (TBP) supply in MOVPE by in situ reflection anisotropy spectroscopy (RAS), X-ray photoelectron spectroscopy (XPS), and low energy electron diffraction (LEED). While XPS reveals a P termination and the presence of carbon on the Ge surface, LEED patterns indicate a disordered surface probably due to by-products of the TBP pyrolysis. However, the TBP annealed Ge(100) surface exhibits a characteristic RA spectrum, which is related to the P termination. RAS allows us to in situ control phosphorus desorption dependent on temperature.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

En los últimos años, el Ge ha ganado de nuevo atención con la finalidad de ser integrado en el seno de las existentes tecnologías de microelectrónica. Aunque no se le considera como un canddato capaz de reemplazar completamente al Si en el futuro próximo, probalemente servirá como un excelente complemento para aumentar las propiedades eléctricas en dispositivos futuros, especialmente debido a su alta movilidad de portadores. Esta integración requiere de un avance significativo del estado del arte en los procesos de fabricado. Técnicas de simulación, como los algoritmos de Monte Carlo cinético (KMC), proporcionan un ambiente atractivo para llevar a cabo investigación y desarrollo en este campo, especialmente en términos de costes en tiempo y financiación. En este estudio se han usado, por primera vez, técnicas de KMC con el fin entender el procesado “front-end” de Ge en su fabricación, específicamente la acumulación de dañado y amorfización producidas por implantación iónica y el crecimiento epitaxial en fase sólida (SPER) de las capas amorfizadas. Primero, simulaciones de aproximación de clisiones binarias (BCA) son usadas para calcular el dañado causado por cada ión. La evolución de este dañado en el tiempo se simula usando KMC sin red, o de objetos (OKMC) en el que sólamente se consideran los defectos. El SPER se simula a través de una aproximación KMC de red (LKMC), siendo capaz de seguir la evolución de los átomos de la red que forman la intercara amorfo/cristalina. Con el modelo de amorfización desarrollado a lo largo de este trabajo, implementado en un simulador multi-material, se pueden simular todos estos procesos. Ha sido posible entender la acumulación de dañado, desde la generación de defectos puntuales hasta la formación completa de capas amorfas. Esta acumulación ocurre en tres regímenes bien diferenciados, empezando con un ritmo lento de formación de regiones de dañado, seguido por una rápida relajación local de ciertas áreas en la fase amorfa donde ambas fases, amorfa y cristalina, coexisten, para terminar en la amorfización completa de capas extensas, donde satura el ritmo de acumulación. Dicha transición ocurre cuando la concentración de dañado supera cierto valor límite, el cual es independiente de las condiciones de implantación. Cuando se implantan los iones a temperaturas relativamente altas, el recocido dinámico cura el dañado previamente introducido y se establece una competición entre la generación de dañado y su disolución. Estos efectos se vuelven especialmente importantes para iones ligeros, como el B, el cual crea dañado más diluido, pequeño y distribuido de manera diferente que el causado por la implantación de iones más pesados, como el Ge. Esta descripción reproduce satisfactoriamente la cantidad de dañado y la extensión de las capas amorfas causadas por implantación iónica reportadas en la bibliografía. La velocidad de recristalización de la muestra previamente amorfizada depende fuertemente de la orientación del sustrato. El modelo LKMC presentado ha sido capaz de explicar estas diferencias entre orientaciones a través de un simple modelo, dominado por una única energía de activación y diferentes prefactores en las frecuencias de SPER dependiendo de las configuraciones de vecinos de los átomos que recristalizan. La formación de maclas aparece como una consecuencia de esta descripción, y es predominante en sustratos crecidos en la orientación (111)Ge. Este modelo es capaz de reproducir resultados experimentales para diferentes orientaciones, temperaturas y tiempos de evolución de la intercara amorfo/cristalina reportados por diferentes autores. Las parametrizaciones preliminares realizadas de los tensores de activación de tensiones son también capaces de proveer una buena correlación entre las simulaciones y los resultados experimentales de velocidad de SPER a diferentes temperaturas bajo una presión hidrostática aplicada. Los estudios presentados en esta tesis han ayudado a alcanzar un mejor entendimiento de los mecanismos de producción de dañado, su evolución, amorfización y SPER para Ge, además de servir como una útil herramienta para continuar el trabajo en este campo. In the recent years, Ge has regained attention to be integrated into existing microelectronic technologies. Even though it is not thought to be a feasible full replacement to Si in the near future, it will likely serve as an excellent complement to enhance electrical properties in future devices, specially due to its high carrier mobilities. This integration requires a significant upgrade of the state-of-the-art of regular manufacturing processes. Simulation techniques, such as kinetic Monte Carlo (KMC) algorithms, provide an appealing environment to research and innovation in the field, specially in terms of time and funding costs. In the present study, KMC techniques are used, for the first time, to understand Ge front-end processing, specifically damage accumulation and amorphization produced by ion implantation and Solid Phase Epitaxial Regrowth (SPER) of the amorphized layers. First, Binary Collision Approximation (BCA) simulations are used to calculate the damage caused by every ion. The evolution of this damage over time is simulated using non-lattice, or Object, KMC (OKMC) in which only defects are considered. SPER is simulated through a Lattice KMC (LKMC) approach, being able to follow the evolution of the lattice atoms forming the amorphous/crystalline interface. With the amorphization model developed in this work, implemented into a multi-material process simulator, all these processes can be simulated. It has been possible to understand damage accumulation, from point defect generation up to full amorphous layers formation. This accumulation occurs in three differentiated regimes, starting at a slow formation rate of the damage regions, followed by a fast local relaxation of areas into the amorphous phase where both crystalline and amorphous phases coexist, ending in full amorphization of extended layers, where the accumulation rate saturates. This transition occurs when the damage concentration overcomes a certain threshold value, which is independent of the implantation conditions. When implanting ions at relatively high temperatures, dynamic annealing takes place, healing the previously induced damage and establishing a competition between damage generation and its dissolution. These effects become specially important for light ions, as B, for which the created damage is more diluted, smaller and differently distributed than that caused by implanting heavier ions, as Ge. This description successfully reproduces damage quantity and extension of amorphous layers caused by means of ion implantation reported in the literature. Recrystallization velocity of the previously amorphized sample strongly depends on the substrate orientation. The presented LKMC model has been able to explain these differences between orientations through a simple model, dominated by one only activation energy and different prefactors for the SPER rates depending on the neighboring configuration of the recrystallizing atoms. Twin defects formation appears as a consequence of this description, and are predominant for (111)Ge oriented grown substrates. This model is able to reproduce experimental results for different orientations, temperatures and times of evolution of the amorphous/crystalline interface reported by different authors. Preliminary parameterizations for the activation strain tensors are able to also provide a good match between simulations and reported experimental results for SPER velocities at different temperatures under the appliance of hydrostatic pressure. The studies presented in this thesis have helped to achieve a greater understanding of damage generation, evolution, amorphization and SPER mechanisms in Ge, and also provide a useful tool to continue research in this field.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The author participated in the 6 th EU Framework Project ―Q-pork Chains (FP6-036245-2)‖ from 2007 to 2009. With understanding of work reports from China and other countries, it is found that compared with other countries, China has great problems in pork quality and safety. By comparing the pork chain management between China and Spain, It is found that the difference in governance structure is one of the main differences in pork chain management between Spain and China. In China, spot-market relationship still dominates governance structure of pork chain, especially between the numerous house-hold pig holders and the great number of small slaughters. While in Spain, chain agents commonly apply cooperatives or integrations to cooperate. It also has been proven by recent studies, that in quality management at the chain level that supply chain integration has a direct effect on quality management practices (Han, 2010). Therefore, the author started to investigate the governance structure choices in supply chain management. And it has been set as the first research objective, which is to explain the governance structure choices process and the influencing factors in supply chain management, analyzing the pork chains cases in Spain and in China. During the further investigation, the author noticed the international trade of pork between Spain and China is not smooth since the signature of bi-lateral agreement on pork trade in 2007. Thus, another objective of the research is to find and solve the problems exist in the international pork chain between Spain and China. For the first objective, to explain the governance structure choices in supply chain management, the thesis conducts research in three main sections. 10 First of all, the thesis gives a literature overview in chapter two on Supply Chain Management (SCM), agri-food chain management and pork chain management. It concludes that SCM is a systems approach to view the supply chains as a whole, and to manage the total flow of goods inventory from the supplier to the ultimate customer. It includes the bi-directional flow of products (materials and services) and information, and the associated managerial and operational activities. And it also is a customer focus to create unique and individual source of customer value with an appropriate use of resources, leading to customer satisfaction and building competitive chain advantages. Agri-food chain management and pork chain management are applications of SCM in agri-food sector and pork sector respectively. Then, the research gives a comparative study in chapter three in the pork chain and pork chain management between Spain and China. Many differences are found, while the main difference is governance structure in pork chain management. Furthermore, the author gives an empirical study on governance structure choice in chapter five. It is concluded that governance structure of supply chain consists of a collection of rules/institutions/constraints structuring the transactions between the various stakeholders. Based on the overview on literatures closely related with governance structure, such as transaction cost economics, transaction value analysis and resource-based view theories, seven hypotheses are proposed, which are: Hypothesis 1: Transaction cost has positive relationship with governance structure choice Hypothesis 2: Uncertainty has positive relationship with transaction cost; higher uncertainty exerts high transaction cost Hypothesis 3: The relationship between asset specificity and transaction cost is positive Hypothesis 4: Collaboration advantages and governance structure choice have positive relationship11 Hypothesis 5: Willingness to collaborate has positive relationship with collaboration advantages Hypothesis 6: Capability to collaborate has positive relationship with collaboration advantages Hypothesis 7: Uncertainty has negative effect on collaboration advantages It is noted that as transaction cost value is negative, the transaction cost mentioned in the hypotheses is its absolute value. To test the seven hypotheses, Structural Equation Model (SEM) is applied and data collected from 350 pork slaughtering and processing companies in Jiangsu, Shandong and Henan Provinces in China is used. Based on the empirical SEM model and its results, the seven hypotheses are proved. The author generates several conclusions accordingly. It is found that the governance structure choice of the chain not only depends on transaction cost, it also depends on collaboration advantages. Exchange partners establish more stable and more intense relationship to reduce transaction cost and to maximize collaboration advantages. ―Collaboration advantages‖ in this thesis is defined as the joint value achieved through transaction (mutual activities) of agents in supply chains. This value forms as improvements, mainly in mutual logistics systems, cash response, information exchange, technological improvements and innovative improvements and quality management improvements, etc. Governance structure choice is jointly decided by transaction cost and collaboration advantages. Chain agents take different governance structures to coordinate in order to decrease their transaction cost and to increase their collaboration advantages. In China´s pork chain case, spot market relationship dominates the governance structure among the numerous backyard pig farmer and small family slaughterhouse 12 as they are connected by acquaintance relationship and the transaction cost in turn is low. Their relationship is reliable as they know each other in the neighborhood; as a result, spot market relationship is suitable for their exchange. However, the transaction between large-scale slaughtering and processing industries and small-scale pig producers is becoming difficult. The information hold back behavior and hold-up behavior of small-scale pig producers increase transaction cost between them and large-scale slaughtering and processing industries. Thus, through the more intense and stable relationship between processing industries and pig producers, processing industries reduce the transaction cost and improve the collaboration advantages with their chain partners, in which quality and safety collaboration advantages be increased, meaning that processing industries are able to provide consumers products with better quality and higher safety. It is also drawn that transaction cost is influenced mainly by uncertainty and asset specificity, which is in line with new institutional economics theories developed by Williamson O. E. In China´s pork chain case, behavioral uncertainty is created by the hold-up behaviors of great numbers of small pig producers, while big slaughtering and processing industries having strong asset specificity. On the other hand, ―collaboration advantages‖ is influenced by chain agents´ willingness to collaborate and chain agents´ capabilities to cooperate. With the fast growth of big scale slaughtering and processing industries, they are more willing to know and make effort to cooperate with their chain members, and they are more capable to create joint value together with other chain agents. Therefore, they are now the main chain agents who drive more intense and stable governance structure in China‘s pork chain. For the other objective, to find and solve the problems in the international pork chain between Spain and China, the research gives an analysis in chapter four on the 13 international pork chain. This study gives explanations why the international trade of pork between Spain and China is not sufficient from the chain perspective. It is found that the first obstacle is the high quality and safety requirement set by Chinese government. It makes the Spanish companies difficult to get authorities to export. Other aspects, such as Spanish pork is not competitive in price compared with other countries such as Denmark, United States, Canada, etc., Chinese consumers do not have sufficient information on Spanish pork products, are also important reasons that Spain does not export great quantity of pork products to China. It is concluded that China´s government has too much concern on the quality and safety requirements to Spanish pork products, which makes trade difficult to complete. The two countries need to establish a more stable and intense trade relationship. They also should make the information exchange sufficient and efficient and try to break trade barriers. Spanish companies should consider proper price strategies to win the Chinese pork market

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A probabilistic safety assessment (PSA) is being developed for a steam-methane reforming hydrogenproduction plant linked to a high-temperature gas-cooled nuclear reactor (HTGR). This work is based on the Japan Atomic Energy Research Institute's (JAERI) High Temperature Engineering Test Reactor (HTTR) prototype in Japan. The objective of this paper is to show how the PSA can be used for improving the design of the coupled plants. A simplified HAZOP study was performed to identify initiating events, based on existing studies. The results of the PSA show that the average frequency of an accident at this complex that could affect the population is 7 × 10−8 year−1 which is divided into the various end states. The dominant sequences are those that result in a methane explosion and occur with a frequency of 6.5 × 10−8 year−1, while the other sequences are much less frequent. The health risk presents itself if there are people in the vicinity who could be affected by the explosion. This analysis also demonstrates that an accident in one of the plants has little effect on the other. This is true given the design base distance between the plants, the fact that the reactor is underground, as well as other safety characteristics of the HTGR.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Supply chain management works to bring the supplier, the distributor, and the customer into one cohesive process. The Supply Chain Council defined supply chain as ‘Supply Chain: The flow and transformation of raw materials into products from suppliers through production and distribution facilities to the ultimate consumer., and then Sunil Chopra and Meindl, (2001) have define Supply chain management as ‘Supply Chain Management involves the flows between and among stages in a supply chain to maximize total profitability.’ After 1950, supply chain management got a boost with the production and manufacturing sector getting highest attention. The inventory became the responsibility of the marketing, accounting and production areas. Order processing was part of accounting and sales. Supply chain management became one of the most powerful engines of business transformation. It is the one area where operational efficiency can be gained. It reduces organizations costs and enhances customer service. With the liberalization of world trade, globalization, and emergence of the new markets, many organizations have customers and competitions throughout the world, either directly or indirectly. Business communities are aware that global competitiveness is the key to the success of a business. Competitiveness is ability to produce, distribute and provide products and services for the open market in competition with others. The supply chain, a critical link between supplier, producer and customer is emerged now as an essential business process and a strategic lever, potential value contributor a differentiator for the success of any business. Supply chain management is the management of all internal and external processes or functions to satisfy a customer’s order (from raw materials through conversion and manufacture through logistics delivery.). Goods-either in raw form or processed, whole sale or retailed distribution, business or technology services, in everyday life- in the business or household- directly or indirectly supply chain is ubiquitously associated in expanding socio-economic development. Supply chain growth competitive performance and supporting strong growth impulse at micro as well as micro economic levels. Keeping the India vision at the core of the objective, the role of supply chain is to take up social economic challenges, improve competitive advantages, develop strategies, built capabilities, enhance value propositions, adapt right technology, collaborate with stakeholders and deliver environmentally sustainable outcomes with minimum resources.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this work, a fiber-based optical powering (or power-by-light) system capable of providing more than 1 W is developed. The prototype was used in order to power a shunt regulator for controlling the activation and deactivation of solar panels in satellites. The work involves the manufacture of a light receiver (a GaAs multiple photovoltaic converter (MPC)), a power conditioning block, and a regulator and the implementation and characterization of the whole system. The MPC, with an active area of just 3.1 mm2, was able to supply 1 W at 5 V with an efficiency of 30%. The maximum measured device efficiency was over 40% at an input power (Pin) of 0.5 W. Open circuit voltage over 7 V was measured for Pin over 0.5 W. A system optoelectronic efficiency (including the optical fiber, connectors, and MPC) of 27% was measured at an output power (Pout) of 1 W. At Pout = 0.2 W, the efficiency was as high as 36%. The power conditioning block and the regulator were successfully powered with the system. The maximum supplied power in steady state was 0.2 W, whereas in transient state, it reached 0.44 W. The paper also describes the characterization of the system within the temperature range going from -70 to +100?°C.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This PhD work is focused on liquid crystal based tunable phase devices with special emphasis on their design and manufacturing. In the course of the work a number of new manufacturing technologies have been implemented in the UPM clean room facilities, leading to an important improvement in the range of devices being manufactured in the laboratory. Furthermore, a number of novel phase devices have been developed, all of them including novel electrodes, and/or alignment layers. The most important manufacturing progress has been the introduction of reactive ion etching as a tool for achieving high resolution photolithography on indium-tin-oxide (ITO) coated glass and quartz substrates. Another important manufacturing result is the successful elaboration of a binding protocol of anisotropic conduction adhesives. These have been employed in high density interconnections between ITO-glass and flexible printed circuits. Regarding material characterization, the comparative study of nonstoichiometric silicon oxide (SiOx) and silica (SiO2) inorganic alignment layers, as well as the relationship between surface layer deposition, layer morphology and liquid crystal electrooptical response must be highlighted, together with the characterization of the degradation of liquid crystal devices in simulated space mission environment. A wide variety of phase devices have been developed, with special emphasis on beam steerers. One of these was developed within the framework of an ESA project, and consisted of a high density reconfigurable 1D blaze grating, with a spatial separation of the controlling microelectronics and the active, radiation exposed, area. The developed devices confirmed the assumption that liquid crystal devices with such a separation of components, are radiation hard, and can be designed to be both vibration and temperature sturdy. In parallel to the above, an evenly variable analog beam steering device was designed, manufactured and characterized, providing a narrow cone diffraction free beam steering. This steering device is characterized by a very limited number of electrodes necessary for the redirection of a light beam. As few as 4 different voltage levels were needed in order to redirect a light beam. Finally at the Wojskowa Akademia Techniczna (Military University of Technology) in Warsaw, Poland, a wedged analog tunable beam steering device was designed, manufactured and characterized. This beam steerer, like the former one, was designed to resist the harsh conditions both in space and in the context of the shuttle launch. Apart from the beam steering devices, reconfigurable vortices and modal lens devices have been manufactured and characterized. In summary, during this work a large number of liquid crystal devices and liquid crystal device manufacturing technologies have been developed. Besides their relevance in scientific publications and technical achievements, most of these new devices have demonstrated their usefulness in the actual work of the research group where this PhD has been completed. El presente trabajo de Tesis se ha centrado en el diseño, fabricación y caracterización de nuevos dispositivos de fase basados en cristal líquido. Actualmente se están desarrollando dispositivos basados en cristal líquido para aplicaciones diferentes a su uso habitual como displays. Poseen la ventaja de que los dispositivos pueden ser controlados por bajas tensiones y no necesitan elementos mecánicos para su funcionamiento. La fabricación de todos los dispositivos del presente trabajo se ha realizado en la cámara limpia del grupo. La cámara limpia ha sido diseñada por el grupo de investigación, es de dimensiones reducidas pero muy versátil. Está dividida en distintas áreas de trabajo dependiendo del tipo de proceso que se lleva a cabo. La cámara limpia está completamente cubierta de un material libre de polvo. Todas las entradas de suministro de gas y agua están selladas. El aire filtrado es constantemente bombeado dentro de la zona limpia, a fin de crear una sobrepresión evitando así la entrada de aire sin filtrar. Las personas que trabajan en esta zona siempre deben de estar protegidas con un traje especial. Se utilizan trajes especiales que constan de: mono, máscara, guantes de látex, gorro, patucos y gafas de protección UV, cuando sea necesario. Para introducir material dentro de la cámara limpia se debe limpiar con alcohol y paños especiales y posteriormente secarlos con nitrógeno a presión. La fabricación debe seguir estrictamente unos pasos determinados, que pueden cambiar dependiendo de los requerimientos de cada dispositivo. Por ello, la fabricación de dispositivos requiere la formulación de varios protocolos de fabricación. Estos protocolos deben ser estrictamente respetados a fin de obtener repetitividad en los experimentos, lo que lleva siempre asociado un proceso de fabricación fiable. Una célula de cristal líquido está compuesta (de forma general) por dos vidrios ensamblados (sándwich) y colocados a una distancia determinada. Los vidrios se han sometido a una serie de procesos para acondicionar las superficies internas. La célula se llena con cristal líquido. De forma resumida, el proceso de fabricación general es el siguiente: inicialmente, se cortan los vidrios (cuya cara interna es conductora) y se limpian. Después se imprimen las pistas sobre el vidrio formando los píxeles. Estas pistas conductoras provienen del vidrio con la capa conductora de ITO (óxido de indio y estaño). Esto se hace a través de un proceso de fotolitografía con una resina fotosensible, y un desarrollo y ataque posterior del ITO sin protección. Más tarde, las caras internas de los vidrios se acondicionan depositando una capa, que puede ser orgánica o inorgánica (un polímero o un óxido). Esta etapa es crucial para el funcionamiento del dispositivo: induce la orientación de las moléculas de cristal líquido. Una vez que las superficies están acondicionadas, se depositan espaciadores en las mismas: son pequeñas esferas o cilindros de tamaño calibrado (pocos micrómetros) para garantizar un espesor homogéneo del dispositivo. Después en uno de los sustratos se deposita un adhesivo (gasket). A continuación, los sustratos se ensamblan teniendo en cuenta que el gasket debe dejar una boca libre para que el cristal líquido se introduzca posteriormente dentro de la célula. El llenado de la célula se realiza en una cámara de vacío y después la boca se sella. Por último, la conexión de los cables a la célula y el montaje de los polarizadores se realizan fuera de la sala limpia (Figura 1). Dependiendo de la aplicación, el cristal líquido empleado y los demás componentes de la célula tendrán unas características particulares. Para el diseño de los dispositivos de este trabajo se ha realizado un estudio de superficies inorgánicas de alineamiento del cristal líquido, que será de gran importancia para la preparación de los dispositivos de fase, dependiendo de las condiciones ambientales en las que vayan a trabajar. Los materiales inorgánicos que se han estudiado han sido en este caso SiOx y SiO2. El estudio ha comprendido tanto los factores de preparación influyentes en el alineamiento, el comportamiento del cristal líquido al variar estos factores y un estudio de la morfología de las superficies obtenidas.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We present a combinatorial decision problem, inspired by the celebrated quiz show called Countdown, that involves the computation of a given target number T from a set of k randomly chosen integers along with a set of arithmetic operations. We find that the probability of winning the game evidences a threshold phenomenon that can be understood in the terms of an algorithmic phase transition as a function of the set size k. Numerical simulations show that such probability sharply transitions from zero to one at some critical value of the control parameter, hence separating the algorithm's parameter space in different phases. We also find that the system is maximally efficient close to the critical point. We derive analytical expressions that match the numerical results for finite size and permit us to extrapolate the behavior in the thermodynamic limit.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Increased globalization and outsourcing to developing countries is fostering the interest in supply chain sustainability. From the academic point of view, while environmental impacts of supply chains have been largely analysed, the research on social issues has been scattered and fragmented. This paper thereby sets out to close this gap. We have identified an emerging sphere of knowledge at the interface between sustainable supply chain management, business strategy and international development literature, which seeks to propose innovative strategies for poverty alleviation. The incorporation of impoverished farmers into supply chains is presented here as one of those strategies, and illustrated through a case study on the integration of these farmers in the Senegalese horticulture supply chain.