42 resultados para self-etching adhesive system
Resumo:
With the rising prices of the retail electricity and the decreasing cost of the PV technology, grid parity with commercial electricity will soon become a reality in Europe. This fact, together with less attractive PV feed-in-tariffs in the near future and incentives to promote self-consumption suggest, that new operation modes for the PV Distributed Generation should be explored; differently from the traditional approach which is only based on maximizing the exported electricity to the grid. The smart metering is experiencing a growth in Europe and the United States but the possibilities of its use are still uncertain, in our system we propose their use to manage the storage and to allow the user to know their electrical power and energy balances. The ADSM has many benefits studied previously but also it has important challenges, in this paper we can observe and ADSM implementation example where we propose a solution to these challenges. In this paper we study the effects of the Active Demand-Side Management (ADSM) and storage systems in the amount of consumed local electrical energy. It has been developed on a prototype of a self-sufficient solar house called “MagicBox” equipped with grid connection, PV generation, lead–acid batteries, controllable appliances and smart metering. We carried out simulations for long-time experiments (yearly studies) and real measures for short and mid-time experiments (daily and weekly studies). Results show the relationship between the electricity flows and the storage capacity, which is not linear and becomes an important design criterion.
Resumo:
In recent decades, there has been an increasing interest in systems comprised of several autonomous mobile robots, and as a result, there has been a substantial amount of development in the eld of Articial Intelligence, especially in Robotics. There are several studies in the literature by some researchers from the scientic community that focus on the creation of intelligent machines and devices capable to imitate the functions and movements of living beings. Multi-Robot Systems (MRS) can often deal with tasks that are dicult, if not impossible, to be accomplished by a single robot. In the context of MRS, one of the main challenges is the need to control, coordinate and synchronize the operation of multiple robots to perform a specic task. This requires the development of new strategies and methods which allow us to obtain the desired system behavior in a formal and concise way. This PhD thesis aims to study the coordination of multi-robot systems, in particular, addresses the problem of the distribution of heterogeneous multi-tasks. The main interest in these systems is to understand how from simple rules inspired by the division of labor in social insects, a group of robots can perform tasks in an organized and coordinated way. We are mainly interested on truly distributed or decentralized solutions in which the robots themselves, autonomously and in an individual manner, select a particular task so that all tasks are optimally distributed. In general, to perform the multi-tasks distribution among a team of robots, they have to synchronize their actions and exchange information. Under this approach we can speak of multi-tasks selection instead of multi-tasks assignment, which means, that the agents or robots select the tasks instead of being assigned a task by a central controller. The key element in these algorithms is the estimation ix of the stimuli and the adaptive update of the thresholds. This means that each robot performs this estimate locally depending on the load or the number of pending tasks to be performed. In addition, it is very interesting the evaluation of the results in function in each approach, comparing the results obtained by the introducing noise in the number of pending loads, with the purpose of simulate the robot's error in estimating the real number of pending tasks. The main contribution of this thesis can be found in the approach based on self-organization and division of labor in social insects. An experimental scenario for the coordination problem among multiple robots, the robustness of the approaches and the generation of dynamic tasks have been presented and discussed. The particular issues studied are: Threshold models: It presents the experiments conducted to test the response threshold model with the objective to analyze the system performance index, for the problem of the distribution of heterogeneous multitasks in multi-robot systems; also has been introduced additive noise in the number of pending loads and has been generated dynamic tasks over time. Learning automata methods: It describes the experiments to test the learning automata-based probabilistic algorithms. The approach was tested to evaluate the system performance index with additive noise and with dynamic tasks generation for the same problem of the distribution of heterogeneous multi-tasks in multi-robot systems. Ant colony optimization: The goal of the experiments presented is to test the ant colony optimization-based deterministic algorithms, to achieve the distribution of heterogeneous multi-tasks in multi-robot systems. In the experiments performed, the system performance index is evaluated by introducing additive noise and dynamic tasks generation over time.
Resumo:
This report presents an overview of the current work performed by us in the context of the efficient parallel implementation of traditional logic programming systems. The work is based on the &-Prolog System, a system for the automatic parallelization and execution of logic programming languages within the Independent And-parallelism model, and the global analysis and parallelization tools which have been developed for this system. In order to make the report self-contained, we first describe the "classical" tools of the &-Prolog system. We then explain in detail the work performed in improving and generalizing the global analysis and parallelization tools. Also, we describe the objectives which will drive our future work in this area.
Resumo:
Salamanca is cataloged as one of the most polluted cities in Mexico. In order to observe the behavior and clarify the influence of wind parameters on the Sulphur Dioxide (SO2) concentrations a Self-Organizing Maps (SOM) Neural Network have been implemented at three monitoring locations for the period from January 1 to December 31, 2006. The maximum and minimum daily values of SO2 concentrations measured during the year of 2006 were correlated with the wind parameters of the same period. The main advantages of the SOM Neural Network is that it allows to integrate data from different sensors and provide readily interpretation results. Especially, it is powerful mapping and classification tool, which others information in an easier way and facilitates the task of establishing an order of priority between the distinguished groups of concentrations depending on their need for further research or remediation actions in subsequent management steps. For each monitoring location, SOM classifications were evaluated with respect to pollution levels established by Health Authorities. The classification system can help to establish a better air quality monitoring methodology that is essential for assessing the effectiveness of imposed pollution controls, strategies, and facilitate the pollutants reduction.
Resumo:
Self-consciousness implies not only self or group recognition, but also real knowledge of one’s own identity. Self-consciousness is only possible if an individual is intelligent enough to formulate an abstract self-representation. Moreover, it necessarily entails the capability of referencing and using this elf-representation in connection with other cognitive features, such as inference, and the anticipation of the consequences of both one’s own and other individuals’ acts. In this paper, a cognitive architecture for self-consciousness is proposed. This cognitive architecture includes several modules: abstraction, self-representation, other individuals'representation, decision and action modules. It includes a learning process of self-representation by direct (self-experience based) and observational learning (based on the observation of other individuals). For model implementation a new approach is taken using Modular Artificial Neural Networks (MANN). For model testing, a virtual environment has been implemented. This virtual environment can be described as a holonic system or holarchy, meaning that it is composed of autonomous entities that behave both as a whole and as part of a greater whole. The system is composed of a certain number of holons interacting. These holons are equipped with cognitive features, such as sensory perception, and a simplified model of personality and self-representation. We explain holons’ cognitive architecture that enables dynamic self-representation. We analyse the effect of holon interaction, focusing on the evolution of the holon’s abstract self-representation. Finally, the results are explained and analysed and conclusions drawn.
Resumo:
In just a few years cloud computing has become a very popular paradigm and a business success story, with storage being one of the key features. To achieve high data availability, cloud storage services rely on replication. In this context, one major challenge is data consistency. In contrast to traditional approaches that are mostly based on strong consistency, many cloud storage services opt for weaker consistency models in order to achieve better availability and performance. This comes at the cost of a high probability of stale data being read, as the replicas involved in the reads may not always have the most recent write. In this paper, we propose a novel approach, named Harmony, which adaptively tunes the consistency level at run-time according to the application requirements. The key idea behind Harmony is an intelligent estimation model of stale reads, allowing to elastically scale up or down the number of replicas involved in read operations to maintain a low (possibly zero) tolerable fraction of stale reads. As a result, Harmony can meet the desired consistency of the applications while achieving good performance. We have implemented Harmony and performed extensive evaluations with the Cassandra cloud storage on Grid?5000 testbed and on Amazon EC2. The results show that Harmony can achieve good performance without exceeding the tolerated number of stale reads. For instance, in contrast to the static eventual consistency used in Cassandra, Harmony reduces the stale data being read by almost 80% while adding only minimal latency. Meanwhile, it improves the throughput of the system by 45% while maintaining the desired consistency requirements of the applications when compared to the strong consistency model in Cassandra.
Resumo:
Within the frame of the HiPER reactor, we propose and study a Self Cooled Lead Lithium blanket with two different cooling arrangements of the system First Wall – Blanket for the HiPER reactor: Integrated First Wall Blanket and Separated First Wall Blanket. We compare the two arrangements in terms of power cycle efficiency, operation flexibility in out-off-normal situations and proper cooling and acceptable corrosion. The Separated First Wall Blanket arrangement is superior in all of them, and it is selected as the advantageous proposal for the HiPER reactor blanket. However, it still has to be improved from the standpoint of proper cooling and corrosion rates
Resumo:
A system for simultaneous 2D estimation of rectangular room and transceiver localization is proposed. The system is based on two radio transceivers, both capable of full duplex operations (simultaneous transmission and reception). This property enables measurements of channel impulse response (CIR) at the same place the signal is transmitted (generated), commonly known as self-to-self CIR. Another novelty of the proposed system is the spatial CIR discrimination that is possible with the receiver antenna design which consists of eight sectorized antennas with 45° aperture in the horizontal plane and total coverage equal to the isotropic one. The dimensions of a rectangular room are reconstructed directly from spatial radio impulse responses by extracting the information regarding round trip time (RTT). Using radar approach estimation of walls and corners positions is derived. Tests using measured data were performed, and the simulation results confirm the feasibility of the approach.
Resumo:
Los sistemas técnicos son cada vez más complejos, incorporan funciones más avanzadas, están más integrados con otros sistemas y trabajan en entornos menos controlados. Todo esto supone unas condiciones más exigentes y con mayor incertidumbre para los sistemas de control, a los que además se demanda un comportamiento más autónomo y fiable. La adaptabilidad de manera autónoma es un reto para tecnologías de control actualmente. El proyecto de investigación ASys propone abordarlo trasladando la responsabilidad de la capacidad de adaptación del sistema de los ingenieros en tiempo de diseño al propio sistema en operación. Esta tesis pretende avanzar en la formulación y materialización técnica de los principios de ASys de cognición y auto-consciencia basadas en modelos y autogestión de los sistemas en tiempo de operación para una autonomía robusta. Para ello el trabajo se ha centrado en la capacidad de auto-conciencia, inspirada en los sistemas biológicos, y se ha explorado la posibilidad de integrarla en la arquitectura de los sistemas de control. Además de la auto-consciencia, se han explorado otros temas relevantes: modelado funcional, modelado de software, tecnología de los patrones, tecnología de componentes, tolerancia a fallos. Se ha analizado el estado de la técnica en los ámbitos pertinentes para las cuestiones de la auto-consciencia y la adaptabilidad en sistemas técnicos: arquitecturas cognitivas, control tolerante a fallos, y arquitecturas software dinámicas y computación autonómica. El marco teórico de ASys existente de sistemas autónomos cognitivos ha sido adaptado para servir de base para este análisis de autoconsciencia y adaptación y para dar sustento conceptual al posterior desarrollo de la solución. La tesis propone una solución general de diseño para la construcción de sistemas autónomos auto-conscientes. La idea central es la integración de un meta-controlador en la arquitectura de control del sistema autónomo, capaz de percibir la estado funcional del sistema de control y, si es necesario, reconfigurarlo en tiempo de operación. Esta solución de metacontrol se ha formalizado en cuatro patrones de diseño: i) el Patrón Metacontrol, que define la integración de un subsistema de metacontrol, responsable de controlar al propio sistema de control a través de la interfaz proporcionada por su plataforma de componentes, ii) el patrón Bucle de Control Epistémico, que define un bucle de control cognitivo basado en el modelos y que se puede aplicar al diseño del metacontrol, iii) el patrón de Reflexión basada en Modelo Profundo propone una solución para construir el modelo ejecutable utilizado por el meta-controlador mediante una transformación de modelo a modelo a partir del modelo de ingeniería del sistema, y, finalmente, iv) el Patrón Metacontrol Funcional, que estructura el meta-controlador en dos bucles, uno para el control de la configuración de los componentes del sistema de control, y otro sobre éste, controlando las funciones que realiza dicha configuración de componentes; de esta manera las consideraciones funcionales y estructurales se desacoplan. La Arquitectura OM y el metamodelo TOMASys son las piezas centrales del marco arquitectónico desarrollado para materializar la solución compuesta de los patrones anteriores. El metamodelo TOMASys ha sido desarrollado para la representación de la estructura y su relación con los requisitos funcionales de cualquier sistema autónomo. La Arquitectura OM es un patrón de referencia para la construcción de una metacontrolador integrando los patrones de diseño propuestos. Este meta-controlador se puede integrar en la arquitectura de cualquier sistema control basado en componentes. El elemento clave de su funcionamiento es un modelo TOMASys del sistema decontrol, que el meta-controlador usa para monitorizarlo y calcular las acciones de reconfiguración necesarias para adaptarlo a las circunstancias en cada momento. Un proceso de ingeniería, complementado con otros recursos, ha sido elaborado para guiar la aplicación del marco arquitectónico OM. Dicho Proceso de Ingeniería OM define la metodología a seguir para construir el subsistema de metacontrol para un sistema autónomo a partir del modelo funcional del mismo. La librería OMJava proporciona una implementación del meta-controlador OM que se puede integrar en el control de cualquier sistema autónomo, independientemente del dominio de la aplicación o de su tecnología de implementación. Para concluir, la solución completa ha sido validada con el desarrollo de un robot móvil autónomo que incorpora un meta-controlador con la Arquitectura OM. Las propiedades de auto-consciencia y adaptación proporcionadas por el meta-controlador han sido validadas en diferentes escenarios de operación del robot, en los que el sistema era capaz de sobreponerse a fallos en el sistema de control mediante reconfiguraciones orquestadas por el metacontrolador. ABSTRACT Technical systems are becoming more complex, they incorporate more advanced functionalities, they are more integrated with other systems and they are deployed in less controlled environments. All this supposes a more demanding and uncertain scenario for control systems, which are also required to be more autonomous and dependable. Autonomous adaptivity is a current challenge for extant control technologies. The ASys research project proposes to address it by moving the responsibility for adaptivity from the engineers at design time to the system at run-time. This thesis has intended to advance in the formulation and technical reification of ASys principles of model-based self-cognition and having systems self-handle at runtime for robust autonomy. For that it has focused on the biologically inspired capability of self-awareness, and explored the possibilities to embed it into the very architecture of control systems. Besides self-awareness, other themes related to the envisioned solution have been explored: functional modeling, software modeling, patterns technology, components technology, fault tolerance. The state of the art in fields relevant for the issues of self-awareness and adaptivity has been analysed: cognitive architectures, fault-tolerant control, and software architectural reflection and autonomic computing. The extant and evolving ASys Theoretical Framework for cognitive autonomous systems has been adapted to provide a basement for this selfhood-centred analysis and to conceptually support the subsequent development of our solution. The thesis proposes a general design solution for building self-aware autonomous systems. Its central idea is the integration of a metacontroller in the control architecture of the autonomous system, capable of perceiving the functional state of the control system and reconfiguring it if necessary at run-time. This metacontrol solution has been formalised into four design patterns: i) the Metacontrol Pattern, which defines the integration of a metacontrol subsystem, controlling the domain control system through an interface provided by its implementation component platform, ii) the Epistemic Control Loop pattern, which defines a modelbased cognitive control loop that can be applied to the design of such a metacontroller, iii) the Deep Model Reflection pattern proposes a solution to produce the online executable model used by the metacontroller by model-to-model transformation from the engineering model, and, finally, iv) the Functional Metacontrol pattern, which proposes to structure the metacontroller in two loops, one for controlling the configuration of components of the controller, and another one on top of the former, controlling the functions being realised by that configuration; this way the functional and structural concerns become decoupled. The OM Architecture and the TOMASys metamodel are the core pieces of the architectural framework developed to reify this patterned solution. The TOMASys metamodel has been developed for representing the structure and its relation to the functional requirements of any autonomous system. The OM architecture is a blueprint for building a metacontroller according to the patterns. This metacontroller can be integrated on top of any component-based control architecture. At the core of its operation lies a TOMASys model of the control system. An engineering process and accompanying assets have been constructed to complete and exploit the architectural framework. The OM Engineering Process defines the process to follow to develop the metacontrol subsystem from the functional model of the controller of the autonomous system. The OMJava library provides a domain and application-independent implementation of an OM Metacontroller than can be used in the implementation phase of OMEP. Finally, the complete solution has been validated in the development of an autonomous mobile robot that incorporates an OM metacontroller. The functional selfawareness and adaptivity properties achieved thanks to the metacontrol system have been validated in different scenarios. In these scenarios the robot was able to overcome failures in the control system thanks to reconfigurations performed by the metacontroller.
Resumo:
This paper presents a novel self-timed multi-purpose sensor especially conceived for Field Programmable Gate Arrays (FPGAs). The aim of the sensor is to measure performance variations during the life-cycle of the device, such as process variability, critical path timing and temperature variations. The proposed topology, through the use of both combinational and sequential FPGA elements, amplifies the time of a signal traversing a delay chain to produce a pulse whose width is the sensor’s measurement. The sensor is fully self-timed, avoiding the need for clock distribution networks and eliminating the limitations imposed by the system clock. One single off- or on-chip time-to-digital converter is able to perform digitization of several sensors in a single operation. These features allow for a simplified approach for designers wanting to intertwine a multi-purpose sensor network with their application logic. Employed as a temperature sensor, it has been measured to have an error of ±0.67 °C, over the range of 20–100 °C, employing 20 logic elements with a 2-point calibration.
Resumo:
The water time constant and mechanical time constant greatly influences the power and speed oscillations of hydro-turbine-generator unit. This paper discusses the turbine power transients in response to different nature and changes in the gate position. The work presented here analyses the characteristics of hydraulic system with an emphasis on changes in the above time constants. The simulation study is based on mathematical first-, second-, third- and fourth-order transfer function models. The study is further extended to identify discrete time-domain models and their characteristic representation without noise and with noise content of 10 & 20 dB signal-to-noise ratio (SNR). The use of self-tuned control approach in minimising the speed deviation under plant parameter changes and disturbances is also discussed.
Resumo:
A través de los años las estructuras de hormigón armado han ido aumentando su cuota de mercado, sustituyendo a las estructuras de fábrica de piedra o ladrillo y restándole participación a las estructuras metálicas. Uno de los primeros problemas que surgieron al ejecutar las estructuras de hormigón armado, era cómo conectar una fase de una estructura de este tipo a una fase posterior o a una modificación posterior. Hasta los años 80-90 las conexiones de una fase de una estructura de hormigón armado, con otra posterior se hacían dejando en la primera fase placas de acero con garrotas embebidas en el hormigón fresco o barras grifadas recubiertas de poliestireno expandido. Una vez endurecido el hormigón se podían conectar nuevas barras, para la siguiente fase mediante soldadura a la placa de la superficie o enderezando las barras grifadas, para embeberlas en el hormigón fresco de la fase siguiente. Estos sistemas requerían conocer la existencia y alcance de la fase posterior antes de hormigonar la fase previa. Además requerían un replanteo muy exacto y complejo de los elementos de conexión. Otro problema existente en las estructuras de hormigón era la adherencia de un hormigón fresco a un hormigón endurecido previamente, ya que la superficie de contacto de ambos hormigones suponía un punto débil, con una adherencia baja. A partir de los años 80, la industria química de la construcción experimentó un gran avance en el desarrollo de productos capaces de generar una buena adherencia sobre el hormigón endurecido. Este avance tecnológico tenía aplicación tanto en la adherencia del hormigón fresco sobre el hormigón endurecido, como en la adherencia de barras post-instaladas en agujeros de hormigón endurecido. Este sistema se denominó “anclajes adherentes de barras de acero en hormigón endurecido”. La forma genérica de ejecutarlos es hacer una perforación cilíndrica en el soporte de hormigón, con una herramienta especifica como un taladro, limpiar la perforación, llenarla del material adherente y finalmente introducir la barra de acero. Los anclajes adherentes se dividen en anclajes cementosos y anclajes químicos, siendo estos últimos los más habituales, fiables, resistentes y fáciles de ejecutar. El uso del anclaje adherente de barras de acero en hormigón endurecido se ha extendido por todo el espectro productivo, siendo muy habitual tanto en construcción de obras de hormigón armado de obra civil y edificación, como en obras industriales, instalaciones o fijación de elementos. La ejecución de un anclaje de una barra de acero en hormigón endurecido depende de numerosas variables, que en su conjunto, o de forma aislada pueden afectar de forma notable a la resistencia del anclaje. Nos referimos a variables de los anclajes, que a menudo no se consideran tales como la dirección de la perforación, la máquina de perforación y el útil de perforación utilizado, la diferencia de diámetros entre el diámetro del taladro y la barra, el tipo de material de anclaje, la limpieza del taladro, la humedad del soporte, la altura del taladro, etc. La utilización en los últimos años de los hormigones Autocompactables, añade una variable adicional, que hasta ahora apenas ha sido estudiada. En línea con lo apuntado, la presente tesis doctoral tiene como objetivo principal el estudio de las condiciones de ejecución en la resistencia de los anclajes en hormigón convencional y autocompactable. Esta investigación se centra principalmente en la evaluación de la influencia de una serie de variables sobre la resistencia de los anclajes, tanto en hormigón convencional como en un hormigón autocompactable. Para este estudio ha sido necesaria la fabricación de dos soportes de hormigón sobre los cuales desarrollar los ensayos. Uno de los bloques se ha fabricado con hormigón convencional y el otro con hormigón autocompactable. En cada pieza de hormigón se han realizado 174 anclajes con barras de acero, variando los parámetros a estudiar, para obtener resultados de todas las variables consideradas. Los ensayos a realizar en ambos bloques son exactamente iguales, para poder comparar la diferencia entre un anclaje en un soporte de hormigón con vibrado convencional (HVC) y un hormigón autocompactante (HAC). De cada tipo de ensayo deseado se harán dos repeticiones en la misma pieza. El ensayo de arrancamiento de las barras se realizara con un gato hidráulico hueco, con un sistema de instrumentación de lectura y registro de datos en tiempo real. El análisis de los resultados, realizado con una potente herramienta estadística, ha permitido determinar y evaluar numéricamente la influencia de los variables consideradas en la resistencia de los anclajes realizados. Así mismo ha permitido diferenciar los resultados obtenidos en los hormigones convencionales y autocompactantes, tanto desde el punto de vista de la resistencia mecánica, como de las deformaciones sufridas en el arrancamiento. Se define la resistencia mecánica de un anclaje, como la fuerza desarrollada en la dirección de la barra, para hacer su arrancamiento del soporte. De la misma forma se considera desplazamiento, a la separación entre un punto fijo de la barra y otro del soporte, en la dirección de la barra. Dichos puntos se determinan cuando se ha terminado el anclaje, en la intersección de la superficie plana del soporte, con la barra. Las conclusiones obtenidas han permitido establecer qué variables afectan a la ejecución de los anclajes y en qué cuantía lo hacen, así como determinar la diferencia entre los anclajes en hormigón vibrado convencional y hormigón autocompactante, con resultados muy interesantes, que permiten valorar la influencia de dichas variables. Dentro de las conclusiones podemos destacar tres grupos, que denominaremos como de alta influencia, baja influencia y sin influencia. En todos los casos hay que hacer el estudio en términos de carga y de desplazamiento. Podemos considerar como de alta influencia, en términos de carga las variables de máquina de perforación y el material de anclaje. En términos de desplazamiento podemos considerar de alta influencia además de la máquina de perforación y el material de anclaje, el diámetro del taladro, así como la limpieza y humedad del soporte. Podemos considerar de baja influencia, en términos de carga las variables de tipo de hormigón, dirección de perforación, limpieza y humedad del soporte. En términos de desplazamiento podemos considerar de baja influencia el tipo de hormigón y la dirección de perforación. Podemos considerar en el apartado de “sin influencia”, en términos de carga las variables de diámetro de perforación y altura del taladro. En términos de desplazamiento podemos considerar como “sin influencia” la variable de altura del taladro. Podemos afirmar que las diferencias entre los valores de carga aumentan de forma muy importante en términos de desplazamiento. ABSTRACT Over the years the concrete structures have been increasing their market share, replacing the masonry structures of stone or brick and subtracting as well the participation of the metallic structures. One of the first problems encountered in the implementing of the reinforced concrete structures was connecting a phase structure of this type at a later stage or a subsequent amendment. Until the 80s and 90s the connections of one phase of a reinforced concrete structure with a subsequent first phase were done by leaving the steel plates embedded in the fresh concrete using hooks or bent bars coated with expanded polystyrene. Once the concrete had hardened new bars could be connected to the next stage by welding them to the surface plate or by straightening the bent bars to embed them in the fresh concrete of the next phase. These systems required a previous knowledge of the existence and scope of the subsequent phase before concreting the previous one. They also required a very precise and complex rethinking of the connecting elements. Another existing problem in the concrete structures was the adhesion of a fresh concrete to a previously hardened concrete, since the contact surface of both concretes leaded to a weak point with low adherence. Since the 80s, the chemicals construction industry experienced a breakthrough in the development of products that generate a good grip on the concrete. This technological advance had its application both in the grip on one hardened fresh concrete and in the adhesion of bar post-installed in holes of hardened concrete. This system was termed as adherent anchors of steel bars in hardened concrete. The generic way of executing this system is by firstly drilling a cylindrical hole in the concrete support using a specific tool such as a drill. Then, cleaning the bore and filling it with bonding material to lastly, introduce the steel bar. These adherent anchors are divided into cement and chemical anchors, the latter being the most common, reliable, durable and easy to run. The use of adhesive anchor of steel bars in hardened concrete has spread across the production spectrum turning itself into a very common solution in both construction of reinforced concrete civil engineering and construction, and industrial works, installations and fixing elements as well. The execution of an anchor of a steel bar in hardened concrete depends on numerous variables which together or as a single solution may significantly affect the strength of the anchor. We are referring to variables of anchors which are often not considered, such as the diameter difference between the rod and the bore, the drilling system, cleansing of the drill, type of anchor material, the moisture of the substrate, the direction of the drill, the drill’s height, etc. During recent years, the emergence of self-compacting concrete adds an additional variable which has hardly been studied so far. According to mentioned this thesis aims to study the main performance conditions in the resistance of conventional and self-compacting concrete anchors. This research is primarily focused on the evaluation of the influence of several variables on the strength of the anchoring, both in conventional concrete and self-compacting concrete. In order to complete this study it has been required the manufacture of two concrete supports on which to develop the tests. One of the blocks has been manufactured with conventional concrete and the other with self-compacting concrete. A total of 174 steel bar anchors have been made in each one of the concrete pieces varying the studied parameters in order to obtain results for all variables considered. The tests to be performed on both blocks are exactly the same in order to compare the difference between an anchor on a stand with vibrated concrete (HVC) and a self-compacting concrete (SCC). Each type of test required two repetitions in the same piece. The pulling test of the bars was made with a hollow jack and with an instrumentation system for reading and recording data in real time. The use of a powerful statistical tool in the analysis of the results allowed to numerically determine and evaluate the influence of the variables considered in the resistance of the anchors made. It has likewise enabled to differentiate the results obtained in the self-compacting and conventional concretes, from both the outlook of the mechanical strength and the deformations undergone by uprooting. The mechanical strength of an anchor is defined as the strength undergone in a direction of the bar to uproot it from the support. Likewise, the movement is defined as the separation between a fixed point of the bar and a fixed point from the support considering the direction of the bar. These points are only determined once the anchor is finished, with the bar, at the intersection in the flat surface of the support. The conclusions obtained have established which variables affect the execution of the anchors and in what quantity. They have also permitted to determine the difference between the anchors in vibrated concrete and selfcompacting concrete with very interesting results that also allow to assess the influence of these mentioned variables. Three groups are highlighted among the conclusions called high influence, low influence and no influence. In every case is necessary to perform the study in terms of loading and movement. In terms of loading, there are considered as high influence two variables: drilling machinery and anchorage material. In terms of movement, there are considered as high influence the drilling diameter and the cleaning and moisture of the support, besides the drilling machinery and the anchorage material. Variables such as type of concrete, drilling direction and cleaning and moisture of the support are considered of low influence in terms of load. In terms of movement, the type of concrete and the direction of the drilling are considered variables of low influence. Within the no influence section in terms of loading, there are included the diameter of the drilling and the height of the drill. In terms of loading, the height of the drill is considered as a no influence variable. We can affirm that the differences among the loading values increase significantly in terms of movement.
Resumo:
Evolvable Hardware (EH) is a technique that consists of using reconfigurable hardware devices whose configuration is controlled by an Evolutionary Algorithm (EA). Our system consists of a fully-FPGA implemented scalable EH platform, where the Reconfigurable processing Core (RC) can adaptively increase or decrease in size. Figure 1 shows the architecture of the proposed System-on-Programmable-Chip (SoPC), consisting of a MicroBlaze processor responsible of controlling the whole system operation, a Reconfiguration Engine (RE), and a Reconfigurable processing Core which is able to change its size in both height and width. This system is used to implement image filters, which are generated autonomously thanks to the evolutionary process. The system is complemented with a camera that enables the usage of the platform for real time applications.
Resumo:
In this paper, an architecture based on a scalable and flexible set of Evolvable Processing arrays is presented. FPGA-native Dynamic Partial Reconfiguration (DPR) is used for evolution, which is done intrinsically, letting the system to adapt autonomously to variable run-time conditions, including the presence of transient and permanent faults. The architecture supports different modes of operation, namely: independent, parallel, cascaded or bypass mode. These modes of operation can be used during evolution time or during normal operation. The evolvability of the architecture is combined with fault-tolerance techniques, to enhance the platform with self-healing features, making it suitable for applications which require both high adaptability and reliability. Experimental results show that such a system may benefit from accelerated evolution times, increased performance and improved dependability, mainly by increasing fault tolerance for transient and permanent faults, as well as providing some fault identification possibilities. The evolvable HW array shown is tailored for window-based image processing applications.
Resumo:
In this paper an on line self-tuned PID controller is proposed for the control of a car whose goal is to follow another one, at distances and speeds typical in urban traffic. The bestknown tuning mechanism is perhaps the MIT rule, due to its ease of implementation. However, as it is well known, this method does not guarantee the stability of the system, providing good results only for constant or slowly varying reference signals and in the absence of noise, which are unrealistic conditions. When the reference input varies with an appreciable rate or in presence of noise, eventually it could result in system instability. In this paper an alternative method is proposed that significantly improves the robustness of the system for varying inputs or in the presence of noise, as demonstrated by simulation.