941 resultados para 091007 Manufacturing Robotics and Mechatronics (excl. Automotive Mechatronics)
Resumo:
Autonomous aerial refueling is a key enabling technology for both manned and unmanned aircraft where extended flight duration or range are required. The results presented within this paper offer one potential vision-based sensing solution, together with a unique test environment. A hierarchical visual tracking algorithm based on direct methods is proposed and developed for the purposes of tracking a drogue during the capture stage of autonomous aerial refueling, and of estimating its 3D position. Intended to be applied in real time to a video stream from a single monocular camera mounted on the receiver aircraft, the algorithm is shown to be highly robust, and capable of tracking large, rapid drogue motions within the frame of reference. The proposed strategy has been tested using a complex robotic testbed and with actual flight hardware consisting of a full size probe and drogue. Results show that the vision tracking algorithm can detect and track the drogue at real-time frame rates of more than thirty frames per second, obtaining a robust position estimation even with strong motions and multiple occlusions of the drogue.
Resumo:
This work describes an acoustic system that allows the automatic detection and location of mechanical impacts on metallic based structures, which is suitable in robotics and industrial applications. The system is based on the time delays of propagation of the acoustic waves along the metallic based structure and it determines the instant and the position when and were the impact has been produced by piezoelectric sensors and an electronic-computerized system. We have obtained that for distance impact of 40 cm and 50 cm the time delay is 2 s and 72 s respectively.
Resumo:
In the last few years there has been a heightened interest in data treatment and analysis with the aim of discovering hidden knowledge and eliciting relationships and patterns within this data. Data mining techniques (also known as Knowledge Discovery in Databases) have been applied over a wide range of fields such as marketing, investment, fraud detection, manufacturing, telecommunications and health. In this study, well-known data mining techniques such as artificial neural networks (ANN), genetic programming (GP), forward selection linear regression (LR) and k-means clustering techniques, are proposed to the health and sports community in order to aid with resistance training prescription. Appropriate resistance training prescription is effective for developing fitness, health and for enhancing general quality of life. Resistance exercise intensity is commonly prescribed as a percent of the one repetition maximum. 1RM, dynamic muscular strength, one repetition maximum or one execution maximum, is operationally defined as the heaviest load that can be moved over a specific range of motion, one time and with correct performance. The safety of the 1RM assessment has been questioned as such an enormous effort may lead to muscular injury. Prediction equations could help to tackle the problem of predicting the 1RM from submaximal loads, in order to avoid or at least, reduce the associated risks. We built different models from data on 30 men who performed up to 5 sets to exhaustion at different percentages of the 1RM in the bench press action, until reaching their actual 1RM. Also, a comparison of different existing prediction equations is carried out. The LR model seems to outperform the ANN and GP models for the 1RM prediction in the range between 1 and 10 repetitions. At 75% of the 1RM some subjects (n = 5) could perform 13 repetitions with proper technique in the bench press action, whilst other subjects (n = 20) performed statistically significant (p < 0:05) more repetitions at 70% than at 75% of their actual 1RM in the bench press action. Rate of perceived exertion (RPE) seems not to be a good predictor for 1RM when all the sets are performed until exhaustion, as no significant differences (p < 0:05) were found in the RPE at 75%, 80% and 90% of the 1RM. Also, years of experience and weekly hours of strength training are better correlated to 1RM (p < 0:05) than body weight. O'Connor et al. 1RM prediction equation seems to arise from the data gathered and seems to be the most accurate 1RM prediction equation from those proposed in literature and used in this study. Epley's 1RM prediction equation is reproduced by means of data simulation from 1RM literature equations. Finally, future lines of research are proposed related to the problem of the 1RM prediction by means of genetic algorithms, neural networks and clustering techniques. RESUMEN En los últimos años ha habido un creciente interés en el tratamiento y análisis de datos con el propósito de descubrir relaciones, patrones y conocimiento oculto en los mismos. Las técnicas de data mining (también llamadas de \Descubrimiento de conocimiento en bases de datos\) se han aplicado consistentemente a lo gran de un gran espectro de áreas como el marketing, inversiones, detección de fraude, producción industrial, telecomunicaciones y salud. En este estudio, técnicas bien conocidas de data mining como las redes neuronales artificiales (ANN), programación genética (GP), regresión lineal con selección hacia adelante (LR) y la técnica de clustering k-means, se proponen a la comunidad del deporte y la salud con el objetivo de ayudar con la prescripción del entrenamiento de fuerza. Una apropiada prescripción de entrenamiento de fuerza es efectiva no solo para mejorar el estado de forma general, sino para mejorar la salud e incrementar la calidad de vida. La intensidad en un ejercicio de fuerza se prescribe generalmente como un porcentaje de la repetición máxima. 1RM, fuerza muscular dinámica, una repetición máxima o una ejecución máxima, se define operacionalmente como la carga máxima que puede ser movida en un rango de movimiento específico, una vez y con una técnica correcta. La seguridad de las pruebas de 1RM ha sido cuestionada debido a que el gran esfuerzo requerido para llevarlas a cabo puede derivar en serias lesiones musculares. Las ecuaciones predictivas pueden ayudar a atajar el problema de la predicción de la 1RM con cargas sub-máximas y son empleadas con el propósito de eliminar o al menos, reducir los riesgos asociados. En este estudio, se construyeron distintos modelos a partir de los datos recogidos de 30 hombres que realizaron hasta 5 series al fallo en el ejercicio press de banca a distintos porcentajes de la 1RM, hasta llegar a su 1RM real. También se muestra una comparación de algunas de las distintas ecuaciones de predicción propuestas con anterioridad. El modelo LR parece superar a los modelos ANN y GP para la predicción de la 1RM entre 1 y 10 repeticiones. Al 75% de la 1RM algunos sujetos (n = 5) pudieron realizar 13 repeticiones con una técnica apropiada en el ejercicio press de banca, mientras que otros (n = 20) realizaron significativamente (p < 0:05) más repeticiones al 70% que al 75% de su 1RM en el press de banca. El ínndice de esfuerzo percibido (RPE) parece no ser un buen predictor del 1RM cuando todas las series se realizan al fallo, puesto que no existen diferencias signifiativas (p < 0:05) en el RPE al 75%, 80% y el 90% de la 1RM. Además, los años de experiencia y las horas semanales dedicadas al entrenamiento de fuerza están más correlacionadas con la 1RM (p < 0:05) que el peso corporal. La ecuación de O'Connor et al. parece surgir de los datos recogidos y parece ser la ecuación de predicción de 1RM más precisa de aquellas propuestas en la literatura y empleadas en este estudio. La ecuación de predicción de la 1RM de Epley es reproducida mediante simulación de datos a partir de algunas ecuaciones de predicción de la 1RM propuestas con anterioridad. Finalmente, se proponen futuras líneas de investigación relacionadas con el problema de la predicción de la 1RM mediante algoritmos genéticos, redes neuronales y técnicas de clustering.
Resumo:
Son muchos los dominios de aplicación que han surgido en los últimos años con motivo de los avances tecnológicos. Algunos como eHealth, Smart Building o Smart Grid están teniendo una gran aceptación por parte de empresas que incrementan sus inversiones en este tipo de campos. Las redes inalámbricas de sensores y actuadores juegan un papel fundamental en el desarrollo de este tipo de aplicaciones. A través de este tipo de redes inalámbricas es posible monitorizar y actuar sobre un entorno gracias a nodos sensores y actuadores de forma cómoda y sencilla. Las WSANs (Wireless Sensors and Actuators Networks) junto con la robótica y M2M (Machine-to-Machine) están forjando el camino hacia el Internet of Things (IoT), un futuro en el que todo esté conectado entre sí. Cada vez aparecen dispositivos más pequeños y autónomos, que junto con el crecimiento de las redes, propician la interconexión de “el todo”. Este Proyecto Fin de Carrera tiene como objetivo contribuir en este avance, desarrollando parcialmente una solución middleware que abstraiga al usuario de la complejidad del hardware, implementando ciertas funcionalidades ofrecidas por el middleware nSOM desarrollado por la UPM. Para conseguir este objetivo se realizará un estudio del Estado del Arte actual y una comparativa de las diferentes plataformas hardware involucradas en las Redes Inalámbricas de Sensores y Actuadores (Wireless Sensor-Actuator Networks). Este estudio tendrá como fin la elección de una de las plataformas hardware para su futuro uso en un despliegue parcial del mencionado middleware nSOM. Posteriormente, se diseñará e implementará un sistema para ejemplificar un caso de uso sobre dicha plataforma integrando la publicación de las características y servicios de cada nodo final y el envío de peticiones y la recepción de respuestas. Finalmente se obtendrá un conjunto de conclusiones a partir de los resultados obtenidos y se detallarán posibles líneas de trabajo. ABSTRACT. There are many applications domains that have arisen because of technological advances in recent years. Some as eHealth, Smart Building or Smart Grid are having a great acceptance by companies that increase their investments in such fields. Wireless sensors and actuators networks play a fundamental role in the development of such applications. By means of this kind of wireless network it is possible to monitor and act upon an environment with the assistance of sensors and actuators nodes, readily. The WSANs (Wireless Sensors and Actuators Networks) together with robotics and M2M (Machine-to-Machine) are forging the way towards the Internet of Things (IoT), a future in which all of them are connected among themselves. Smaller and more autonomous devices are appearing that, along with the growth of networks, foster the interconnection of ‘the whole’. This Degree Final Project aims to contribute to this breakthrough, developing partially a middleware solution that abstracts the user from the complexity of hardware, implementing certain functionalities offered by the nSOM middleware solution carried out by UPM. To achieve this objective a study of the current state of the art and a comparison of the different hardware platforms involved in the Wireless and Actuators Sensor Networks (Wireless Sensor-Actuator Networks) will be performed. This study will aim the election of one of the hardware platforms for its future use in a partial deployment of the mentioned middleware nSOM. Subsequently, a system will be designed and implemented to exemplify a use case on the platform mentioned before integrating the publication of the features and services of each end node and sending requests and receiving responses. Finally a set of conclusions from the results will be stated and possible lines of future works will be detailed.
Resumo:
We have developed new analytical expressions for designing liquid crystal (LC) microlenses. These equations are based on a novel equivalent electric circuit and can be used to create an optimum design for the LC lenses in which the lens diameter ranges from a few micrometers to several millimeters. Thus far, only experimental studies have been conducted on the LC lenses. The analytical expressions developed in this letter depend on various manufacturing parameters and can be used to design lenses with specific focal lengths and a parabolic phase profile. The required driving scheme (modal or hole-patterned) can be predicted. The LC microlenses were manufactured and electrooptically characterized: the measurements were compared using an analytical approach.
Resumo:
The monkey anterior intraparietal area (AIP) encodes visual information about three-dimensional object shape that is used to shape the hand for grasping. In robotics a similar role has been played by modules that fit point cloud data to the superquadric family of shapes and its various extensions. We developed a model of shape tuning in AIP based on cosine tuning to superquadric parameters. However, the model did not fit the data well, and we also found that it was difficult to accurately reproduce these parameters using neural networks with the appropriate inputs (modelled on the caudal intraparietal area, CIP). The latter difficulty was related to the fact that there are large discontinuities in the superquadric parameters between very similar shapes. To address these limitations we adopted an alternative shape parameterization based on an Isomap nonlinear dimension reduction. The Isomap was built using gradients and curvatures of object surface depth. This alternative parameterization was low-dimensional (like superquadrics), but data-driven (similar to an alternative clustering approach that is also sometimes used in robotics) and lacked large discontinuities. Isomaps with 16 or more dimensions reproduced the AIP data fairly well. Moreover, we found that the Isomap parameters could be approximated from CIP-like input much more accurately than the superquadric parameters. We conclude that Isomaps, or perhaps alternative dimension reductions of CIP signals, provide a promising model of AIP tuning. We have now started to integrate our model with a robot hand, to explore the efficacy of Isomap shape reductions in grasp planning. Future work will consider dynamics of spike responses and integration with related visual and motor area models.
Resumo:
This paper presents a communication interface between supervisory low-cost mobile robots and domestic Wireless Sensor Network (WSN) based on the Zig Bee protocol from different manufacturers. The communication interface allows control and communication with other network devices using the same protocol. The robot can receive information from sensor devices (temperature, humidity, luminosity) and send commands to actuator devices (lights, shutters, thermostats) from different manufacturers. The architecture of the system, the interfaces and devices needed to establish the communication are described in the paper.
Resumo:
Graphene and carbon nanotubes are promising materials for supercapacitor electrodes because of their high specific surface area and excellent electrical, thermal, and mechanical properties. However, these materials suffer from a high manufacturing cost and some aggregation of graphene layers or the presence of toxic residual metallic impurities of carbon nanotubes.
Resumo:
Mosaicing is a technique that allows obtaining a large high resolution image by stitching several images together. These base images are usually acquired from an elevated point of view. Until recently, low-altitude image acquisition has been performed typically by using using airplanes, as well as other manned platforms. However, mini unmanned aerial vehicles (MUAV) endowed with a camera have lately made this task more available for small for cicil applications, for example for small farmers in order to obtain accurate agronomic information about their crop fields. The stitching orientation, or the image acquisition orientation usually coincides with the aircraft heading assuming a downwards orientation of the camera. In this paper, the efect of the image orientation in the eficiency of the aerial coverage path planning is studied. Moreover, an algorithm to compute an optimal stitching orientation angle is proposed and results are numerically compared with classical approaches.
Resumo:
Los sistemas transaccionales tales como los programas informáticos para la planificación de recursos empresariales (ERP software) se han implementado ampliamente mientras que los sistemas analíticos para la gestión de la cadena de suministro (SCM software) no han tenido el éxito deseado por la industria de tecnología de información (TI). Aunque se documentan beneficios importantes derivados de las implantaciones de SCM software, las empresas industriales son reacias a invertir en este tipo de sistemas. Por una parte esto es debido a la falta de métodos que son capaces de detectar los beneficios por emplear esos sistemas, y por otra parte porque el coste asociado no está identificado, detallado y cuantificado suficientemente. Los esquemas de coordinación basados únicamente en sistemas ERP son alternativas válidas en la práctica industrial siempre que la relación coste-beneficio esta favorable. Por lo tanto, la evaluación de formas organizativas teniendo en cuenta explícitamente el coste debido a procesos administrativos, en particular por ciclos iterativos, es de gran interés para la toma de decisiones en el ámbito de inversiones en TI. Con el fin de cerrar la brecha, el propósito de esta investigación es proporcionar métodos de evaluación que permitan la comparación de diferentes formas de organización y niveles de soporte por sistemas informáticos. La tesis proporciona una amplia introducción, analizando los retos a los que se enfrenta la industria. Concluye con las necesidades de la industria de SCM software: unas herramientas que facilitan la evaluación integral de diferentes propuestas de organización. A continuación, la terminología clave se detalla centrándose en la teoría de la organización, las peculiaridades de inversión en TI y la tipología de software de gestión de la cadena de suministro. La revisión de la literatura clasifica las contribuciones recientes sobre la gestión de la cadena de suministro, tratando ambos conceptos, el diseño de la organización y su soporte por las TI. La clasificación incluye criterios relacionados con la metodología de la investigación y su contenido. Los estudios empíricos en el ámbito de la administración de empresas se centran en tipologías de redes industriales. Nuevos algoritmos de planificación y esquemas de coordinación innovadoras se desarrollan principalmente en el campo de la investigación de operaciones con el fin de proponer nuevas funciones de software. Artículos procedentes del área de la gestión de la producción se centran en el análisis de coste y beneficio de las implantaciones de sistemas. La revisión de la literatura revela que el éxito de las TI para la coordinación de redes industriales depende en gran medida de características de tres dimensiones: la configuración de la red industrial, los esquemas de coordinación y las funcionalidades del software. La literatura disponible está enfocada sobre todo en los beneficios de las implantaciones de SCM software. Sin embargo, la coordinación de la cadena de suministro, basándose en el sistema ERP, sigue siendo la práctica industrial generalizada, pero el coste de coordinación asociado no ha sido abordado por los investigadores. Los fundamentos de diseño organizativo eficiente se explican en detalle en la medida necesaria para la comprensión de la síntesis de las diferentes formas de organización. Se han generado varios esquemas de coordinación variando los siguientes parámetros de diseño: la estructura organizativa, los mecanismos de coordinación y el soporte por TI. Las diferentes propuestas de organización desarrolladas son evaluadas por un método heurístico y otro basado en la simulación por eventos discretos. Para ambos métodos, se tienen en cuenta los principios de la teoría de la organización. La falta de rendimiento empresarial se debe a las dependencias entre actividades que no se gestionan adecuadamente. Dentro del método heurístico, se clasifican las dependencias y se mide su intensidad basándose en factores contextuales. A continuación, se valora la idoneidad de cada elemento de diseño organizativo para cada dependencia específica. Por último, cada forma de organización se evalúa basándose en la contribución de los elementos de diseño tanto al beneficio como al coste. El beneficio de coordinación se refiere a la mejora en el rendimiento logístico - este concepto es el objeto central en la mayoría de modelos de evaluación de la gestión de la cadena de suministro. Por el contrario, el coste de coordinación que se debe incurrir para lograr beneficios no se suele considerar en detalle. Procesos iterativos son costosos si se ejecutan manualmente. Este es el caso cuando SCM software no está implementada y el sistema ERP es el único instrumento de coordinación disponible. El modelo heurístico proporciona un procedimiento simplificado para la clasificación sistemática de las dependencias, la cuantificación de los factores de influencia y la identificación de configuraciones que indican el uso de formas organizativas y de soporte de TI más o menos complejas. La simulación de eventos discretos se aplica en el segundo modelo de evaluación utilizando el paquete de software ‘Plant Simulation’. Con respecto al rendimiento logístico, por un lado se mide el coste de fabricación, de inventario y de transporte y las penalizaciones por pérdida de ventas. Por otro lado, se cuantifica explícitamente el coste de la coordinación teniendo en cuenta los ciclos de coordinación iterativos. El método se aplica a una configuración de cadena de suministro ejemplar considerando diversos parámetros. Los resultados de la simulación confirman que, en la mayoría de los casos, el beneficio aumenta cuando se intensifica la coordinación. Sin embargo, en ciertas situaciones en las que se aplican ciclos de planificación manuales e iterativos el coste de coordinación adicional no siempre conduce a mejor rendimiento logístico. Estos resultados inesperados no se pueden atribuir a ningún parámetro particular. La investigación confirma la gran importancia de nuevas dimensiones hasta ahora ignoradas en la evaluación de propuestas organizativas y herramientas de TI. A través del método heurístico se puede comparar de forma rápida, pero sólo aproximada, la eficiencia de diferentes formas de organización. Por el contrario, el método de simulación es más complejo pero da resultados más detallados, teniendo en cuenta parámetros específicos del contexto del caso concreto y del diseño organizativo. ABSTRACT Transactional systems such as Enterprise Resource Planning (ERP) systems have been implemented widely while analytical software like Supply Chain Management (SCM) add-ons are adopted less by manufacturing companies. Although significant benefits are reported stemming from SCM software implementations, companies are reluctant to invest in such systems. On the one hand this is due to the lack of methods that are able to detect benefits from the use of SCM software and on the other hand associated costs are not identified, detailed and quantified sufficiently. Coordination schemes based only on ERP systems are valid alternatives in industrial practice because significant investment in IT can be avoided. Therefore, the evaluation of these coordination procedures, in particular the cost due to iterations, is of high managerial interest and corresponding methods are comprehensive tools for strategic IT decision making. The purpose of this research is to provide evaluation methods that allow the comparison of different organizational forms and software support levels. The research begins with a comprehensive introduction dealing with the business environment that industrial networks are facing and concludes highlighting the challenges for the supply chain software industry. Afterwards, the central terminology is addressed, focusing on organization theory, IT investment peculiarities and supply chain management software typology. The literature review classifies recent supply chain management research referring to organizational design and its software support. The classification encompasses criteria related to research methodology and content. Empirical studies from management science focus on network types and organizational fit. Novel planning algorithms and innovative coordination schemes are developed mostly in the field of operations research in order to propose new software features. Operations and production management researchers realize cost-benefit analysis of IT software implementations. The literature review reveals that the success of software solutions for network coordination depends strongly on the fit of three dimensions: network configuration, coordination scheme and software functionality. Reviewed literature is mostly centered on the benefits of SCM software implementations. However, ERP system based supply chain coordination is still widespread industrial practice but the associated coordination cost has not been addressed by researchers. Fundamentals of efficient organizational design are explained in detail as far as required for the understanding of the synthesis of different organizational forms. Several coordination schemes have been shaped through the variation of the following design parameters: organizational structuring, coordination mechanisms and software support. The different organizational proposals are evaluated using a heuristic approach and a simulation-based method. For both cases, the principles of organization theory are respected. A lack of performance is due to dependencies between activities which are not managed properly. Therefore, within the heuristic method, dependencies are classified and their intensity is measured based on contextual factors. Afterwards the suitability of each organizational design element for the management of a specific dependency is determined. Finally, each organizational form is evaluated based on the contribution of the sum of design elements to coordination benefit and to coordination cost. Coordination benefit refers to improvement in logistic performance – this is the core concept of most supply chain evaluation models. Unfortunately, coordination cost which must be incurred to achieve benefits is usually not considered in detail. Iterative processes are costly when manually executed. This is the case when SCM software is not implemented and the ERP system is the only available coordination instrument. The heuristic model provides a simplified procedure for the classification of dependencies, quantification of influence factors and systematic search for adequate organizational forms and IT support. Discrete event simulation is applied in the second evaluation model using the software package ‘Plant Simulation’. On the one hand logistic performance is measured by manufacturing, inventory and transportation cost and penalties for lost sales. On the other hand coordination cost is explicitly considered taking into account iterative coordination cycles. The method is applied to an exemplary supply chain configuration considering various parameter settings. The simulation results confirm that, in most cases, benefit increases when coordination is intensified. However, in some situations when manual, iterative planning cycles are applied, additional coordination cost does not always lead to improved logistic performance. These unexpected results cannot be attributed to any particular parameter. The research confirms the great importance of up to now disregarded dimensions when evaluating SCM concepts and IT tools. The heuristic method provides a quick, but only approximate comparison of coordination efficiency for different organizational forms. In contrast, the more complex simulation method delivers detailed results taking into consideration specific parameter settings of network context and organizational design.
Resumo:
Recientemente, ha surgido un interés por aprender a programar, debido a las oportunidades profesionales que da este tipo de estudios universitarios. Es fácil de entender porque el número de trabajos para programadores e ingenieros informáticos está creciendo rápidamente. Por otro lado, un amplio grupo de psicólogos opinan que el pensamiento computacional es una destreza fundamental para cualquiera, no sólo para los ingenieros informáticos. Para leer, escribir y realizar operaciones aritméticas, deberíamos utilizar el pensamiento computacional y por lo tanto, para desarrollar todas las habilidades analíticas de los niños. Es necesario cambiar los requerimientos de las destrezas necesarias para trabajar, los nuevos trabajadores necesitarán destrezas más sofisticadas en ciencias, matemáticas, ingeniería y tecnología. Consecuentemente, los contenidos sobre Tecnología de la Información tales como electrónica, programación, robótica y control se incrementan en la educación tecnológica en enseñanza secundaria. El desarrollo y utilización de los Laboratorios Virtuales de Control y Robótica ayuda a alcanzar este objetivo. Nos vamos a centrar en control y robótica porque un proyecto de control y robótica incluye contenidos de otras tecnologías tales como electrónica, programación, … Se ha implementado un sitio web con Laboratorios Virtuales de Control y Robótica. En este trabajo, se muestran seis grupos de laboratorios virtuales para la enseñanza del control y la robótica a niveles preuniversitarios. Estos laboratorios virtuales han sido usados para la docencia de alumnos de enseñanza secundaria. Las estadísticas del proceso de enseñanza-aprendizaje permiten validar ciertos aspectos de dicho trabajo. Se describen dichos laboratorios y la mejora del aprendizaje en cuanto a conocimientos procedimentales y conceptuales, así como la mejora de la interactividad respecto al aprendizaje con análogas aplicaciones con objetivos de aprendizaje idénticos, pero careciendo de la componente de laboratorio virtual. Se explican algunas de las experiencias realizadas con los alumnos. Los resultados sugieren, que dentro de la educación tecnológica de la educación secundaria, los laboratorios virtuales pueden ser explotados como un efectivo y motivacional entorno de aprendizaje. ABSTRACT Recently, there has been a surge of interest in learning to code, focusing especially on career opportunities. It is easy to understand why: the number of jobs for programmers and computer scientists is growing rapidly. On the other hand, the psychologists think that computational thinking is a fundamental skill for everyone, not just for computer scientists. To reading, writing, and arithmetic, we should add computational thinking to every child’s analytical ability. It is necessary to change workforce requirements mean that new workers will need ever more sophisticated skills in science, mathematics, engineering and technology. Consequently, the contents about Information Technology as well as electronics, coding, robotics and control increase in Technology Education in High School . The development and utilization of the Virtual Laboratories of Control and Robotics help to achieve this goal. We focus on control and robotics because a control and robotics project includes other technologies contents like electronics, coding,... A web site with Virtual Laboratories of Control and Robotics was implemented. In this work, six groups of virtual laboratories for teaching control and robotics in preuniversity level are shown. These Virtual Laboratories were used for teaching students at high school. The statistics of teaching-learning process allow to check some issues of this work. The laboratories, the improvement of learning (concepts and procedures) and interactivity are described and are compared to similar applications. They share identical learning objectives but they lack the virtual laboratory aspect. Some experiences with students are explained too. The results suggest that within high school technology education, virtual laboratories can be exploited as effective and motivational learning environments.
Resumo:
Nowadays robots have made their way into real applications that were prohibitive and unthinkable thirty years ago. This is mainly due to the increase in power computations and the evolution in the theoretical field of robotics and control. Even though there is plenty of information in the current literature on this topics, it is not easy to find clear concepts of how to proceed in order to design and implement a controller for a robot. In general, the design of a controller requires of a complete understanding and knowledge of the system to be controlled. Therefore, for advanced control techniques the systems must be first identified. Once again this particular objective is cumbersome and is never straight forward requiring of great expertise and some criteria must be adopted. On the other hand, the particular problem of designing a controller is even more complex when dealing with Parallel Manipulators (PM), since their closed-loop structures give rise to a highly nonlinear system. Under this basis the current work is developed, which intends to resume and gather all the concepts and experiences involve for the control of an Hydraulic Parallel Manipulator. The main objective of this thesis is to provide a guide remarking all the steps involve in the designing of advanced control technique for PMs. The analysis of the PM under study is minced up to the core of the mechanism: the hydraulic actuators. The actuators are modeled and experimental identified. Additionally, some consideration regarding traditional PID controllers are presented and an adaptive controller is finally implemented. From a macro perspective the kinematic and dynamic model of the PM are presented. Based on the model of the system and extending the adaptive controller of the actuator, a control strategy for the PM is developed and its performance is analyzed with simulation.
Resumo:
Es notoria la importancia de la naturaleza en la evolución de la arquitectura. Hasta la gran eclosión de la industria manufacturera y la mejora en las infraestructuras de comunicación, se podría decir que ambas corrían paralelas, para a partir de ahí dejar “congelada” la sabiduría popular y dar un salto a la globalidad, generando una situación de libertad arquitectónica prácticamente total, que independiza la construcción de su ubicación. "El biomimetismo es esencialmente un campo de investigación interdisciplinar, una serie de colaboraciones entre botánicos, físicos, matemáticos, ingenieros y zoólogos; donde la rígida división entre disciplinas «puras» cede lugar a un área de investigación que apunta a generar tecnología inteligente (smarttechnologies), utilizando materiales o procesos que sean de alguna manera sensibles al medio ambiente." (M. Weinstock, 1998). “La morfología de las plantas en los diferentes climas parece tener cierta analogía con la edificación, ya que algunas de las tensiones que inciden en su forma (tales como las variaciones de temperatura) corresponden de manera similar a las necesidades humanas.” (V. Olgyay 1963). En el presente trabajo se han estudiado las especies endémicas que nos rodean, para poder leer a través de ellas millones de años de supervivencia en este entorno, con el fin de mimetizar sus respuestas. También se han estudiado diferentes tipologías de arquitectura vernácula y su simulación energética, con el propósito de evaluar la demanda energética optima exigible. A partir de una ubicación específica, la orientación, compacidad, perforación y las características de la envolvente son los elementos que más influyen en la demanda energética de una edificación. Tanto la forma como los materiales pueden ser mimetizados con la naturaleza. En esta Tesis se han cuantificado los parámetros de diseño formales tomando como referencia las especies vegetales o la arquitectura vernácula, sin perder de vista los objetivos buscados por normativas o institutos en la reducción del consumo energético vinculado a la calefacción y ventilación. ABSTRACT The importance of the nature in the evolution of the architecture is well-known. Until the great burst of the manufacturing industry and the improvement in communication infrastructures, it would be possible to be said that both ran parallel, stops there from leaving “frozen” the popular wisdom and jump to the globalization, creating a situation almost complete architectural freedom, that it frees the construction of its location. "Biomimicry is essentially an interdisciplinary field of research, a series of collaborations among botanists, physicists, mathematicians, engineers and zoologists; where the rigid division between "pure" disciplines gives way to an area of research that aims to generate intelligent technology (smarttechnologies), using materials and processes that are in some environmentally sensitive manner. "(M. Weinstock, 1998). “The morphology of the plants in different climates seems to have some analogy with the building, as some of the tensions that affect their form (such as the temperature variations) are similar to the human necessities.” (V. Olgyay 1963). In the present work, the endemic species that surround to us have been studied, to be able to read through them millions of years of survival in this environment, in order to mimic their answers. Also different types or popular architecture and their energy simulation have been studied, in order to evaluate the rate of energy optimum demand. Orientation, compactness, perforation and characteristics of the envelope are the elements that influence more in the energy demand of a building. The shape and materials can be mimic with nature. Each of them has been quantified in this work by reference plant species or popular architecture, without losing sight of the objectives sought by regulations or institutes about reduction in energy consumption.
Resumo:
Esse trabalho de pesquisa apresenta um estudo detalhado sobre guias de ondas integrados ao substrato (SIW) operando em frequências de micro-ondas com base na teoria de guias de ondas retangulares (RWG). O estudo sobre guias SIW associa equações apresentadas na literatura e utiliza simulações eletromagnéticas para desenvolver um procedimento de projeto bem definido. É considerada a integração entre guias SIW e linhas de transmissão de microfita, projetando-se transições entre essas duas estruturas com o propósito de prover casamento de impedância e de viabilizar a caracterização em frequências de micro-ondas. São apresentadas considerações sobre processos de fabricação de circuitos SIW em substratos constituídos por laminados de alta frequência. Uma vez estabelecidos os procedimentos de fabricação e os critérios de projeto, a tecnologia SIW é aplicada ao projeto de três guias de ondas SIW nas bandas S e X, que foram fabricados empregando laminados de alta-frequência. Foram projetados dois filtros SIW passa-faixa empregando cavidades ressonantes e postes metálicos indutivos. Os dois filtros operam na frequência central de 10,61 GHz, sendo que um deles tem banda de passagem de 7,5%e é de 3ª ordem e o outro filtro tem banda de passagem de 15%, sendo de 5ª ordem. Foram realizadas comparações entre o desempenho simulado e experimental das estruturas SIW projetadas. Os resultados de simulações eletromagnéticas e experimentais demonstraram boa concordância. Os projetos em tecnologia SIW apresentados neste trabalho de pesquisa possuem perdas de retorno melhores que 10 dB na banda de operação e perdas por inserção de 1,0 dB a 1,5 dB. É apresentada a análise da sensibilidade do desempenho dos guias de ondas e filtros SIW projetados a desvios dimensionais típicos do processo de fabricação por microfresagem mecânica. Com os resultados experimentais e de simulação foi possível validar os procedimentos de projeto e de fabricação de circuitos SIW operando em frequências de micro-ondas.
Resumo:
The semantic localization problem in robotics consists in determining the place where a robot is located by means of semantic categories. The problem is usually addressed as a supervised classification process, where input data correspond to robot perceptions while classes to semantic categories, like kitchen or corridor. In this paper we propose a framework, implemented in the PCL library, which provides a set of valuable tools to easily develop and evaluate semantic localization systems. The implementation includes the generation of 3D global descriptors following a Bag-of-Words approach. This allows the generation of fixed-dimensionality descriptors from any type of keypoint detector and feature extractor combinations. The framework has been designed, structured and implemented to be easily extended with different keypoint detectors, feature extractors as well as classification models. The proposed framework has also been used to evaluate the performance of a set of already implemented descriptors, when used as input for a specific semantic localization system. The obtained results are discussed paying special attention to the internal parameters of the BoW descriptor generation process. Moreover, we also review the combination of some keypoint detectors with different 3D descriptor generation techniques.