16 resultados para Supervision of employees
em Universidad Politécnica de Madrid
Resumo:
The coagulation of milk is the fundamental process in cheese-making, based on a gel formation as consequence of physicochemical changes taking place in the casein micelles, the monitoring the whole process of milk curd formation is a constant preoccupation for dairy researchers and cheese companies (Lagaude et al., 2004). In addition to advances in composition-based applications of near infrared spectroscopy (NIRS), innovative uses of this technology are pursuing dynamic applications that show promise, especially in regard to tracking a sample in situ during food processing (Bock and Connelly, 2008). In this way the literature describes cheese making process applications of NIRS for curd cutting time determination, which conclude that NIRS would be a suitable method of monitoring milk coagulation, as shown i.e. the works published by Fagan et al. (Fagan et al., 2008; Fagan et al., 2007), based in the use of the commercial CoAguLite probe (with a LED at 880nm and a photodetector for light reflectance detection).
Resumo:
The European Higher Education Area (EHEA) has leaded to a change in the way the subjects are taught. One of the more important aspects of the EHEA is to support the autonomous study of the students. Taking into account this new approach, the virtual laboratory of the subject Mechanisms of the Aeronautical studies at the Technical University of Madrid is being migrated to an on-line scheme. This virtual laboratory consist on two practices: the design of cam-follower mechanisms and the design of trains of gears. Both practices are software applications that, in the current situation, need to be installed on each computer and the students carry out the practice at the computer classroom of the school under the supervision of a teacher. During this year the design of cam-follower mechanisms practice has been moved to a web application using Java and the Google Development Toolkit. In this practice the students has to design and study the running of a cam to perform a specific displacement diagram with a selected follower taking into account that the mechanism must be able to work properly at high speed regime. The practice has maintained its objectives in the new platform but to take advantage of the new methodology and try to avoid the inconveniences that the previous version had shown. Once the new practice has been ready, a pilot study has been carried out to compare both approaches: on-line and in-lab. This paper shows the adaptation of the cam and follower practice to an on-line methodology. Both practices are described and the changes that has been done to the initial one are shown. They are compared and the weak and strong points of each one are analyzed. Finally we explain the pilot study carried out, the students impression and the results obtained.
Resumo:
The European Higher Education Area (EHEA) has leaded to a change in the way the subjects are taught. One of the more important aspects of the EHEA is to support the autonomous study of the students. Taking into account this new approach, the virtual laboratory of the subject Mechanisms of the Aeronautical studies at the Technical University of Madrid is being migrated to an on-line scheme. This virtual laboratory consist on two practices: the design of cam-follower mechanisms and the design of trains of gears. Both practices are software applications that, in the current situation, need to be installed on each computer and the students carry out the practice at the computer classroom of the school under the supervision of a teacher. During this year the design of cam-follower mechanisms practice has been moved to a web application using Java and the Google Development Toolkit. In this practice the students has to design and study the running of a cam to perform a specific displacement diagram with a selected follower taking into account that the mechanism must be able to work properly at high speed regime. The practice has maintained its objectives in the new platform but to take advantage of the new methodology and try to avoid the inconveniences that the previous version had shown. Once the new practice has been ready, a pilot study has been carried out to compare both approaches: on-line and in-lab. This paper shows the adaptation of the cam and follower practice to an on-line methodology. Both practices are described and the changes that has been done to the initial one are shown. They are compared and the weak and strong points of each one are analyzed. Finally we explain the pilot study carried out, the students impression and the results obtained.
Resumo:
The present research is focused on the application of hyperspectral images for the supervision of quality deterioration in ready to use leafy spinach during storage (Spinacia oleracea). Two sets of samples of packed leafy spinach were considered: (a) a first set of samples was stored at 20 °C (E-20) in order to accelerate the degradation process, and these samples were measured the day of reception in the laboratory and after 2 days of storage; (b) a second set of samples was kept at 10 °C (E-10), and the measurements were taken throughout storage, beginning the day of reception and repeating the acquisition of Images 3, 6 and 9 days later. Twenty leaves per test were analyzed. Hyperspectral images were acquired with a push-broom CCD camera equipped with a spectrograph VNIR (400–1000 nm). Calibration set of spectra was extracted from E-20 samples, containing three classes of degradation: class A (optimal quality), class B and class C (maximum deterioration). Reference average spectra were defined for each class. Three models, computed on the calibration set, with a decreasing degree of complexity were compared, according to their ability for segregating leaves at different quality stages (fresh, with incipient and non-visible symptoms of degradation, and degraded): spectral angle mapper distance (SAM), partial least squares discriminant analysis models (PLS-DA), and a non linear index (Leafy Vegetable Evolution, LEVE) combining five wavelengths were included among the previously selected by CovSel procedure. In sets E-10 and E-20, artificial images of the membership degree according to the distance of each pixel to the reference classes, were computed assigning each pixel to the closest reference class. The three methods were able to show the degradation of the leaves with storage time.
Resumo:
The aim of this project is to create a website which is useful both employees and students of a university, so employees can add information, if they log in with username and password access, and students can view this information . Employees may modify and display information such as title, room, or their faculty (from a list defined by the administrator), and most importantly, their schedule, whether class, tutoring, free time, or any of tasks that the administrator define. There will be a manager, responsible for managing employees, the availables faculties and the types of tasks that employees can use on their schedule. Students may see the employees schedules and rooms on the homepage. They differentiate between differents tasks of employees, because these are in different colors. They can also filter information for faculty, employee or day. To achieve our goal, we decided to program in Java using Servlets, which we will use to generate response to requests coming from users from the website. We will also use JSP, allowing us to create different websites files. We use JSP files and not HTML, because we need that the pages are dynamic, since not only want to show specific information, we like that information can change depending on user requests. The JSP file allows us to generate HTML, but also using JAVA language, which is necessary for our purpose. As the information we store is not fixed. We want this information can be modified at any time by employees and admin, so we need a database, which can be accessed from anywhere. We decided SQLite databases because are integrated quite well in our application, and offer a quick response. To access the database from our program, we simply connect it to the database, and with very few lines of code, add, delete or modify entries in different tables that owns the database. To facilitate the initial creation of the database, and the first tables, we use a Mozilla Firefox browser plugin, called SQLite Manager, which allows us to do so from a more friendly interface. Finally, we need a server that supports and implements specifications Servlets and JSP. We decided on the TomCat server, which is a container Servlets, because is free, easy to use, and compatible with our program. We realized all the project with Eclipse environment, also free program that allows integrating database, server and program the JSP and Servlets. Once submitted all the tools we used, we must first organize the structure of the web, relating each Servlets with JSP files. Next, create the database and the different Servlets, and adjust the database accesses to make sure we do it right. From here simply is to build up the page step by step, showing in each place we need, and redirect to different pages. In this way, we can build a complex website, free, and without being an expert in the field. RESUMEN. El objetivo de este proyecto, es crear una página web que sirva tanto a empleados como a alumnos de una universidad, de tal manera que los empleados podrán añadir información, mediante el acceso con usuario y contraseña, y los alumnos podrán visualizar está información. Los empleados podrán modificar y mostrar información como su título, despacho, facultad a la que pertenecen (de entre una lista definida por el administrador), y lo más importante, sus horarios, ya sean de clase, tutorías, tiempo libre, o cualquiera de las tareas que el administrador defina. Habrá un administrador, encargado de gestionar los empleados existentes, las facultades disponibles y los tipos de tareas que podrán usar los empleados en su horario. Los alumnos, podrán visualizar los horarios y despacho de los empleados en la página principal. Diferenciarán entre las distintas tareas de los profesores, porque estas se encuentran en colores diferentes. Además, podrán filtrar la información, por facultad, empleado o día de la semana. Para conseguir nuestro objetivo, hemos decidido programar en Java, mediante el uso de Servlets, los cuales usaremos para generar respuesta antes las peticiones que llegan de los usuarios desde la página web. También usaremos archivos JSP, que nos permitirán crear las diferentes páginas webs. Usamos archivos JSP y no HTML, porque necesitamos que las diferentes páginas sean dinámicas, ya que no solo queremos mostrar una información concreta, si no que esta información puede variar en función de las peticiones de usuario. El archivo JSP nos permite generar HTML, pero a la vez usar lenguaje JAVA, algo necesario para nuestro cometido. Como la información que queremos almacenar no es fija, si no que en todo momento debe poder ser modificada por empleados y administrador, necesitamos una base de datos, a la que podamos acceder desde la web. Nos hemos decidido por bases SQLite, ya que se integran bastante bien en nuestra aplicación, y además ofrecen una rápida respuesta. Para acceder a la base de datos desde nuestro programa, simplemente debemos conectar el mismo a la base de datos, y con muy pocas líneas de código, añadir, eliminar o modificar entradas de las diferentes tablas que posee la base de datos. Para facilitar la creación inicial de la base de datos, y de las primeras tablas, usamos un complemento del navegador Mozilla Firefox, llamado SQLite Manager, que nos permite hacerlo desde una interfaz más amigable. Por último, necesitamos de un servidor que soporte e implemente las especificaciones de los Servlets y JSP. Nos decidimos por el servidor TomCat, que es un contenedor de Servlets gratuito, de fácil manejo, y compatible con nuestro programa. Todo el desarrollo del proyecto, lo realizamos desde el entorno Eclipse, programa también gratuito, que permite integrar la base de datos, el servidor y programar los JSP y Servlets. Una vez presentadas todas las herramientas que hemos utilizado, primero debemos organizar la estructura de la web, relacionando cada archivo JSP con los Servlets a los que debe acceder. A continuación creamos la base de datos y los diferentes Servlets, y ajustamos bien los accesos a la base de datos para comprobar que lo hacemos correctamente. A partir de aquí, simplemente es ir construyendo la página paso a paso, mostrando en cada lugar lo que necesitemos, y redirigiendo a las diferentes páginas. De esta manera, podremos construir una página web compleja, de manera gratuita, y sin ser un experto en la materia.
Resumo:
Automatic visual object counting and video surveillance have important applications for home and business environments, such as security and management of access points. However, in order to obtain a satisfactory performance these technologies need professional and expensive hardware, complex installations and setups, and the supervision of qualified workers. In this paper, an efficient visual detection and tracking framework is proposed for the tasks of object counting and surveillance, which meets the requirements of the consumer electronics: off-the-shelf equipment, easy installation and configuration, and unsupervised working conditions. This is accomplished by a novel Bayesian tracking model that can manage multimodal distributions without explicitly computing the association between tracked objects and detections. In addition, it is robust to erroneous, distorted and missing detections. The proposed algorithm is compared with a recent work, also focused on consumer electronics, proving its superior performance.
Resumo:
La presente tesis doctoral describe los desarrollos realizados, y finalmente materializados en patentes con registro de la propiedad intelectual, para la integración de las nuevas tecnologías de documentación fotogramétrica y las bases de datos de los barredores láser terrestres, en los procesos de elaboración, redacción y ejecución de proyectos de restauración y rehabilitación arquitectónicos. Los avances tecnológicos aparecidos en control métrico, junto con las técnicas de imagen digital y los desarrollos fotogramétricos, pueden aportar mejoras significativas en el proceso proyectual y permiten aplicar nuevos procedimientos de extracción de datos para generar de forma sencilla, bajo el control directo y supervisión de los responsables del proyecto, la información métrica y documental más adecuada. Se establecen como principios, y por tanto como base para el diseño de dicha herramienta, que los desarrollos aparecidos sí han producido el uso extendido del sistema CAD (como instrumento de dibujo) así como el uso de la imagen digital como herramienta de documentación. La herramienta a diseñar se fundamenta por tanto en la imagen digital (imágenes digitales, imágenes rectificadas, ortofotografías, estéreo- modelos, estereo- ortofotografías) así como su integración en autocad para un tratamiento interactivo. En la aplicación de la fotogrametría a la disciplina arquitectónica, se considera de interés estructurar aplicaciones con carácter integrador que, con mayores capacidades de interactuación y a partir de información veraz y rigurosa, permitan completar o elaborar documentos de interés proyectual, ABSTRACT This doctoral thesis explains the developments carried out, and finally patented with intellectual property rights, for the integration of the new photogrammetric technology documentation and terrestrial scanner databases in the preparation, documentation and implementation processes of restoration projects and architectural renovation. The technological advances in metric control, as well as the digital image techniques and photogrammetric developments, can together bring a significant improvement to the projecting process, and, under the direct control and supervision of those in charge of the project, can allow new data extraction processes to be applied in order to easily generate the most appropriate metric information and documentation. The principles and, therefore, the basis for the design of this tool are that the developments have indeed produced the widespread use of the CAD system (as a drawing instrument) and the use of digital images as a documentation tool. The tool to be designed is therefore based on digital images (rectify images, orthophotos, stereomodels, stereo-orthophotos) as well as its integration in Autocad for interactive processing. In the application of photogrammetry to the architectural discipline, what interests us is to structure applications of an integrative nature which, with a greater capacity for interaction and from accurate and thorough information, enable the completion or elaboration of documents that are of interest to the project.
Resumo:
Different procedures for monitoring the evolution of leafy vegetables, under plastic covers during cold storage, have been studied. Fifteen spinach leaves were put inside Petri dishes covered with three different plastic films and stored at 4 °C for 21 days. Hyperspectral images were taken during this storage. A radiometric correction is proposed in order to avoid the variation in transmittance of the plastic films during time in the hyperspectral images. Afterwards, three spectral pre-processing procedures (no pre-process, Savitsky–Golay and Standard Normal Variate, combined with Principal Component Analysis) were applied to obtain different models. The corresponding artificial images of scores were studied by means of Analysis of Variance to compare their ability to sense the aging of the leaves. All models were able to monitor the aging through storage. Radiometric correction seemed to work properly and could allow the supervision of shelf-life in leafy vegetables through commercial transparent films.
Resumo:
This article presents the results of the review of the concept and classification of SMEs, by analyzing the state of the art of the past 6 years, the definitions applied in the countries of Central and South America and other international organizations. It was a powerful analytical documentary research, whose purpose was to reflect on the definition of this business category. The results suggest using the criteria: number of employees and turnover. These criteria, future research could be applied to uniform ranges all the American countries, minimizing the potential barriers to promotion policies and existing funding, regardless of sector or country to which they belong.
Resumo:
Fresh-cut or minimally processed fruit and vegetables have been physically modified from its original form (by peeling, trimming, washing and cutting) to obtain a 100% edible product that is subsequently packaged (usually under modified atmosphere packaging –MAP) and kept in refrigerated storage. In fresh-cut products, physiological activity and microbiological spoilage, determine their deterioration and shelf-life. The major preservation techniques applied to delay spoilage are chilling storage and MAP, combined with chemical treatments antimicrobial solutions antibrowning, acidulants, antioxidants, etc.). The industry looks for safer alternatives. Consequently, the sector is asking for innovative, fast, cheap and objective techniques to evaluate the overall quality and safety of fresh-cut products in order to obtain decision tools for implementing new packaging materials and procedures. In recent years, hyperspectral imaging technique has been regarded as a tool for analyses conducted for quality evaluation of food products in research, control and industries. The hyperspectral imaging system allows integrating spectroscopic and imaging techniques to enable direct identification of different components or quality characteristics and their spatial distribution in the tested sample. The objective of this work is to develop hyperspectral image processing methods for the supervision through plastic films of changes related to quality deterioration in packed readyto-use leafy vegetables during shelf life. The evolutions of ready-to-use spinach and watercress samples covered with three different common transparent plastic films were studied. Samples were stored at 4 ºC during the monitoring period (until 21 days). More than 60 hyperspectral images (from 400 to 1000 nm) per species were analyzed using ad hoc routines and commercial toolboxes of MatLab®. Besides common spectral treatments for removing additive and multiplicative effects, additional correction, previously to any other correction, was performed in the images of leaves in order to avoid the modification in their spectra due to the presence of the plastic transparent film. Findings from this study suggest that the developed images analysis system is able to deal with the effects caused in the images by the presence of plastic films in the supervision of shelf-life in leafy vegetables, in which different stages of quality has been identified.
Resumo:
Esta tesis que tiene por título "Contribución a los arrays de antenas activos en banda X", ha sido desarrollada por el estudiante de doctorado Gonzalo Expósito Domínguez, ingeniero de telecomunicación en el Grupo de Radiación del Departamento de Señales, Sistemas y Radiocomunicaciones de la ETSI de Telecomunicación de la Universidad Politécnica de Madrid bajo la dirección de los doctores Manuel Sierra Castañer y José Manuel Fernández González. Esta tesis contiene un profundo estudio del arte en materia de antenas activas en el campo de apuntamiento electrónico. Este estudio comprende desde los fundamentos de este tipo de antenas, problemas de operación y limitaciones hasta los sistemas actuales más avanzados. En ella se identifican las partes críticas en el diseño y posteriormente se llevan a la práctica con el diseño, simulación y construcción de un subarray de una antena integrada en el fuselaje de un avión para comunicaciones multimedia por satélite que funciona en banda X. El prototipo consta de una red de distribución multihaz de banda ancha y una antena planar. El objetivo de esta tesis es el de aplicar nuevas técnicas al diseño de antenas de apuntamiento electrónico. Es por eso que las contribuciones originales son la aplicación de barreras electromagnéticas entre elementos radiantes para reducir los acoplamientos mutuos en arrays de exploración electrónica y el diseño de redes desfasadoras sencillas en las que no son necesarios complejos desfasadores para antenas multihaz. Hasta la fecha, las barreras electromagnéticas, Electronic Band Gap (EBG), se construyen en sustratos de permitividad alta con el fin de aumentar el espacio disponible entre elementos radiantes y reducir el tamaño de estas estructuras. Sin embargo, la utilización de sustratos de alta permitividad aumenta la propagación por ondas de superficie y con ellas el acoplo mutuo. Utilizando sustratos multicapa y colocando la vía de las estructuras en su borde, en vez de en su centro, se consigue reducir el tamaño sin necesidad de usar sustratos de alta permitividad, reducir la eficiencia de radiación de la antena o aumentar la propagación por ondas de superficie. La última parte de la tesis se dedica a las redes conmutadoras y desfasadoras para antenas multihaz. El diseño de las redes de distribución para antenas son una parte crítica ya que se comportan como un atenuador a la entrada de la cadena receptora, modificando en gran medida la figura de ruido del sistema. Las pérdidas de un desfasador digital varían con el desfase introducido, por ese motivo es necesario caracterizar y calibrar los dispositivos correctamente. Los trabajos presentados en este manuscrito constan de un desfasador reflectivo con un conmutador doble serie paralelo para igualar las pérdidas de inserción en los dos estados y también un conmutador de una entrada y dos salidas cuyos puertos están adaptados en todo momento independientemente del camino del conmutador para evitar las reflexiones y fugas entre redes o elementos radiantes. El tomo finaliza con un resumen de las publicaciones en revistas científicas y ponencias en congresos, nacionales e internacionales, el marco de trabajo en el que se ha desarrollado, las colaboraciones que se han realizado y las líneas de investigación futuras. ABSTRACT This thesis was carried out in the Radiation Group of the Signals, Systems and Radiocomunications department of ETSI de Telecomunicación from Technical University of Madrid. Its title is "Contribution to active array antennas at X band" and it is developed by Gonzalo Expósito Domínguez, Electrical Engineer MsC. under the supervision of Prof. Dr. Manuel Sierra Castañer and Dr. José Manuel Fernández González. This thesis is focused on active antennas, specifically multibeam and electronic steering antenas. In the first part of the thesis a thorough description of the state of the art is presented. This study compiles the fundamentals of this antennas, operation problems and limits, up to the breakthrough applications. The critical design problems are described to use them eventually in the design, simulation and prototyping of an airborne steering array antenna for satellite communication at X band. The main objective of this thesis is to apply new techniques to the design of electronically steering antennas. Therefore the new original contributions are the application of Electromagnetic Band Gap materials (EBG) between radiating elements to reduce the mutual coupling when phase shift between elements exist and phase shifting networks where special characteristics are required. So far, the EBG structures have been constructed with high permitivity substrates in order to increase the available space between radiating elements and reduce the size of the structures. However, the surface wave propagation modes are enhanced and therefore the mutual coupling increases when high permitivity substrates are used. By using multilayered substrates and edge location via, the size is reduced meanwhile low permitivity substrates are used without reducing the radiation efficiency or enhancing the surface propagation modes. The last part of the thesis is focused on the phase shifting distribution networks for multibeam antennas. This is a critical part in the antenna design because the insertion loss in the distribution network behaves as an attenuator located in the first place in a receiver chain. The insertion loss will affect directly to the receiver noise figure and the insertion loss in a phase shifter vary with the phase shift. Therefore the devices must be well characterized and calibrated in order to obtain a properly operation. The work developed in this thesis are a reflective phase shifter with a series-shunt switch in order to make symmetrical the insertion loss for the two states and a complex Single Pole Double Through (SPDT) with matched ports in order to reduce the reflections and leakage between feeding networks and radiating elements. The end of this Ph D. dissertation concludes with a summary of the publications in national and international conferences and scientific journals, the collaborations carried out along the thesis and the future research lines.
Resumo:
Nowadays, it is urgent to renovate a great number of residential buildings. The necessity of improving energy efficiency must also be considered as an opportunity to improve indoor comfort. To achieve this goal, it is essential to develop tools to be used in the decision-making process, aiming to refurbish buildings in an integrated, efficient and sustainable way. The integrated system developed is based on a set of indicators. Sustainability indicators are useful to synthesize and organize complex information. They can provide data to evaluate a process in different stages: evaluation, diagnosis, comparison and tracing. The set of proposed indicators aims to accomplish the holistic approach pursued by sustainable development. So, these indicators are divided into three groups: environmental, social and economic. However, the main innovation of the system of indicators is the social ones. The sustainable refurbishment system aims to be a user-focused one. Therefore, the starting point is the needs of the user and social indicators are developed around this. The system tackles the sustainable refurbishment of buildings beyond energy problems. It proposes incorporating users in the decision-making process involving them in the refurbishment and so, contributing to the success of the renovation. In order to achieve this target, three social indicators are used, divided into 10 sub-indicators, and a ?Questionnaire about Sustainable Refurbishment? is drawn up. This research has been carried out in the framework of ?Sustainable Refurbishment? Research and Development Project, an integrated project under the supervision of the Centro para el Desarrollo Tecnológico e Industrial (CDTI) from the Spanish Government, in which University and the Construction Industry collaborate. This research project aims to develop an integrated system for the retrofitting of existing buildings to improve their energy efficiency. Accordingly, an additional objective of the project is to improve quality of life of residents.
Resumo:
En un mundo en el que cada vez son más las personas que utilizan Internet para llevar a cabo sus comunicaciones y relaciones, tanto a nivel personal como profesional, las Redes Sociales se han convertido en una herramienta, para muchos imprescindible, de interacción social. A lo largo de este proyecto se realizará un análisis exhaustivo de las características del ya conocido como Fenómeno Social Media, para posteriormente centrar toda la atención en las ventajas y desventajas de dicha actividad en el ámbito empresarial. Dado que hasta el momento no se ha encontrado ningún otro estudio que trate el Fenómeno Social Media en la Pequeña Empresa del sector TIC español, el objetivo principal del proyecto se centra en conocer la influencia que las Redes Sociales están generando en la estrategia empresarial de dicho sector de empresas, cuyas características y particularidades hacen de ellas un modelo de negocio único y diferente y cuyo número se ha visto incrementado en los últimos años, debido quizás, a la situación de crisis que atraviesa el país. Para ello, se analizarán las ventajas e inconvenientes que se derivan de su uso, como afecta a la creación de valor, y, finalmente, cual es el nivel de implicación y/o inversión que la Pequeña Empresa ha adoptado al respecto. Dentro de este marco, se pretende realizar un énfasis especial en el concepto de cocreación de valor y sus variantes (Crowdsourcing, Open Innovation e Innovación Orientada por las Personas), que permita determinar el grado de importancia del usuario como generador de valor. Posteriormente, en la parte práctica, se llevará a cabo un amplio estudio de mercado, a partir de una encuesta online realizada a un conjunto de empresas, cuidadosamente seleccionadas de entre más de 100.000 compañías, según su sector de actividad, número de empleados y volumen de negocios anual, y cuyos resultados serán analizados con la intención de obtener una visión clara del comportamiento global de las empresas seleccionadas. Por último, tomaremos de referencia las dos últimas publicaciones anuales del “Observatorio sobre el uso de las Redes Sociales en las Pymes españolas” (2011 y 2013), estudios dedicados íntegramente a Redes Sociales y Pymes, y cuya información, junto con los resultados obtenidos de nuestro estudio, nos permitirán determinar la tendencia actual y futura, y en consecuencia, obtener las conclusiones oportunas que nos permitan dar respuesta a las cuestiones planteadas durante la realización del proyecto, y determinar si nos encontramos o no ante una verdadera fuente de éxito empresarial. ABSTRACT. In a world where more and more people are using the Internet to perform their communications and relationships, both personally and professionally, Social Networks have become a social interaction tool, indispensable for many people. Throughout this project we will perform a thorough analysis of the characteristics of the Social Media Phenomenon, focusing on the advantages and disadvantages of such activity in a business scope. Given that we have not found no study on the Social Media Phenomenon in Small Business of the Spanish ICT sector, the main objective of this project is to establish the influence that Social Networks are generating in the business strategy of this business sector, the characteristics and peculiarities that make them a unique and different business model and whose number has increased in recent years, perhaps due to the crisis the country is facing. To do this, we will discuss the advantages and disadvantages derived from its use, how it affects value creation, and finally, what is the level of involvement and / or investment that the Small Business has taken in this regard. In this context, we intend to make a special emphasis on the concept of value co-creation and its variants (Crowdsourcing, Open Innovation and People Oriented Innovation), to determine the degree of importance of the user to value generator. Subsequently, on the practical section, out a comprehensive market study will be carried, based on an online test to a group of enterprises, carefully selected from more than 100,000 companies, according to their sector of activity, number of employees and annual turnover, and the results will be analyzed with the intention to get a clear view of the selected company's overall performance. Finally, we will take as reference the last two annual publications of the "Observatory on the use of Social Networks in Spanish SMEs" (2011 and 2013), entirely dedicated to Social Media and SMEs studies, and whose information, together with the results obtained in our study will allow us to determine the current and future trends, and consequently, obtain the appropriate conclusions to respond to the raised questions throughout the project, and determine whether or not they are a real source of business success.
Resumo:
La presente investigación tiene como objetivo principal diseñar un Modelo de Gestión de Riesgos Operacionales (MGRO) según las Directrices de los Acuerdos II y III del Comité de Supervisión Bancaria de Basilea del Banco de Pagos Internacionales (CSBB-BPI). Se considera importante realizar un estudio sobre este tema dado que son los riesgos operacionales (OpR) los responsables en gran medida de las últimas crisis financieras mundiales y por la dificultad para detectarlos en las organizaciones. Se ha planteado un modelo de gestión subdividido en dos vías de influencias. La primera acoge el paradigma holístico en el que se considera que hay múltiples maneras de percibir un proceso cíclico, así como las herramientas para observar, conocer y entender el objeto o sujeto percibido. La segunda vía la representa el paradigma totalizante, en el que se obtienen datos tanto cualitativos como cuantitativos, los cuales son complementarios entre si. Por otra parte, este trabajo plantea el diseño de un programa informático de OpR Cualitativo, que ha sido diseñado para determinar la raíz de los riesgos en las organizaciones y su Valor en Riesgo Operacional (OpVaR) basado en el método del indicador básico. Aplicando el ciclo holístico al caso de estudio, se obtuvo el siguiente diseño de investigación: no experimental, univariable, transversal descriptiva, contemporánea, retrospectiva, de fuente mixta, cualitativa (fenomenológica y etnográfica) y cuantitativa (descriptiva y analítica). La toma de decisiones y recolección de información se realizó en dos fases en la unidad de estudio. En la primera se tomó en cuenta la totalidad de la empresa Corpoelec-EDELCA, en la que se presentó un universo estadístico de 4271 personas, una población de 2390 personas y una unidad de muestreo de 87 personas. Se repitió el proceso en una segunda fase, para la Central Hidroeléctrica Simón Bolívar, y se determinó un segundo universo estadístico de 300 trabajadores, una población de 191 personas y una muestra de 58 profesionales. Como fuentes de recolección de información se utilizaron fuentes primarias y secundarias. Para recabar la información primaria se realizaron observaciones directas, dos encuestas para detectar las áreas y procesos con mayor nivel de riesgos y se diseñó un cuestionario combinado con otra encuesta (ad hoc) para establecer las estimaciones de frecuencia y severidad de pérdidas operacionales. La información de fuentes secundarias se extrajo de las bases de datos de Corpoelec-EDELCA, de la IEA, del Banco Mundial, del CSBB-BPI, de la UPM y de la UC at Berkeley, entre otras. Se establecieron las distribuciones de frecuencia y de severidad de pérdidas operacionales como las variables independientes y el OpVaR como la variable dependiente. No se realizó ningún tipo de seguimiento o control a las variables bajo análisis, ya que se consideraron estas para un instante especifico y solo se determinan con la finalidad de establecer la existencia y valoración puntual de los OpR en la unidad de estudio. El análisis cualitativo planteado en el MGRO, permitió detectar que en la unidad de investigación, el 67% de los OpR detectados provienen de dos fuentes principales: procesos (32%) y eventos externos (35%). Adicionalmente, la validación del MGRO en Corpoelec-EDELCA, permitió detectar que el 63% de los OpR en la organización provienen de tres categorías principales, siendo los fraudes externos los presentes con mayor regularidad y severidad de pérdidas en la organización. La exposición al riesgo se determinó fundamentándose en la adaptación del concepto de OpVaR que generalmente se utiliza para series temporales y que en el caso de estudio presenta la primicia de aplicarlo a datos cualitativos transformados con la escala Likert. La posibilidad de utilizar distribuciones de probabilidad típicas para datos cuantitativos en distribuciones de frecuencia y severidad de pérdidas con datos de origen cualitativo fueron analizadas. Para el 64% de los OpR estudiados se obtuvo que la frecuencia tiene un comportamiento semejante al de la distribución de probabilidad de Poisson y en un 55% de los casos para la severidad de pérdidas se obtuvo a las log-normal como las distribuciones de probabilidad más comunes, con lo que se concluyó que los enfoques sugeridos por el BCBS-BIS para series de tiempo son aplicables a los datos cualitativos. Obtenidas las distribuciones de frecuencia y severidad de pérdidas, se convolucionaron estas implementando el método de Montecarlo, con lo que se obtuvieron los enfoques de distribuciones de pérdidas (LDA) para cada uno de los OpR. El OpVaR se dedujo como lo sugiere el CSBB-BPI del percentil 99,9 o 99% de cada una de las LDA, obteniéndose que los OpR presentan un comportamiento similar al sistema financiero, resultando como los de mayor peligrosidad los que se ubican con baja frecuencia y alto impacto, por su dificultad para ser detectados y monitoreados. Finalmente, se considera que el MGRO permitirá a los agentes del mercado y sus grupos de interés conocer con efectividad, fiabilidad y eficiencia el status de sus entidades, lo que reducirá la incertidumbre de sus inversiones y les permitirá establecer una nueva cultura de gestión en sus organizaciones. ABSTRACT This research has as main objective the design of a Model for Operational Risk Management (MORM) according to the guidelines of Accords II and III of the Basel Committee on Banking Supervision of the Bank for International Settlements (BCBS- BIS). It is considered important to conduct a study on this issue since operational risks (OpR) are largely responsible for the recent world financial crisis and due to the difficulty in detecting them in organizations. A management model has been designed which is divided into two way of influences. The first supports the holistic paradigm in which it is considered that there are multiple ways of perceiving a cyclical process and contains the tools to observe, know and understand the subject or object perceived. The second way is the totalizing paradigm, in which both qualitative and quantitative data are obtained, which are complementary to each other. Moreover, this paper presents the design of qualitative OpR software which is designed to determine the root of risks in organizations and their Operational Value at Risk (OpVaR) based on the basic indicator approach. Applying the holistic cycle to the case study, the following research design was obtained: non- experimental, univariate, descriptive cross-sectional, contemporary, retrospective, mixed-source, qualitative (phenomenological and ethnographic) and quantitative (descriptive and analytical). Decision making and data collection was conducted in two phases in the study unit. The first took into account the totality of the Corpoelec-EDELCA company, which presented a statistical universe of 4271 individuals, a population of 2390 individuals and a sampling unit of 87 individuals. The process was repeated in a second phase to the Simon Bolivar Hydroelectric Power Plant, and a second statistical universe of 300 workers, a population of 191 people and a sample of 58 professionals was determined. As sources of information gathering primary and secondary sources were used. To obtain the primary information direct observations were conducted and two surveys to identify the areas and processes with higher risks were designed. A questionnaire was combined with an ad hoc survey to establish estimates of frequency and severity of operational losses was also considered. The secondary information was extracted from the databases of Corpoelec-EDELCA, IEA, the World Bank, the BCBS-BIS, UPM and UC at Berkeley, among others. The operational loss frequency distributions and the operational loss severity distributions were established as the independent variables and OpVaR as the dependent variable. No monitoring or control of the variables under analysis was performed, as these were considered for a specific time and are determined only for the purpose of establishing the existence and timely assessment of the OpR in the study unit. Qualitative analysis raised in the MORM made it possible to detect that in the research unit, 67% of detected OpR come from two main sources: external processes (32%) and external events (35%). Additionally, validation of the MORM in Corpoelec-EDELCA, enabled to estimate that 63% of OpR in the organization come from three main categories, with external fraud being present more regularly and greater severity of losses in the organization. Risk exposure is determined basing on adapting the concept of OpVaR generally used for time series and in the case study it presents the advantage of applying it to qualitative data transformed with the Likert scale. The possibility of using typical probability distributions for quantitative data in loss frequency and loss severity distributions with data of qualitative origin were analyzed. For the 64% of OpR studied it was found that the frequency has a similar behavior to that of the Poisson probability distribution and 55% of the cases for loss severity it was found that the log-normal were the most common probability distributions. It was concluded that the approach suggested by the BCBS-BIS for time series can be applied to qualitative data. Once obtained the distributions of loss frequency and severity have been obtained they were subjected to convolution implementing the Monte Carlo method. Thus the loss distribution approaches (LDA) were obtained for each of the OpR. The OpVaR was derived as suggested by the BCBS-BIS 99.9 percentile or 99% of each of the LDA. It was determined that the OpR exhibits a similar behavior to the financial system, being the most dangerous those with low frequency and high impact for their difficulty in being detected and monitored. Finally, it is considered that the MORM will allows market players and their stakeholders to know with effectiveness, efficiency and reliability the status of their entities, which will reduce the uncertainty of their investments and enable them to establish a new management culture in their organizations.
Resumo:
La presente Tesis está orientada al análisis de la supervisión multidistribuida de tres procesos agroalimentarios: el secado solar, el transporte refrigerado y la fermentación de café, a través de la información obtenida de diferentes dispositivos de adquisición de datos, que incorporan sensores, así como el desarrollo de metodologías de análisis de series temporales, modelos y herramientas de control de procesos para la ayuda a la toma de decisiones en las operaciones de estos entornos. En esta tesis se han utilizado: tarjetas RFID (TemTrip®) con sistema de comunicación por radiofrecuencia y sensor de temperatura; el registrador (i-Button®), con sensor integrado de temperatura y humedad relativa y un tercer prototipo empresarial, módulo de comunicación inalámbrico Nlaza, que integra un sensor de temperatura y humedad relativa Sensirion®. Estos dispositivos se han empleado en la conformación de redes multidistribuidas de sensores para la supervisión de: A) Transportes de producto hortofrutícola realizados en condiciones comerciales reales, que son: dos transportes terrestre de producto de IV gama desde Murcia a Madrid; transporte multimodal (barco-barco) de limones desde Montevideo (Uruguay) a Cartagena (España) y transporte multimodal (barco-camión) desde Montevideo (Uruguay) a Verona (Italia). B) dos fermentaciones de café realizadas en Popayán (Colombia) en un beneficiadero. Estas redes han permitido registrar la dinámica espacio-temporal de temperaturas y humedad relativa de los procesos estudiados. En estos procesos de transporte refrigerado y fermentación la aplicación de herramientas de visualización de datos y análisis de conglomerados, han permitido identificar grupos de sensores que presentan patrones análogos de sus series temporales, caracterizando así zonas con dinámicas similares y significativamente diferentes del resto y permitiendo definir redes de sensores de menor densidad cubriendo las diferentes zonas identificadas. Las metodologías de análisis complejo de las series espacio-temporales (modelos psicrométricos, espacio de fases bidimensional e interpolaciones espaciales) permitieron la cuantificación de la variabilidad del proceso supervisado tanto desde el punto de vista dinámico como espacial así como la identificación de eventos. Constituyendo así herramientas adicionales de ayuda a la toma de decisiones en el control de los procesos. Siendo especialmente novedosa la aplicación de la representación bidimensional de los espacios de fases en el estudio de las series espacio-temporales de variables ambientales en aplicaciones agroalimentarias, aproximación que no se había realizado hasta el momento. En esta tesis también se ha querido mostrar el potencial de un sistema de control basado en el conocimiento experto como es el sistema de lógica difusa. Se han desarrollado en primer lugar, los modelos de estimación del contenido en humedad y las reglas semánticas que dirigen el proceso de control, el mejor modelo se ha seleccionado mediante un ensayo de secado realizado sobre bolas de hidrogel como modelo alimentario y finalmente el modelo se ha validado mediante un ensayo en el que se deshidrataban láminas de zanahoria. Los resultados sugirieron que el sistema de control desarrollado, es capaz de hacer frente a dificultades como las variaciones de temperatura día y noche, consiguiendo un producto con buenas características de calidad comparables a las conseguidas sin aplicar ningún control sobre la operación y disminuyendo así el consumo energético en un 98% con respecto al mismo proceso sin control. La instrumentación y las metodologías de análisis de datos implementadas en esta Tesis se han mostrado suficientemente versátiles y transversales para ser aplicadas a diversos procesos agroalimentarios en los que la temperatura y la humedad relativa sean criterios de control en dichos procesos, teniendo una aplicabilidad directa en el sector industrial ABSTRACT This thesis is focused on the analysis of multi-distributed supervision of three agri-food processes: solar drying, refrigerated transport and coffee fermentation, through the information obtained from different data acquisition devices with incorporated sensors, as well as the development of methodologies for analyzing temporary series, models and tools to control processes in order to help in the decision making in the operations within these environments. For this thesis the following has been used: RFID tags (TemTrip®) with a Radiofrequency ID communication system and a temperature sensor; the recorder (i-Button®), with an integrated temperature and relative humidity and a third corporate prototype, a wireless communication module Nlaza, which has an integrated temperature and relative humidity sensor, Sensirion®. These devices have been used in creating three multi-distributed networks of sensors for monitoring: A) Transport of fruits and vegetables made in real commercial conditions, which are: two land trips of IV range products from Murcia to Madrid; multimodal transport (ship - ship) of lemons from Montevideo (Uruguay) to Cartagena (Spain) and multimodal transport (ship - truck) from Montevideo (Uruguay) to Verona (Italy). B) Two coffee fermentations made in Popayan (Colombia) in a coffee processing plant. These networks have allowed recording the time space dynamics of temperatures and relative humidity of the processed under study. Within these refrigerated transport and fermentation processes, the application of data display and cluster analysis tools have allowed identifying sensor groups showing analogical patterns of their temporary series; thus, featuring areas with similar and significantly different dynamics from the others and enabling the definition of lower density sensor networks covering the different identified areas. The complex analysis methodologies of the time space series (psychrometric models, bi-dimensional phase space and spatial interpolation) allowed quantifying the process variability of the supervised process both from the dynamic and spatial points of view; as well as the identification of events. Thus, building additional tools to aid decision-making on process control brought the innovative application of the bi-dimensional representation of phase spaces in the study of time-space series of environmental variables in agri-food applications, an approach that had not been taken before. This thesis also wanted to show the potential of a control system based on specialized knowledge such as the fuzzy logic system. Firstly, moisture content estimation models and semantic rules directing the control process have been developed, the best model has been selected by an drying assay performed on hydrogel beads as food model; and finally the model has been validated through an assay in which carrot sheets were dehydrated. The results suggested that the control system developed is able to cope with difficulties such as changes in temperature daytime and nighttime, getting a product with good quality features comparable to those features achieved without applying any control over the operation and thus decreasing consumption energy by 98% compared to the same uncontrolled process. Instrumentation and data analysis methodologies implemented in this thesis have proved sufficiently versatile and cross-cutting to apply to several agri-food processes in which the temperature and relative humidity are the control criteria in those processes, having a direct effect on the industry sector.