57 resultados para categorization IT PFC computational neuroscience model HMAX
em Universidad Politécnica de Madrid
Resumo:
The monkey anterior intraparietal area (AIP) encodes visual information about three-dimensional object shape that is used to shape the hand for grasping. In robotics a similar role has been played by modules that fit point cloud data to the superquadric family of shapes and its various extensions. We developed a model of shape tuning in AIP based on cosine tuning to superquadric parameters. However, the model did not fit the data well, and we also found that it was difficult to accurately reproduce these parameters using neural networks with the appropriate inputs (modelled on the caudal intraparietal area, CIP). The latter difficulty was related to the fact that there are large discontinuities in the superquadric parameters between very similar shapes. To address these limitations we adopted an alternative shape parameterization based on an Isomap nonlinear dimension reduction. The Isomap was built using gradients and curvatures of object surface depth. This alternative parameterization was low-dimensional (like superquadrics), but data-driven (similar to an alternative clustering approach that is also sometimes used in robotics) and lacked large discontinuities. Isomaps with 16 or more dimensions reproduced the AIP data fairly well. Moreover, we found that the Isomap parameters could be approximated from CIP-like input much more accurately than the superquadric parameters. We conclude that Isomaps, or perhaps alternative dimension reductions of CIP signals, provide a promising model of AIP tuning. We have now started to integrate our model with a robot hand, to explore the efficacy of Isomap shape reductions in grasp planning. Future work will consider dynamics of spike responses and integration with related visual and motor area models.
Resumo:
En los últimos años la externalización de TI ha ganado mucha importancia en el mercado y, por ejemplo, el mercado externalización de servicios de TI sigue creciendo cada año. Ahora más que nunca, las organizaciones son cada vez más los compradores de las capacidades necesarias mediante la obtención de productos y servicios de los proveedores, desarrollando cada vez menos estas capacidades dentro de la empresa. La selección de proveedores de TI es un problema de decisión complejo. Los gerentes que enfrentan una decisión sobre la selección de proveedores de TI tienen dificultades en la elaboración de lo que hay que pensar, además en sus discursos. También de acuerdo con un estudio del SEI (Software Engineering Institute) [40], del 20 al 25 por ciento de los grandes proyectos de adquisición de TI fracasan en dos años y el 50 por ciento fracasan dentro de cinco años. La mala gestión, la mala definición de requisitos, la falta de evaluaciones exhaustivas, que pueden ser utilizadas para llegar a los mejores candidatos para la contratación externa, la selección de proveedores y los procesos de contratación inadecuados, la insuficiencia de procedimientos de selección tecnológicos, y los cambios de requisitos no controlados son factores que contribuyen al fracaso del proyecto. La mayoría de los fracasos podrían evitarse si el cliente aprendiese a comprender los problemas de decisión, hacer un mejor análisis de decisiones, y el buen juicio. El objetivo principal de este trabajo es el desarrollo de un modelo de decisión para la selección de proveedores de TI que tratará de reducir la cantidad de fracasos observados en las relaciones entre el cliente y el proveedor. La mayor parte de estos fracasos son causados por una mala selección, por parte del cliente, del proveedor. Además de estos problemas mostrados anteriormente, la motivación para crear este trabajo es la inexistencia de cualquier modelo de decisión basado en un multi modelo (mezcla de modelos adquisición y métodos de decisión) para el problema de la selección de proveedores de TI. En el caso de estudio, nueve empresas españolas fueron analizadas de acuerdo con el modelo de decisión para la selección de proveedores de TI desarrollado en este trabajo. Dos softwares se utilizaron en este estudio de caso: Expert Choice, y D-Sight. ABSTRACT In the past few years IT outsourcing has gained a lot of importance in the market and, for example, the IT services outsourcing market is still growing every year. Now more than ever, organizations are increasingly becoming acquirers of needed capabilities by obtaining products and services from suppliers and developing less and less of these capabilities in-house. IT supplier selection is a complex and opaque decision problem. Managers facing a decision about IT supplier selection have difficulty in framing what needs to be thought about further in their discourses. Also according to a study from SEI (Software Engineering Institute) [40], 20 to 25 percent of large information technology (IT) acquisition projects fail within two years and 50 percent fail within five years. Mismanagement, poor requirements definition, lack of comprehensive evaluations, which can be used to come up with the best candidates for outsourcing, inadequate supplier selection and contracting processes, insufficient technology selection procedures, and uncontrolled requirements changes are factors that contribute to project failure. The majority of project failures could be avoided if the acquirer learns how to understand the decision problems, make better decision analysis, and good judgment. The main objective of this work is the development of a decision model for IT supplier selection that will try to decrease the amount of failures seen in the relationships between the client-supplier. Most of these failures are caused by a not well selection of the supplier. Besides these problems showed above, the motivation to create this work is the inexistence of any decision model based on multi model (mixture of acquisition models and decision methods) for the problem of IT supplier selection. In the case study, nine different Spanish companies were analyzed based on the IT supplier selection decision model developed in this work. Two software products were used in this case study, Expert Choice and D-Sight.
Resumo:
It is easy to get frustrated at spoken conversational agents (SCAs), perhaps because they seem to be callous. By and large, the quality of human-computer interaction is affected due to the inability of the SCAs to recognise and adapt to user emotional state. Now with the mass appeal of artificially-mediated communication, there has been an increasing need for SCAs to be socially and emotionally intelligent, that is, to infer and adapt to their human interlocutors’ emotions on the fly, in order to ascertain an affective, empathetic and naturalistic interaction. An enhanced quality of interaction would reduce users’ frustrations and consequently increase their satisfactions. These reasons have motivated the development of SCAs towards including socio-emotional elements, turning them into affective and socially-sensitive interfaces. One barrier to the creation of such interfaces has been the lack of methods for modelling emotions in a task-independent environment. Most emotion models for spoken dialog systems are task-dependent and thus cannot be used “as-is” in different applications. This Thesis focuses on improving this, in which it concerns computational modeling of emotion, personality and their interrelationship for task-independent autonomous SCAs. The generation of emotion is driven by needs, inspired by human’s motivational systems. The work in this Thesis is organised in three stages, each one with its own contribution. The first stage involved defining, integrating and quantifying the psychological-based motivational and emotional models sourced from. Later these were transformed into a computational model by implementing them into software entities. The computational model was then incorporated and put to test with an existing SCA host, a HiFi-control agent. The second stage concerned automatic prediction of affect, which has been the main challenge towards the greater aim of infusing social intelligence into the HiFi agent. In recent years, studies on affect detection from voice have moved on to using realistic, non-acted data, which is subtler. However, it is more challenging to perceive subtler emotions and this is demonstrated in tasks such as labelling and machine prediction. In this stage, we attempted to address part of this challenge by considering the roles of user satisfaction ratings and conversational/dialog features as the respective target and predictors in discriminating contentment and frustration, two types of emotions that are known to be prevalent within spoken human-computer interaction. The final stage concerned the evaluation of the emotional model through the HiFi agent. A series of user studies with 70 subjects were conducted in a real-time environment, each in a different phase and with its own conditions. All the studies involved the comparisons between the baseline non-modified and the modified agent. The findings have gone some way towards enhancing our understanding of the utility of emotion in spoken dialog systems in several ways; first, an SCA should not express its emotions blindly, albeit positive. Rather, it should adapt its emotions to user states. Second, low performance in an SCA may be compensated by the exploitation of emotion. Third, the expression of emotion through the exploitation of prosody could better improve users’ perceptions of an SCA compared to exploiting emotions through just lexical contents. Taken together, these findings not only support the success of the emotional model, but also provide substantial evidences with respect to the benefits of adding emotion in an SCA, especially in mitigating users’ frustrations and ultimately improving their satisfactions. Resumen Es relativamente fácil experimentar cierta frustración al interaccionar con agentes conversacionales (Spoken Conversational Agents, SCA), a menudo porque parecen ser un poco insensibles. En general, la calidad de la interacción persona-agente se ve en cierto modo afectada por la incapacidad de los SCAs para identificar y adaptarse al estado emocional de sus usuarios. Actualmente, y debido al creciente atractivo e interés de dichos agentes, surge la necesidad de hacer de los SCAs unos seres cada vez más sociales y emocionalmente inteligentes, es decir, con capacidad para inferir y adaptarse a las emociones de sus interlocutores humanos sobre la marcha, de modo que la interacción resulte más afectiva, empática y, en definitiva, natural. Una interacción mejorada en este sentido permitiría reducir la posible frustración de los usuarios y, en consecuencia, mejorar el nivel de satisfacción alcanzado por los mismos. Estos argumentos justifican y motivan el desarrollo de nuevos SCAs con capacidades socio-emocionales, dotados de interfaces afectivas y socialmente sensibles. Una de las barreras para la creación de tales interfaces ha sido la falta de métodos de modelado de emociones en entornos independientes de tarea. La mayoría de los modelos emocionales empleados por los sistemas de diálogo hablado actuales son dependientes de tarea y, por tanto, no pueden utilizarse "tal cual" en diferentes dominios o aplicaciones. Esta tesis se centra precisamente en la mejora de este aspecto, la definición de modelos computacionales de las emociones, la personalidad y su interrelación para SCAs autónomos e independientes de tarea. Inspirada en los sistemas motivacionales humanos en el ámbito de la psicología, la tesis propone un modelo de generación/producción de la emoción basado en necesidades. El trabajo realizado en la presente tesis está organizado en tres etapas diferenciadas, cada una con su propia contribución. La primera etapa incluyó la definición, integración y cuantificación de los modelos motivacionales de partida y de los modelos emocionales derivados a partir de éstos. Posteriormente, dichos modelos emocionales fueron plasmados en un modelo computacional mediante su implementación software. Este modelo computacional fue incorporado y probado en un SCA anfitrión ya existente, un agente con capacidad para controlar un equipo HiFi, de alta fidelidad. La segunda etapa se orientó hacia el reconocimiento automático de la emoción, aspecto que ha constituido el principal desafío en relación al objetivo mayor de infundir inteligencia social en el agente HiFi. En los últimos años, los estudios sobre reconocimiento de emociones a partir de la voz han pasado de emplear datos actuados a usar datos reales en los que la presencia u observación de emociones se produce de una manera mucho más sutil. El reconocimiento de emociones bajo estas condiciones resulta mucho más complicado y esta dificultad se pone de manifiesto en tareas tales como el etiquetado y el aprendizaje automático. En esta etapa, se abordó el problema del reconocimiento de las emociones del usuario a partir de características o métricas derivadas del propio diálogo usuario-agente. Gracias a dichas métricas, empleadas como predictores o indicadores del grado o nivel de satisfacción alcanzado por el usuario, fue posible discriminar entre satisfacción y frustración, las dos emociones prevalentes durante la interacción usuario-agente. La etapa final corresponde fundamentalmente a la evaluación del modelo emocional por medio del agente Hifi. Con ese propósito se llevó a cabo una serie de estudios con usuarios reales, 70 sujetos, interaccionando con diferentes versiones del agente Hifi en tiempo real, cada uno en una fase diferente y con sus propias características o capacidades emocionales. En particular, todos los estudios realizados han profundizado en la comparación entre una versión de referencia del agente no dotada de ningún comportamiento o característica emocional, y una versión del agente modificada convenientemente con el modelo emocional propuesto. Los resultados obtenidos nos han permitido comprender y valorar mejor la utilidad de las emociones en los sistemas de diálogo hablado. Dicha utilidad depende de varios aspectos. En primer lugar, un SCA no debe expresar sus emociones a ciegas o arbitrariamente, incluso aunque éstas sean positivas. Más bien, debe adaptar sus emociones a los diferentes estados de los usuarios. En segundo lugar, un funcionamiento relativamente pobre por parte de un SCA podría compensarse, en cierto modo, dotando al SCA de comportamiento y capacidades emocionales. En tercer lugar, aprovechar la prosodia como vehículo para expresar las emociones, de manera complementaria al empleo de mensajes con un contenido emocional específico tanto desde el punto de vista léxico como semántico, ayuda a mejorar la percepción por parte de los usuarios de un SCA. Tomados en conjunto, los resultados alcanzados no sólo confirman el éxito del modelo emocional, sino xv que constituyen además una evidencia decisiva con respecto a los beneficios de incorporar emociones en un SCA, especialmente en cuanto a reducir el nivel de frustración de los usuarios y, en última instancia, mejorar su satisfacción.
Finite Element Analysis Model of a Contactless Transformer for Battery Chargers in Electric Vehicles
Resumo:
A contactless transformer model is proposed in this paper using Finite Element Analysis (FEA). This model can be used to simulate Inductive Coupling Power Transfer (ICPT) systems with good accuracy of the transformer and reduce the fabrication time of these systems. The model not only takes into account the geometry of the windings but also the frequency effects in them. As the transformer does not have a magnetic core, it is complicated to model because the flux is expanded in the area around the windings. In order to obtain a very accurate model, it is necessary to use a 2D/3D field solver.
Resumo:
En el presente proyecto se ha realizado un estudio sobre las condiciones acústicas de la iglesia Santa María del Castillo, ubicada en la localidad de Campo Real, al sureste de Madrid. Se trata de una iglesia construida entre los siglos XIV y XVII en diferentes fases, rica en características arquitectónicas correspondientes a varios estilos, tales como el gótico, el renacentista y el barroco. Reconocida en 1981 por sus valores arquitectónicos como Monumento Histórico–Artístico. A partir de unas completas mediciones del interior de la iglesia, se ha realizado un modelo tridimensional del mismo como base para la simulación mediante el software de simulación acústica EASE versión 4.3. Para conseguir que este modelo se asemeje a la realidad, se han realizado medidas del ruido de fondo en el interior de la iglesia en diferentes condiciones ambientales. Además se han creado mediante el software los coeficientes de absorción correspondientes a cada material presente en el interior de la iglesia y se han tenido en cuenta las características de los altavoces utilizados en la megafonía del recinto. El modelo en 3D obtenido caracteriza completamente las condiciones acústicas de la iglesia Santa María del Castillo, y nos sirve para valorar cómo es el sonido en el interior de la misma. Para ello obtenemos valores de diferentes parámetros acústico realizando simulaciones. Parámetros como el tiempo de reverberación y el nivel de presión sonora nos dan una idea general de cómo es el campo sonoro en el interior del recinto. Otros parámetros como el ALCons y el STI nos dan información sobre la inteligibilidad de la palabra en el recinto en el que se está realizando el estudio. Finalmente basándonos en los resultados obtenidos de la simulación se sacan conclusiones sobre las características acústicas de este recinto. La iglesia estudiada no es un recinto apropiado para la palabra y/o la música, además el predominio del campo reverberante sobre el campo directo es claro, esto es debido a las dimensiones del recinto y la poca absorción de los diferentes materiales empleados en su construcción, que son bastante reflexivos al sonido. ABSTRACT The present project undertakes the acoustic study of the church Santa María del Castillo. The church is the main temple of Campo Real, in the south-east of Madrid. It was built over different phases between the 14th and the 17th centuries and therefore, the presence of several architectural styles makes the church of Campo Real an interesting aim for this study. The building was recognised as Historic-Artistic Monument for its architectural value in 1981. Complete measurements from inside of the church were taken to build a computational 3D model which has been used to perform acoustic simulations of the church with the software EASE (Version 4.3). Noise measurements have been taken inside the church at different ambient conditions and they have been used to improve the reliability of the computational model. Furthermore, the model has been provided with software generated absorption coefficients and the characteristics of the actual loudspeakers have been taken into account. The 3D model created characterises all the acoustic conditions of the church Santa María del Castillo and allows the study of the sound properties inside the temple. Parameters such as reverberation time and sound pressure level were calculated performing simulations so the sound field inside the building can be described. Other parameters such as the Articulation Loss of Consonants (ALCons) and the Speech Transmission Index (STI) were studied to derive information about intelligibility inside the church. Finally, based on the results obtained by the simulation, I expose my conclusions about the acoustic characteristics of the building. The main conclusion derived from the present study is that the temple is not an appropriate place for voice or music listening due to the dimensions and the characteristics of the materials used in the construction since they are highly reflective to sound. The reverberant field predominates over the whole audience area in comparison with the direct field.
Resumo:
Esta tesis doctoral está encuadrada dentro del marco general de la ingeniería biomédica aplicada al tratamiento de las enfermedades cardiovasculares, enfermedades que provocan alrededor de 1.9 millones (40%) de muertes al año en la Unión Europea. En este contexto surge el proyecto europeo SCATh-Smart Catheterization, cuyo objetivo principal es mejorar los procedimientos de cateterismo aórtico introduciendo nuevas tecnologías de planificación y navegación quirúrgica y minimizando el uso de fluoroscopía. En particular, esta tesis aborda el modelado y diagnóstico de aneurismas aórticos abdominales (AAA) y del trombo intraluminal (TIL), allí donde esté presente, así como la segmentación de estas estructuras en imágenes preoperatorias de RM. Los modelos físicos específicos del paciente, construidos a partir de imágenes médicas preoperatorias, tienen múltiples usos, que van desde la evaluación preoperatoria de estructuras anatómicas a la planificación quirúrgica para el guiado de catéteres. En el diagnóstico y tratamiento de AAA, los modelos físicos son útiles a la hora de evaluar diversas variables biomecánicas y fisiológicas de las estructuras vasculares. Existen múltiples técnicas que requieren de la generación de modelos físicos que representen la anatomía vascular. Una de las principales aplicaciones de los modelos físicos es el análisis de elementos finitos (FE). Las simulaciones de FE para AAA pueden ser específicas para el paciente y permiten modelar estados de estrés complejos, incluyendo los efectos provocados por el TIL. La aplicación de métodos numéricos de análisis tiene como requisito previo la generación de una malla computacional que representa la geometría de interés mediante un conjunto de elementos poliédricos, siendo los hexaédricos los que presentan mejores resultados. En las estructuras vasculares, generar mallas hexaédricas es un proceso especialmente exigente debido a la compleja anatomía 3D ramificada. La mayoría de los AAA se encuentran situados en la bifurcación de la arteria aorta en las arterias iliacas y es necesario modelar de manera fiel dicha bifurcación. En el caso de que la sangre se estanque en el aneurisma provocando un TIL, éste forma una estructura adyacente a la pared aórtica. De este modo, el contorno externo del TIL es el mismo que el contorno interno de la pared, por lo que las mallas resultantes deben reflejar esta particularidad, lo que se denomina como "mallas conformadas". El fin último de este trabajo es modelar las estructuras vasculares de modo que proporcionen nuevas herramientas para un mejor diagnóstico clínico, facilitando medidas de riesgo de rotura de la arteria, presión sistólica o diastólica, etc. Por tanto, el primer objetivo de esta tesis es diseñar un método novedoso y robusto para generar mallas hexaédricas tanto de la pared aórtica como del trombo. Para la identificación de estas estructuras se utilizan imágenes de resonancia magnética (RM). Deben mantenerse sus propiedades de adyacencia utilizando elementos de alta calidad, prestando especial atención al modelado de la bifurcación y a que sean adecuadas para el análisis de FE. El método tiene en cuenta la evolución de la línea central del vaso en el espacio tridimensional y genera la malla directamente a partir de las imágenes segmentadas, sin necesidad de reconstruir superficies triangulares. Con el fin de reducir la intervención del usuario en el proceso de generación de las mallas, es también objetivo de esta tesis desarrollar un método de segmentación semiautomática de las distintas estructuras de interés. Las principales contribuciones de esta tesis doctoral son: 1. El diseño, implementación y evaluación de un algoritmo de generación de mallas hexaédricas conformadas de la pared y el TIL a partir de los contornos segmentados en imágenes de RM. Se ha llevado a cabo una evaluación de calidad que determine su aplicabilidad a métodos de FE. Los resultados demuestran que el algoritmo desarrollado genera mallas conformadas de alta calidad incluso en la región de la bifurcación, que son adecuadas para su uso en métodos de análisis de FE. 2. El diseño, implementación y evaluación de un método de segmentación automático de las estructuras de interés. La luz arterial se segmenta de manera semiautomática utilizando un software disponible a partir de imágenes de RM con contraste. Los resultados de este proceso sirven de inicialización para la segmentación automática de las caras interna y externa de la pared aórtica utilizando métodos basado en modelos de textura y forma a partir de imágenes de RM sin contraste. Los resultados demuestran que el algoritmo desarrollado proporciona segmentaciones fieles de las distintas estructuras de interés. En conclusión, el trabajo realizado en esta tesis doctoral corrobora las hipótesis de investigación postuladas, y pretende servir como aportación para futuros avances en la generación de modelos físicos de geometrías biológicas. ABSTRACT The frame of this PhD Thesis is the biomedical engineering applied to the treatment of cardiovascular diseases, which cause around 1.9 million deaths per year in the European Union and suppose about 40% of deaths per year. In this context appears the European project SCATh-Smart Catheterization. The main objective of this project is creating a platform which improves the navigation of catheters in aortic catheterization minimizing the use of fluoroscopy. In the framework of this project, the specific field of this PhD Thesis is the diagnosis and modeling of abdominal aortic aneurysm (AAAs) and the intraluminal thrombus (ILT) whenever it is present. Patient-specific physical models built from preoperative imaging are becoming increasingly important in the area of minimally invasive surgery. These models can be employed for different purposes, such as the preoperatory evaluation of anatomic structures or the surgical planning for catheter guidance. In the specific case of AAA diagnosis and treatment, physical models are especially useful for evaluating pressures over vascular structures. There are multiple techniques that require the generation of physical models which represent the target anatomy. Finite element (FE) analysis is one the principal applications for physical models. FE simulations for AAA may be patient-specific and allow modeling biomechanical and physiological variables including those produced by ILT, and also the segmentation of those anatomical structures in preoperative MR images. Applying numeric methods requires the generation of a proper computational mesh. These meshes represent the patient anatomy using a set of polyhedral elements, with hexahedral elements providing better results. In the specific case of vascular structures, generating hexahedral meshes is a challenging task due to the complex 3D branching anatomy. Each patient’s aneurysm is unique, characterized by its location and shape, and must be accurately represented for subsequent analyses to be meaningful. Most AAAs are located in the region where the aorta bifurcates into the iliac arteries and it is necessary to model this bifurcation precisely and reliably. If blood stagnates in the aneurysm and forms an ILT, it exists as a conforming structure with the aortic wall, i.e. the ILT’s outer contour is the same as the wall’s inner contour. Therefore, resulting meshes must also be conforming. The main objective of this PhD Thesis is designing a novel and robust method for generating conforming hexahedral meshes for the aortic wall and the thrombus. These meshes are built using largely high-quality elements, especially at the bifurcation, that are suitable for FE analysis of tissue stresses. The method accounts for the evolution of the vessel’s centerline which may develop outside a single plane, and generates the mesh directly from segmented images without the requirement to reconstruct triangular surfaces. In order to reduce the user intervention in the mesh generation process is also a goal of this PhD. Thesis to develop a semiautomatic segmentation method for the structures of interest. The segmentation is performed from magnetic resonance image (MRI) sequences that have tuned to provide high contrast for the arterial tissue against the surrounding soft tissue, so that we determine the required information reliably. The main contributions of this PhD Thesis are: 1. The design, implementation and evaluation of an algorithm for generating hexahedral conforming meshes of the arterial wall and the ILT from the segmented contours. A quality inspection has been applied to the meshes in order to determine their suitability for FE methods. Results show that the developed algorithm generates high quality conforming hexahedral meshes even at the bifurcation region. Thus, these meshes are suitable for FE analysis. 2. The design, implementation and evaluation of a semiautomatic segmentation method for the structures of interest. The lumen is segmented in a semiautomatic way from contrast filled MRI using an available software. The results obtained from this process are used to initialize the automatic segmentation of the internal and external faces of the aortic wall. These segmentations are performed by methods based on texture and shape models from MRI with no contrast. The results show that the algorithm provides faithful segmentations of the structures of interest requiring minimal user intervention. In conclusion, the work undertaken in this PhD. Thesis verifies the investigation hypotheses. It intends to serve as basis for future physical model generation of proper biological anatomies used by numerical methods.
Resumo:
University education in Peru is based on models of teacher-centered teaching and a conception of knowledge which is closed and static and under the dominance of an information model now overwhelmed by multiple factors hastened by international change. The worlds most prestigious universities have chosen cultural diversity as a sign of quality and are hence interested in the mobility of teachers and students through exchange and cooperation with foreign educational institutions. These universities respond more effectively to pressure from the international business sector, better satisfy training demands, introduce new information and communication technologies into education and research and have improved administration and management structures. While there is progress, the university system in Peru is a planning model defined "as a discipline that seeks to respond to the needs of an organization defined by new cultural and social models" (A. Cazorla, et al 2007).This paper studies the non-Euclidean thinking of planning and development of John Friedmann (2001). Based on the four domains of social practice, it proposes a planning model for Peruvian universities that meets international requirements.
Resumo:
The monkey anterior intraparietal area (AIP) encodes visual information about three-dimensional object shape that is used to shape the hand for grasping. We modeled shape tuning in visual AIP neurons and its relationship with curvature and gradient information from the caudal intraparietal area (CIP). The main goal was to gain insight into the kinds of shape parameterizations that can account for AIP tuning and that are consistent with both the inputs to AIP and the role of AIP in grasping. We first experimented with superquadric shape parameters. We considered superquadrics because they occupy a role in robotics that is similar to AIP , in that superquadric fits are derived from visual input and used for grasp planning. We also experimented with an alternative shape parameterization that was based on an Isomap dimension reduction of spatial derivatives of depth (i.e., distance from the observer to the object surface). We considered an Isomap-based model because its parameters lacked discontinuities between similar shapes. When we matched the dimension of the Isomap to the number of superquadric parameters, the superquadric model fit the AIP data somewhat more closely. However, higher-dimensional Isomaps provided excellent fits. Also, we found that the Isomap parameters could be approximated much more accurately than superquadric parameters by feedforward neural networks with CIP-like inputs. We conclude that Isomaps, or perhaps alternative dimension reductions of visual inputs to AIP, provide a promising model of AIP electrophysiology data. Further work is needed to test whether such shape parameterizations actually provide an effective basis for grasp control.
Resumo:
Abstract Interneuron classification is an important and long-debated topic in neuroscience. A recent study provided a data set of digitally reconstructed interneurons classified by 42 leading neuroscientists according to a pragmatic classification scheme composed of five categorical variables, namely, of the interneuron type and four features of axonal morphology. From this data set we now learned a model which can classify interneurons, on the basis of their axonal morphometric parameters, into these five descriptive variables simultaneously. Because of differences in opinion among the neuroscientists, especially regarding neuronal type, for many interneurons we lacked a unique, agreed-upon classification, which we could use to guide model learning. Instead, we guided model learning with a probability distribution over the neuronal type and the axonal features, obtained, for each interneuron, from the neuroscientists’ classification choices. We conveniently encoded such probability distributions with Bayesian networks, calling them label Bayesian networks (LBNs), and developed a method to predict them. This method predicts an LBN by forming a probabilistic consensus among the LBNs of the interneurons most similar to the one being classified. We used 18 axonal morphometric parameters as predictor variables, 13 of which we introduce in this paper as quantitative counterparts to the categorical axonal features. We were able to accurately predict interneuronal LBNs. Furthermore, when extracting crisp (i.e., non-probabilistic) predictions from the predicted LBNs, our method outperformed related work on interneuron classification. Our results indicate that our method is adequate for multi-dimensional classification of interneurons with probabilistic labels. Moreover, the introduced morphometric parameters are good predictors of interneuron type and the four features of axonal morphology and thus may serve as objective counterparts to the subjective, categorical axonal features.
Resumo:
Interneuron classification is an important and long-debated topic in neuroscience. A recent study provided a data set of digitally reconstructed interneurons classified by 42 leading neuroscientists according to a pragmatic classification scheme composed of five categorical variables, namely, of the interneuron type and four features of axonal morphology. From this data set we now learned a model which can classify interneurons, on the basis of their axonal morphometric parameters, into these five descriptive variables simultaneously. Because of differences in opinion among the neuroscientists, especially regarding neuronal type, for many interneurons we lacked a unique, agreed-upon classification, which we could use to guide model learning. Instead, we guided model learning with a probability distribution over the neuronal type and the axonal features, obtained, for each interneuron, from the neuroscientists’ classification choices. We conveniently encoded such probability distributions with Bayesian networks, calling them label Bayesian networks (LBNs), and developed a method to predict them. This method predicts an LBN by forming a probabilistic consensus among the LBNs of the interneurons most similar to the one being classified. We used 18 axonal morphometric parameters as predictor variables, 13 of which we introduce in this paper as quantitative counterparts to the categorical axonal features. We were able to accurately predict interneuronal LBNs. Furthermore, when extracting crisp (i.e., non-probabilistic) predictions from the predicted LBNs, our method outperformed related work on interneuron classification. Our results indicate that our method is adequate for multi-dimensional classification of interneurons with probabilistic labels. Moreover, the introduced morphometric parameters are good predictors of interneuron type and the four features of axonal morphology and thus may serve as objective counterparts to the subjective, categorical axonal features.
Resumo:
With the growing body of research on traumatic brain injury and spinal cord injury, computational neuroscience has recently focused its modeling efforts on neuronal functional deficits following mechanical loading. However, in most of these efforts, cell damage is generally only characterized by purely mechanistic criteria, function of quantities such as stress, strain or their corresponding rates. The modeling of functional deficits in neurites as a consequence of macroscopic mechanical insults has been rarely explored. In particular, a quantitative mechanically based model of electrophysiological impairment in neuronal cells has only very recently been proposed (Jerusalem et al., 2013). In this paper, we present the implementation details of Neurite: the finite difference parallel program used in this reference. Following the application of a macroscopic strain at a given strain rate produced by a mechanical insult, Neurite is able to simulate the resulting neuronal electrical signal propagation, and thus the corresponding functional deficits. The simulation of the coupled mechanical and electrophysiological behaviors requires computational expensive calculations that increase in complexity as the network of the simulated cells grows. The solvers implemented in Neurite-explicit and implicit-were therefore parallelized using graphics processing units in order to reduce the burden of the simulation costs of large scale scenarios. Cable Theory and Hodgkin-Huxley models were implemented to account for the electrophysiological passive and active regions of a neurite, respectively, whereas a coupled mechanical model accounting for the neurite mechanical behavior within its surrounding medium was adopted as a link between lectrophysiology and mechanics (Jerusalem et al., 2013). This paper provides the details of the parallel implementation of Neurite, along with three different application examples: a long myelinated axon, a segmented dendritic tree, and a damaged axon. The capabilities of the program to deal with large scale scenarios, segmented neuronal structures, and functional deficits under mechanical loading are specifically highlighted.
Resumo:
Present research is framed within the project MODIFICA (MODelo predictivo - edIFIcios - Isla de Calor urbanA) aimed at developing a predictive model for dwelling energy performance under the urban heat island effect in order to implement it in the evaluation of real energy demand and consumption of dwellings as well as in the selection of energy retrofitting strategies. It is funded by Programa de I+D+i orientada a los retos de la sociedad 'Retos Investigación' 2013. Despite great advances on building energy performance have been achieved during the last years, available climate data is derived from weather stations placed in the outskirts of the city. Hence, urban heat island effect is not considered in energy simulations, which implies an important lack of accuracy. Since 1980's several international studies have been conducted on the urban heat island (UHI) phenomena, which modifies the atmospheric conditions of the urban centres due to urban agglomeration [1][2][3][4]. In the particular case of Madrid, multiple maps haven been generated using different methodologies during the last two decades [5][6][7]. These maps allow us to study the UHI phenomena from a wide perspective, offering however an static representation of it. Consequently a dynamic model for Madrid UHI is proposed, in order to evaluate it in a continuous way, and to be able to integrate it in building energy simulations.
Resumo:
Complexity has always been one of the most important issues in distributed computing. From the first clusters to grid and now cloud computing, dealing correctly and efficiently with system complexity is the key to taking technology a step further. In this sense, global behavior modeling is an innovative methodology aimed at understanding the grid behavior. The main objective of this methodology is to synthesize the grid's vast, heterogeneous nature into a simple but powerful behavior model, represented in the form of a single, abstract entity, with a global state. Global behavior modeling has proved to be very useful in effectively managing grid complexity but, in many cases, deeper knowledge is needed. It generates a descriptive model that could be greatly improved if extended not only to explain behavior, but also to predict it. In this paper we present a prediction methodology whose objective is to define the techniques needed to create global behavior prediction models for grid systems. This global behavior prediction can benefit grid management, specially in areas such as fault tolerance or job scheduling. The paper presents experimental results obtained in real scenarios in order to validate this approach.
Resumo:
In recent years, interest has been drawn towards new inclusive business models designed to generate opportunities for human development and thus improve the living standards of the poorest. Social enterprise, defined as an organization that aims to create social value by developing an economic activity, can influence change and therefore serve as a model to create sustainable organizations in the coming years. This paper will describe the concept of social enterprise from two different points of view ? that of the traditional European social economy and the developing economies approach. In addition, it will propose a model for the organizational analysis and will apply the suggested model in the case of Grameen-Danone.
Resumo:
La presente tesis doctoral estudia las características de un transductor acústico bioinspirado en la estructura del maxilar inferior de un Zifio de Couvier (Ziphius cavirostris). El mecanismo de funcionamiento del sensor se basa en las características de un sistema acoplado formado por los distintos componentes acústicos identificados en el maxilar. Para analizar las características del sensor se propone un modelo simplificado 2D que consta de una cavidad cerrada con forma de bocina acoplada a una lengüeta. Una parte de la lengüeta se encuentra dentro de la cavidad y otra en el exterior. Dicha lengüeta detecta los cambios de presión acústica y las vibraciones generadas por el sonido y las transmite con ondas de flexión al interior de la cavidad. La excitación prolongada sobre la placa puede provocar la activación de los modos propios del sistema acoplado. Dichos modos se caracterizan porque presentan un máximo de presión en el cono de la bocina la cual a su vez actúa como un amplificador acústico. Mediante el Método de los elementos Finitos se analizan las características acústicas del sensor y se construye un prototipo experimental para validar los resultados evaluados en el modelo numérico. Se propone una metodología numérica que permite desarrollar y validar un elemento tetraédrico para caracterizar el comportamiento isotrópico de los medios porosos. La metodología permite construir elementos de línea y bidimensionales. A partir de esta metodología se desarrolla un elemento plano ortotrópico. Se realiza un modelo de la cavidad en el que una de las paredes de la bocina está constituida por material poroso y se une la parte exterior de la lengüeta a dicho material para que constituya una viga sobre un apoyo elástico. Se calcula la respuesta modal y se discuten los efectos del material poroso en la eficiencia del transductor y las posibles mejoras a introducir en el mismo. SUMMARY This Thesis studies the characteristics of an acoustic transducer bioinspired by the structure of the lower maxilla of an odontoceto. In this case a Cuvier’s beaked whale (Ziphius cavirostris). The transducer working mechanism is based in a coupled system, with components identified in the maxilla. To analyze the transducer a simplified 2D model composed by a horn shaped closed cavity is modeled. The cavity is coupled with a flat belt. The belt has one part inside the cavity and the other part outside of it. The belt traverses the cavity wall and it is in charge to pick the vibrations from outside and introduce it inside the cavity. The transmission is obtained through the belt bending. A sustained external load with the right frequency contents will allow the system to reach a stationary pressure intensity distribution inside the cavity. Frequencies with modal shapes that show an important intensity increase at the horn tip are of special interest because of the signal amplification. A finite element model is constructed to study the transducer coupled modes and a prototype is constructed to validate the numerical results. A numerical methodology to construct and validate a tetrahedral finite element for isotropic porous materials is presented. The methodology allows constructing linear and 2D elements. It is extended to model orthotropic porous materials behavior. At the end, one of the horn walls is made of an orthotropic material and the external belt is glued to it in order to configure a belt over an elastic foundation. Modal response is evaluated and the porous material effects in the transducer efficiency and further improvements are discussed.