898 resultados para Vision, colour and visual optics


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Oxygen isotopic (d18O) climatic stratigraphy and radiocarbon chronology, at high resolution, have been used to establish an age model for Ocean Drilling Program Hole 1017E, a continuous 25-m sequence of hemipelagic sediments from the continental slope (956 m water depth), east of Point Arguella, Southern California. The upper part of Hole 1017E from ~33 ka (7.445 mbsf) was dated using 13 calendar-corrected radiocarbon ages of mixed planktonic foraminiferal assemblages. Benthic oxygen isotopic stratigraphy records a continuous 130-k.y. sequence ranging from marine isotope Stage 6 to the present day. The benthic d18O curve, representing the last two interglacial and glacial cycles, closely resembles the well-dated, deep-sea reference sequence, providing a detailed chronologic framework. Sedimentation rates remained relatively constant throughout the sequence at ~18 cm/k.y. and were sufficiently rapid to provide considerable potential for high-resolution paleoceanographic/paleoclimatic investigations. Planktonic foraminiferal oxygen isotopic stratigraphy based on the surface-dwelling form Globigerina bulloides defines an almost complete sequence of interstadial/stadial oscillations (Dansgaard/Oeschger cycles [D/O]). Combined use of radiocarbon chronology, deep-sea oxygen isotopic datums, and visual pattern matching has enabled us to identify the sequence of D/O cycles as described for the Greenland (GRIP2) ice core. This has strengthened the stratigraphic framework for the last 60 k.y. in the sequence as a basis for further paleoenvironmental investigations.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In order to document changes in Holocene glacier extent and activity in NE Greenland (~73° N) we study marine sediment records that extend from the fjords (PS2631 and PS2640), across the shelf (PS2623 and PS2641), to the Greenland Sea (JM07-174GC). The primary bedrock geology of the source areas is the Caledonian sediment outcrop, including Devonian red beds, plus early Neoproterozoic gneisses and early Tertiary volcanics. We examine the variations in colour (CIE*), grain size, and bulk mineralogy (from X-ray diffraction of the <2 mm sediment fraction). Fjord core PS2640 in Sofia Sund, with a marked red hue, is distinct in grain size, colour and mineralogy from the other fjord and shelf cores. Five distinct grain-size modes are distinguished of which only one is associated with a coarse ice-rafting signal - this mode is rare in the mid- and late Holocene. A sediment unmixing program (SedUnMixMC) is used to characterize down-core changes in sediment composition based on the upper late Holocene sediments from cores PS2640 (Sofia Sund), PS2631 (Kaiser Franz Joseph Fjord) and PS2623 (south of Shannon Is), and surface samples from the Kara Sea (as an indicator of transport from the Russian Arctic shelves). Major changes in mineral composition are noted in all cores with possible coeval shifts centred c. 2.5, 4.5 and 7.5 cal. ka BP (±0.5 ka) but are rarely linked with changes in the grain-size spectra. Coarse IRD (>2 mm) and IRD-grain-size spectra are rare in the last 9-10 cal. ka BP and, in contrast with areas farther south (~68° N), there is no distinct IRD signal at the onset of neoglaciation. Our paper demonstrates the importance of the quantitative analysis of sediment properties in clarifying source to sink changes in glacial marine environments.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Hydrocarbon seepage is a widespread process at the continental margins of the Gulf of Mexico. We used a multidisciplinary approach, including multibeam mapping and visual seafloor observations with different underwater vehicles to study the extent and character of complex hydrocarbon seepage in the Bay of Campeche, southern Gulf of Mexico. Our observations showed that seafloor asphalt deposits previously only known from the Chapopote Knoll also occur at numerous other knolls and ridges in water depths from 1230 to 3150 m. In particular the deeper sites (Chapopopte and Mictlan knolls) were characterized by asphalt deposits accompanied by extrusion of liquid oil in form of whips or sheets, and in some places (Tsanyao Yang, Mictlan, and Chapopote knolls) by gas emission and the presence of gas hydrates in addition. Molecular and stable carbon isotopic compositions of gaseous hydrocarbons suggest their primarily thermogenic origin. Relatively fresh asphalt structures were settled by chemosynthetic communities including bacterial mats and vestimentiferan tube worms, whereas older flows appeared largely inert and devoid of corals and anemones at the deep sites. The gas hydrates at Tsanyao Yang and Mictlan Knolls were covered by a 5-to-10 cm-thick reaction zone composed of authigenic carbonates, detritus, and microbial mats, and were densely colonized by 1-2 m-long tube worms, bivalves, snails, and shrimps. This study increased knowledge on the occurrences and dimensions of asphalt fields and associated gas hydrates at the Campeche Knolls. The extent of all discovered seepage structure areas indicates that emission of complex hydrocarbons is a widespread, thus important feature of the southern Gulf of Mexico.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

PURPOSE To report acute/subacute vision loss and paracentral scotomata in patients with idiopathic multifocal choroiditis/punctate inner choroidopathy due to large zones of acute photoreceptor attenuation surrounding the chorioretinal lesions. METHODS Multimodal imaging case series. RESULTS Six women and 2 men were included (mean age, 31.5 ± 5.8 years). Vision ranged from 20/20-1 to hand motion (mean, 20/364). Spectral domain optical coherence tomography demonstrated extensive attenuation of the external limiting membrane, ellipsoid and interdigitation zones, adjacent to the visible multifocal choroiditis/punctate inner choroidopathy lesions. The corresponding areas were hyperautofluorescent on fundus autofluorescence and were associated with corresponding visual field defects. Full-field electroretinogram (available in three cases) showed markedly decreased cone/rod response, and multifocal electroretinogram revealed reduced amplitudes and increased implicit times in two cases. Three patients received no treatment, the remaining were treated with oral corticosteroids (n = 4), oral acyclovir/valacyclovir (n = 2), intravitreal/posterior subtenon triamcinolone acetate (n = 3), and anti-vascular endothelial growth factor (n = 2). Visual recovery occurred in only three cases of whom two were treated. Varying morphological recovery was found in six cases, associated with decrease in hyperautofluorescence on fundus autofluorescence. CONCLUSION Multifocal choroiditis/punctate inner choroidopathy can present with transient or permanent central photoreceptor attenuation/loss. This presentation is likely a variant of multifocal choroiditis/punctate inner choroidopathy with chorioretinal atrophy. Associated changes are best evaluated using multimodal imaging.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Recently, broadcasted 3D video content has reached households with the first generation of 3DTV. However, few studies have been done to analyze the Quality of Experience (QoE) perceived by the end-users in this scenario. This paper studies the impact of trans- mission errors in 3DTV, considering that the video is delivered in side-by-side format over a conventional packet-based network. For this purpose, a novel evaluation methodology based on standard sin- gle stimulus methods and with the aim of keeping as close as pos- sible the home environment viewing conditions has been proposed. The effects of packet losses in monoscopic and stereoscopic videos are compared from the results of subjective assessment tests. Other aspects were also measured concerning 3D content as naturalness, sense of presence and visual fatigue. The results show that although the final perceived QoE is acceptable, some errors cause important binocular rivalry, and therefore, substantial visual discomfort.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Los cambios percibidos hacia finales del siglo XX y a principios del nuevo milenio, nos ha mostrado que la crisis cultural de la que somos participes refleja también una crisis de los modelos universales. Nuestra situación contemporánea, parece indicar que ya no es posible formular un sistema estético para atribuirle una vigencia universal e intemporal más allá de su estricta eficacia puntual. La referencia organizada, delimitada, invariable y específica que ofrecía cualquier emplazamiento, en tanto preexistencia, reflejaba una jerarquía del sistema formal basado en lo extensivo: la medida, las normas, el movimiento, el tiempo, la modulación, los códigos y las reglas. Sin embargo, actualmente, algunos aspectos que permanecían latentes sobre lo construido, emergen bajo connotaciones intensivas, transgrediendo la simple manifestación visual y expresiva, para centrase en las propiedades del comportamiento de la materia y la energía como determinantes de un proceso de adaptación en el entorno. A lo largo del todo el siglo XX, el desarrollo de la relación del proyecto sobre lo construido ha sido abordado, casi en exclusiva, entre acciones de preservación o intervención. Ambas perspectivas, manifestaban esfuerzos por articular un pensamiento que diera una consistencia teórica, como soporte para la producción de la acción aditiva. No obstante, en las últimas décadas de finales de siglo, la teoría arquitectónica terminó por incluir pensamientos de otros campos que parecen contaminar la visión sesgada que nos refería lo construido. Todo este entramado conceptual previo, aglomeraba valiosos intentos por dar contenido a una teoría que pudiese ser entendida desde una sola posición argumental. Es así, que en 1979 Ignasi Solá-Morales integró todas las imprecisiones que referían una actuación sobre una arquitectura existente, bajo el termino de “intervención”, el cual fue argumentado en dos sentidos: El primero referido a cualquier tipo de actuación que se puede hacer en un edificio, desde la defensa, preservación, conservación, reutilización, y demás acciones. Se trata de un ámbito donde permanece latente el sentido de intensidad, como factor común de entendimiento de una misma acción. En segundo lugar, más restringido, la idea de intervención se erige como el acto crítico a las ideas anteriores. Ambos representan en definitiva, formas de interpretación de un nuevo discurso. “Una intervención, es tanto como intentar que el edificio vuelva a decir algo o lo diga en una determinada dirección”. A mediados de 1985, motivado por la corriente de revisión historiográfica y la preocupación del deterioro de los centros históricos que recorría toda Europa, Solá-Morales se propone reflexionar sobre “la relación” entre una intervención de nueva arquitectura y la arquitectura previamente existente. Relación condicionada estrictamente bajo consideraciones lingüísticas, a su entender, en sintonía con toda la producción arquitectónica de todo el siglo XX. Del Contraste a la Analogía, resumirá las transformaciones en la concepción discursiva de la intervención arquitectónica, como un fenómeno cambiante en función de los valores culturales, pero a su vez, mostrando una clara tendencia dialógica entres dos categorías formales: El Contraste, enfatizando las posibilidades de la novedad y la diferencia; y por otro lado la emergente Analogía, como una nueva sensibilidad de interpretación del edificio antiguo, donde la semejanza y la diversidad se manifiestan simultáneamente. El aporte reflexivo de los escritos de Solá-Morales podría ser definitivo, si en las últimas décadas antes del fin de siglo, no se hubiesen percibido ciertos cambios sobre la continuidad de la expresión lingüística que fomentaba la arquitectura, hacia una especie de hipertrofia figurativa. Entre muchos argumentos: La disolución de la consistencia compositiva y el estilo unitario, la incorporación volumétrica del proyecto como dispositivo reactivo, y el cambio de visión desde lo retrospectivo hacia lo prospectivo que sugiere la nueva conservación. En este contexto de desintegración, el proyecto, en tanto incorporación o añadido sobre un edificio construido, deja de ser considerado como un apéndice volumétrico subordinado por la reglas compositivas y formales de lo antiguo, para ser considerado como un organismo de orden reactivo, que produce en el soporte existente una alteración en su conformación estructural y sistémica. La extensión, antes espacial, se considera ahora una extensión sensorial y morfológica con la implementación de la tecnología y la hiper-información, pero a su vez, marcados por una fuerte tendencia de optimización energética en su rol operativo, ante el surgimiento del factor ecológico en la producción contemporánea. En una sociedad, como la nuestra, que se está modernizando intensamente, es difícil compartir una adecuada sintonía con las formas del pasado. Desde 1790, fecha de la primera convención francesa para la conservación de monumentos, la escala de lo que se pretende preservar es cada vez más ambiciosa, tanto es así, que al día de hoy el repertorio de lo que se conserva incluye prácticamente todas las tipologías del entorno construido. Para Koolhaas, el intervalo entre el objeto y el momento en el cual se decide su conservación se ha reducido, desde dos milenios en 1882 a unas décadas hoy en día. En breve este lapso desaparecerá, demostrando un cambio radical desde lo retrospectivo hacia lo prospectivo, es decir, que dentro de poco habrá que decidir que es lo que se conserva antes de construir. Solá-Morales, en su momento, distinguió la relación entre lo nuevo y lo antiguo, entre el contraste y la analogía. Hoy casi tres décadas después, el objetivo consiste en evaluar si el modelo de intervención arquitectónica sobre lo construido se ha mantenido desde entonces o si han aparecido nuevas formas de posicionamiento del proyecto sobre lo construido. Nuestro trabajo pretende demostrar el cambio de enfoque proyectual con la preexistencia y que éste tiene estrecha relación con la incorporación de nuevos conceptos, técnicas, herramientas y necesidades que imprimen el contexto cultural, producido por el cambio de siglo. Esta suposición nos orienta a establecer un paralelismo arquitectónico entre los modos de relación en que se manifiesta lo nuevo, entre una posición comúnmente asumida (Tópica), genérica y ortodoxa, fundamentada en lo visual y expresivo de las últimas décadas del siglo XX, y una realidad emergente (Heterotópica), extraordinaria y heterodoxa que estimula lo inmaterial y que parece emerger con creciente intensidad en el siglo XXI. Si a lo largo de todo el siglo XX, el proyecto de intervención arquitectónico, se debatía entre la continuidad y discontinuidad de las categorías formales marcadas por la expresión del edificio preexistente, la nueva intervención contemporánea, como dispositivo reactivo en el paisaje y en el territorio, demanda una absoluta continuidad, ya no visual, expresiva, ni funcional, sino una continuidad fisiológica de adaptación y cambio con la propia dinámica del territorio, bajo nuevas reglas de juego y desplegando planes y estrategias operativas (proyectivas) desde su propia lógica y contingencia. El objeto de esta investigación es determinar los nuevos modos de continuidad y las posibles lógicas de producción que se manifiestan dentro de la Intervención Arquitectónica, intentando superar lo aparente de su relación física y visual, como resultado de la incorporación del factor operativo desplegado por el nuevo dispositivo contemporáneo. Creemos que es acertado mantener la senda connotativa que marca la denominación intervención arquitectónica, por aglutinar conceptos y acercamientos teóricos previos que han ido evolucionando en el tiempo. Si bien el término adolece de mayor alcance operativo desde su formulación, una cualidad que infieren nuestras lógicas contemporáneas, podría ser la reformulación y consolidación de un concepto de intervención más idóneo con nuestros tiempos, anteponiendo un procedimiento lógico desde su propia necesidad y contingencia. Finalmente, nuestro planteamiento inicial aspira a constituir un nueva forma de reflexión que nos permita comprender las complejas implicaciones que infiere la nueva arquitectura sobre la preexistencia, motivada por las incorporación de factores externos al simple juicio formal y expresivo preponderante a finales del siglo XX. Del mismo modo, nuestro camino propuesto, como alternativa, permite proyectar posibles sendas de prospección, al considerar lo preexistente como un ámbito que abarca la totalidad del territorio con dinámicas emergentes de cambio, y con ellas, sus lógicas de intervención.Abstract The perceived changes towards the end of the XXth century and at the beginning of the new milennium have shown us that the cultural crisis in which we participate also reflects a crisis of the universal models. The difference between our contemporary situation and the typical situations of modern orthodoxy and post-modernistic fragmentation, seems to indicate that it is no longer possible to formulate a valid esthetic system, to assign a universal and eternal validity to it beyond its strictly punctual effectiveness; which is even subject to questioning because of the continuous transformations that take place in time and in the sensibility of the subject itself every time it takes over the place. The organised reference that any location offered, limited, invariable and specific, while pre-existing, reflected a hierarchy of the formal system based on the applicable: measure, standards, movement, time, modulation, codes and rules. Authors like Marshall Mc Luhan, Paul Virilio, or Marc Augé anticipated a reality where the conventional system already did not seem to respond to the new architectural requests in which information, speed, disappearance and the virtual had blurred the traditional limits of place; pre-existence did no longer possess a specific delimitation and, on the contrary, they expect to reach a global scale. Currently, some aspects that stayed latent relating to the constructed, surface from intensive connotations, transgressing the simple visual and expressive manifestation in order to focus on the traits of the behaviour of material and energy as determinants of a process of adaptation to the surroundings. Throughout the entire Century, the development of the relation of the project relating to the constructed has been addressed, almost exclusively, in preservational or interventianal actions. Both perspectives showed efforts in order to express a thought that would give a theoretical consistency as a base for the production of the additive action. Nevertheless, the last decades of the Century, architectural theory ended up including thoughts from other fields that seem to contaminate the biased vision 15 which the constructed related us. Ecology, planning, philosophy, global economy, etc, suggest new approaches to the construction of the contemporary city; but this time with a determined idea of change and continuous transformation, that enriches the panorama of thought and architectural practice, at the same time, according to some, it puts disciplinary specification at risk, given that there is no architecture without destruction, the constructed organism requires mutation in order to adjust to the change of shape. All of this previous conceptual framework gathered valuable intents to give importance to a theory that could be understood solely from an argumental position. Thusly, in 1979 Ignasi Solá-Morales integrated all of the imprecisions that referred to an action in existing architecture under the term of “Intervention”, which was explained in two ways: The first referring to any type of intervention that can be carried out in a building, regarding protection, conservation, reuse, etc. It is about a scope where the meaning of intensity stays latent as a common factor of the understanding of a single action. Secondly, more limitedly, the idea of intervention is established as the critical act to the other previous ideas such as restauration, conservation, reuse, etc. Both ultimately represent ways of interpretation of a new speech. “An intervention, is as much as trying to make the building say something again or that it be said in a certain direction”. Mid 1985, motivated by the current of historiographical revision and the concerns regarding the deterioration of historical centres that traversed Europe, Solá-Morales decides to reflect on “the relationship” between an intervention of the new architecture and the previously existing architecture. A relationship determined strictly by linguistic considerations, to his understanding, in harmony with all of the architectural production of the XXth century. From Contrast to Analogy would summarise transformations in the discursive perception of architectural intervention, as a changing phenomenon depending on cultural values, but at the same time, showing a clear dialogical tendency between two formal categories: Contrast, emphasising the possibilities of novelty and difference; and on the other hand the emerging Analogy, as a new awareness of interpretation of the ancient building, where the similarity and diversity are manifested simultaneously. For Solá-Morales the analogical procedure is not based on the visible simultaneity of formal orders, but on associations that the subject establishes throughout time. Through analogy it is tried to overcome the simple visual relationship with the antique, to focus on its spacial, physical and geographical nature. If the analogical attempt guides an opening towards a new continuity; it still persists in the connection of dimensional, typological and figurative factors, subordinate to the formal hierarchy of the preexisting subjects. 16 The reflexive contribution of Solá-Morales’ works could be final, if in the last decades before the end of the century there had not been certain changes regarding linguistic expression, encouraged by architecture, towards a kind of figurative hypertrophy, amongst many arguments we are in this case interested in three moments: The dissolution of the compositional consistency and the united style, the volumetric incorporation of the project as a reactive mechanism, and the change of the vision from retrospective towards prospective that the new conservation suggests. The recurrence to the history of architecture and its recognisable forms, as a way of perpetuating memory and establishing a reference, dissolved any instinct of compositive unity and style, provoking permanent relationships to tend to disappear. The composition and coherence lead to suppose a type of discontinuity of isolated objects in which only possible relationships could appear; no longer as an order of certain formal and compositive rules, but as a special way of setting elements in a specific work. The new globalised field required new forms of consistency between the project and the pre-existent subject, motivated amongst others by the higher pace of market evolution, increase of consumer tax and the level of information and competence between different locations; aspects which finally made stylistic consistence inefficient. In this context of disintegration, the project, in incorporation as well as added to a constructed building, stops being considered as a volumetric appendix subordinate to compositive and formal rules of old, to be considered as an organism of reactive order, that causes a change in the structural and systematic configuration of the existing foundation. The extension, previsouly spatial, is now considered a sensorial and morphological extension, with the implementation of technology and hyper-information, but at the same time, marked by a strong tendency of energetic optimization in its operational role, facing the emergence of the ecological factor in contemporary production. The technological world turns into a new nature, a nature that should be analysed from ecological terms; in other words, as an event of transition in the continuous redistribution of energy. In this area, effectiveness is not only determined by the capacity of adaptation to changing conditions, but also by its transforming capacity “expressly” in order to change an environment. In a society, like ours, that is modernising intensively, it is difficult to share an adecuate agreement with the forms of the past. From 1790, the date of the first French convention for the conservation of monuments, the scale of what is expexted to be preserved is more and more ambitious, so much so that nowadays the repertoire of that what is conserved includes practically all typologies of the constructed surroundings. For Koolhaas, the ínterval between the object and the moment when its conservation is decided has been reduced, from two 17 milennia in 1882 to a few decades nowadays. Shortly this lapse will disappear, showing a radical change of retrospective towards prospective, that is to say, that soon it will be necessary to decide what to conserve before constructing. The shapes of cities are the result of the continuous incorporation of architecture, and perhaps that only through architecture the response to the universe can be understood, the continuity of what has already been constructed. Our work is understood also within that system, modifying the field of action and leaving the road ready for the next movement of those that will follow after us. Continuity does not mean conservatism, continuity means being conscient of the transitory value of our answers to specific needs, accepting the change that we have received. That what has been constructed to remain and last, should cause future interventions to be integrated in it. It is necessary to accept continuity as a rule. Solá-Morales, in his time, distinguished between the relationship with new and old, between contrast and analogy. Today, almost three decades later, the objective consists of evaluating whether the model of architectural intervention in the constructed has been maintained since then or if new ways of positioning the project regarding the constructed have appeared. Our work claims to show the change of the approach of projects with pre-existing subjects and that this has got a close relation to the incorporation of new concepts, techniques, tools and necessities that impress the cultural context, caused by the change of centuries. This assumption guides us to establish a parallelism between the forms of connection where that what is new is manifested between a commonly assumed (topical), generic and orthodox position, based on that what is visual and expressive in the last decades of the XXth century, and an emerging (heterotopical), extraordinary and heterodox reality that stimulates the immaterial and that seems to emerge with growing intensity in the XXIst century. If throughout the XXth century the project of architectural intervention was considered from the continuity and discontinuity of formal categories, marked by the expression of the pre-existing building, the new contemporary intervention, as a reactive device in the landscape and territory, demands an absolute continuity. No longer a visual, expressive or functional one but a morphological continuity of adaptation and change with its own territorial dynamics, under new game rules and unfolding new operative (projective) strategies from its own logic and contingency. 18 The aim of this research is to determine new forms of continuity and the possible logic of production that are expressed in the Architectural Intervention, trying to overcome the obviousness of its physical and visual relationship, at the beginning of this new century, as a result of the incorporation of the operative factor that the new architectural device unfolds. We think it is correct to maintain the connotative path that marks the name architectural intervention by bringing previous concepts and theorical approaches that have been evolving through time together. If the name suffers from a wider operational range because of its formulation, a quality that our contemporary logic provokes, the reformulation and consolidation of an interventional concept could be more suitable for our times, giving preference to a logical method from its own necessity and contingency. It seems that now time shapes the topics, it is no longer about materialising a certain time but about expressing the changes that its new temporality generates. Finally, our initial approach aspires to form a new way of reflection that permits us to understand the complex implications that the new architecture submits the pre-existing subject to, motivated by the incorporation of factors external to simple formal and expressive judgement, prevailing at the end of the XXth century. In the same way, our set road, as an alternative, permits the contemplation of possible research paths, considering that what is pre-existing as an area that spans the whole territory with emerging changing dynamics and, with them, their interventional logics.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

There is clear evidence that investment in intelligent transportation system technologies brings major social and economic benefits. Technological advances in the area of automatic systems in particular are becoming vital for the reduction of road deaths. We here describe our approach to automation of one the riskiest autonomous manœuvres involving vehicles – overtaking. The approach is based on a stereo vision system responsible for detecting any preceding vehicle and triggering the autonomous overtaking manœuvre. To this end, a fuzzy-logic based controller was developed to emulate how humans overtake. Its input is information from the vision system and from a positioning-based system consisting of a differential global positioning system (DGPS) and an inertial measurement unit (IMU). Its output is the generation of action on the vehicle’s actuators, i.e., the steering wheel and throttle and brake pedals. The system has been incorporated into a commercial Citroën car and tested on the private driving circuit at the facilities of our research center, CAR, with different preceding vehicles – a motorbike, car, and truck – with encouraging results.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

El diseño y desarrollo de sistemas de suspensión para vehículos se basa cada día más en el diseño por ordenador y en herramientas de análisis por ordenador, las cuales permiten anticipar problemas y resolverlos por adelantado. El comportamiento y las características dinámicas se calculan con precisión, bajo coste, y recursos y tiempos de cálculo reducidos. Sin embargo, existe una componente iterativa en el proceso, que requiere la definición manual de diseños a través de técnicas “prueba y error”. Esta Tesis da un paso hacia el desarrollo de un entorno de simulación eficiente capaz de simular, analizar y evaluar diseños de suspensiones vehiculares, y de mejorarlos hacia la solución optima mediante la modificación de los parámetros de diseño. La modelización mediante sistemas multicuerpo se utiliza aquí para desarrollar un modelo de autocar con 18 grados de libertad, de manera detallada y eficiente. La geometría y demás características de la suspensión se ajustan a las del vehículo real, así como los demás parámetros del modelo. Para simular la dinámica vehicular, se utiliza una formulación multicuerpo moderna y eficiente basada en las ecuaciones de Maggi, a la que se ha incorporado un visor 3D. Así, se consigue simular maniobras vehiculares en tiempos inferiores al tiempo real. Una vez que la dinámica está disponible, los análisis de sensibilidad son cruciales para una optimización robusta y eficiente. Para ello, se presenta una técnica matemática que permite derivar las variables dinámicas dentro de la formulación, de forma algorítmica, general, con la precisión de la maquina, y razonablemente eficiente: la diferenciación automática. Este método propaga las derivadas con respecto a las variables de diseño a través del código informático y con poca intervención del usuario. En contraste con otros enfoques en la bibliografía, generalmente particulares y limitados, se realiza una comparación de librerías, se desarrolla una formulación híbrida directa-automática para el cálculo de sensibilidades, y se presentan varios ejemplos reales. Finalmente, se lleva a cabo la optimización de la respuesta dinámica del vehículo citado. Se analizan cuatro tipos distintos de optimización: identificación de parámetros, optimización de la maniobrabilidad, optimización del confort y optimización multi-objetivo, todos ellos aplicados al diseño del autocar. Además de resultados analíticos y gráficos, se incluyen algunas consideraciones acerca de la eficiencia. En resumen, se mejora el comportamiento dinámico de vehículos por medio de modelos multicuerpo y de técnicas de diferenciación automática y optimización avanzadas, posibilitando un ajuste automático, preciso y eficiente de los parámetros de diseño. ABSTRACT Each day, the design and development of vehicle suspension systems relies more on computer-aided design and computer-aided engineering tools, which allow anticipating the problems and solving them ahead of time. Dynamic behavior and characteristics are thus simulated accurately and inexpensively with moderate computational times and resources. There is, however, an iterative component in the process, which involves the manual definition of designs in a trialand-error manner. This Thesis takes a step towards the development of an efficient simulation framework capable of simulating, analyzing and evaluating vehicle suspension designs, and automatically improving them by varying the design parameters towards the optimal solution. The multibody systems approach is hereby used to model a three-dimensional 18-degrees-of-freedom coach in a comprehensive yet efficient way. The suspension geometry and characteristics resemble the ones from the real vehicle, as do the rest of vehicle parameters. In order to simulate vehicle dynamics, an efficient, state-of-the-art multibody formulation based on Maggi’s equations is employed, and a three-dimensional graphics viewer is developed. As a result, vehicle maneuvers can be simulated faster than real-time. Once the dynamics are ready, a sensitivity analysis is crucial for a robust optimization. To that end, a mathematical technique is introduced, which allows differentiating the dynamic variables within the multibody formulation in a general, algorithmic, accurate to machine precision, and reasonably efficient way: automatic differentiation. This method propagates the derivatives with respect to the design parameters throughout the computer code, with little user interaction. In contrast with other attempts in the literature, mostly not generalpurpose, a benchmarking of libraries is carried out, a hybrid direct-automatic differentiation approach for the computation of sensitivities is developed, and several real-life examples are analyzed. Finally, a design optimization process of the aforementioned vehicle is carried out. Four different types of dynamic response optimization are presented: parameter identification, handling optimization, ride comfort optimization and multi-objective optimization; all of which are applied to the design of the coach example. Together with analytical and visual proof of the results, efficiency considerations are made. In summary, the dynamic behavior of vehicles is improved by using the multibody systems approach, along with advanced differentiation and optimization techniques, enabling an automatic, accurate and efficient tuning of design parameters.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In developing instrumentation for the measurement of fruit quality, there is the need for fast and non-destructive devices, based on sensors, to be installed on-line. In the case of some fruits, like peaches, post-harvest ripeness, which is closely related to high quality for the consumer, is a priority. During ripening, external appearance (colour) and internal mechanical (firmness) and chemical (sugars and acids) quality are main features that evolve rapidly from and unripe to a ripe (high quality) stage. When considering the evolution of fruit quality in this scheme, external colour and firmness are shown to evolve in a parallel pattern, if monitored from the time of harvest to full consumer ripeness ( Rood, 1957; Crisosto et al, 1995; Kader, 1996). The visible (VIS) reflectance spectrum is a fast and easy reference that can be used to estimate quality of peaches, if we could show it to be reliably correlated with peach ripening rate during postharvest (Genard et al. 1994; Moras, 1995; Delwiche and Baumgartner, 1983; Delwiche et al. 1987; Slaughter, 1995; Lleo et al., 1998). Taste, described as an expert acceptance score, improves with ripeness (firmness and colour evolution), when considering the fruits on the tree, and also post-harvest.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In recent years, remote sensing imaging systems for the measurement of oceanic sea states have attracted renovated attention. Imaging technology is economical, non-invasive and enables a better understanding of the space-time dynamics of ocean waves over an area rather than at selected point locations of previous monitoring methods (buoys, wave gauges, etc.). We present recent progress in space-time measurement of ocean waves using stereo vision systems on offshore platforms, which focus on sea states with wavelengths in the range of 0.01 m to 1 m. Both traditional disparity-based systems and modern elevation-based ones are presented in a variational optimization framework: the main idea is to pose the stereoscopic reconstruction problem of the surface of the ocean in a variational setting and design an energy functional whose minimizer is the desired temporal sequence of wave heights. The functional combines photometric observations as well as spatial and temporal smoothness priors. Disparity methods estimate the disparity between images as an intermediate step toward retrieving the depth of the waves with respect to the cameras, whereas elevation methods estimate the ocean surface displacements directly in 3-D space. Both techniques are used to measure ocean waves from real data collected at offshore platforms in the Black Sea (Crimean Peninsula, Ukraine) and the Northern Adriatic Sea (Venice coast, Italy). Then, the statistical and spectral properties of the resulting observed waves are analyzed. We show the advantages and disadvantages of the presented stereo vision systems and discuss future lines of research to improve their performance in critical issues such as the robustness of the camera calibration in spite of undesired variations of the camera parameters or the processing time that it takes to retrieve ocean wave measurements from the stereo videos, which are very large datasets that need to be processed efficiently to be of practical usage. Multiresolution and short-time approaches would improve efficiency and scalability of the techniques so that wave displacements are obtained in feasible times.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Remote sensing imaging systems for the measurement of oceanic sea states have recently attracted renovated attention. Imaging technology is economical, non-invasive and enables a better understanding of the space-time dynamics of ocean waves over an area rather than at selected point locations of previous monitoring methods (buoys, wave gauges, etc.). We present recent progress in space-time measurement of ocean waves using stereo vision systems on offshore platforms. Both traditional disparity-based systems and modern elevation-based ones are presented in a variational optimization framework: the main idea is to pose the stereoscopic reconstruction problem of the surface of the ocean in a variational setting and design an energy functional whose minimizer is the desired temporal sequence of wave heights. The functional combines photometric observations as well as spatial and temporal smoothness priors. Disparity methods estimate the disparity between images as an intermediate step toward retrieving the depth of the waves with respect to the cameras, whereas elevation methods estimate the ocean surface displacements directly in 3-D space. Both techniques are used to measure ocean waves from real data collected at offshore platforms in the Black Sea (Crimean Peninsula, Ukraine) and the Northern Adriatic Sea (Venice coast, Italy). Then, the statistical and spectral properties of the resulting observed waves are analyzed. We show the advantages and disadvantages of the presented stereo vision systems and discuss the improvement of their performance in critical issues such as the robustness of the camera calibration in spite of undesired variations of the camera parameters.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Crystallization and grain growth technique of thin film silicon are among the most promising methods for improving efficiency and lowering cost of solar cells. A major advantage of laser crystallization and annealing over conventional heating methods is its ability to limit rapid heating and cooling to thin surface layers. Laser energy is used to heat the amorphous silicon thin film, melting it and changing the microstructure to polycrystalline silicon (poly-Si) as it cools. Depending on the laser density, the vaporization temperature can be reached at the center of the irradiated area. In these cases ablation effects are expected and the annealing process becomes ineffective. The heating process in the a-Si thin film is governed by the general heat transfer equation. The two dimensional non-linear heat transfer equation with a moving heat source is solve numerically using the finite element method (FEM), particularly COMSOL Multiphysics. The numerical model help to establish the density and the process speed range needed to assure the melting and crystallization without damage or ablation of the silicon surface. The samples of a-Si obtained by physical vapour deposition were irradiated with a cw-green laser source (Millennia Prime from Newport-Spectra) that delivers up to 15 W of average power. The morphology of the irradiated area was characterized by confocal laser scanning microscopy (Leica DCM3D) and Scanning Electron Microscopy (SEM Hitachi 3000N). The structural properties were studied by micro-Raman spectroscopy (Renishaw, inVia Raman microscope).

Relevância:

100.00% 100.00%

Publicador:

Resumo:

El interés por los sistemas fotovoltaicos de concentración (CPV) ha resurgido en los últimos años amparado por el desarrollo de células multiunión de muy alta eficiencia basadas en semiconductores de los grupos III-V. Estas células han permitido obtener módulos de concentración con eficiencias que prácticamente duplican las del panel plano y que llegan al 35% en los módulos récord. Esta tesis está dedicada al diseño y la implementación experimental de nuevos conceptos que permitan obtener módulos CPV que no sólo alcancen una eficiencia alta en condiciones estándar sino que, además, sean lo suficientemente tolerantes a errores de montaje, seguimiento, temperatura y variaciones espectrales para que la energía que producen a lo largo del año sea máxima. Una de las primeras cuestiones que se abordan es el diseño de elementos ópticos secundarios para sistemas cuyo primario es una lente de Fresnel y que permiten, para una concentración fija, aumentar el ángulo de aceptancia y la tolerancia del sistema. Varios secundarios reflexivos y refractivos han sido diseñados y analizados mediante trazado de rayos. En particular, utilizando óptica anidólica y basándose en el diseño de una sola etapa conocido como ‘concentrador dieléctrico que funciona por reflexión total interna‘, se ha diseñado, fabricado y caracterizado un secundario con salida cuadrada que, usado junto con una lente de Fresnel, permite alcanzar simultáneamente una elevada eficiencia, concentración y aceptancia. Además, se ha propuesto y prototipado un método alternativo de fabricación para otro de los secundarios, denominado domo, consistente en el sobremoldeo de silicona sobre células solares. Una de las características que impregna todo el trabajo realizado en esta tesis es la aproximación holística en el diseño de módulos CPV, es decir, se ha prestado especial atención al diseño conjunto de la célula y la óptica para garantizar que el sistema total alcance la mayor eficiencia posible. En este sentido muchos sistemas ópticos desarrollados en esta tesis han sido diseñados, caracterizados y optimizados teniendo en cuenta que el ajuste de corriente entre las distintas subcélulas que comprenden la célula multiunión bajo el concentrador sea muy próximo a uno. La capa antirreflectante sobre la célula funciona, en cierto modo, como interfaz entre la óptica y la célula, por lo que se ha diseñado un método de optimización de capas antirreflectantes que considera no sólo el amplio rango de longitudes de onda para el que las células multiunión son sensibles sino también la distribución angular de intensidad sobre la célula creada por la óptica de concentración. Además, la cuestión de la falta de uniformidad también se ha abordado mediante la comparación de las distribuciones espectrales y espaciales de irradiancia que crean diferentes ópticas (simuladas mediante trazado de rayos y fotografiadas) y las pérdidas de eficiencia que experimentan las células iluminadas por dichas ópticas de concentración medidas experimentalmente. El efecto de la temperatura en la óptica de concentración también ha sido objeto de estudio de esta tesis. En particular, mediante simulaciones de elementos finitos se han dado los primeros pasos para el análisis de las deformaciones que sufren los dientes de las lentes de Fresnel híbridas (vidrio-silicona), así como el cambio de índice de refracción con la temperatura y la influencia de ambos efectos sobre el funcionamiento de los sistemas. Se ha implementado un modelo que tiene por objeto considerar las variaciones ambientales, principalmente temperatura y contenido espectral de la radiación directa, así como las sensibilidades térmica y espectral de los sistemas CPV, con el fin de maximizar la energía producida por un módulo de concentración a lo largo de un año en un emplazamiento determinado. Los capítulos 5 y 6 de este libro están dedicados al diseño, fabricación y caracterización de un nuevo concepto de módulo fotovoltaico denominado FluidReflex y basado en una única etapa reflexiva con dieléctrico fluido. En este nuevo concepto la presencia del fluido aporta algunas ventajas significativas como son: un aumento del producto concentración por aceptancia (CAP, en sus siglas en inglés) alcanzable al rodear la célula con un medio cuyo índice de refracción es mayor que uno, una mejora de la eficiencia óptica al disminuir las pérdidas por reflexión de Fresnel en varias interfaces, una mejora de la disipación térmica ya que el calor que se concentra junto a la célula se trasmite por convección natural y conducción en el fluido y un aislamiento eléctrico mejorado. Mediante la construcción y medida de varios prototipos de unidad elemental se ha demostrado que no existe ninguna razón fundamental que impida la implementación práctica del concepto teórico alcanzando una elevada eficiencia. Se ha realizado un análisis de fluidos candidatos probando la existencia de al menos dos de ellos que cumplen todos los requisitos (en particular el de estabilidad bajo condiciones de luz concentrada) para formar parte del sistema de concentración FluidReflex. Por ´ultimo, se han diseñado, fabricado y caracterizado varios prototipos preindustriales de módulos FluidReflex para lo cual ha sido necesario optimizar el proceso de fabricación de la óptica multicavidad a fin de mantener el buen comportamiento óptico obtenido en la fabricación de la unidad elemental. Los distintos prototipos han sido medidos, tanto en el laboratorio como bajo el sol real, analizando el ajuste de corriente de la célula iluminada por el concentrador FluidReflex bajo diferentes distribuciones espectrales de la radiación incidente así como el excelente comportamiento térmico del módulo. ABSTRACT A renewed interest in concentrating photovoltaic (CPV) systems has emerged in recent years encouraged by the development of high-efficiency multijunction solar cells based in IIIV semiconductors that have led to CPV module efficiencies which practically double that of flat panel PV and which reach 35% for record modules. This thesis is devoted to the design and experimental implementation of new concepts for obtaining CPV modules that not only achieve high efficiency under standard conditions but also have such a wide tolerance to assembly errors, tracking, temperature and spectral variations, that the energy generated by them throughout the year is maximized. One of the first addressed issues is the design of secondary optical elements whose primary optics is a Fresnel lens and which, for a fixed concentration, allow an increased acceptance angle and tolerance of the system. Several reflective and refractive secondaries have been designed and analyzed using ray tracing. In particular, using nonimaging optics and based on the single-stage design known as ‘dielectric totally internally reflecting concentrator’, a secondary with square output has been designed, fabricated and characterized. Used together with a Fresnel lens, the secondary can simultaneously achieve high efficiency, concentration and acceptance. Furthermore, an alternative method has been proposed and prototyped for the fabrication of the secondary named dome. The optics is manufactured by direct overmolding of silicone over the solar cells. One characteristic that permeates all the work done in this thesis is the holistic approach in the design of CPV modules, meaning that special attention has been paid to the joint design of the solar cell and the optics to ensure that the total system achieves the highest attainable efficiency. In this regard, many optical systems developed in the thesis have been designed, characterized and optimized considering that the current matching among the subcells within the multijunction solar cell beneath the optics must be close to one. Antireflective coating over the cell acts, somehow, as an interface between the optics and the cell. Consequently, a method has been designed to optimize antireflective coatings that takes into account not only the broad wavelength range that multijunction solar cells are sensitive to but also the angular intensity distribution created by the concentrating optics. In addition, the issue of non-uniformity has also been addressed by comparing the spectral and spatial distributions of irradiance created by different optics (simulated by ray tracing and photographed) and the efficiency losses experienced by cells illuminated by those concentrating optics experimentally determined. The effect of temperature on the concentrating optics has also been studied in this thesis. In particular, finite element simulations have been use to analyze the deformations experienced by the facets of hybrid (silicon-glass) Fresnel lenses, the change of refractive index with temperature and the influence of both effects on the system performance. A model has been implemented which take into consideration atmospheric variations, mainly temperature and spectral content of the direct normal irradiance, as well as thermal and spectral sensitivities of systems, with the aim of maximizing the energy harvested by a CPV module throughout the year in a particular location. Chapters 5 and 6 of this book are devoted to the design, fabrication, and characterization of a new concentrator concept named FluidReflex and based on a single-stage reflective optics with fluid dielectric. In this new concept, the presence of the fluid provides some significant advantages such as: an increased concentration acceptance angle product (CAP) achievable by surrounding the cell with a medium whose refractive index is greater than one, an improvement of the optical efficiency by reducing losses due to Fresnel reflection at several interfaces, an improvement in heat dissipation as the heat concentrated near the cell is transmitted by natural convection and conduction in the fluid, and an improved electrical insulation. By fabricating and characterizing several elementary-unit prototypes it was shown that there is no fundamental reason that prevents the practical implementation of this theoretical concept reaching high efficiency. Several fluid candidates were investigated proving the existence of at least to fluids that meet all the requirements (including the stability under concentrated light) to become part of the FluidReflex concentrator. Finally, several pre-industrial FluidReflex module prototypes have been designed and fabricated. An optimization process for the manufacturing of the multicavity optics was necessary to attain such an optics quality as the one achieved by the single unit. The module prototypes have been measured, both indoors and outdoors, analyzing the current matching of the solar cells beneath the concentrator for different spectral distribution of the incident irradiance. Additionally, the module showed an excellent thermal performance.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this paper we propose an innovative approach to tackle the problem of traffic sign detection using a computer vision algorithm and taking into account real-time operation constraints, trying to establish intelligent strategies to simplify as much as possible the algorithm complexity and to speed up the process. Firstly, a set of candidates is generated according to a color segmentation stage, followed by a region analysis strategy, where spatial characteristic of previously detected objects are taken into account. Finally, temporal coherence is introduced by means of a tracking scheme, performed using a Kalman filter for each potential candidate. Taking into consideration time constraints, efficiency is achieved two-fold: on the one side, a multi-resolution strategy is adopted for segmentation, where global operation will be applied only to low-resolution images, increasing the resolution to the maximum only when a potential road sign is being tracked. On the other side, we take advantage of the expected spacing between traffic signs. Namely, the tracking of objects of interest allows to generate inhibition areas, which are those ones where no new traffic signs are expected to appear due to the existence of a TS in the neighborhood. The proposed solution has been tested with real sequences in both urban areas and highways, and proved to achieve higher computational efficiency, especially as a result of the multi-resolution approach.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Esta tesis se ha desarrollado en el contexto del proyecto Cajal Blue Brain, una iniciativa europea dedicada al estudio del cerebro. Uno de los objetivos de esta iniciativa es desarrollar nuevos métodos y nuevas tecnologías que simplifiquen el análisis de datos en el campo neurocientífico. El presente trabajo se ha centrado en diseñar herramientas que combinen información proveniente de distintos canales sensoriales con el fin de acelerar la interacción y análisis de imágenes neurocientíficas. En concreto se estudiará la posibilidad de combinar información visual con información háptica. Las espinas dendríticas son pequeñas protuberancias que recubren la superficie dendrítica de muchas neuronas del cerebro. A día de hoy, se cree que tienen un papel clave en la transmisión de señales neuronales. Motivo por el cual, el interés por parte de la comunidad científica por estas estructuras ha ido en aumento a medida que las técnicas de adquisición de imágenes mejoraban hasta alcanzar una calidad suficiente para analizar dichas estructuras. A menudo, los neurocientíficos utilizan técnicas de microscopía con luz para obtener los datos que les permitan analizar estructuras neuronales tales como neuronas, dendritas y espinas dendríticas. A pesar de que estas técnicas ofrezcan ciertas ventajas frente a su equivalente electrónico, las técnicas basadas en luz permiten una menor resolución. En particular, estructuras pequeñas como las espinas dendríticas pueden capturarse de forma incorrecta en las imágenes obtenidas, impidiendo su análisis. En este trabajo, se presenta una nueva técnica, que permite editar imágenes volumétricas, mediante un dispositivo háptico, con el fin de reconstruir de los cuellos de las espinas dendríticas. Con este objetivo, en un primer momento se desarrolló un algoritmo que proporciona retroalimentación háptica en datos volumétricos, completando la información que provine del canal visual. Dicho algoritmo de renderizado háptico permite a los usuarios tocar y percibir una isosuperficie en el volumen de datos. El algoritmo asegura un renderizado robusto y eficiente. Se utiliza un método basado en las técnicas de “marching tetrahedra” para la extracción local de una isosuperficie continua, lineal y definida por intervalos. La robustez deriva tanto de una etapa de detección de colisiones continua de la isosuperficie extraída, como del uso de técnicas eficientes de renderizado basadas en un proxy puntual. El método de “marching tetrahedra” propuesto garantiza que la topología de la isosuperficie extraída coincida con la topología de una isosuperficie equivalente determinada utilizando una interpolación trilineal. Además, con el objetivo de mejorar la coherencia entre la información háptica y la información visual, el algoritmo de renderizado háptico calcula un segundo proxy en la isosuperficie pintada en la pantalla. En este trabajo se demuestra experimentalmente las mejoras en, primero, la etapa de extracción de isosuperficie, segundo, la robustez a la hora de mantener el proxy en la isosuperficie deseada y finalmente la eficiencia del algoritmo. En segundo lugar, a partir del algoritmo de renderizado háptico propuesto, se desarrolló un procedimiento, en cuatro etapas, para la reconstrucción de espinas dendríticas. Este procedimiento, se puede integrar en los cauces de segmentación automática y semiautomática existentes como una etapa de pre-proceso previa. El procedimiento está diseñando para que tanto la navegación como el proceso de edición en sí mismo estén controlados utilizando un dispositivo háptico. Se han diseñado dos experimentos para evaluar esta técnica. El primero evalúa la aportación de la retroalimentación háptica y el segundo se centra en evaluar la idoneidad del uso de un háptico como dispositivo de entrada. En ambos casos, los resultados demuestran que nuestro procedimiento mejora la precisión de la reconstrucción. En este trabajo se describen también dos casos de uso de nuestro procedimiento en el ámbito de la neurociencia: el primero aplicado a neuronas situadas en la corteza cerebral humana y el segundo aplicado a espinas dendríticas situadas a lo largo de neuronas piramidales de la corteza del cerebro de una rata. Por último, presentamos el programa, Neuro Haptic Editor, desarrollado a lo largo de esta tesis junto con los diferentes algoritmos ya mencionados. ABSTRACT This thesis took place within the Cajal Blue Brain project, a European initiative dedicated to the study of the brain. One of the main goals of this project is the development of new methods and technologies simplifying data analysis in neuroscience. This thesis focused on the development of tools combining information originating from distinct sensory channels with the aim of accelerating both the interaction with neuroscience images and their analysis. In concrete terms, the objective is to study the possibility of combining visual information with haptic information. Dendritic spines are thin protrusions that cover the dendritic surface of numerous neurons in the brain and whose function seems to play a key role in neural circuits. The interest of the neuroscience community toward those structures kept increasing as and when acquisition methods improved, eventually to the point that the produced datasets enabled their analysis. Quite often, neuroscientists use light microscopy techniques to produce the dataset that will allow them to analyse neuronal structures such as neurons, dendrites and dendritic spines. While offering some advantages compared to their electronic counterpart, light microscopy techniques achieve lower resolutions. Particularly, small structures such as dendritic spines might suffer from a very low level of fluorescence in the final dataset, preventing further analysis. This thesis introduces a new technique enabling the edition of volumetric datasets in order to recreate dendritic spine necks using a haptic device. In order to fulfil this objective, we first presented an algorithm to provide haptic feedback directly from volumetric datasets, as an aid to regular visualization. The haptic rendering algorithm lets users perceive isosurfaces in volumetric datasets, and it relies on several design features that ensure a robust and efficient rendering. A marching tetrahedra approach enables the dynamic extraction of a piecewise linear continuous isosurface. Robustness is derived using a Continuous Collision Detection step coupled with acknowledged proxy-based rendering methods over the extracted isosurface. The introduced marching tetrahedra approach guarantees that the extracted isosurface will match the topology of an equivalent isosurface computed using trilinear interpolation. The proposed haptic rendering algorithm improves the coherence between haptic and visual cues computing a second proxy on the isosurface displayed on screen. Three experiments demonstrate the improvements on the isosurface extraction stage as well as the robustness and the efficiency of the complete algorithm. We then introduce our four-steps procedure for the complete reconstruction of dendritic spines. Based on our haptic rendering algorithm, this procedure is intended to work as an image processing stage before the automatic segmentation step giving the final representation of the dendritic spines. The procedure is designed to allow both the navigation and the volume image editing to be carried out using a haptic device. We evaluated our procedure through two experiments. The first experiment concerns the benefits of the force feedback and the second checks the suitability of the use of a haptic device as input. In both cases, the results shows that the procedure improves the editing accuracy. We also report two concrete cases where our procedure was employed in the neuroscience field, the first one concerning dendritic spines in the human cortex, the second one referring to an ongoing experiment studying dendritic spines along dendrites of mouse cortical pyramidal neurons. Finally, we present the software program, Neuro Haptic Editor, that was built along the development of the different algorithms implemented during this thesis, and used by neuroscientists to use our procedure.