7 resultados para Visual studio
em Universidad Politécnica de Madrid
Resumo:
The integration of correlation processes in design systems has as a target measurements in 3D directly and according to the users criteria in order to generate the required database for the development of the project. In the phase of photogrammetric works, internal and external orientation parameters are calculated and stereo models are created from standard images. The aforementioned are integrated in the system where the measurement of the selected items is done by applying developed correlation algorithms. The processing period has the tools to carry out the calculations in an easy and automatic way, as well as image measurement techniques to acquire the most correct information. The proposed software development is done on Visual Studio platforms for PC, applying the most apt codes and symbols according to the terms of reference required for the design. The results of generating the data base in an interactive way with the geometric study of the structures, facilitates and improves the quality of the works in the projects.
Resumo:
El presente proyecto fin de carrera, realizado por el ingeniero técnico en telecomunicaciones Pedro M. Matamala Lucas, es la fase final de desarrollo de un proyecto de mayor magnitud correspondiente al software de vídeo forense SAVID. El propósito del proyecto en su totalidad es la creación de una herramienta informática capacitada para realizar el análisis de ficheros de vídeo, codificados y comprimidos por el sistema DV –Digital Video-. El objetivo del análisis, es aportar información acerca de si la cinta magnética presenta indicios de haber sido manipulada con una edición posterior a su grabación original, además, de mostrar al usuario otros datos de interés como las especificaciones técnicas de la señal de vídeo y audio. Por lo tanto, se facilitará al usuario, analista de vídeo forense, información que le ayude a valorar la originalidad del contenido del soporte que es sujeto del análisis. El objetivo específico de esta fase final, es la creación de la interfaz de usuario del software, que informa tanto del código binario de los sectores significativos, como de su interpretación tras el análisis. También permitirá al usuario el reporte de los resultados, además de otras funcionalidades que le permitan la navegación por los sectores del código que han sido modificados como efecto colateral de la edición de la cinta magnética original. Otro objetivo importante del proyecto ha sido la investigación de metodologías y técnicas de desarrollo de software para su posterior implementación, buscando con esto, una mayor eficiencia en la gestión del tiempo y una mayor calidad de software con el fin de garantizar su evolución y sostenibilidad en el futuro. Se ha hecho hincapié en las metodologías ágiles que han ido ganando relevancia en el sector de las tecnologías de la información en las últimas décadas, sustituyendo a metodologías clásicas como el desarrollo en cascada. Su flexibilidad durante el ciclo de vida del software, permite obtener mejores resultados cuando las especificaciones no están del todo definidas, ajustándose de este modo a las condiciones del proyecto. Resumiendo las especificaciones técnicas del software, C++ es el lenguaje de programación orientado a objetos con el que se ha desarrollado, utilizándose la tecnología MFC -Microsoft Foundation Classes- para la implementación. Es un proyecto MFC de tipo cuadro de dialogo,creado, compilado y publicado, con la herramienta de desarrollo integrado Microsoft Visual Studio 2010. La arquitectura con la que se ha estructurado es la arquetípica de tres capas, compuesta por la interfaz de usuario, capa de negocio y capa de acceso a datos. Se ha visto necesario configurar el proyecto con compatibilidad con CLR –Common Languages Runtime- para poder implementar la funcionalidad de creación de reportes. Acompañando a la aplicación informática, se presenta la memoria del proyecto y sus anexos correspondientes a los documentos EDRF –Especificaciones Detalladas de Requisitos funcionales-, EIU –Especificaciones de Interfaz de Usuario , DT -Diseño Técnico- y Guía de Usuario. SUMMARY. This dissertation, carried out by the telecommunications engineer Pedro M. Matamala Lucas, is in its final stage and is part of a larger project for the software of forensic video called SAVID. The purpose of the entire project is the creation of a software tool capable of analyzing video files that are coded and compressed by the DV -Digital Video- System. The objective of the analysis is to provide information on whether the magnetic tape shows signs of having been tampered with after the editing of the original recording, and also to show the user other relevant data and technical specifications of the video signal and audio. Therefore the user, forensic video analyst, will have information to help assess the originality of the content of the media that is subject to analysis. The specific objective of this final phase is the creation of the user interface of the software that provides information about the binary code of the significant sectors and also its interpretation after analysis. It will also allow the user to report the results, and other features that will allow browsing through the sections of the code that have been modified as a secondary effect of the original magnetic tape being tampered. Another important objective of the project is the investigation of methodologies and software development techniques to be used in deployment, with the aim of greater efficiency in time management and enhanced software quality in order to ensure its development and maintenance in the future. Agile methodologies, which have become important in the field of information technology in recent decades, have been used during the execution of the project, replacing classical methodologies such as Waterfall Development. The flexibility, as the result of using by agile methodologies, during the software life cycle, produces better results when the specifications are not fully defined, thus conforming to the initial conditions of the project. Summarizing the software technical specifications, C + + the programming language – which is object oriented and has been developed using technology MFC- Microsoft Foundation Classes for implementation. It is a project type dialog box, created, compiled and released with the integrated development tool Microsoft Visual Studio 2010. The architecture is structured in three layers: the user interface, business layer and data access layer. It has been necessary to configure the project with the support CLR -Common Languages Runtime – in order to implement the reporting functionality. The software application is submitted with the project report and its annexes to the following documents: Functional Requirements Specifications - Detailed User Interface Specifications, Technical Design and User Guide.
Resumo:
En este proyecto se trata la simulación numérica de un fenómeno dinámico, basado en el comportamiento de una onda transmitida a lo largo de una cuerda elástica de un instrumento musical, cuyos extremos se encuentran anclados. El fenómeno físico, se desarrolla utilizando una ecuación en derivadas parciales hiperbólicas con variables espacial y temporal, acompañada por unas condiciones de contorno tipo Dirichlet en los extremos y por más condiciones iniciales que dan comienzo al proceso. Posteriormente se han generado algoritmos para el método numérico empleado (Diferencias finitas centrales y progresivas) y la programación del problema aproximado con su consistencia, estabilidad y convergencia, obteniéndose unos resultados acordes con la solución analítica del problema matemático. La programación y salida de resultados se ha realizado con Visual Studio 8.0. y la programación de objetos con Visual Basic .Net In this project the topic is the numerical simulation of a dynamic phenomenon, based on the behavior of a transmitted wave along an elastic string of a musical instrument, whose ends are anchored. The physical phenomenon is developed using a hyperbolic partial differential equation with spatial and temporal variables, accompanied by a Dirichlet boundary conditions at the ends and more initial conditions that start the process. Subsequently generated algorithms for the numerical method used (central and forward finite differences) and the programming of the approximate problem with consistency, stability and convergence, yielding results in line with the analytical solution of the mathematical problem. Programming and output results has been made with Visual Studio 8.0. and object programming with Visual Basic. Net
Resumo:
El presente proyecto trata sobre uno de los campos más problemáticos de la inteligencia artificial, el reconocimiento facial. Algo tan sencillo para las personas como es reconocer una cara conocida se traduce en complejos algoritmos y miles de datos procesados en cuestión de segundos. El proyecto comienza con un estudio del estado del arte de las diversas técnicas de reconocimiento facial, desde las más utilizadas y probadas como el PCA y el LDA, hasta técnicas experimentales que utilizan imágenes térmicas en lugar de las clásicas con luz visible. A continuación, se ha implementado una aplicación en lenguaje C++ que sea capaz de reconocer a personas almacenadas en su base de datos leyendo directamente imágenes desde una webcam. Para realizar la aplicación, se ha utilizado una de las librerías más extendidas en cuanto a procesado de imágenes y visión artificial, OpenCV. Como IDE se ha escogido Visual Studio 2010, que cuenta con una versión gratuita para estudiantes. La técnica escogida para implementar la aplicación es la del PCA ya que es una técnica básica en el reconocimiento facial, y además sirve de base para soluciones mucho más complejas. Se han estudiado los fundamentos matemáticos de la técnica para entender cómo procesa la información y en qué se datos se basa para realizar el reconocimiento. Por último, se ha implementado un algoritmo de testeo para poder conocer la fiabilidad de la aplicación con varias bases de datos de imágenes faciales. De esta forma, se puede comprobar los puntos fuertes y débiles del PCA. ABSTRACT. This project deals with one of the most problematic areas of artificial intelligence, facial recognition. Something so simple for human as to recognize a familiar face becomes into complex algorithms and thousands of data processed in seconds. The project begins with a study of the state of the art of various face recognition techniques, from the most used and tested as PCA and LDA, to experimental techniques that use thermal images instead of the classic visible light images. Next, an application has been implemented in C + + language that is able to recognize people stored in a database reading images directly from a webcam. To make the application, it has used one of the most outstretched libraries in terms of image processing and computer vision, OpenCV. Visual Studio 2010 has been chosen as the IDE, which has a free student version. The technique chosen to implement the software is the PCA because it is a basic technique in face recognition, and also provides a basis for more complex solutions. The mathematical foundations of the technique have been studied to understand how it processes the information and which data are used to do the recognition. Finally, an algorithm for testing has been implemented to know the reliability of the application with multiple databases of facial images. In this way, the strengths and weaknesses of the PCA can be checked.
Resumo:
In recent decades, full electric and hybrid electric vehicles have emerged as an alternative to conventional cars due to a range of factors, including environmental and economic aspects. These vehicles are the result of considerable efforts to seek ways of reducing the use of fossil fuel for vehicle propulsion. Sophisticated technologies such as hybrid and electric powertrains require careful study and optimization. Mathematical models play a key role at this point. Currently, many advanced mathematical analysis tools, as well as computer applications have been built for vehicle simulation purposes. Given the great interest of hybrid and electric powertrains, along with the increasing importance of reliable computer-based models, the author decided to integrate both aspects in the research purpose of this work. Furthermore, this is one of the first final degree projects held at the ETSII (Higher Technical School of Industrial Engineers) that covers the study of hybrid and electric propulsion systems. The present project is based on MBS3D 2.0, a specialized software for the dynamic simulation of multibody systems developed at the UPM Institute of Automobile Research (INSIA). Automobiles are a clear example of complex multibody systems, which are present in nearly every field of engineering. The work presented here benefits from the availability of MBS3D software. This program has proven to be a very efficient tool, with a highly developed underlying mathematical formulation. On this basis, the focus of this project is the extension of MBS3D features in order to be able to perform dynamic simulations of hybrid and electric vehicle models. This requires the joint simulation of the mechanical model of the vehicle, together with the model of the hybrid or electric powertrain. These sub-models belong to completely different physical domains. In fact the powertrain consists of energy storage systems, electrical machines and power electronics, connected to purely mechanical components (wheels, suspension, transmission, clutch…). The challenge today is to create a global vehicle model that is valid for computer simulation. Therefore, the main goal of this project is to apply co-simulation methodologies to a comprehensive model of an electric vehicle, where sub-models from different areas of engineering are coupled. The created electric vehicle (EV) model consists of a separately excited DC electric motor, a Li-ion battery pack, a DC/DC chopper converter and a multibody vehicle model. Co-simulation techniques allow car designers to simulate complex vehicle architectures and behaviors, which are usually difficult to implement in a real environment due to safety and/or economic reasons. In addition, multi-domain computational models help to detect the effects of different driving patterns and parameters and improve the models in a fast and effective way. Automotive designers can greatly benefit from a multidisciplinary approach of new hybrid and electric vehicles. In this case, the global electric vehicle model includes an electrical subsystem and a mechanical subsystem. The electrical subsystem consists of three basic components: electric motor, battery pack and power converter. A modular representation is used for building the dynamic model of the vehicle drivetrain. This means that every component of the drivetrain (submodule) is modeled separately and has its own general dynamic model, with clearly defined inputs and outputs. Then, all the particular submodules are assembled according to the drivetrain configuration and, in this way, the power flow across the components is completely determined. Dynamic models of electrical components are often based on equivalent circuits, where Kirchhoff’s voltage and current laws are applied to draw the algebraic and differential equations. Here, Randles circuit is used for dynamic modeling of the battery and the electric motor is modeled through the analysis of the equivalent circuit of a separately excited DC motor, where the power converter is included. The mechanical subsystem is defined by MBS3D equations. These equations consider the position, velocity and acceleration of all the bodies comprising the vehicle multibody system. MBS3D 2.0 is entirely written in MATLAB and the structure of the program has been thoroughly studied and understood by the author. MBS3D software is adapted according to the requirements of the applied co-simulation method. Some of the core functions are modified, such as integrator and graphics, and several auxiliary functions are added in order to compute the mathematical model of the electrical components. By coupling and co-simulating both subsystems, it is possible to evaluate the dynamic interaction among all the components of the drivetrain. ‘Tight-coupling’ method is used to cosimulate the sub-models. This approach integrates all subsystems simultaneously and the results of the integration are exchanged by function-call. This means that the integration is done jointly for the mechanical and the electrical subsystem, under a single integrator and then, the speed of integration is determined by the slower subsystem. Simulations are then used to show the performance of the developed EV model. However, this project focuses more on the validation of the computational and mathematical tool for electric and hybrid vehicle simulation. For this purpose, a detailed study and comparison of different integrators within the MATLAB environment is done. Consequently, the main efforts are directed towards the implementation of co-simulation techniques in MBS3D software. In this regard, it is not intended to create an extremely precise EV model in terms of real vehicle performance, although an acceptable level of accuracy is achieved. The gap between the EV model and the real system is filled, in a way, by introducing the gas and brake pedals input, which reflects the actual driver behavior. This input is included directly in the differential equations of the model, and determines the amount of current provided to the electric motor. For a separately excited DC motor, the rotor current is proportional to the traction torque delivered to the car wheels. Therefore, as it occurs in the case of real vehicle models, the propulsion torque in the mathematical model is controlled through acceleration and brake pedal commands. The designed transmission system also includes a reduction gear that adapts the torque coming for the motor drive and transfers it. The main contribution of this project is, therefore, the implementation of a new calculation path for the wheel torques, based on performance characteristics and outputs of the electric powertrain model. Originally, the wheel traction and braking torques were input to MBS3D through a vector directly computed by the user in a MATLAB script. Now, they are calculated as a function of the motor current which, in turn, depends on the current provided by the battery pack across the DC/DC chopper converter. The motor and battery currents and voltages are the solutions of the electrical ODE (Ordinary Differential Equation) system coupled to the multibody system. Simultaneously, the outputs of MBS3D model are the position, velocity and acceleration of the vehicle at all times. The motor shaft speed is computed from the output vehicle speed considering the wheel radius, the gear reduction ratio and the transmission efficiency. This motor shaft speed, somehow available from MBS3D model, is then introduced in the differential equations corresponding to the electrical subsystem. In this way, MBS3D and the electrical powertrain model are interconnected and both subsystems exchange values resulting as expected with tight-coupling approach.When programming mathematical models of complex systems, code optimization is a key step in the process. A way to improve the overall performance of the integration, making use of C/C++ as an alternative programming language, is described and implemented. Although this entails a higher computational burden, it leads to important advantages regarding cosimulation speed and stability. In order to do this, it is necessary to integrate MATLAB with another integrated development environment (IDE), where C/C++ code can be generated and executed. In this project, C/C++ files are programmed in Microsoft Visual Studio and the interface between both IDEs is created by building C/C++ MEX file functions. These programs contain functions or subroutines that can be dynamically linked and executed from MATLAB. This process achieves reductions in simulation time up to two orders of magnitude. The tests performed with different integrators, also reveal the stiff character of the differential equations corresponding to the electrical subsystem, and allow the improvement of the cosimulation process. When varying the parameters of the integration and/or the initial conditions of the problem, the solutions of the system of equations show better dynamic response and stability, depending on the integrator used. Several integrators, with variable and non-variable step-size, and for stiff and non-stiff problems are applied to the coupled ODE system. Then, the results are analyzed, compared and discussed. From all the above, the project can be divided into four main parts: 1. Creation of the equation-based electric vehicle model; 2. Programming, simulation and adjustment of the electric vehicle model; 3. Application of co-simulation methodologies to MBS3D and the electric powertrain subsystem; and 4. Code optimization and study of different integrators. Additionally, in order to deeply understand the context of the project, the first chapters include an introduction to basic vehicle dynamics, current classification of hybrid and electric vehicles and an explanation of the involved technologies such as brake energy regeneration, electric and non-electric propulsion systems for EVs and HEVs (hybrid electric vehicles) and their control strategies. Later, the problem of dynamic modeling of hybrid and electric vehicles is discussed. The integrated development environment and the simulation tool are also briefly described. The core chapters include an explanation of the major co-simulation methodologies and how they have been programmed and applied to the electric powertrain model together with the multibody system dynamic model. Finally, the last chapters summarize the main results and conclusions of the project and propose further research topics. In conclusion, co-simulation methodologies are applicable within the integrated development environments MATLAB and Visual Studio, and the simulation tool MBS3D 2.0, where equation-based models of multidisciplinary subsystems, consisting of mechanical and electrical components, are coupled and integrated in a very efficient way.
Resumo:
La expansión experimentada por la informática, las nuevas tecnologías e internet en los últimos años, no solo viene dada por la evolución del hardware subyacente, sino por la evolución del desarrollo de software y del crecimiento del número de desarrolladores. Este incremento ha hecho evolucionar el software de unos sistemas de gestión basados en ficheros, prácticamente sin interfaz gráfico y de unos pocos miles de líneas a grandes sistemas distribuidos multiplataforma. El desarrollo de estos grandes sistemas, requiere gran cantidad de personas involucradas en el desarrollo, y que las herramientas de desarrollo hayan crecido también para facilitar su análisis, diseño, codificación, pruebas, implantación y mantenimiento. La base de estas herramientas software las proveen las propias plataformas de desarrollo, pero la experiencia de los desarrolladores puede aportar un sinfín de utilidades y de técnicas que agilicen los desarrollos y cumplan los requisitos del software en base a la reutilización de soluciones lo suficientemente probadas y optimizadas. Dichas herramientas se agrupan ordenadamente, creando así frameworks personalizados, con herramientas de todo tipo, clases, controles, interfaces, patrones de diseño, de tal manera que se dan soluciones personalizadas a un amplio número de problemas para emplearlas cuantas veces se quiera, bien marcando directrices de desarrollo mediante el uso de patrones, bien con la encapsulación de complejidades de tal modo que los desarrolladores ya dispongan de componentes que asuman cierta lógica o cierta complejidad aliviando así la fase de construcción. En este trabajo se abordan temas sobre las tecnologías base y plataformas de desarrollo para poder acometer la creación de un framework personalizado, necesidades a evaluar antes de acometerlo, y técnicas a emplear para la consecución del mismo, orientadas a la documentación, mantenimiento y extensión del framework. La exposición teórica consiste en mostrar y evaluar los requisitos para crear un framework, requisitos de la plataforma de desarrollo, y explicar cómo funcionan las grandes plataformas de desarrollo actuales, que elementos los componen y su funcionamiento, así como marcar ciertas pautas de estructuración y nomenclatura que el desarrollo de un framework debe contemplar para su mantenimiento y extensión. En la parte metodológica se ha usado un subconjunto de Métrica V3, ya que para el desarrollo de controles no aplica dicha metodología en su totalidad, pero contempla el catálogo de requisitos, los casos de uso, diagramas de clase, diagramas de secuencia, etc… Aparte de los conceptos teóricos, se presenta un caso práctico con fines didácticos de cómo parametrizar y configurar el desarrollo bajo la plataforma .NET. Dicho caso práctico consiste en la extensión de un control de usuario genérico de la plataforma .NET, de tal modo que se aplican conceptos más allá del hecho de crear funciones como las funcionalidades que puede brindar un API. Conceptos sobre como extender y modificar controles ya existentes, que interactúan por medio de eventos con otros controles, con vistas a que ese nuevo control forme parte de una biblioteca de controles de usuario personalizados ampliamente divulgada. Los controles de usuario son algo que no solo tienen una parte funcional, sino que también tienen una parte visual, y definiciones funcionales distintas de las típicas del software de gestión, puesto que han de controlar eventos, visualizaciones mientras se dan estos eventos y requisitos no funcionales de optimización de rendimiento, etc… Para el caso práctico se toma como herramienta la plataforma de desarrollo .Net Framework, en todas sus versiones, ya que el control a extender es el control ListView y hacerlo editable. Este control está presente en todas las versiones de .NET framework y con un alto grado de reutilización. Esta extensión muestra además como se puede migrar fácilmente este tipo de extensiones sobre todos los frameworks. Los entornos de desarrollo usados son varias versiones de Visual Studio para el mostrar dicha compatibilidad, aunque el desarrollo que acompaña este documento esté realizado sobre Visual Studio 2013. ABSTRACT The expansion in computer science, new technologies and the Internet in recent years, not only is given by the evolution of the underlying hardware, but for the evolution of software development and the growing number of developers. This increase has evolved software from management systems based on files almost without graphical interface and a few thousand of code lines, to large multiplatform distributed systems. The development of these large systems, require lots of people involved in development, and development tools have also grown to facilitate analysis, design, coding, testing, deployment and maintenance. The basis of these software tools are providing by their own development platforms, but the experience of the developers can bring a lot of utilities and techniques to speed up developments and meet the requirements of software reuse based on sufficiently proven solutions and optimized. These tools are grouped neatly, creating in this way custom frameworks, with tools of all types, classes, controls, interfaces, design patterns,… in such a way that they provide customized solutions to a wide range of problems to use them many times as you want to occur, either by dialing development guidelines by using patterns or along with the encapsulation of complexities, so that developers already have components that take some logic or some complexity relieving the construction phase. This paper cover matters based on technologies and development platforms to undertake the creation of a custom framework, needs to evaluate before rush it and techniques to use in order to achieve it, a part from techniques oriented to documentation, maintenance and framework extension. The theoretical explanation consists in to demonstrate and to evaluate the requirements for creating a framework, development platform requirements, and explain how large current development platforms work, which elements compose them and their operation work, as well as mark certain patterns of structure and nomenclature that the development of a framework should include for its maintenance and extension. In the methodological part, a subset of Métrica V3 has been used, because of, for the development of custom controls this methodology does not apply in its entirety, but provides a catalogue of requirements, use cases, class diagrams, sequence diagrams, etc ... Apart from the theoretical concepts, a study case for teaching purposes about how to parameterize and configure the development under the .NET platform is presented. This study case involves the extension of a generic user control of the .NET platform, so that concepts apply beyond the fact of creating functions as the functionalities that can provide an API. Concepts on how to extend and modify existing controls that interact through events with other controls, overlooking that new control as a part of a custom user controls library widely publicized. User controls are something that not only have a functional part, but also have a visual part, and various functional definitions of typical management software, since that they have to control events, visualizations while these events are given and not functional of performance optimization requirements, etc ... For the study case the development platform .Net Framework is taken as tool, in all its versions, considering that control to extend is the ListView control and make it editable. This control is present in all versions of .NET framework and with a high degree of reuse. This extension also shows how you can easily migrate these extensions on all frameworks. The used development environments are several versions of Visual Studio to show that compatibility, although the development that accompanies this document is done on Visual Studio 2013.
Luz industrial e imagen tecnificada: de Moholy Nagy al C.A.V.S. (Center for Advanced Visual Studies)
Resumo:
El desarrollo de la tecnología de la luz implicará la transformación de la vida social, cultural y económica. Tanto las consideraciones espaciales del Movimiento Moderno, como los efectos producidos por la segunda Guerra Mundial, tendrán efectos visibles en las nuevas configuraciones espaciales y en la relación simbiótica y recíproca que se dará entre ideología y tecnología. La transformación en la comprensión de la articulación espacial, asociada al desarrollo tecnológico, afectará al modo en que este espacio es experimentado y percibido. El espacio expositivo y el espacio escénico se convertirán en laboratorio práctico donde desarrollar y hacer comprensible todo el potencial ilusorio de la luz, la proyección y la imagen, como parámetros modificadores y dinamizadores del espacio arquitectónico. Esta experimentación espacial estará precedida por la investigación y creación conceptual en el mundo plástico, donde los nuevos medios mecánicos serán responsables de la construcción de una nueva mirada moderna mediatizada por los elementos técnicos. La experimentación óptica, a través de la fotografía, el cine, o el movimiento de la luz y su percepción, vinculada a nuevos modos de representación y comunicación, se convertirá en elemento fundamental en la configuración espacial. Este ámbito de experimentación se hará patente en la Escuela de la Bauhaus, de la mano de Gropius, Schlemmer o Moholy Nagy entre otros; tanto en reflexiones teóricas como en el desarrollo de proyectos expositivos, arquitectónicos o teatrales, que evolucionarán en base a la tecnología y la modificación de la relación con el espectador. El espacio expositivo y el espacio escénico se tomarán como oportunidad de investigación espacial y de análisis de los modos de percepción, convirtiéndose en lugares de experimentación básicos para el aprendizaje. El teatro se postula como punto de encuentro entre el arte y la técnica, cobrando especial importancia la intersección con otras disciplinas en la definición espacial. Las múltiples innovaciones técnicas ligadas a los nuevos fundamentos teatrales en la modificación de la relación con la escena, que se producen a principios del siglo XX, tendrán como consecuencia la transformación del espacio en un espacio dinámico, tanto física como perceptivamente, que dará lugar a nuevas concepciones espaciales, muchas de ellas utópicas. La luz, la proyección y la creación de ilusión en base a estímulos visuales y sonoros, aparecen como elementos proyectuales efímeros e inmateriales, que tendrán una gran incidencia en el espacio y su modo de ser experimentado. La implicación de la tecnología en el arte conllevará modificaciones en la visualización, así como en la configuración espacial de los espacios destinados a esta. Destacaremos como propuesta el Teatro Total de Walter Gropius, en cuyo desarrollo se recogen de algún modo las experiencias espaciales y las investigaciones desarrolladas sobre la estructura formal de la percepción realizadas por Moholy Nagy, además de los conceptos acerca del espacio escénico desarrollados en el taller de Teatro de la Bauhaus por Oskar Schlemmer. En el Teatro Total, Gropius incorporará su propia visión de cuestiones que pertenecen a la tradición de la arquitectura teatral y las innovaciones conceptuales que estaban teniendo lugar desde finales del s.XIX, tales como la participación activa del público o la superación entre escena y auditorio, estableciendo en el proyecto una nueva relación perceptual entre sala, espectáculo y espectador; aumentando la sensación de inmersión, a través del uso de la física, la óptica, y la acústica, creando una energía concéntrica capaz de extenderse en todas direcciones. El Teatro Total será uno de los primeros ejemplos en los que desde el punto de partida del proyecto, se conjuga la imagen como elemento comunicativo con la configuración espacial. Las nuevas configuraciones escénicas tendrán como premisa de desarrollo la capacidad de transformación tanto perceptiva, como física. En la segunda mitad del s.XX, la creación de centros de investigación como el CAVS (The Center for Advanced Visual Studies,1967), o el EAT (Experiments in Art and Technology, 1966), favorecerán la colaboración interdisciplinar entre arte y ciencia, implicando a empresas de carácter tecnológico, como Siemens, HP, IBM o Philips, facilitando soporte técnico y económico para el desarrollo de nuevos sistemas. Esta colaboración interdisciplinar dará lugar a una serie de intervenciones espaciales que tendrán su mayor visibilidad en algunas Exposiciones Universales. El resultado será, en la mayoría de los casos, la creación de espacios de carácter inmersivo, donde se establecerá una relación simbiótica entre espacio, imagen, sonido, y espectador. La colocación del espectador en el centro de la escena y la disposición dinámica de imagen y sonido, crearán una particular narrativa espacial no lineal, concebida para la experiencia. Desde las primeras proyecciones de cine a la pantalla múltiple de los Eames, las técnicas espaciales de difusión del sonido en Stockhausen, o los experimentos con el movimiento físico interactivo, la imagen, la luz en movimiento y el sonido, quedan inevitablemente convertidos en material arquitectónico. ABSTRACT. Light technology development would lead to a social, cultural and economic transformation. Both spatial consideration of “Modern Movement” and Second World War effects on technology, would have a visible aftereffect on spatial configuration and on the symbiotic and mutual relationship between ideology & technology. Comprehension adjustment on the articulation of space together with technology development, would impact on how space is perceived and felt. Exhibition space and scenic space would turn into a laboratory where developing and making comprehensive all illusory potential of light, projection and image. These new parameters would modify and revitalize the architectonic space. as modifying and revitalizing parameters of architectonic space. Spatial experimentation would be preceded by conceptual creation and investigation on the sculptural field, where new mechanic media would be responsible for a fresh and modern look influenced by technical elements. Optical experimentation, through photography, cinema or light movement and its perception, would turn into essential components for spatial arrangement linked to new ways of performance and communication. This experimentation sphere would be clear at The Bauhaus School, by the hand of Gropius, Schlemmer or Moholy Nag among others; in theoretical, theatrical or architectural performance’s projects, that would evolve based on technology and also based on the transformation of the relationship with the observer. Exhibition and perfor-mance areas would be taken as opportunities of spatial investigation and for the analysis of the different ways of perception, thus becoming key places for learning. Theater is postulated as a meeting point between art and technique, taking on a new significance at its intersection with other disciplines working with spatial definition too. The multiple innovation techniques linked to the new foundations for the theater regarding stage relation, would have as a consequence the regeneration of the space. Space would turn dynamic, both physically and perceptibly, bringing innovative spatial conceptions, many of them unrealistic. Light, projection and illusory creation based on sound and visual stimulus would appear as intangible and momentary design components, which would have a great impact on the space and on the way it is experienced. Implication of technology in art would bring changes on the observer as well as on the spatial configuration of the art spaces2. It would stand out as a proposal Walter Groupis Total Theater, whose development would include somehow the spatial experiments and studies about formal structure of perception accomplished by Moholy Nagy besides the concepts regarding stage space enhanced at the Bauhaus Theater Studio by Oskar Schlemmer. Within Total Theater, Groupis would incorporate his own view about traditional theatric architecture and conceptual innovations that were taking place since the end of the nineteenth century, such as active audience participation or the diffusing limits between scene and audience, establishing a new perception relationship between auditorium, performance and audience, improving the feeling of immersion through the use of physics, optics and acoustics, creating a concentric energy capable of spreading in all directions. Total Theater would be one of the first example in which, from the beginning of the Project, image is combined as a communicating element with the spatial configuration. As a premise of development, new stage arrangement would have the capacity of transformation, both perceptive and physically. During the second half or the twentieth century, the creation of investigation centers such as CAVS (Center for Advanced Visual Studies, 1967) or EAT (Experiments in Art and Technology, 1966), would help to the interdisciplinary collaboration between art and science, involving technology companies like Siemens, HP, IBM or Philips, providing technical and economic support to the development of new systems. This interdisciplinary collaboration would give room to a series of spatial interventions which would have visibility in some Universal Exhibitions. The result would be, in most cases, the creation of immersive character spaces, where a symbiotic relationship would be stablished between space, image, sound and audience. The new location of the audience in the middle of the display, together with the dynamic arrangement of sound and image would create a particular, no lineal narrative conceived to be experienced. Since the first cinema projections, the multiple screen of Eames, the spatial techniques for sound dissemination at Stockhausen or the interactive physical movement experimentation, image, motion light and sound would turn inevitably into architectural material.