71 resultados para Analysis software
em Universidad Politécnica de Madrid
Resumo:
The analysis of the interdependence between time series has become an important field of research, mainly as a result of advances in the characterization of dynamical systems from the signals they produce, and the introduction of concepts such as Generalized (GS) and Phase synchronization (PS). This increase in the number of approaches to tackle the existence of the so-called functional (FC) and effective connectivity (EC) (Friston 1994) between two, (or among many) neural networks, along with their mathematical complexity, makes it desirable to arrange them into a unified toolbox, thereby allowing neuroscientists, neurophysiologists and researchers from related fields to easily access and make use of them.
Resumo:
Traditional identification of ore minerals with reflected light microscopy relies heavily on the experience of the observer. Qualified observers have become a rarity, as ore microscopy is often neglected in today’s university training, but since it furnishes necessary and inexpensive information, innovative alternatives are needed, especially for quantification. Many of the diagnostic optical properties of ores defy quantification, but recent developments in electronics and optics allow new insights into the reflectance and colour properties of ores. Preliminary results for the development of an expert system aimed at the automatic identification of ores based on their reflectance properties are presented. The discriminatory capacity of the system is enhanced by near IR reflectance measures, while UV filters tested to date are unreliable. Interaction with image analysis software through a wholly automated microscope, to furnish quantitative and morphological information for geometallurgy, relies on automated identification of the ores based on the measured spectra. This methodology increases enormously the performance of the microscopist; nevertheless supervision by an expert is always needed.
Resumo:
Este Proyecto Fin de Carrera trata sobre el estudio de la precisión de los modelos de transistores de radiofrecuencia y el efecto que la misma produce en un circuito real, como es la variabilidad de diversas magnitudes tales como la ganancia y el punto de compresión a 1 decibelio. Para ello se ha construido un circuito de pruebas que ha sido sometido a diversas simulaciones y sobre el que se han realizado numerosas medidas que han sido analizadas mediante software de análisis estadístico. This Final Year Project is about the study of the precision of models of radio frequency transistors and the effect it produces in a real circuit, as is the variability of various magnitudes such as gain and one‐decibel compression point. For this we have built a test circuit that has been subjected to various simulations and on which there have been numerous measures that have been analyzed by statistical analysis software.
Resumo:
The energy and specific energy absorbed in the main cell compartments (nucleus and cytoplasm) in typical radiobiology experiments are usually estimated by calculations as they are not accessible for a direct measurement. In most of the work, the cell geometry is modelled using the combination of simple mathematical volumes. We propose a method based on high resolution confocal imaging and ion beam analysis (IBA) in order to import realistic cell nuclei geometries in Monte-Carlo simulations and thus take into account the variety of different geometries encountered in a typical cell population. Seventy-six cell nuclei have been imaged using confocal microscopy and their chemical composition has been measured using IBA. A cellular phantom was created from these data using the ImageJ image analysis software and imported in the Geant4 Monte-Carlo simulation toolkit. Total energy and specific energy distributions in the 76 cell nuclei have been calculated for two types of irradiation protocols: a 3 MeV alpha particle microbeam used for targeted irradiation and a 239Pu alpha source used for large angle random irradiation. Qualitative images of the energy deposited along the particle tracks have been produced and show good agreement with images of DNA double strand break signalling proteins obtained experimentally. The methodology presented in this paper provides microdosimetric quantities calculated from realistic cellular volumes. It is based on open-source oriented software that is publicly available.
Resumo:
En el presente proyecto se ha realizado un estudio de las condiciones acústicas y del refuerzo sonoro de la sala de conciertos Ritmo y Compás (Madrid), repasando las principales magnitudes de calidad acústica. El proyecto combina medidas y simulaciones. Con las medidas se obtienen los parámetros acústicos de la sala, que permiten cuantificar la sensación que producen en la audiencia las cualidades acústicas de un espacio. Mediante software de simulación electroacústica (EASE) se ha generado el modelo geométrico y acústico de la sala, ajustado en base a los resultados experimentales. El objetivo fundamental del trabajo es generar un modelo virtual que refleje fielmente las condiciones reales de la sala de conciertos, de forma que se puedan estudiar en profundidad las características acústicas y electroacústicas del recinto. En el proceso de análisis se ha querido resaltar la importancia de la acústica en salas de conciertos de música amplificada, muchas veces relegada a un segundo plano por el diseño electroacústico. La calidad de una sala de este tipo depende estrechamente de la sonorización, pero su inevitable relación con las cualidades acústicas del recinto obliga a establecer una serie de criterios acústicos mínimos que aseguren las mejores condiciones para el sistema sonoro. Dado el peso de las simulaciones, en el proyecto se presentan los resultados obtenidos con distintos métodos y recursos de análisis software para apoyar el estudio, con una información completa que muestre, con la máxima claridad, el potencial de la sala Ritmo y Compás como sala de conciertos. SUMMARY. This project is about the acoustical and electroacustical studio of “Ritmo y Compás”, an important concert hall in Madrid, analyzing the main magnitudes of acoustic quality. The project combines measurements and simulations. With the measurements it is obtained the acoustic parameters of the hall, allowing the quantification of the sensation produced in the audience by the acoustic attributes of the hall. With the acoustic simulation software (Ease), based on the experimental results, the geometric and acoustic model has been created. The main purpose of this study is to generate a virtual model that accurately reflects the real conditions of the concert hall, allowing the deep study of the acoustic and electroacoustic features of the hall. In the process of the analysis, the importance of the acoustic characteristic in the amplified music concert halls was emphasized, which is often underestimate because of the electroacoustic design. The quality of this kind of hall strictly depends on the sound electrical system, but its inevitable relation with the acoustic characteristics of the hall, forces to establish a series of minimum acoustic rules that assures the best conditions for the sound system. Due to the importance of the simulations in the project, the results are presented with different methods and analysis software resources to back up the study with complete information that shows the maximum quality and clarity of the potential of the enclosure Ritmo y Compás as a concert hall.
Resumo:
A high-power high-efficiency laser power transmission system at 100m based on an optimized multi-cell GaAs converter capable of supplying 9.7W of electricity is demonstrated. An I-V testing system integrated with a data acquisition circuit and an analysis software is designed to measure the efficiency and the I-V characteristics of the laser power converter (LPC). The dependencies of the converter’s efficiency with respect to wavelength, laser intensity and temperature are analyzed. A diode laser with 793nm of wavelength and 24W of power is used to test the LPC and the software. The maximum efficiency of the LPC is 48.4% at an input laser power of 8W at room temperature. When the input laser power is 24W (laser intensity of 60000W/m2), the efficiency is 40.4% and the output voltage is 4 V. The overall efficiency from electricity to electricity is 11.6%.
Resumo:
The study of lateral dynamics of running trains on bridges is of importance mainly for the safety of the traffic, and may be relevant for laterally compliant bridges. These studies require threedimensional coupled vehicle-bridge models, wheree consideration of wheel to rail contact is a key aspect. Furthermore, an adequate evaluation of safety of rail traffic requires nonlinear models. A nonlinear coupled model is proposed here for vehicle-structure vertical and lateral dynamics. Vehicles are considered as fully three-dimensional multibody systems including gyroscopic terms and large rotation effects. The bridge structure is modeled by means of finite elements which may be of beam, shell or continuum type and may include geometric or material nonlinearities. The track geometry includes distributed track alignment irregularities. Both subsystems (bridge and vehicles) are described with coordinates in absolute reference frames, as opposed to alternative approaches which describe the multibody system with coordinates relative to the base bridge motion. The wheelrail contact employed is a semi-Hertzian model based on realistic wheel-rail profiles. It allows a detailed geometrical description of the contact patch under each wheel including multiple-point contact, flange contact and uplift. Normal and tangential stresses in each contact are integrated at each time-step to obtain the resultant contact forces. The models have been implemented within an existing finite element analysis software with multibody capabilities, Abaqus (Simulia Ltd., 2010). Further details of the model are presented in Antolín et al. (2012). Representative applications are presented for railway vehicles under lateral wind action on laterally compliant viaducts, showing the relevance of the nonlinear wheel-rail contact model as well as the interaction between bridge and vehicle.
Resumo:
El objetivo del presente trabajo es realizar una evaluación realista de la incertidumbre en los ensayos de ruido transmitido al interior de una sala según el Anexo IV del Real Decreto 1367/2007, contemplando todas las posibles causas que intervienen en la incertidumbre de forma que las imprecisiones observadas en los ejercicios de intercomparación se cubran con un único ensayo. Otra parte fundamental del trabajo es cuantificar la fuente de incertidumbre generada en la elección del punto de medición. En primer lugar se realiza la medición del nivel continuo equivalente ponderado A en habitaciones de distinto tamaño, a la que llegan los distintos tipos de ruido emitido desde el exterior de las mismas. El siguiente paso es realizar un análisis de los niveles medidos, tanto en su distribución espacial como en su evolución temporal, incidiendo en los valores máximos, LAeq,5s. A continuación se comprueba si la distribución de los niveles medidos se ajusta a una distribución normal mediante el software de análisis estadístico STATGRAPHICS. Determinando, en base al tamaño de las muestras escogidas, por medio de análisis estadísticos, la aportación a la incertidumbre generada en la elección del punto de medida, cuantificando su valor. Por último se realiza una medición del ruido de actividades según el Real Decreto 1367/2007, aportando una evaluación de la incertidumbre, teniendo en cuenta todas las fuentes que la generan, mediante el enfoque clásico de la GUM. ABSTRACT. The objective of this work is to carry out a realistic assessment of the uncertainty in the trials of noise transmitted into the interior of a room according to Annex IV of the Royal Decree 136772997, considering all of the causes involved in uncertainty in such a way that the inaccuracies observed in intercomparison exercises are covered with a single trial. Another fundamental part of the work is to quantify the source of uncertainty in the choice of the point of measurement. First is the weighted equivalent of the continuous level measurement in rooms of different sizes, which reach different types of noise emitted from outside of them. The next step is to perform an analysis of the measured levels, both in their spatial distribution and their temporal evolution, influencing the maximum values, LAeq,5s. Then it is checked to see whether the distribution of the measured levels conforms to a normal distribution using the STATGRAPHICS statistical analysis software. Determining therefore, based on the size of the selected samples, through statistical analysis, the contribution to the uncertainty generated in the choice of the measurement point, quantifying its value. At last is the measure of the noise of activities according to the Royal Decree 1367/2007, providing an assessment of the uncertainty, taking into account all sources that generate it, using the classical approach to the GUM.
Resumo:
Dentro del análisis y diseño estructural surgen frecuentemente problemas de ingeniería donde se requiere el análisis dinámico de grandes modelos de elementos finitos que llegan a millones de grados de libertad y emplean volúmenes de datos de gran tamaño. La complejidad y dimensión de los análisis se dispara cuando se requiere realizar análisis paramétricos. Este problema se ha abordado tradicionalmente desde diversas perspectivas: en primer lugar, aumentando la capacidad tanto de cálculo como de memoria de los sistemas informáticos empleados en los análisis. En segundo lugar, se pueden simplificar los análisis paramétricos reduciendo su número o detalle y por último se puede recurrir a métodos complementarios a los elementos .nitos para la reducción de sus variables y la simplificación de su ejecución manteniendo los resultados obtenidos próximos al comportamiento real de la estructura. Se propone el empleo de un método de reducción que encaja en la tercera de las opciones y consiste en un análisis simplificado que proporciona una solución para la respuesta dinámica de una estructura en el subespacio modal complejo empleando un volumen de datos muy reducido. De este modo se pueden realizar análisis paramétricos variando múltiples parámetros, para obtener una solución muy aproximada al objetivo buscado. Se propone no solo la variación de propiedades locales de masa, rigidez y amortiguamiento sino la adición de grados de libertad a la estructura original para el cálculo de la respuesta tanto permanente como transitoria. Adicionalmente, su facilidad de implementación permite un control exhaustivo sobre las variables del problema y la implementación de mejoras como diferentes formas de obtención de los autovalores o la eliminación de las limitaciones de amortiguamiento en la estructura original. El objetivo del método se puede considerar similar a los que se obtienen al aplicar el método de Guyan u otras técnicas de reducción de modelos empleados en dinámica estructural. Sin embargo, aunque el método permite ser empleado en conjunción con otros para obtener las ventajas de ambos, el presente procedimiento no realiza la condensación del sistema de ecuaciones, sino que emplea la información del sistema de ecuaciones completa estudiando tan solo la respuesta en las variables apropiadas de los puntos de interés para el analista. Dicho interés puede surgir de la necesidad de obtener la respuesta de las grandes estructuras en unos puntos determinados o de la necesidad de modificar la estructura en zonas determinadas para cambiar su comportamiento (respuesta en aceleraciones, velocidades o desplazamientos) ante cargas dinámicas. Por lo tanto, el procedimiento está particularmente indicado para la selección del valor óptimo de varios parámetros en grandes estructuras (del orden de cientos de miles de modos) como pueden ser la localización de elementos introducidos, rigideces, masas o valores de amortiguamientos viscosos en estudios previos en los que diversas soluciones son planteadas y optimizadas, y que en el caso de grandes estructuras, pueden conllevar un número de simulaciones extremadamente elevado para alcanzar la solución óptima. Tras plantear las herramientas necesarias y desarrollar el procedimiento, se propone un caso de estudio para su aplicación al modelo de elementos .nitos del UAV MILANO desarrollado por el Instituto Nacional de Técnica Aeroespacial. A dicha estructura se le imponen ciertos requisitos al incorporar un equipo en aceleraciones en punta de ala izquierda y desplazamientos en punta de ala derecha en presencia de la sustentación producida por una ráfaga continua de viento de forma sinusoidal. La modificación propuesta consiste en la adición de un equipo en la punta de ala izquierda, bien mediante un anclaje rígido, bien unido mediante un sistema de reducción de la respuesta dinámica con propiedades de masa, rigidez y amortiguamiento variables. El estudio de los resultados obtenidos permite determinar la optimización de los parámetros del sistema de atenuación por medio de múltiples análisis dinámicos de forma que se cumplan de la mejor forma posible los requisitos impuestos con la modificación. Se comparan los resultados con los obtenidos mediante el uso de un programa comercial de análisis por el método de los elementos .nitos lográndose soluciones muy aproximadas entre el modelo completo y el reducido. La influencia de diversos factores como son el amortiguamiento modal de la estructura original, el número de modos retenidos en la truncatura o la precisión proporcionada por el barrido en frecuencia se analiza en detalle para, por último, señalar la eficiencia en términos de tiempo y volumen de datos de computación que ofrece el método propuesto en comparación con otras aproximaciones. Por lo tanto, puede concluirse que el método propuesto se considera una opción útil y eficiente para el análisis paramétrico de modificaciones locales en grandes estructuras. ABSTRACT When developing structural design and analysis some projects require dynamic analysis of large finite element models with millions of degrees of freedom which use large size data .les. The analysis complexity and size grow if a parametric analysis is required. This problem has been approached traditionally in several ways: one way is increasing the power and the storage capacity of computer systems involved in the analysis. Other obvious way is reducing the total amount of analyses and their details. Finally, complementary methods to finite element analysis can also be employed in order to limit the number of variables and to reduce the execution time keeping the results as close as possible to the actual behaviour of the structure. Following this third option, we propose a model reduction method that is based in a simplified analysis that supplies a solution for the dynamic response of the structure in the complex modal space using few data. Thereby, parametric analysis can be done varying multiple parameters so as to obtain a solution which complies with the desired objetive. We propose not only mass, stiffness and damping variations, but also addition of degrees of freedom to the original structure in order to calculate the transient and steady-state response. Additionally, the simple implementation of the procedure allows an in-depth control of the problem variables. Furthermore, improvements such as different ways to obtain eigenvectors or to remove damping limitations of the original structure are also possible. The purpose of the procedure is similar to that of using the Guyan or similar model order reduction techniques. However, in our method we do not perform a true model order reduction in the traditional sense. Furthermore, additional gains, which we do not explore herein, can be obtained through the combination of this method with traditional model-order reduction procedures. In our procedure we use the information of the whole system of equations is used but only those nodes of interest to the analyst are processed. That interest comes from the need to obtain the response of the structure at specific locations or from the need to modify the structure at some suitable positions in order to change its behaviour (acceleration, velocity or displacement response) under dynamic loads. Therefore, the procedure is particularly suitable for parametric optimization in large structures with >100000 normal modes such as position of new elements, stiffness, mass and viscous dampings in previous studies where different solutions are devised and optimized, and in the case of large structures, can carry an extremely high number of simulations to get the optimum solution. After the introduction of the required tools and the development of the procedure, a study case is proposed with use the finite element model (FEM) of the MILANO UAV developed by Instituto Nacional de Técnica Aeroespacial. Due to an equipment addition, certain acceleration and displacement requirements on left wing tip and right wing tip, respectively, are imposed. The structure is under a continuous sinusoidal wind gust which produces lift. The proposed modification consists of the addition of an equipment in left wing tip clamped through a rigid attachment or through a dynamic response reduction system with variable properties of mass, stiffness and damping. The analysis of the obtained results allows us to determine the optimized parametric by means of multiple dynamic analyses in a way such that the imposed requirements have been accomplished in the best possible way. The results achieved are compared with results from a commercial finite element analysis software, showing a good correlation. Influence of several factors such as the modal damping of the original structure, the number of modes kept in the modal truncation or the precission given by the frequency sweep is analyzed. Finally, the efficiency of the proposed method is addressed in tems of computational time and data size compared with other approaches. From the analyses performed, we can conclude that the proposed method is a useful and efficient option to perform parametric analysis of possible local modifications in large structures.
Resumo:
El autor de este proyecto es miembro reciente de la asociación SoloBoulder, dedicada a la modalidad de escalada boulder, noticias y actualidad, contenido multimedia, promoción de un equipo de escaladores y defensa de valores medioambientales en la montaña. El principal canal de distribución de contenidos es una página web existente previa a este proyecto. La asociación ha detectado una escasez y mala calidad de recursos en internet en cuanto a guías de zonas donde poder practicar el boulder. Tal circunstancia impulsa la iniciativa de este proyecto fin de carrera. El objetivo general es el desarrollo de una nueva aplicación que proporcione a los usuarios a nivel mundial una guía interactiva de boulder y otros puntos de interés, una red social que permita la creación cooperativa y orgánica de contenido, y servicios web para el consumo de la información desde otras plataformas u organizaciones. El nuevo software desarrollado es independiente de la página web de SoloBoulder previa. No obstante, ambas partes se integran bajo el mismo domino web y aspecto. La nueva aplicación ofrece a escaladores y turistas un servicio informativo e interactivo de calidad, con el que se espera aumentar el número de visitas en todo el sitio web y poder ampliar la difusión de valores medioambientales, diversificar las zonas de boulder y regular las masificadas, favorecer el deporte y brindar al escalador una oportunidad de autopromoción personal. Una gran motivación para el autor también es el proceso de investigación y formación en tecnologías, patrones arquitecturales de diseño y metodologías de trabajo adaptadas a las tendencias actuales en la ingeniería de software, con especial curiosidad hacia el mundo web. A este respecto podemos destacar: metodología de trabajo en proyectos, análisis de proyectos, arquitecturas de software, diseño de software, bases de datos, programación y buenas prácticas, seguridad, interfaz gráfica web, diseño gráfico, Web Performance Optimization, Search Engine Optimization, etc. En resumen, este proyecto constituye un aprendizaje y puesta en práctica de diversos conocimientos adquiridos durante la ejecución del mismo, así como afianzamiento de materias estudiadas en la carrera. Además, el producto desarrollado ofrece un servicio de calidad a los usuarios y favorece el deporte y la autopromoción del escalador. ABSTRACT. The author of this Project is recent member of the association SoloBoulder, dedicated to a rock climbing discipline called bouldering, news, multimedia content, promotion of a team of climbers and defense of environmental values in the mountain. The main content distribution channel is a web page existing previous to this project. The association has detected scarcity and bad quality of resources on the internet about guides of bouldering areas. This circumstance motivates the initiative of this project. The general objective is the development of a new application which provides a worldwide, interactive bouldering guide, including other points of interest, a social network which allows the cooperative and organic creation of content, and web services for consumption of information from other platforms or organizations. The new software developed is independent of the previous SoloBoulder web page. However, both parts are integrated under the same domain and appearance. The new application offers to climbers and tourists a quality informative and interactive service, with which we hope to increase the number of visits in the whole web site and be able to expand the dissemination of environmental values, diversify boulder areas and regulate the overcrowded ones, encourage sport and offer to the climber an opportunity of self-promotion. A strong motivation for the author is also the process of investigation and education in technologies, architectural design patterns and working methodologies adapted to the actual trends in software engineering, with special curiosity about the web world. In this regard we could highlight: project working methodologies, project analysis, software architectures, software design, data bases, programming and good practices, security, graphic web interface, graphic design, Web Performance Optimization, Search Engine Optimization, etc. To sum up, this project constitutes learning and practice of diverse knowledge acquired during its execution, as well as consolidation of subjects studied in the degree. In addition, the product developed offers a quality service to the users and favors the sport and the selfpromotion of the climber.
Resumo:
This paper describes a case study in WCET analysis of an on-board spacecraft software system. The attitude control system of UPMSat-2, an experimental micro-satellite which is scheduled to be launched in 2013, is used for an experiment on analysing the worst-case execution time of code automatically generated from a Simulink model. In order to properly test the code, a hardware-in-the-loop configuration with a simulation model of the spacecraft environment has been used as a test bench. The code has been analysed with RapiTime, with some modifications to the original instrumentation routines, in order to take into account the particularities of the test configuration. Results from the experiment are described and commented in the paper.
Resumo:
This paper describes the authors? experience with static analysis of both WCET and stack usage of a satellite on-board software subsystem. The work is a continuation of a previous case study that used a dynamic WCET analysis tool on an earlier version of the same software system. In particular, the AbsInt aiT tool has been evaluated by analysing both C and Ada code generated by Simulink within the UPMSat-2 project. Some aspects of the aiT tool, specifically those dealing with SPARC register windows, are compared to another static analysis tool, Bound-T. The results of the analysis are discussed, and some conclusions on the use of static WCET analysis tools on the SPARC architecture are commented in the paper.
Resumo:
Background Gray scale images make the bulk of data in bio-medical image analysis, and hence, the main focus of many image processing tasks lies in the processing of these monochrome images. With ever improving acquisition devices, spatial and temporal image resolution increases, and data sets become very large. Various image processing frameworks exists that make the development of new algorithms easy by using high level programming languages or visual programming. These frameworks are also accessable to researchers that have no background or little in software development because they take care of otherwise complex tasks. Specifically, the management of working memory is taken care of automatically, usually at the price of requiring more it. As a result, processing large data sets with these tools becomes increasingly difficult on work station class computers. One alternative to using these high level processing tools is the development of new algorithms in a languages like C++, that gives the developer full control over how memory is handled, but the resulting workflow for the prototyping of new algorithms is rather time intensive, and also not appropriate for a researcher with little or no knowledge in software development. Another alternative is in using command line tools that run image processing tasks, use the hard disk to store intermediate results, and provide automation by using shell scripts. Although not as convenient as, e.g. visual programming, this approach is still accessable to researchers without a background in computer science. However, only few tools exist that provide this kind of processing interface, they are usually quite task specific, and don’t provide an clear approach when one wants to shape a new command line tool from a prototype shell script. Results The proposed framework, MIA, provides a combination of command line tools, plug-ins, and libraries that make it possible to run image processing tasks interactively in a command shell and to prototype by using the according shell scripting language. Since the hard disk becomes the temporal storage memory management is usually a non-issue in the prototyping phase. By using string-based descriptions for filters, optimizers, and the likes, the transition from shell scripts to full fledged programs implemented in C++ is also made easy. In addition, its design based on atomic plug-ins and single tasks command line tools makes it easy to extend MIA, usually without the requirement to touch or recompile existing code. Conclusion In this article, we describe the general design of MIA, a general purpouse framework for gray scale image processing. We demonstrated the applicability of the software with example applications from three different research scenarios, namely motion compensation in myocardial perfusion imaging, the processing of high resolution image data that arises in virtual anthropology, and retrospective analysis of treatment outcome in orthognathic surgery. With MIA prototyping algorithms by using shell scripts that combine small, single-task command line tools is a viable alternative to the use of high level languages, an approach that is especially useful when large data sets need to be processed.
Resumo:
According to the PMBOK (Project Management Body of Knowledge), project management is “the application of knowledge, skills, tools, and techniques to project activities to meet the project requirements” [1]. Project Management has proven to be one of the most important disciplines at the moment of determining the success of any project [2][3][4]. Given that many of the activities covered by this discipline can be said that are “horizontal” for any kind of domain, the importance of acknowledge the concepts and practices becomes even more obvious. The specific case of the projects that fall in the domain of Software Engineering are not the exception about the great influence of Project Management for their success. The critical role that this discipline plays in the industry has come to numbers. A report by McKinsey & Co [4] shows that the establishment of programs for the teaching of critical skills of project management can improve the performance of the project in time and costs. As an example of the above, the reports exposes: “One defense organization used these programs to train several waves of project managers and leaders who together administered a portfolio of more than 1,000 capital projects ranging in Project management size from $100,000 to $500 million. Managers who successfully completed the training were able to cut costs on most projects by between 20 and 35 percent. Over time, the organization expects savings of about 15 percent of its entire baseline spending”. In a white paper by the PMI (Project Management Institute) about the value of project management [5], it is stated that: “Leading organizations across sectors and geographic borders have been steadily embracing project management as a way to control spending and improve project results”. According to the research made by the PMI for the paper, after the economical crisis “Executives discovered that adhering to project management methods and strategies reduced risks, cut costs and improved success rates—all vital to surviving the economic crisis”. In every elite company, a proper execution of the project management discipline has become a must. Several members of the software industry have putted effort into achieving ways of assuring high quality results from projects; many standards, best practices, methodologies and other resources have been produced by experts from different fields of expertise. In the industry and the academic community, there is a continuous research on how to teach better software engineering together with project management [4][6]. For the general practices of Project Management the PMI produced a guide of the required knowledge that any project manager should have in their toolbox to lead any kind of project, this guide is called the PMBOK. On the side of best practices 10 and required knowledge for the Software Engineering discipline, the IEEE (Institute of Electrical and Electronics Engineers) developed the SWEBOK (Software Engineering Body of Knowledge) in collaboration with software industry experts and academic researchers, introducing into the guide many of the needed knowledge for a 5-year expertise software engineer [7]. The SWEBOK also covers management from the perspective of a software project. This thesis is developed to provide guidance to practitioners and members of the academic community about project management applied to software engineering. The way used in this thesis to get useful information for practitioners is to take an industry-approved guide for software engineering professionals such as the SWEBOK, and compare the content to what is found in the PMBOK. After comparing the contents of the SWEBOK and the PMBOK, what is found missing in the SWEBOK is used to give recommendations on how to enrich project management skills for a software engineering professional. Recommendations for members of the academic community on the other hand, are given taking into account the GSwE2009 (Graduated Software Engineering 2009) standard [8]. GSwE2009 is often used as a main reference for software engineering master programs [9]. The standard is mostly based on the content of the SWEBOK, plus some contents that are considered to reinforce the education of software engineering. Given the similarities between the SWEBOK and the GSwE2009, the results of comparing SWEBOK and PMBOK are also considered valid to enrich what the GSwE2009 proposes. So in the end the recommendations for practitioners end up being also useful for the academic community and their strategies to teach project management in the context of software engineering.
Resumo:
To our knowledge, no current software development methodology explicitly describes how to transit from the analysis model to the software architecture of the application. This paper presents a method to derive the software architecture of a system from its analysis model. To do this, we are going to use MDA. Both the analysis model and the architectural model are PIMs described with UML 2. The model type mapping designed consists of several rules (expressed using OCL and natural language) that, when applied to the analysis artifacts, generate the software architecture of the application. Specifically the rules act on elements of the UML 2 metamodel (metamodel mapping). We have developed a tool (using Smalltalk) that permits the automatic application of these rules to an analysis model defined in RoseTM to generate the application architecture expressed in the architectural style C2.