386 resultados para Plugin eclipse


Relevância:

10.00% 10.00%

Publicador:

Resumo:

A new version of the TomoRebuild data reduction software package is presented, for the reconstruction of scanning transmission ion microscopy tomography (STIMT) and particle induced X-ray emission tomography (PIXET) images. First, we present a state of the art of the reconstruction codes available for ion beam microtomography. The algorithm proposed here brings several advantages. It is a portable, multi-platform code, designed in C++ with well-separated classes for easier use and evolution. Data reduction is separated in different steps and the intermediate results may be checked if necessary. Although no additional graphic library or numerical tool is required to run the program as a command line, a user friendly interface was designed in Java, as an ImageJ plugin. All experimental and reconstruction parameters may be entered either through this plugin or directly in text format files. A simple standard format is proposed for the input of experimental data. Optional graphic applications using the ROOT interface may be used separately to display and fit energy spectra. Regarding the reconstruction process, the filtered backprojection (FBP) algorithm, already present in the previous version of the code, was optimized so that it is about 10 times as fast. In addition, Maximum Likelihood Expectation Maximization (MLEM) and its accelerated version Ordered Subsets Expectation Maximization (OSEM) algorithms were implemented. A detailed user guide in English is available. A reconstruction example of experimental data from a biological sample is given. It shows the capability of the code to reduce noise in the sinograms and to deal with incomplete data, which puts a new perspective on tomography using low number of projections or limited angle.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

El objetivo de este proyecto es investigar la viabilidad del almacenamiento de CO2 en un acuífero salino profundo ubicado en el margen suroccidental de la Cuenca del Guadalquivir. Este proyecto está destinado a una operación industrial con tasas de emisión de CO2 superiores a medio millón de toneladas anuales. Se ha construido un modelo geológico de la formación almacén en Petrel y se ha simulado la inyección utilizando la versión composicional de ECLIPSE. El objetivo es inyectar CO2 manteniendo una tasa de inyección constante durante 30 años, el máximo periodo permitido por la legislación española sobre almacenamiento geológico de CO2. La cantidad de CO2 inyectada en cada uno de los casos ha sido determinada. Los resultados parecen indicar que la inyección de CO2 a escala industrial podría ser viable, aunque la viabilidad del proyecto podría verse comprometida por la escasa profundidad a la que se encuentra el contacto entre la formación almacén y el sello lateral. Antes de seguir adelante con el desarrollo del proyecto sería conveniente determinar mejor la continuidad lateral de la formación almacén y sus condiciones de sello. ABSTRACT The aim of this project is to investigate the feasibility of CO2 geological storage in a deep saline aquifer located onshore in the southwestern margin of Guadalquivir Basin. The project is addressed for an industrial scale operation with CO2 emission rates higher than half a million tons per year. A geological model of the target reservoir was built in Petrel and injection simulations were performed with the compositional version of ECLIPSE. The purpose is to inject CO2 at constant rate during 30 years, the maximum period allowed by the Spanish law on carbon dioxide geological storage. The amount of CO2 injected in each studied scenario has been determined. Results suggest that CO2 injection at industrial scale could be viable, but the project feasibility could be endangered by the shallow depth of the contact between the target reservoir and the lateral seal. Prior to injection, further work should include ascertaining the reservoir’s lateral continuity and better determination of its sealing conditions

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Simulation of satellite subsystems behaviour is extramely important in the design at early stages. The subsystems are normally simulated in the both ways : isolated and as part of more complex simulation that takes into account imputs from other subsystems (concurrent design). In the present work, a simple concurrent simulation of the power subsystem of a microsatellite, UPMSat-2, is described. The aim of the work is to obtain the performance profile of the system (battery charging level, power consumption by the payloads, power supply from solar panels....). Different situations such as battery critical low or high level, effects of high current charging due to the low temperature of solar panels after eclipse,DoD margins..., were analysed, and different safety strategies studied using the developed tool (simulator) to fulfil the mission requirements. Also, failure cases were analysed in order to study the robustness of the system. The mentioned simulator has been programed taking into account the power consumption performances (average and maximum consumptions per orbit/day) of small part of the subsystem (SELEX GALILEO SPVS modular generators built with Azur Space solar cells, SAFT VES16 6P4S Li-ion battery, SSBV magnetometers, TECNOBIT and DATSI/UPM On Board Data Handling -OBDH-...). The developed tool is then intended to be a modular simulator, with the chance of use any other components implementing some standard data.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Actualmente las redes sociales son muy utilizadas en todo el mundo, existen diferentes tipos de redes sociales con las que podemos conectar con amigos, ampliar nuestra red de contactos profesionales, aprender cosas nuevas, etc. Su elevado uso las ha convertido en uno de los negocios más rentables en internet generando fortunas para sus creadores, principalmente, a través de la publicidad. Muchas redes sociales son creadas por empresas que cuentan con un equipo experto, pero otras muchas han sido creadas por personas comunes, con bajos conocimientos informáticos, muchos motivados en sus aficiones o carreras profesionales, que al no encontrar nada en la red que les resulte útil, han decidido desarrollar ellos mismos sus propias redes sociales con la ayuda de herramientas informáticas. Una de esas herramientas son los sistemas de gestión de contenidos (CMS), con los cuales ahorraremos mucho tiempo de desarrollo y no necesitaremos invertir grandes cantidades de dinero. Este proyecto trata, principalmente, de cómo crear redes sociales haciendo uso de estas herramientas y tiene el objetivo de ser lo suficientemente claro para que cualquier persona, sin importar su nivel de conocimientos técnicos, sea capaz de desarrollar sus ideas. En la primera parte del proyecto se habla sobre las redes sociales en general y el impacto de éstas en la sociedad actual, donde se ve que, actualmente, debido a la cantidad de redes sociales y a la accesibilidad móvil, el uso de las redes sociales es una cotidianidad. También se explican algunos métodos para obtener beneficios económicos de una red social y las ventajas que presentan las redes sociales específicas frente a las generalistas, concluyendo que las redes sociales específicas van ganando mayor protagonismo con el paso del tiempo. Además, se habla sobre la crítica de las redes sociales desde el punto de vista del usuario de la red, donde se resalta el tema de la administración de la publicidad y la privatización que han supuesto las redes sociales. Posteriormente se presenta una base teórica sobre las herramientas antes mencionadas, los CMS. Se explica su funcionamiento, clasificación y las ventajas que obtenemos con el uso de este software en nuestros proyectos, de los cuales destacan el corto tiempo de desarrollo y el bajo coste. Al final se eligen los CMS de estudio en este proyecto principalmente en base a tres criterios: licencia, cuota de uso y características de red social. En la segunda parte del proyecto se habla acerca de los CMS elegidos: WordPress con su plugin BuddyPress, Elgg, y Joomla con su plugin JomSocial. Se explican las características de cada uno de ellos y se muestran ejemplos de redes sociales reales hechas con estos CMS. En esta parte del proyecto se hace un uso práctico de estos CMS y se detalla paso a paso todo el proceso de creación de una red social (instalación, configuración y personalización de la red social) para cada CMS. El resultado son tres redes sociales hechas con distintos CMS, de los cuales se hace una valoración en base a la experiencia obtenida con el uso de los mismos, concluyéndose que JomSocial es una buena opción para redes sociales de uso generalista, pero para redes específicas son mejores tanto Elgg como BuddyPress, presentando este último una ligera ventaja por tener una gran comunidad en español. ABSTRACT. Nowadays social networks are widely used throughout the world, there are different types of social networks where you can connect with friends, expand your network of professional contacts, learn new things, etc. Its high usage has turned them into one of the most profitable businesses on Internet generating fortunes to its developers, mainly through advertising. Many social networks are developed by companies that count on an expert team, but many others have been created by ordinary people, with low computer skills, many of them motivated in their hobbies or careers, that did not find anything useful on Internet and decide to develop their own social networks with the help of software tools. One of those tools is a content management system (CMS), which will help us to save a lot of development time and we will not need to invest large amounts of money. This project is, mainly, about how to create social networks using these tools and aims to be clear enough to help anyone, regardless of their computer skills, to develop their ideas. The first part of the project is about social networks in general and the impact on today's society, where we can see that, due to the number of social networks and mobile accessibility, the use of social networks is daily. Also it explains some ways to obtain economic benefits from a social network and the advantages of specific social networks against generalist social networks, concluding that specific social networks are gaining more prominence with the passage of time. In addition, it refers on social networks critique from the point of view of social network users, where it highlights the issue of the advertising administration and privatization which have brought social networks. Subsequently, it presents a theoretical base of the above mentioned tools, CMS. Explains their operation, classification and the advantages we get with the use of this software in our projects, where the short development time and lower cost are highlighted. At the end the CMS studied in this project are chosen mainly based on three criteria: license, community size and social network features. The second part of the project is about the chosen CMS: WordPress with its plugin BuddyPress, Elgg, and Joomla with its plugin JomSocial. It explains features of all of them and shows examples of real social networks developed with these CMS. This part of the project is a practical use of these CMS and detailed step by step throughout the process of creating a social network (installation, configuration and customization of the social network) for each CMS. The result are three social networks made with different CMS, from which is made an assessment on the basis of the experience gained with the use of these software, concluding that JomSocial is a good choice to develop generalist social networks, but for specific social networks are better Elgg and BuddyPress, presenting the latter a slight advantage by having a large community in Spanish.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Los debates contemporáneos sobre el eclipse del espacio público suelen ofrecer una visión idealizada de su pasado, impidiendo una correcta reconsideración del papel que las técnicas urbanísticas han tenido en la producción y evolución del mismo. Contribuyendo al desarrollo de una historia crítica que subsane estas lagunas, Central Park se presenta aquí como un dispositivo gubernamental ideado para sustituir el régimen de uso del espacio público habitual en las calles de Manhattan a mediados del siglo XIX, por un nuevo conjunto de prácticas espaciales definidas y monitorizadas por el Estado. Tras una descripción de las formas de apropiaciones espontáneas de la calle por parte de las clases populares, se analizan los distintos niveles proyectuales en los cuales se articulaba esta estrategia de domesticación del espacio público: del propio diseño espacial y concepción de la red de lugares del parque, a la regulación normativa del uso y comportamiento de los visitantes, al ejercicio activo de vigilancia y castigo de conductas y sujetos indeseables.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Context. This thesis is framed in experimental software engineering. More concretely, it addresses the problems arisen when assessing process conformance in test-driven development experiments conducted by UPM's Experimental Software Engineering group. Process conformance was studied using the Eclipse's plug-in tool Besouro. It has been observed that Besouro does not work correctly in some circumstances. It creates doubts about the correction of the existing experimental data which render it useless. Aim. The main objective of this work is the identification and correction of Besouro's faults. A secondary goal is fixing the datasets already obtained in past experiments to the maximum possible extent. This way, existing experimental results could be used with confidence. Method. (1) Testing Besouro using different sequences of events (creation methods, assertions etc..) to identify the underlying faults. (2) Fix the code and (3) fix the datasets using code specially created for this purpose. Results. (1) We confirmed the existence of several fault in Besouro's code that affected to Test-First and Test-Last episode identification. These faults caused the incorrect identification of 20% of episodes. (2) We were able to fix Besouro's code. (3) The correction of existing datasets was possible, subjected to some restrictions (such us the impossibility of tracing code size increase to programming time. Conclusion. The results of past experiments dependent upon Besouro's data could no be trustable. We have the suspicion that more faults remain in Besouro's code, whose identification requires further analysis.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

PAMELA (Phased Array Monitoring for Enhanced Life Assessment) SHMTM System is an integrated embedded ultrasonic guided waves based system consisting of several electronic devices and one system manager controller. The data collected by all PAMELA devices in the system must be transmitted to the controller, who will be responsible for carrying out the advanced signal processing to obtain SHM maps. PAMELA devices consist of hardware based on a Virtex 5 FPGA with a PowerPC 440 running an embedded Linux distribution. Therefore, PAMELA devices, in addition to the capability of performing tests and transmitting the collected data to the controller, have the capability of perform local data processing or pre-processing (reduction, normalization, pattern recognition, feature extraction, etc.). Local data processing decreases the data traffic over the network and allows CPU load of the external computer to be reduced. Even it is possible that PAMELA devices are running autonomously performing scheduled tests, and only communicates with the controller in case of detection of structural damages or when programmed. Each PAMELA device integrates a software management application (SMA) that allows to the developer downloading his own algorithm code and adding the new data processing algorithm to the device. The development of the SMA is done in a virtual machine with an Ubuntu Linux distribution including all necessary software tools to perform the entire cycle of development. Eclipse IDE (Integrated Development Environment) is used to develop the SMA project and to write the code of each data processing algorithm. This paper presents the developed software architecture and describes the necessary steps to add new data processing algorithms to SMA in order to increase the processing capabilities of PAMELA devices.An example of basic damage index estimation using delay and sum algorithm is provided.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

La conciencia de la crisis de la modernidad -que comienza ya a finales del siglo XIX- ha cobrado más experiencia debido al conocimiento de los límites del desarrollo económico, ya que como parecía razonable pensar, también los recursos naturales son finitos. En 1972, el Club de Roma analizó las distintas opciones disponibles para conseguir armonizar el desarrollo sostenible y las limitaciones medioambientales. Fue en 1987 cuando la Comisión Mundial para el Medio Ambiente y el Desarrollo de la ONU definía por primera vez el concepto de desarrollo sostenible. Definición que posteriormente fue incorporada en todos los programas de la ONU y sirvió de eje, por ejemplo, a la Cumbre de la Tierra celebrada en Río de Janeiro en 1992. Parece evidente que satisfacer la demanda energética, fundamentalmente desde la Revolución Industrial en el s XIX, trajo consigo un creciente uso de los combustibles fósiles, con la consiguiente emisión de los gases de efecto invernadero (GEI) y el aumento de la temperatura global media terrestre. Esta temperatura se incrementó en los últimos cien años en una media de 0.74ºC. La mayor parte del incremento observado desde la mitad del siglo XX en esta temperatura media se debe, con una probabilidad de al menos el 90%, al aumento observado en los GEI antropogénicos, siendo uno de ellos el CO2 que proviene de la transformación del carbono de los combustibles fósiles durante su combustión. Ante el creciente uso de los combustibles fósiles, los proyectos CAC, proyectos de captura, transporte y almacenamiento, se presentan como una contribución al desarrollo sostenible ya que se trata de una tecnología que permite mitigar el cambio climático. Para valorar si la tecnología CAC es sostenible, habrá que comprobar si existe o no capacidad para almacenar el CO2 en una cantidad mayor a la de producción y durante el tiempo necesario que impone la evolución de la concentración de CO2 en la atmósfera para mantenerla por debajo de las 450ppmv (concentración de CO2 que propone el Panel Intergubernamental para el Cambio Climático). El desarrollo de los proyectos CAC completos pasa por la necesaria selección de adecuados almacenes de CO2 que sean capaces de soportar los efectos de las presiones de inyección, así como asegurar la capacidad de dichos almacenes y la estanqueidad del CO2 en los mismos. La caracterización geológica de un acuífero susceptible de ser almacén de CO2 debe conducir a determinar las propiedades que dicho almacén posee para asegurar un volumen adecuado de almacenamiento, una inyectabilidad del CO2 en el mismo a un ritmo adecuado y la estanqueidad del CO2 en dicho acuífero a largo plazo. El presente trabajo pretende estudiar los parámetros que tienen influencia en el cálculo de la capacidad del almacén, para lo que en primer lugar se ha desarrollado la tecnología necesaria para llevar a cabo la investigación mediante ensayos de laboratorio. Así, se ha desarrollado una patente, "ATAP, equipo para ensayos petrofísicos (P201231913)", con la que se ha llevado a cabo la parte experimental de este trabajo para la caracterización de los parámetros que tienen influencia en el cálculo de la capacidad del almacén. Una vez desarrollada la tecnología, se aborda el estudio de los distintos parámetros que tienen influencia en la capacidad del almacén realizando ensayos con ATAP. Estos ensayos definen el volumen del almacenamiento, llegándose a la conclusión de que en la determinación de este volumen, juegan un papel importante el alcance de los mecanismos trampa, físicos o químicos, del CO2 en el almacén. Ensayos que definen la capacidad del almacén de "aceptar" o "rechazar" el CO2 inyectado, la inyectabilidad, y por último, ensayos encaminados a determinar posibles fugas que se pueden dar a través de los pozos de inyección, definidos estos como caminos preferenciales de fugas en un almacén subterráneo de CO2. Queda de este modo caracterizada la estanqueidad del CO2 en el acuífero a largo plazo y su influencia obvia en la determinación de la capacidad del almacén. Unido al propósito de la estimación de la capacidad del almacén, se encuentra el propósito de asegurar la estanqueidad de dichos almacenes en el tiempo, y adelantarse a la evolución de la pluma de CO2 en el interior de dichos almacenes. Para cumplir este propósito, se ha desarrollado un modelo dinámico a escala de laboratorio, mediante el programa ECLIPSE 300, con el fin de establecer una metodología para el cálculo de la capacidad estimada del almacén, así como el estudio de la evolución de la pluma de CO2 dentro del acuífero a lo largo del tiempo, partiendo de los resultados obtenidos en los ensayos realizados en ATAP y con la modelización de la probeta de roca almacén empleada en dichos ensayos. Presentamos por tanto un trabajo que establece las bases metodológicas para el estudio de la influencia de distintos parámetros petrofísicos en el cálculo de la capacidad del almacén unidos al desarrollo tecnológico de ATAP y su utilización para la determinación de dichos parámetros aplicables a cada acuífero concreto de estudio. ABSTRACT The crisis of modernity –which begins at the end of 19th Century- has been more important due to the knowledge of the limits of economic development, since it appeared to be thought reasonable, the natural resources are finite. In 1972, The Club of Rome analyzed the different options available in order to harmonize the sustainability and the environment development. It was in 1987 when The Global Commission on The Environment and the Development of UN, defined for the first time the concept of Sustainable Development. This definition that was fully incorporated in all the UN programs and it was useful as an axis, for example, in La Cumbre de la Tierra summit in Río de Janeiro in 1992. It seems obvious to satisfy energetic demand, basically after The Industrial Revolution in 19th Century, which represented an increasing use of fossil fuels, therefore greenhouse gases emission and the increasing of global average temperature. This temperature increased in the last 100 years up to 0.74ºC. The major part of the temperature increase is due to the increase observed in Greenhouse gases with human origin, at least with 90% of probability. The most important gas is the CO2 because of its quantity. In the face of the increasing use of fossil fuels, the CCS projects, Carbon Capture and Storage projects, appear as a contribution of sustainable development since it is a technology for avoiding the climate change. In order to evaluate if CCS technology is sustainable, it will be necessary to prove if the capacity for CO2 storage is available or not in a quantity greater than the production one and during the time necessary to keep the CO2 concentration in the atmosphere lower than 450ppmv (concentration imposed by IPCC). The development of full CCS projects goes through the selection of good CO2 storages that are able to support the effects of pressure injection, and assure the capacity of such storages and the watertightness of CO2. The geological characterization of the aquifer that could be potential CO2 storage should lead to determine the properties that such storage has in order to assure the adequate storage volume, the CO2 injectivity in a good rate, and the watertightness of the CO2 in the long term. The present work aims to study the parameters that have influence on the calculation of storage capacity, and for that purpose the appropriate technology has been developed for carrying out the research by mean of laboratory tests. Thus, a patent has been developed, "ATAP, equipo para ensayos petrofísicos (P201231913)", that has been used for developing the experimental part of this work. Once the technology has been developed, the study of different parameters, that have influence on the capacity of the storage, has been addressed developing different tests in ATAP. These tests define the storage volume which is related to the scope of different CO2 trap mechanisms, physical or chemical, in the storage. Tests that define the capacity of the storage to “accept” or “reject” the injected CO2, the injectivity, and tests led to determine possible leakages through injection wells. In this way we could talk about the watertightness in the aquifer in the long term and its influence on the storage capacity estimation. Together with the purpose of the storage capacity estimation, is the purpose of assuring the watertightness of such storages in the long term and anticipating the evolution of CO2 plume inside such aquifers. In order to fulfill this purpose, a dynamic model has been developed with ECLIPSE 300, for stablishing the methodology for the calculation of storage capacity estimation and the evolution of the CO2 plume, starting out with the tests carried out in ATAP. We present this work that establishes the methodology bases for the study of the influence of different petrophysics parameters in the calculation of the capacity of the storage together with the technological development of ATAP and its utilization for the determination of such parameters applicable to each aquifer.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

La evolución de las redes eléctricas se dirige hacia lo que se conoce como “Smart Grids” o “Redes Eléctricas Inteligentes”. Estas “Smart Grids” se componen de subestaciones eléctricas, que a su vez se componen de unos dispositivos llamados IEDs (Dispositivos Electrónicos Inteligentes – Intelligent Electronic Devices). El diseño de IEDs se encuentra definido en la norma IEC 61850, que especifica además un Lenguaje de Configuración de Subestaciones (Substation Configuration Language SCL) para la definición de la configuración de subestaciones y sus IEDs. Hoy en día, este estándar internacional no sólo se utiliza para diseñar correctamente IEDs y asegurar su interoperabilidad, sino que también se utiliza para el diseño de otros dispositivos de la red eléctrica, como por ejemplo, medidores inteligentes. Sin embargo, aunque existe una tendencia cada vez mayor del uso de este estándar, la comprensión y el manejo del mismo resulta difícil debido al gran volumen de información que lo compone y del nivel de detalle que utiliza, por lo que su uso para el diseño de IEDs se hace tedioso sin la ayuda de un soporte software. Es por ello que, para facilitar la aplicación del estándar IEC 61850 en el diseño de IEDs se han desarrollado herramientas como “Visual SCL”, “SCL Explorer” o “61850 SCLVisual Design Tool”. En concreto, “61850 SCLVisual Design Tool” es una herramienta gráfica para el modelado de subestaciones electricas, generada mediante el uso de los frameworks Eclipse Modeling Framework (EMF) y Epsilon Generative Modeling Technologies (GMT) y desarrollada por el grupo de investigación SYST de la UPM. El objetivo de este proyecto es añadir una nueva funcionalidad a la herramienta “61850 Visual SCL DesignTool”. Esta nueva funcionalidad consiste en la generación automática de un fichero de configuración de subestaciones eléctricas según el estándar IEC 61850 a partir de de una herramienta de diseño gráfico. Este fichero, se denomina SCD (Substation Configuration Description), y se trata de un fichero XML conforme a un esquema XSD (XML Schema Definition) mediante el que se define el lenguaje de configuración de subestaciones SCL del IEC 61850. Para el desarrollo de este proyecto, es necesario el estudio del lenguaje para la configuración de subestaciones SCL, así como del lenguaje gráfico específico de dominio definido por la herramienta “61850 SCLVisual Design Tool”, la estructura de los ficheros SCD, y finalmente, del lenguaje EGL (Epsilon Generation Language) para la transformación y generación automática de código a partir de modelos EMF. ABSTRACT Electrical networks are evolving to “Smart Grids”. Smart Grids are composed of electrical substations that in turn are composed of devices called IEDs (Intelligent Electronic Devices). The design of IEDs is defined by the IEC 61850 standard, which also specifies a Substation Configuration Languaje (SCL) used to define the configuration of substations and their IEDs. Nowadays, this international standard is not only used to design properly IEDs and guarantee their interoperability, but it is also used to design different electrical network devices, such as, smart meters. However, although the use of this standard is growing, its compression as well as its management, is still difficult due to its large volume of information and its level of detail. As a result, designing IEDs becomes a tedious task without a software support. As a consequence of this, in order to make easier the application of the IEC 61850 standard while designing IEDs, some software tools have been developed, such as: “Visual SCL”, “SCL Explorer” or “61850 SCLVisual Design Tool”. In particular, “61850 SCLVisual Design Tool” is a graphical tool used to make electrical substations models, and developed with the Eclipse Modeling Framework (EMF) and Epsilon Generative Modeling Technologies (GMT) by the research group SYST of the UPM. The aim of this project is to add a new functionality to “61850 Visual SCL DesignTool”. This new functionality consists of the automatic code generation of a substation configuration file according to the IEC 61850 standard. This file is called SCD (Substation Configuration Description), and it is a XML file that follows a XSD (XML Schema Definition) that defines the Substation Configuration Language (SCL) of the IEC 61850. In order to develop this project, it is necessary to study the Substation Configuration Language (SCL), the domain-specific graphical languaje defined by the tool “61850 SCLVisual Design Tool”, the structure of a SCD file, and the Epsilon Generation Language (EGL) used for the automatic code generation from EMF models

Relevância:

10.00% 10.00%

Publicador:

Resumo:

El proyecto “Aplicación móvil y web para la gestión de lugares geolocalizados (www.midiez.com)” tiene como objetivo principal crear un repositorio de listas categorizadas de sitios para su uso en el ámbito personal o comercial. Tanto la aplicación web como la aplicación móvil desarrollada en Android tienen el propósito de gestionar listas de lugares de interés (Restaurantes, tiendas,..) o con propósitos específicos (Organización de viajes) o simplemente como una forma de anotar aquellos sitios que nos comentan y que nos gustaría visitar. El desarrollo de este proyecto además permitirá contrastar las distintas alternativas y la evolución de las distintas herramientas que se han ido desarrollando para la gestión del ocio en los últimos años desde el sistema Android y plataformas web. Todo el proyecto ha sido realizado usando software libre (PHP para el lenguaje web servidor y Java para la programación móvil). La principal finalidad desde el punto de vista del desarrollador es: aprovechar las sinergias de la programación móvil y la programación web de manera que las mismas capas de negocio de Datos sean usadas por ambas plataformas. Asimismo crear una aplicación distribuida y fácilmente escalable. Las herramientas que se han usado para desarrollar han sido: la SDK proporcionada por Google, una JDK de Java y un IDE de desarrollo Java como es Eclipse y otro similar para el desarrollo de la parte PHP. La BBDD elegida ha sido MySQL. El proyecto pretende mostrar el potencial de las aplicaciones móviles geolocalizadas desde el punto de vista del ocio y compararlas con el estado del arte actual. Por lo tanto la mayor parte del tiempo dedicado al proyecto ha sido empleado en el desarrollo de la aplicación web, la aplicación móvil y en la base de datos pero también he dedicado una pequeña parte del trabajo para realizar un estudio sobre las consecuencias que esta tecnología está teniendo en nuestros cerebros. ABSTRACT The project "Web and Mobile App for managing geolocation places” has as main objective managing of places lists in order to use them in the leisure time scope. Nowadays the use of GPS is being a constant in mobile applications so that is already part of our daily life. We used to know where we are always and at the same time we can find locations using the technology of our mobile phones. Now it is very difficult to get lost outside but also is difficult to explain somebody how to get to anywhere without using Google Maps. Google Maps, Geolocation, gps navigators, … all that kind of stuff are making our life easier and less complicated but also are making our brains lazier. Furthermore, the development of this project will use the potential of locate places into maps to avoid annotate every spot we would like to visit or a brand new restaurant. The project itself shows the location features of Google Maps combined with an places data base in order to create, and manage places lists and use them to get to them as well as to share those places with our contacts. Also, the main purpose from the point of view of the developer is to combine different programming languages and use the resulting synergies in a easily scalable and portable environment. The tools that have been used to develop are: the SDK provided by Google, one JDK Java and Java development IDE such as Eclipse and similar to the development of the PHP part. The DB has been chosen MySQL. Finally, this project aims to show, from an educational point of view, the use and potential of this technology. Thus, it has been devoted a large amount of time of the project (and, consequently, its documentation) on develop the Android app, the data base and the web app but also but also to highlight the consequences of using technology.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Reconhecendo, a partir da constatação empírica, a multiplicidade de escolhas de crenças no Mundo e em particular na periferia urbana paulistana, reconhecemos, também, a emergência criativa de novas possibilidades de crer e não crer. Tal amplitude não apenas aponta para o crer (segundo as ofertas de um sem número de religiões) e o não crer (ateu e agnóstico), mas para uma escolha que poderia vir a ser silenciada e esquecida, neste binômio arcaico e obsoleto, quando alguém se dá à liberdade crer sem ter religião. Reconhecer interessadamente os sem-religião nas periferias urbanas paulistanas é dar-se conta das violências a que estes indivíduos estão submetidos: violência econômica, violência da cidadania (vulnerabilidade) e proveniente da armas (grupos x Estado). Tanto quanto a violência do esquecimento e silenciamento. A concomitância espaço-temporal dos sem-religião nas periferias, levou-nos buscar referências em teorias de secularização e de laicidade, e, a partir destas, traçar uma história do poder violento, cuja pretensão é a inelutabilidade, enquanto suas fissuras são abertas em espaços de resistências. A história da legitimação do poder que se quer único, soberano, de caráter universal, enquanto fragmenta a sociedade em indivíduos atomizados, fragilizando vínculos horizontais, e a dos surgimentos de resistências não violentas questionadoras da totalidade trágica, ao reconhecer a liberdade de ser com autonomia, enquanto se volta para a produção de partilha de bens comuns. Propomos reconhecer a igual liberdade de ser (expressa na crença da filiação divina) e de partilhar o bem comum em reconhecimentos mútuos (expressa pela ação social), uma expressão de resistência não violenta ao poder que requer a igual abdicação da liberdade pela via da fragmentação individualizante e submissão inquestionável à ordem totalizante. Os sem-religião nas periferias urbanas, nossos contemporâneos, partilhariam uma tal resistência, ao longo da história, com as melissas gregas, os profetas messiânicos hebreus, os hereges cristãos e os ateus modernos, cuja pretensão não é o poder, mas a partilha igual da liberdade e dos bens comuns. Estes laicos, de fato, seriam agentes de resistências de reconhecimento mútuos, em espaços de multiplicidade crescente, ao poder violento real na história.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

A necessidade de obter solução de grandes sistemas lineares resultantes de processos de discretização de equações diferenciais parciais provenientes da modelagem de diferentes fenômenos físicos conduz à busca de técnicas numéricas escaláveis. Métodos multigrid são classificados como algoritmos escaláveis.Um estimador de erros deve estar associado à solução numérica do problema discreto de modo a propiciar a adequada avaliação da solução obtida pelo processo de aproximação. Nesse contexto, a presente tese caracteriza-se pela proposta de reutilização das estruturas matriciais hierárquicas de operadores de transferência e restrição dos métodos multigrid algébricos para acelerar o tempo de solução dos sistemas lineares associados à equação do transporte de contaminantes em meio poroso saturado. Adicionalmente, caracteriza-se pela implementação das estimativas residuais para os problemas que envolvem dados constantes ou não constantes, os regimes de pequena ou grande advecção e pela proposta de utilização das estimativas residuais associadas ao termo de fonte e à condição inicial para construir procedimentos adaptativos para os dados do problema. O desenvolvimento dos códigos do método de elementos finitos, do estimador residual e dos procedimentos adaptativos foram baseados no projeto FEniCS, utilizando a linguagem de programação PYTHONR e desenvolvidos na plataforma Eclipse. A implementação dos métodos multigrid algébricos com reutilização considera a biblioteca PyAMG. Baseado na reutilização das estruturas hierárquicas, os métodos multigrid com reutilização com parâmetro fixo e automática são propostos, e esses conceitos são estendidos para os métodos iterativos não-estacionários tais como GMRES e BICGSTAB. Os resultados numéricos mostraram que o estimador residual captura o comportamento do erro real da solução numérica, e fornece algoritmos adaptativos para os dados cuja malha retornada produz uma solução numérica similar à uma malha uniforme com mais elementos. Adicionalmente, os métodos com reutilização são mais rápidos que os métodos que não empregam o processo de reutilização de estruturas. Além disso, a eficiência dos métodos com reutilização também pode ser observada na solução do problema auxiliar, o qual é necessário para obtenção das estimativas residuais para o regime de grande advecção. Esses resultados englobam tanto os métodos multigrid algébricos do tipo SA quanto os métodos pré-condicionados por métodos multigrid algébrico SA, e envolvem o transporte de contaminantes em regime de pequena e grande advecção, malhas estruturadas e não estruturadas, problemas bidimensionais, problemas tridimensionais e domínios com diferentes escalas.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper introduces the Sm4RIA Extension for OIDE, which implements the Sm4RIA approach in OIDE (OOH4RIA Integrated Development Environment). The application, based on the Eclipse framework, supports the design of the Sm4RIA models as well as the model-to-model and model-to-text transformation processes that facilitate the generation of Semantic Rich Internet Applications, i.e., RIA applications capable of sharing data as Linked data and consuming external data from other sources in the same manner. Moreover, the application implements mechanisms for the creation of RIA interfaces from ontologies and the automatic generation of administration interfaces for a previously design application.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We present analysis of 100 ks contiguous XMM-Newton data of the prototypical wind accretor Vela X-1. The observation covered eclipse egress between orbital phases 0.134 and 0.265, during which a giant flare took place, enabling us to study the spectral properties both outside and during the flare. This giant flare with a peak luminosity of 3.92+0.42-0.09 × 1037 erg s-1 allows estimates of the physical parameters of the accreted structure with a mass of ~1021 g. We have been able to model several contributions to the observed spectrum with a phenomenological model formed by three absorbed power laws plus three emission lines. After analysing the variations with orbital phase of the column density of each component, as well as those in the Fe and Ni fluorescence lines, we provide a physical interpretation for each spectral component. Meanwhile, the first two components are two aspects of the principal accretion component from the surface of the neutron star, the third component seems to be the X-ray light echo formed in the stellar wind of the companion.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Vela X–1 is the prototype of the class of wind-fed accreting pulsars in high-mass X-ray binaries hosting a supergiant donor. We have analysed in a systematic way 10 years of INTEGRAL data of Vela X–1 (22–50 keV) and we found that when outside the X-ray eclipse, the source undergoes several luminosity drops where the hard X-rays luminosity goes below ∼3 × 1035 erg s−1, becoming undetected by INTEGRAL. These drops in the X-ray flux are usually referred to as ‘off-states’ in the literature. We have investigated the distribution of these off-states along the Vela X–1 ∼ 8.9 d orbit, finding that their orbital occurrence displays an asymmetric distribution, with a higher probability to observe an off-state near the pre-eclipse than during the post-eclipse. This asymmetry can be explained by scattering of hard X-rays in a region of ionized wind, able to reduce the source hard X-ray brightness preferentially near eclipse ingress. We associate this ionized large-scale wind structure with the photoionization wake produced by the interaction of the supergiant wind with the X-ray emission from the neutron star. We emphasize that this observational result could be obtained thanks to the accumulation of a decade of INTEGRAL data, with observations covering the whole orbit several times, allowing us to detect an asymmetric pattern in the orbital distribution of off-states in Vela X–1.