960 resultados para equivalent web thickness method
Resumo:
Context: This paper addresses one of the major end-user development (EUD) challenges, namely, how to pack today?s EUD support tools with composable elements. This would give end users better access to more components which they can use to build a solution tailored to their own needs. The success of later end-user software engineering (EUSE) activities largely depends on how many components each tool has and how adaptable components are to multiple problem domains. Objective: A system for automatically adapting heterogeneous components to a common development environment would offer a sizeable saving of time and resources within the EUD support tool construction process. This paper presents an automated adaptation system for transforming EUD components to a standard format. Method: This system is based on the use of description logic. Based on a generic UML2 data model, this description logic is able to check whether an end-user component can be transformed to this modeling language through subsumption or as an instance of the UML2 model. Besides it automatically finds a consistent, non-ambiguous and finite set of XSLT mappings to automatically prepare data in order to leverage the component as part of a tool that conforms to the target UML2 component model. Results: The proposed system has been successfully applied to components from four prominent EUD tools. These components were automatically converted to a standard format. In order to validate the proposed system, rich internet applications (RIA) used as an operational support system for operators at a large services company were developed using automatically adapted standard format components. These RIAs would be impossible to develop using each EUD tool separately. Conclusion: The positive results of applying our system for automatically adapting components from current tool catalogues are indicative of the system?s effectiveness. Use of this system could foster the growth of web EUD component catalogues, leveraging a vast ecosystem of user-centred SaaS to further current EUSE trends.
Resumo:
This paper presents a Focused Crawler in order to Get Semantic Web Resources (CSR). Structured data web are available in formats such as Extensible Markup Language (XML), Resource Description Framework (RDF) and Ontology Web Language (OWL) that can be used for processing. One of the main challenges for performing a manual search and download semantic web resources is that this task consumes a lot of time. Our research work propose a focused crawler which allow to download these resources automatically and store them on disk in order to have a collection that will be used for data processing. CRS consists of three layers: (a) The User Interface Layer, (b) The Focus Crawler Layer and (c) The Base Crawler Layer. CSR uses as a selection policie the Shark-Search method. CSR was conducted with two experiments. The first one starts on December 15 2012 at 7:11 am and ends on December 16 2012 at 4:01 were obtained 448,123,537 bytes of data. The CSR ends by itself after to analyze 80,4375 seeds with an unlimited depth. CSR got 16,576 semantic resources files where the 89 % was RDF, the 10 % was XML and the 1% was OWL. The second one was based on the Web Data Commons work of the Research Group Data and Web Science at the University of Mannheim and the Institute AIFB at the Karlsruhe Institute of Technology. This began at 4:46 am of June 2 2013 and 1:37 am June 9 2013. After 162.51 hours of execution the result was 285,279 semantic resources where predominated the XML resources with 99 % and OWL and RDF with 1 % each one.
Resumo:
Laser peening has recently emerged as a useful technique to overcome detrimental effects associated to another well-known surface modification processes such as shot peening or grit blasting used in the biomedical field. It is worth to notice that besides the primary residual stress effect, thermally induced effects might also cause subtle surface and subsurface microstructural changes that might influence corrosion resistance. Moreover, since maximum loads use to occur at the surface, they could also play a critical role in the fatigue strength. In this work, plates of Ti-6Al-4V alloy of 7 mm in thickness were modified by laser peening without using a sacrificial outer layer. Irradiation by a Q-switched Nd-YAG laser (9.4 ns pulse length) working in fundamental harmonic at 2.8 J/pulse and with water as confining medium was used. Laser pulses with a 1.5 mm diameter at an equivalent overlapping density (EOD) of 5000 cm-2 were applied. Attempts to analyze the global induced effects after laser peening were addressed by using the contacting and non-contacting thermoelectric power (TEP) techniques. It was demonstrated that the thermoelectric method is entirely insensitive to surface topography while it is uniquely sensitive to subtle variations in thermoelectric properties, which are associated with the different material effects induced by different surface modification treatments. These results indicate that the stress-dependence of the thermoelectric power in metals produces sufficient contrast to detect and quantitatively characterize regions under compressive residual stress based on their thermoelectric power contrast with respect to the surrounding intact material. However, further research is needed to better separate residual stress effects from secondary material effects, especially in the case of low-conductivity engineering materials like titanium alloys.
Resumo:
El empleo de los micropilotes en la ingeniería civil ha revolucionado las técnicas de estabilización de terraplenes a media ladera, ya que aunque los pilotes pueden ser la opción más económica, el uso de micropilotes permite llegar a sitios inaccesibles con menor coste de movimientos de tierras, realización de plataformas de trabajo de dimensiones reducidas, maquinaria necesaria es mucho más pequeña, liviana y versátil en su uso, incluyendo la posibilidad de situar la fabricación de morteros o lechadas a distancias de varias decenas de metros del elemento a ejecutar. Sin embargo, realizando una revisión de la documentación técnica que se tiene en el ámbito ingenieril, se comprobó que los sistemas de diseño de algunos casos (micropilotes en terraplenes a media ladera, micropilotes en pantallas verticales, micropilotes como “paraguas” en túneles, etc.) eran bastante deficientes o poco desarrollados. Premisa que permite concluir que el constructor ha ido por delante (como suele ocurrir en ingeniería geotécnica) del cálculo o de su análisis teórico. Del mismo modo se determinó que en su mayoría los micropilotes se utilizan en labores de recalce o como nueva solución de cimentación en condiciones de difícil acceso, casos en los que el diseño de los micropilotes viene definido por cargas axiales, de compresión o de tracción, consideraciones que se contemplan en reglamentaciones como la “Guía para el proyecto y la ejecución de micropilotes en obras de carretera” del Ministerio de Fomento. En los micropilotes utilizados para estabilizar terraplenes a media ladera y micropilotes actuando como muros pantalla, en los que éstos trabajan a esfuerzo cortante y flexión, no se dispone de sistemas de análisis fiables o no se introduce adecuadamente el problema de interacción terreno-micropilote. Además en muchos casos, los parámetros geotécnicos que se utilizan no tienen una base técnico-teórica adecuada por lo que los diseños pueden quedar excesivamente del lado de la seguridad, en la mayoría de los casos, o todo lo contrario. Uno de los objetivos principales de esta investigación es estudiar el comportamiento de los micropilotes que están sometidos a esfuerzos de flexión y cortante, además de otros objetivos de gran importancia que se describen en el apartado correspondiente de esta tesis. Cabe indicar que en este estudio no se ha incluido el caso de micropilotes quasi-horizontales trabajando a flexion (como los “paraguas” en túneles), por considerarse que estos tienen un comportamiento y un cálculo diferente, que está fuera del alcance de esta investigación. Se ha profundizado en el estudio del empleo de micropilotes en taludes, presentando casos reales de obras ejecutadas, datos estadísticos, problemas de diseño y ejecución, métodos de cálculo simplificados y modelación teórica en cada caso, efectuada mediante el empleo de elementos finitos con el Código Plaxis 2D. Para llevar a cabo los objetivos que se buscan con esta investigación, se ha iniciado con el desarrollo del “Estado del Arte” que ha permitido establecer tipología, aplicaciones, características y cálculo de los micropilotes que se emplean habitualmente. Seguidamente y a efectos de estudiar el problema dentro de un marco geotécnico real, se ha seleccionado una zona española de actuación, siendo ésta Andalucía, en la que se ha utilizado de manera muy importante la técnica de micropilotes sobre todo en problemas de estabilidad de terraplenes a media ladera. A partir de ahí, se ha realizado un estudio de las propiedades geotécnicas de los materiales (principalmente suelos y rocas muy blandas) que están presentes en esta zona geográfica, estudio que ha sido principalmente bibliográfico o a partir de la experiencia en la zona del Director de esta tesis. Del análisis realizado se han establecido ordenes de magnitud de los parámetros geotécnicos, principalmente la cohesión y el ángulo de rozamiento interno (además del módulo de deformación aparente o de un módulo de reacción lateral equivalente) para los diversos terrenos andaluces. Con el objeto de conocer el efecto de la ejecución de un micropilote en el terreno (volumen medio real del micropilote, una vez ejecutado; efecto de la presión aplicada en las propiedades del suelo circundante, etc.) se ha realizado una encuesta entre diversas empresas españolas especializadas en la técnica de los micropilotes, a efectos de controlar los volúmenes de inyección y las presiones aplicadas, en función de la deformabilidad del terreno circundante a dichos micropilotes, con lo que se ha logrado definir una rigidez a flexión equivalente de los mismos y la definición y características de una corona de terreno “mejorado” lograda mediante la introducción de la lechada y el efecto de la presión alrededor del micropilote. Con las premisas anteriores y a partir de los parámetros geotécnicos determinados para los terrenos andaluces, se ha procedido a estudiar la estabilidad de terraplenes apoyados sobre taludes a media ladera, mediante el uso de elementos finitos con el Código Plaxis 2D. En el capítulo 5. “Simulación del comportamiento de micropilotes estabilizando terraplenes”, se han desarrollado diversas simulaciones. Para empezar se simplificó el problema simulando casos similares a algunos reales en los que se conocía que los terraplenes habían llegado hasta su situación límite (de los que se disponía información de movimientos medidos con inclinómetros), a partir de ahí se inició la simulación de la inestabilidad para establecer el valor de los parámetros de resistencia al corte del terreno (mediante un análisis retrospectivo – back-análisis) comprobando a su vez que estos valores eran similares a los deducidos del estudio bibliográfico. Seguidamente se han introducido los micropilotes en el borde de la carretera y se ha analizado el comportamiento de éstos y del talud del terraplén (una vez construidos los micropilotes), con el objeto de establecer las bases para su diseño. De este modo y adoptando los distintos parámetros geotécnicos establecidos para los terrenos andaluces, se simularon tres casos reales (en Granada, Málaga y Ceuta), comparando los resultados de dichas simulaciones numéricas con los resultados de medidas reales de campo (desplazamientos del terreno, medidos con inclinómetros), obteniéndose una reproducción bastante acorde a los movimientos registrados. Con las primeras simulaciones se concluye que al instalar los micropilotes la zona más insegura de la ladera es la de aguas abajo. La superficie de rotura ya no afecta a la calzada que protegen los micropilotes. De ahí que se deduzca que esta solución sea válida y se haya aplicado masivamente en Andalucía. En esas condiciones, podría decirse que no se está simulando adecuadamente el trabajo de flexión de los micropilotes (en la superficie de rotura, ya que no les corta), aunque se utilicen elementos viga. Por esta razón se ha realizado otra simulación, basada en las siguientes hipótesis: − Se desprecia totalmente la masa potencialmente deslizante, es decir, la que está por delante de la fila exterior de micros. − La estratigrafía del terreno es similar a la considerada en las primeras simulaciones. − La barrera de micropilotes está constituida por dos elementos inclinados (uno hacia dentro del terraplén y otro hacia fuera), con inclinación 1(H):3(V). − Se puede introducir la rigidez del encepado. − Los micros están separados 0,556 m ó 1,00 m dentro de la misma alineación. − El empotramiento de los micropilotes en el sustrato resistente puede ser entre 1,5 y 7,0 m. Al “anular” el terreno que está por delante de los micropilotes, a lo largo del talud, estos elementos empiezan claramente a trabajar, pudiendo deducirse los esfuerzos de cortante y de flexión que puedan actuar sobre ellos (cota superior pero prácticamente muy cerca de la solución real). En esta nueva modelación se ha considerado tanto la rigidez equivalente (coeficiente ϴ) como la corona de terreno tratado concéntrico al micropilote. De acuerdo a esto último, y gracias a la comparación de estas modelaciones con valores reales de movimientos en laderas instrumentadas con problemas de estabilidad, se ha verificado que existe una similitud bastante importante entre los valores teóricos obtenidos y los medidos en campo, en relación al comportamiento de los micropilotes ejecutados en terraplenes a media ladera. Finalmente para completar el análisis de los micropilotes trabajando a flexión, se ha estudiado el caso de micropilotes dispuestos verticalmente, trabajando como pantallas discontinuas provistas de anclajes, aplicado a un caso real en la ciudad de Granada, en la obra “Hospital de Nuestra Señora de la Salud”. Para su análisis se utilizó el código numérico CYPE, basado en que la reacción del terreno se simula con muelles de rigidez Kh o “módulo de balasto” horizontal, introduciendo en la modelación como variables: a) Las diferentes medidas obtenidas en campo; b) El espesor de terreno cuaternario, que por lo que se pudo determinar, era variable, c) La rigidez y tensión inicial de los anclajes. d) La rigidez del terreno a través de valores relativos de Kh, recopilados en el estudio de los suelos de Andalucía, concretamente en la zona de Granada. Dicha pantalla se instrumentó con 4 inclinómetros (introducidos en los tubos de armadura de cuatro micropilotes), a efectos de controlar los desplazamientos horizontales del muro de contención durante las excavaciones pertinentes, a efectos de comprobar la seguridad del conjunto. A partir del modelo de cálculo desarrollado, se ha comprobado que el valor de Kh pierde importancia debido al gran número de niveles de anclajes, en lo concerniente a las deformaciones horizontales de la pantalla. Por otro lado, los momentos flectores son bastante sensibles a la distancia entre anclajes, al valor de la tensión inicial de los mismos y al valor de Kh. Dicho modelo también ha permitido reproducir de manera fiable los valores de desplazamientos medidos en campo y deducir los parámetros de deformabilidad del terreno, Kh, con valores del orden de la mitad de los medidos en el Metro Ligero de Granada, pero visiblemente superiores a los deducibles de ábacos que permiten obtener Kh para suelos granulares con poca cohesión (gravas y cuaternario superior de Sevilla) como es el caso del ábaco de Arozamena, debido, a nuestro juicio, a la cementación de los materiales presentes en Granada. En definitiva, de las anteriores deducciones se podría pensar en la optimización del diseño de los micropilotes en las obras que se prevean ejecutar en Granada, con similares características al caso de la pantalla vertical arriostrada mediante varios niveles de anclajes y en las que los materiales de emplazamiento tengan un comportamiento geotécnico similar a los estudiados, con el consiguiente ahorro económico. Con todo ello, se considera que se ha hecho una importante aportación para el diseño de futuras obras de micropilotes, trabajando a flexión y cortante, en obras de estabilización de laderas o de excavaciones. Using micropiles in civil engineering has transformed the techniques of stabilization of embankments on the natural or artificial slopes, because although the piles may be the cheapest option, the use of micropiles can reach inaccessible places with lower cost of earthworks, carrying out small work platforms. Machinery used is smaller, lightweight and versatile, including the possibility of manufacturing mortars or cement grouts over distances of several tens of meters of the element to build. However, making a review of the technical documentation available in the engineering field, it was found that systems designed in some cases (micropiles in embankments on the natural slopes, micropiles in vertical cut-off walls, micropiles like "umbrella" in tunnels, etc.) were quite poor or underdeveloped. Premise that concludes the builder has gone ahead (as usually happen in geotechnical engineering) of calculation or theoretical analysis. In the same way it was determined that most of the micropiles are used in underpinning works or as a new foundation solution in conditions of difficult access, in which case the design of micropiles is defined by axial, compressive or tensile loads, considered in regulations as the " Handbook for the design and execution of micropiles in road construction" of the Ministry of Development. The micropiles used to stabilize embankments on the slopes and micropiles act as retaining walls, where they work under shear stress and bending moment, there are not neither reliable systems analysis nor the problem of soil-micropile interaction are properly introduced. Moreover, in many cases, the geotechnical parameters used do not have a proper technical and theoretical basis for what designs may be excessively safe, or the opposite, in most cases. One of the main objectives of this research is to study the behavior of micro piles which are subjected to bending moment and shear stress, as well as other important objectives described in the pertinent section of this thesis. It should be noted that this study has not included the case of quasi-horizontal micropiles working bending moment (as the "umbrella" in tunnels), because it is considered they have a different behavior and calculation, which is outside the scope of this research. It has gone in depth in the study of using micropiles on slopes, presenting real cases of works made, statistics, problems of design and implementation, simplified calculation methods and theoretical modeling in each case, carried out by using FEM (Finite Element Method) Code Plaxis 2D. To accomplish the objectives of this research, It has been started with the development of the "state of the art" which stipulate types, applications, characteristics and calculation of micropiles that are commonly used. In order to study the problem in a real geotechnical field, it has been selected a Spanish zone of action, this being Andalusia, in which it has been used in a very important way, the technique of micropiles especially in embankments stability on natural slopes. From there, it has made a study of the geotechnical properties of the materials (mainly very soft soils and rocks) that are found in this geographical area, which has been mainly a bibliographic study or from the experience in the area of the Director of this thesis. It has been set orders of magnitude of the geotechnical parameters from analyzing made, especially the cohesion and angle of internal friction (also apparent deformation module or a side reaction module equivalent) for various typical Andalusian ground. In order to determine the effect of the implementation of a micropile on the ground (real average volume of micropile once carried out, effect of the pressure applied on the properties of the surrounding soil, etc.) it has conducted a survey among various skilled companies in the technique of micropiles, in order to control injection volumes and pressures applied, depending on the deformability of surrounding terrain such micropiles, whereby it has been possible to define a bending stiffness and the definition and characteristics of a crown land "improved" achieved by introducing the slurry and the effect of the pressure around the micropile. With the previous premises and from the geotechnical parameters determined for the Andalusian terrain, we proceeded to study the stability of embankments resting on batters on the slope, using FEM Code Plaxis 2D. In the fifth chapter "Simulation of the behavior of micropiles stabilizing embankments", there were several different numerical simulations. To begin the problem was simplified simulating similar to some real in which it was known that the embankments had reached their limit situation (for which information of movements measured with inclinometers were available), from there the simulation of instability is initiated to set the value of the shear strength parameters of the ground (by a retrospective analysis or back-analysis) checking these values were similar to those deduced from the bibliographical study Then micropiles have been introduced along the roadside and its behavior was analyzed as well as the slope of embankment (once micropiles were built ), in order to establish the basis for its design. In this way and taking the different geotechnical parameters for the Andalusian terrain, three real cases (in Granada, Malaga and Ceuta) were simulated by comparing the results of these numerical simulations with the results of real field measurements (ground displacements measured with inclinometers), getting quite consistent information according to registered movements. After the first simulations it has been concluded that after installing the micropiles the most insecure area of the natural slope is the downstream. The failure surface no longer affects the road that protects micropiles. Hence it is inferred that this solution is acceptable and it has been massively applied in Andalusia. Under these conditions, one could say that it is not working properly simulating the bending moment of micropiles (on the failure surface, and that does not cut them), although beam elements are used. Therefore another simulation was performed based on the following hypotheses: − The potentially sliding mass is totally neglected, that is, which is ahead of the outer row of micropiles. − Stratigraphy field is similar to the one considered in the first simulations. − Micropiles barrier is constituted by two inclined elements (one inward and one fill out) with inclination 1 (H): 3 (V). − You can enter the stiffness of the pile cap. − The microlies lines are separated 0.556 m or 1.00 m in the same alignment. − The embedding of the micropiles in the tough substrate can be between 1.5 and 7.0 m. To "annul" the ground that is in front of the micro piles, along the slope, these elements clearly start working, efforts can be inferred shear stress and bending moment which may affect them (upper bound but pretty close to the real) solution. In this new modeling it has been considered both equivalent stiffness coefficient (θ) as the treated soil crown concentric to the micropile. According to the latter, and by comparing these values with real modeling movements on field slopes instrumented with stability problems, it was verified that there is quite a significant similarity between the obtained theoretical values and the measured field in relation to the behavior of micropiles executed in embankments along the natural slope. Finally to complete the analysis of micropiles working in bending conditions, we have studied the case of micropiles arranged vertically, working as discontinued cut-off walls including anchors, applied to a real case in the city of Granada, in the play "Hospital of Our Lady of the Health ". CYPE numeric code, based on the reaction of the ground is simulated spring stiffness Kh or "subgrade" horizontal, introduced in modeling was used as variables for analysis: a) The different measurements obtained in field; b) The thickness of quaternary ground, so that could be determined, was variable, c) The stiffness and the prestress of the anchors. d) The stiffness of the ground through relative values of Kh, collected in the study of soils in Andalusia, particularly in the area of Granada. (previously study of the Andalusia soils) This cut-off wall was implemented with 4 inclinometers (introduced in armor tubes four micropiles) in order to control the horizontal displacements of the retaining wall during the relevant excavations, in order to ensure the safety of the whole. From the developed model calculation, it was found that the value of Kh becomes less important because a large number of anchors levels, with regard to the horizontal deformation of the cut-off wall. On the other hand, the bending moments are quite sensitive to the distance between anchors, the initial voltage value thereof and the value of Kh. This model has also been reproduced reliably displacement values measured in the field and deduce parameters terrain deformability, Kh, with values around half the measured Light Rail in Granada, but visibly higher than deductible of abacuses which can obtain Kh for granular soils with low cohesion (upper Quaternary gravels and Sevilla) such as Abacus Arozamena, because, in our view, to cementing materials in Granada. In short, previous deductions you might think on optimizing the design of micropiles in the works that are expected to perform in Granada, with similar characteristics to the case of the vertical cut-off wall braced through several levels of anchors and in which materials location have a geotechnical behavior similar to those studied, with the consequent economic savings. With all this, it is considered that a significant contribution have been made for the design of future works of micropiles, bending moment and shear stress working in slope stabilization works or excavations.
Resumo:
Se plantea desarrollar una herramienta que ofrezca un soporte eficiente para la creación y el diseño de protocolos biológicos a los investigadores en biología sintética. Partiendo de este objetivo, se definen dos cometidos principales: Realizar un estudio de las herramientas existentes que ofrezcan soporte al diseño y aquellas pensadas para diseñar protocolos biológicos, el fin de este estudio es descubrir las funcionalidades que implementan estas herramientas para mejorarlas. Además, se ha de desarrollar una herramienta web que, mediante un lenguaje visual, permita diseñar y crear protocolos de biología sintética, guardándolos en un formato de archivo independiente del lenguaje. En este documento se encuentra, en primer lugar, la definición de objetivos y la descripción del método de desarrollo seguido durante la implementación del proyecto; después, el marco teórico, donde se exponen las herramientas estudiadas y las similitudes y diferencias con la idea que se tiene de la aplicación, y también las herramientas de desarrollo web con las que se va a implementar el proyecto. A continuación, se muestran los resultados obtenidos, mediante la definición de requisitos, así como una exposición de la propia herramienta. Por último, se encuentra la estrategia de validación que se ha seguido en el desarrollo del proyecto y se exponen las conclusiones obtenidas de estas validaciones; también se incluyen al final las conclusiones del proyecto y las líneas futuras de desarrollo.---ABSTRACT---It is planned to develop a tool that provides efficient support for the creation and design of biological protocols researchers in synthetic biology. Based on this goal, two main tasks are defined: Conduct a study of existing tools that provide design support and those intended to design biological protocols, the purpose of this study is to discover the functionalities that implement these tools to improve them. Furthermore, it has to develop a web tool that, through a visual language, allowing design and create synthetic biology protocols, storing them in an independent language file format. In this document is located, first, the definition of objectives and description of the development method followed during project implementation; then the theoretical framework where tools and studied the similarities and differences with the idea we have of the application are discussed, and development tools with which they will implement the project. Then the results obtained, by defining requirements as well as an exhibition of the own tool. Finally, the validation strategy that has been followed in the development of the project and the conclusions drawn from these validations are exposed; also, it is included at the end of the project conclusions and future lines of development.
Resumo:
Hoy en día, el proceso de un proyecto sostenible persigue realizar edificios de elevadas prestaciones que son, energéticamente eficientes, saludables y económicamente viables utilizando sabiamente recursos renovables para minimizar el impacto sobre el medio ambiente reduciendo, en lo posible, la demanda de energía, lo que se ha convertido, en la última década, en una prioridad. La Directiva 2002/91/CE "Eficiencia Energética de los Edificios" (y actualizaciones posteriores) ha establecido el marco regulatorio general para el cálculo de los requerimientos energéticos mínimos. Desde esa fecha, el objetivo de cumplir con las nuevas directivas y protocolos ha conducido las políticas energéticas de los distintos países en la misma dirección, centrándose en la necesidad de aumentar la eficiencia energética en los edificios, la adopción de medidas para reducir el consumo, y el fomento de la generación de energía a través de fuentes renovables. Los edificios de energía nula o casi nula (ZEB, Zero Energy Buildings ó NZEB, Net Zero Energy Buildings) deberán convertirse en un estándar de la construcción en Europa y con el fin de equilibrar el consumo de energía, además de reducirlo al mínimo, los edificios necesariamente deberán ser autoproductores de energía. Por esta razón, la envolvente del edifico y en particular las fachadas son importantes para el logro de estos objetivos y la tecnología fotovoltaica puede tener un papel preponderante en este reto. Para promover el uso de la tecnología fotovoltaica, diferentes programas de investigación internacionales fomentan y apoyan soluciones para favorecer la integración completa de éstos sistemas como elementos arquitectónicos y constructivos, los sistemas BIPV (Building Integrated Photovoltaic), sobre todo considerando el próximo futuro hacia edificios NZEB. Se ha constatado en este estudio que todavía hay una falta de información útil disponible sobre los sistemas BIPV, a pesar de que el mercado ofrece una interesante gama de soluciones, en algunos aspectos comparables a los sistemas tradicionales de construcción. Pero por el momento, la falta estandarización y de una regulación armonizada, además de la falta de información en las hojas de datos técnicos (todavía no comparables con las mismas que están disponibles para los materiales de construcción), hacen difícil evaluar adecuadamente la conveniencia y factibilidad de utilizar los componentes BIPV como parte integrante de la envolvente del edificio. Organizaciones internacionales están trabajando para establecer las normas adecuadas y procedimientos de prueba y ensayo para comprobar la seguridad, viabilidad y fiabilidad estos sistemas. Sin embargo, hoy en día, no hay reglas específicas para la evaluación y caracterización completa de un componente fotovoltaico de integración arquitectónica de acuerdo con el Reglamento Europeo de Productos de la Construcción, CPR 305/2011. Los productos BIPV, como elementos de construcción, deben cumplir con diferentes aspectos prácticos como resistencia mecánica y la estabilidad; integridad estructural; seguridad de utilización; protección contra el clima (lluvia, nieve, viento, granizo), el fuego y el ruido, aspectos que se han convertido en requisitos esenciales, en la perspectiva de obtener productos ambientalmente sostenibles, saludables, eficientes energéticamente y económicamente asequibles. Por lo tanto, el módulo / sistema BIPV se convierte en una parte multifuncional del edificio no sólo para ser física y técnicamente "integrado", además de ser una oportunidad innovadora del diseño. Las normas IEC, de uso común en Europa para certificar módulos fotovoltaicos -IEC 61215 e IEC 61646 cualificación de diseño y homologación del tipo para módulos fotovoltaicos de uso terrestre, respectivamente para módulos fotovoltaicos de silicio cristalino y de lámina delgada- atestan únicamente la potencia del módulo fotovoltaico y dan fe de su fiabilidad por un período de tiempo definido, certificando una disminución de potencia dentro de unos límites. Existe también un estándar, en parte en desarrollo, el IEC 61853 (“Ensayos de rendimiento de módulos fotovoltaicos y evaluación energética") cuyo objetivo es la búsqueda de procedimientos y metodologías de prueba apropiados para calcular el rendimiento energético de los módulos fotovoltaicos en diferentes condiciones climáticas. Sin embargo, no existen ensayos normalizados en las condiciones específicas de la instalación (p. ej. sistemas BIPV de fachada). Eso significa que es imposible conocer las efectivas prestaciones de estos sistemas y las condiciones ambientales que se generan en el interior del edificio. La potencia nominal de pico Wp, de un módulo fotovoltaico identifica la máxima potencia eléctrica que éste puede generar bajo condiciones estándares de medida (STC: irradición 1000 W/m2, 25 °C de temperatura del módulo y distribución espectral, AM 1,5) caracterizando eléctricamente el módulo PV en condiciones específicas con el fin de poder comparar los diferentes módulos y tecnologías. El vatio pico (Wp por su abreviatura en inglés) es la medida de la potencia nominal del módulo PV y no es suficiente para evaluar el comportamiento y producción del panel en términos de vatios hora en las diferentes condiciones de operación, y tampoco permite predecir con convicción la eficiencia y el comportamiento energético de un determinado módulo en condiciones ambientales y de instalación reales. Un adecuado elemento de integración arquitectónica de fachada, por ejemplo, debería tener en cuenta propiedades térmicas y de aislamiento, factores como la transparencia para permitir ganancias solares o un buen control solar si es necesario, aspectos vinculados y dependientes en gran medida de las condiciones climáticas y del nivel de confort requerido en el edificio, lo que implica una necesidad de adaptación a cada contexto específico para obtener el mejor resultado. Sin embargo, la influencia en condiciones reales de operación de las diferentes soluciones fotovoltaicas de integración, en el consumo de energía del edificio no es fácil de evaluar. Los aspectos térmicos del interior del ambiente o de iluminación, al utilizar módulos BIPV semitransparentes por ejemplo, son aún desconocidos. Como se dijo antes, la utilización de componentes de integración arquitectónica fotovoltaicos y el uso de energía renovable ya es un hecho para producir energía limpia, pero también sería importante conocer su posible contribución para mejorar el confort y la salud de los ocupantes del edificio. Aspectos como el confort, la protección o transmisión de luz natural, el aislamiento térmico, el consumo energético o la generación de energía son aspectos que suelen considerarse independientemente, mientras que todos juntos contribuyen, sin embargo, al balance energético global del edificio. Además, la necesidad de dar prioridad a una orientación determinada del edificio, para alcanzar el mayor beneficio de la producción de energía eléctrica o térmica, en el caso de sistemas activos y pasivos, respectivamente, podría hacer estos últimos incompatibles, pero no necesariamente. Se necesita un enfoque holístico que permita arquitectos e ingenieros implementar sistemas tecnológicos que trabajen en sinergia. Se ha planteado por ello un nuevo concepto: "C-BIPV, elemento fotovoltaico consciente integrado", esto significa necesariamente conocer los efectos positivos o negativos (en términos de confort y de energía) en condiciones reales de funcionamiento e instalación. Propósito de la tesis, método y resultados Los sistemas fotovoltaicos integrados en fachada son a menudo soluciones de vidrio fácilmente integrables, ya que por lo general están hechos a medida. Estos componentes BIPV semitransparentes, integrados en el cerramiento proporcionan iluminación natural y también sombra, lo que evita el sobrecalentamiento en los momentos de excesivo calor, aunque como componente estático, asimismo evitan las posibles contribuciones pasivas de ganancias solares en los meses fríos. Además, la temperatura del módulo varía considerablemente en ciertas circunstancias influenciada por la tecnología fotovoltaica instalada, la radiación solar, el sistema de montaje, la tipología de instalación, falta de ventilación, etc. Este factor, puede suponer un aumento adicional de la carga térmica en el edificio, altamente variable y difícil de cuantificar. Se necesitan, en relación con esto, más conocimientos sobre el confort ambiental interior en los edificios que utilizan tecnologías fotovoltaicas integradas, para abrir de ese modo, una nueva perspectiva de la investigación. Con este fin, se ha diseñado, proyectado y construido una instalación de pruebas al aire libre, el BIPV Env-lab "BIPV Test Laboratory", para la caracterización integral de los diferentes módulos semitransparentes BIPV. Se han definido también el método y el protocolo de ensayos de caracterización en el contexto de un edificio y en condiciones climáticas y de funcionamiento reales. Esto ha sido posible una vez evaluado el estado de la técnica y la investigación, los aspectos que influyen en la integración arquitectónica y los diferentes tipos de integración, después de haber examinado los métodos de ensayo para los componentes de construcción y fotovoltaicos, en condiciones de operación utilizadas hasta ahora. El laboratorio de pruebas experimentales, que consiste en dos habitaciones idénticas a escala real, 1:1, ha sido equipado con sensores y todos los sistemas de monitorización gracias a los cuales es posible obtener datos fiables para evaluar las prestaciones térmicas, de iluminación y el rendimiento eléctrico de los módulos fotovoltaicos. Este laboratorio permite el estudio de tres diferentes aspectos que influencian el confort y consumo de energía del edificio: el confort térmico, lumínico, y el rendimiento energético global (demanda/producción de energía) de los módulos BIPV. Conociendo el balance de energía para cada tecnología solar fotovoltaica experimentada, es posible determinar cuál funciona mejor en cada caso específico. Se ha propuesto una metodología teórica para la evaluación de estos parámetros, definidos en esta tesis como índices o indicadores que consideran cuestiones relacionados con el bienestar, la energía y el rendimiento energético global de los componentes BIPV. Esta metodología considera y tiene en cuenta las normas reglamentarias y estándares existentes para cada aspecto, relacionándolos entre sí. Diferentes módulos BIPV de doble vidrio aislante, semitransparentes, representativos de diferentes tecnologías fotovoltaicas (tecnología de silicio monocristalino, m-Si; de capa fina en silicio amorfo unión simple, a-Si y de capa fina en diseleniuro de cobre e indio, CIS) fueron seleccionados para llevar a cabo una serie de pruebas experimentales al objeto de demostrar la validez del método de caracterización propuesto. Como resultado final, se ha desarrollado y generado el Diagrama Caracterización Integral DCI, un sistema gráfico y visual para representar los resultados y gestionar la información, una herramienta operativa útil para la toma de decisiones con respecto a las instalaciones fotovoltaicas. Este diagrama muestra todos los conceptos y parámetros estudiados en relación con los demás y ofrece visualmente toda la información cualitativa y cuantitativa sobre la eficiencia energética de los componentes BIPV, por caracterizarlos de manera integral. ABSTRACT A sustainable design process today is intended to produce high-performance buildings that are energy-efficient, healthy and economically feasible, by wisely using renewable resources to minimize the impact on the environment and to reduce, as much as possible, the energy demand. In the last decade, the reduction of energy needs in buildings has become a top priority. The Directive 2002/91/EC “Energy Performance of Buildings” (and its subsequent updates) established a general regulatory framework’s methodology for calculation of minimum energy requirements. Since then, the aim of fulfilling new directives and protocols has led the energy policies in several countries in a similar direction that is, focusing on the need of increasing energy efficiency in buildings, taking measures to reduce energy consumption, and fostering the use of renewable sources. Zero Energy Buildings or Net Zero Energy Buildings will become a standard in the European building industry and in order to balance energy consumption, buildings, in addition to reduce the end-use consumption should necessarily become selfenergy producers. For this reason, the façade system plays an important role for achieving these energy and environmental goals and Photovoltaic can play a leading role in this challenge. To promote the use of photovoltaic technology in buildings, international research programs encourage and support solutions, which favors the complete integration of photovoltaic devices as an architectural element, the so-called BIPV (Building Integrated Photovoltaic), furthermore facing to next future towards net-zero energy buildings. Therefore, the BIPV module/system becomes a multifunctional building layer, not only physically and functionally “integrated” in the building, but also used as an innovative chance for the building envelope design. It has been found in this study that there is still a lack of useful information about BIPV for architects and designers even though the market is providing more and more interesting solutions, sometimes comparable to the existing traditional building systems. However at the moment, the lack of an harmonized regulation and standardization besides to the non-accuracy in the technical BIPV datasheets (not yet comparable with the same ones available for building materials), makes difficult for a designer to properly evaluate the fesibility of this BIPV components when used as a technological system of the building skin. International organizations are working to establish the most suitable standards and test procedures to check the safety, feasibility and reliability of BIPV systems. Anyway, nowadays, there are no specific rules for a complete characterization and evaluation of a BIPV component according to the European Construction Product Regulation, CPR 305/2011. BIPV products, as building components, must comply with different practical aspects such as mechanical resistance and stability; structural integrity; safety in use; protection against weather (rain, snow, wind, hail); fire and noise: aspects that have become essential requirements in the perspective of more and more environmentally sustainable, healthy, energy efficient and economically affordable products. IEC standards, commonly used in Europe to certify PV modules (IEC 61215 and IEC 61646 respectively crystalline and thin-film ‘Terrestrial PV Modules-Design Qualification and Type Approval’), attest the feasibility and reliability of PV modules for a defined period of time with a limited power decrease. There is also a standard (IEC 61853, ‘Performance Testing and Energy Rating of Terrestrial PV Modules’) still under preparation, whose aim is finding appropriate test procedures and methodologies to calculate the energy yield of PV modules under different climate conditions. Furthermore, the lack of tests in specific conditions of installation (e.g. façade BIPV devices) means that it is difficult knowing the exact effective performance of these systems and the environmental conditions in which the building will operate. The nominal PV power at Standard Test Conditions, STC (1.000 W/m2, 25 °C temperature and AM 1.5) is usually measured in indoor laboratories, and it characterizes the PV module at specific conditions in order to be able to compare different modules and technologies on a first step. The “Watt-peak” is not enough to evaluate the panel performance in terms of Watt-hours of various modules under different operating conditions, and it gives no assurance of being able to predict the energy performance of a certain module at given environmental conditions. A proper BIPV element for façade should take into account thermal and insulation properties, factors as transparency to allow solar gains if possible or a good solar control if necessary, aspects that are linked and high dependent on climate conditions and on the level of comfort to be reached. However, the influence of different façade integrated photovoltaic solutions on the building energy consumption is not easy to assess under real operating conditions. Thermal aspects, indoor temperatures or luminance level that can be expected using building integrated PV (BIPV) modules are not well known. As said before, integrated photovoltaic BIPV components and the use of renewable energy is already a standard for green energy production, but would also be important to know the possible contribution to improve the comfort and health of building occupants. Comfort, light transmission or protection, thermal insulation or thermal/electricity power production are aspects that are usually considered alone, while all together contribute to the building global energy balance. Besides, the need to prioritize a particular building envelope orientation to harvest the most benefit from the electrical or thermal energy production, in the case of active and passive systems respectively might be not compatible, but also not necessary. A holistic approach is needed to enable architects and engineers implementing technological systems working in synergy. A new concept have been suggested: “C-BIPV, conscious integrated BIPV”. BIPV systems have to be “consciously integrated” which means that it is essential to know the positive and negative effects in terms of comfort and energy under real operating conditions. Purpose of the work, method and results The façade-integrated photovoltaic systems are often glass solutions easily integrable, as they usually are custommade. These BIPV semi-transparent components integrated as a window element provides natural lighting and shade that prevents overheating at times of excessive heat, but as static component, likewise avoid the possible solar gains contributions in the cold months. In addition, the temperature of the module varies considerably in certain circumstances influenced by the PV technology installed, solar radiation, mounting system, lack of ventilation, etc. This factor may result in additional heat input in the building highly variable and difficult to quantify. In addition, further insights into the indoor environmental comfort in buildings using integrated photovoltaic technologies are needed to open up thereby, a new research perspective. This research aims to study their behaviour through a series of experiments in order to define the real influence on comfort aspects and on global energy building consumption, as well as, electrical and thermal characteristics of these devices. The final objective was to analyze a whole set of issues that influence the global energy consumption/production in a building using BIPV modules by quantifying the global energy balance and the BIPV system real performances. Other qualitative issues to be studied were comfort aspect (thermal and lighting aspects) and the electrical behaviour of different BIPV technologies for vertical integration, aspects that influence both energy consumption and electricity production. Thus, it will be possible to obtain a comprehensive global characterization of BIPV systems. A specific design of an outdoor test facility, the BIPV Env-lab “BIPV Test Laboratory”, for the integral characterization of different BIPV semi-transparent modules was developed and built. The method and test protocol for the BIPV characterization was also defined in a real building context and weather conditions. This has been possible once assessed the state of the art and research, the aspects that influence the architectural integration and the different possibilities and types of integration for PV and after having examined the test methods for building and photovoltaic components, under operation conditions heretofore used. The test laboratory that consists in two equivalent test rooms (1:1) has a monitoring system in which reliable data of thermal, daylighting and electrical performances can be obtained for the evaluation of PV modules. The experimental set-up facility (testing room) allows studying three different aspects that affect building energy consumption and comfort issues: the thermal indoor comfort, the lighting comfort and the energy performance of BIPV modules tested under real environmental conditions. Knowing the energy balance for each experimented solar technology, it is possible to determine which one performs best. A theoretical methodology has been proposed for evaluating these parameters, as defined in this thesis as indices or indicators, which regard comfort issues, energy and the overall performance of BIPV components. This methodology considers the existing regulatory standards for each aspect, relating them to one another. A set of insulated glass BIPV modules see-through and light-through, representative of different PV technologies (mono-crystalline silicon technology, mc-Si, amorphous silicon thin film single junction, a-Si and copper indium selenide thin film technology CIS) were selected for a series of experimental tests in order to demonstrate the validity of the proposed characterization method. As result, it has been developed and generated the ICD Integral Characterization Diagram, a graphic and visual system to represent the results and manage information, a useful operational tool for decision-making regarding to photovoltaic installations. This diagram shows all concepts and parameters studied in relation to each other and visually provides access to all the results obtained during the experimental phase to make available all the qualitative and quantitative information on the energy performance of the BIPV components by characterizing them in a comprehensive way.
Resumo:
En este documento se van a desarrollar las técnicas para la creación de páginas web utilizando uno de los sistemas de gestión de contenido, también llamado CMS, por sus siglas en inglés, Content Manager System, más potentes del mercado, y open source, es decir de código abierto, como es el caso de Drupal. Se usará este sistema de gestión en su versión 7, y la actualización 37, o lo que es lo mismo, se usará Drupal 7.37, aunque se actualizará a la versión 38, puesto que esta versión fue publicada a mitad de los desarrollos y la versión 8 todavía está en versión beta y no es suficientemente estable como para crear una página web robusta con esa versión del CMS. Estas explicaciones están basadas en dos páginas web creadas en dicho gestor de contenidos, una de ellas ya en producción, www.lrl.es, que también podrá ser accedida desde www.luqueagentescomerciales.es, página web diseñada para un agente comercial de herramientas. La segunda, todavía no está en producción pero se puede acceder a ella desde el servidor wamp montado en local, 81.37.0.251/epi, que es una mejora de la página www.elpactoinvisible.es, diseñada para una galería de arte de Málaga. A lo largo del documento se verán distintas opciones que se pueden llegar a utilizar gracias a Drupal para obtener el resultado esperado y se darán las explicaciones pertinentes para saber cuál de ellas usar para focalizar distintos aspectos según se decida ir por una vía u otra de desarrollo, y se mostraran, con código cuando sea necesario, las soluciones utilizadas en las distintas páginas, de las que se ha hablado en el párrafo precedente, y el porqué de usar esa solución en vez de otras alternativas que también son viables para dichos objetivos. ABSTRACT This paper will develop techniques for creating web pages using one of the systems content management, also called CMS, for its acronym in English, Content Manager System, and the most powerful in the market, and open source, as is the case of Drupal. This management system in its version 7, and updating 37, or as it is called, Drupal 7.37 is used, although it will be updated to version 38, because this version was published in the middle of the developments and version 8 is still in beta and is not stable enough to create a robust site with this version of CMS. These explanations are based on two websites created in this CMS, one of them already in production, www.lrl.es, which also can be accessed from www.luqueagentescomerciales.es, web page designed for a tools commercial agent. The second one, is not yet in production but can be accessed from the WAMP server installed locally, 81.37.0.251/epi, which is an improvement to www.elpactoinvisible.es page, designed for an art gallery in Malaga. Throughout the document you can get several options to be used thanks to Drupal in order to achieve result and explanations will be given to indicate one of theme to use for targeting different aspects and which developing method or path to follow and will be shown, in code if it should be necessary, the solutions used in different pages which have been discussed in the preceding paragraph, and why to use that solution instead of other alternatives are also viable for such purposes.
Application of the Boundary Method to the determination of the properties of the beam cross-sections
Resumo:
Using the 3-D equations of linear elasticity and the asylllptotic expansion methods in terms of powers of the beam cross-section area as small parameter different beam theories can be obtained, according to the last term kept in the expansion. If it is used only the first two terms of the asymptotic expansion the classical beam theories can be recovered without resort to any "a priori" additional hypotheses. Moreover, some small corrections and extensions of the classical beam theories can be found and also there exists the possibility to use the asymptotic general beam theory as a basis procedure for a straightforward derivation of the stiffness matrix and the equivalent nodal forces of the beam. In order to obtain the above results a set of functions and constants only dependent on the cross-section of the beam it has to be computed them as solutions of different 2-D laplacian boundary value problems over the beam cross section domain. In this paper two main numerical procedures to solve these boundary value pf'oblems have been discussed, namely the Boundary Element Method (BEM) and the Finite Element Method (FEM). Results for some regular and geometrically simple cross-sections are presented and compared with ones computed analytically. Extensions to other arbitrary cross-sections are illustrated.
Resumo:
We describe a method to design dominant-negative proteins (D-N) to the basic helix–loop–helix–leucine zipper (B-HLHZip) family of sequence-specific DNA binding transcription factors. The D-Ns specifically heterodimerize with the B-HLHZip dimerization domain of the transcription factors and abolish DNA binding in an equimolar competition. Thermal denaturation studies indicate that a heterodimer between a Myc B-HLHZip domain and a D-N consisting of a 12-amino acid sequence appended onto the Max dimerization domain (A-Max) is −6.3 kcal⋅mol−1 more stable than the Myc:Max heterodimer. One molar equivalent of A-Max can totally abolish the DNA binding activity of a Myc:Max heterodimer. This acidic extension also has been appended onto the dimerization domain of the B-HLHZip protein Mitf, a member of the transcription factor enhancer binding subfamily, to produce A-Mitf. The heterodimer between A-Mitf and the B-HLHZip domain of Mitf is −3.7 kcal⋅mol−1 more stable than the Mitf homodimer. Cell culture studies show that A-Mitf can inhibit Mitf-dependent transactivation both in acidic extension and in a dimerization-dependent manner. A-Max can inhibit Myc-dependent foci formation twice as well as the Max dimerization domain (HLHZip). This strategy of producing D-Ns may be applicable to other B-HLHZip or B-HLH proteins because it provides a method to inhibit the DNA binding of these transcription factors in a dimerization-specific manner.
Resumo:
Accurate and automated methods for measuring the thickness of human cerebral cortex could provide powerful tools for diagnosing and studying a variety of neurodegenerative and psychiatric disorders. Manual methods for estimating cortical thickness from neuroimaging data are labor intensive, requiring several days of effort by a trained anatomist. Furthermore, the highly folded nature of the cortex is problematic for manual techniques, frequently resulting in measurement errors in regions in which the cortical surface is not perpendicular to any of the cardinal axes. As a consequence, it has been impractical to obtain accurate thickness estimates for the entire cortex in individual subjects, or group statistics for patient or control populations. Here, we present an automated method for accurately measuring the thickness of the cerebral cortex across the entire brain and for generating cross-subject statistics in a coordinate system based on cortical anatomy. The intersubject standard deviation of the thickness measures is shown to be less than 0.5 mm, implying the ability to detect focal atrophy in small populations or even individual subjects. The reliability and accuracy of this new method are assessed by within-subject test–retest studies, as well as by comparison of cross-subject regional thickness measures with published values.
Resumo:
Melanina é um polímero constituído por uma grande heterogeneidade de monômeros tendo como característica comum a presença de grupos indóis. Por outro lado, a eumelanina produzida pela oxidação enzimática da tirosina é um polímero mais simples constituído principalmente de monômeros 5,6-dihidroxindol (DHI) e de indol-5,6-quinona (IQ). Tirosinase é a enzima chave na produção de melanina, sendo que a sua atividade cinética é medida em função da formação do intermediário dopacroma. Nanopartículas (NPs) de sílica são partículas nanométricas compostas de oxido de silício e são obtidas pelo processo sol-gel desenvolvido por Stöber de hidrólise e condensação de tetraetilortosilicato (TEOS), usando etanol como solvente em meio alcalino. As NPs foram funcionalizadas com 3-Aminopropiltrietoxissilano (ATPES) e depois com glutaraldeído. Este último permitiu a imobilização da tirosinase na superfície da sílica. Caracterizamos as NPs antes e após a reação da enzima, a atividade catalítica da enzima ligada à NP e o mecanismos de formação de melanina na superfície da sílica. As NPs foram caracterizadas por espectrofotometria de absorção e de reflectância, termogravimetria e microscopia eletrônica. A síntese da NP de sílica retornou partículas esféricas com 55nm de diâmetro e a funcionalização da partícula mostrou modificar eficientemente a sua superfície. A imobilização da tirosinase por ligação covalente foi de 99,5% contra 0,5% da adsorção física. A atividade da tirosinase foi caracterizada pela formação de dopacroma. O Km da enzima imobilizada não sofreu alteração em comparação com a tirosinase livre, mas a eficiência catalítica - que considera a eficiência recuperada - foi de apenas 1/3 para a enzima ligada covalentemente, significando que 2/3 das enzimas ligadas não estão ativas. Obtivemos NPs revestidas com melanina a partir de oxidação de tirosina solubilizada em duas preparações: NP com tirosinase ligada covalentemente na superfície e NP funcionalizada com glutaraldeido dispersa em solução de DHI e IQ. O revestimento de melanina foi na forma de um filme fino com espessura ~1,9nm, conferindo perfil de absorção luminosa equivalente ao da própria melanina. Mostramos que o mecanismo de polimerização passa pela oxidação da tirosina pela tirosinase, que gera intermediários oxidados (principalmente DHI e IQ) que vão para solução (mesmo quando a tirosinase está ligada covalentemente na sílica). Estes intermediários ligam-se ao glutaraldeido e a superfície da sílica passa a funcionar como ambiente de polimerização da melanina.
Resumo:
This dissertation introduces an approach to generate tests to test fail-safe behavior for web applications. We apply the approach to a commercial web application. We build models for both behavioral and mitigation requirements. We create mitigation tests from an existing functional black box test suite by determining failure type and points of failure in the test suite and weaving required mitigation based on weaving rules to generate a test suite that tests proper mitigation of failures. A genetic algorithm (GA) is used to determine points of failure and type of failure that needs to be tested. Mitigation test paths are woven into the behavioral test at the point of failure based on failure specific weaving rules. A simulator was developed to evaluate choice of parameters for the genetic algorithm. We showed how to tune the fitness function and performed tuning experiments for GA to determine what values to use for exploration weight and prospecting weight. We found that higher defect densities make prospecting and mining more successful, while lower mitigation defect densities need more exploration. We compare efficiency and effectiveness of the approach. First, the GA approach is compared to random selection. The results show that the GA performance was better than random selection and that the approach was robust when the search space increased. Second, we compare the GA against four coverage criteria. The results of comparison show that test requirements generated by a genetic algorithm (GA) are more efficient than three of the four coverage criteria for large search spaces. They are equally effective. For small search spaces, the genetic algorithm is less effective than three of the four coverage criteria. The fourth coverage criteria is too weak and unable to find all defects in almost all cases. We also present a large case study of a mortgage system at one of our industrial partners and show how we formalize the approach. We evaluate the use of a GA to create test requirements. The evaluation includes choice of initial population, multiplicity of runs and a discussion of the cost of evaluating fitness. Finally, we build a selective regression testing approach based on types of changes (add, delete, or modify) that could occur in the behavioral model, the fault model, the mitigation models, the weaving rules, and the state-event matrix. We provide a systematic method by showing the formalization steps for each type of change to the various models.
Resumo:
Current model-driven Web Engineering approaches (such as OO-H, UWE or WebML) provide a set of methods and supporting tools for a systematic design and development of Web applications. Each method addresses different concerns using separate models (content, navigation, presentation, business logic, etc.), and provide model compilers that produce most of the logic and Web pages of the application from these models. However, these proposals also have some limitations, especially for exchanging models or representing further modeling concerns, such as architectural styles, technology independence, or distribution. A possible solution to these issues is provided by making model-driven Web Engineering proposals interoperate, being able to complement each other, and to exchange models between the different tools. MDWEnet is a recent initiative started by a small group of researchers working on model-driven Web Engineering (MDWE). Its goal is to improve current practices and tools for the model-driven development of Web applications for better interoperability. The proposal is based on the strengths of current model-driven Web Engineering methods, and the existing experience and knowledge in the field. This paper presents the background, motivation, scope, and objectives of MDWEnet. Furthermore, it reports on the MDWEnet results and achievements so far, and its future plan of actions.
Empirical study on the maintainability of Web applications: Model-driven Engineering vs Code-centric
Resumo:
Model-driven Engineering (MDE) approaches are often acknowledged to improve the maintainability of the resulting applications. However, there is a scarcity of empirical evidence that backs their claimed benefits and limitations with respect to code-centric approaches. The purpose of this paper is to compare the performance and satisfaction of junior software maintainers while executing maintainability tasks on Web applications with two different development approaches, one being OOH4RIA, a model-driven approach, and the other being a code-centric approach based on Visual Studio .NET and the Agile Unified Process. We have conducted a quasi-experiment with 27 graduated students from the University of Alicante. They were randomly divided into two groups, and each group was assigned to a different Web application on which they performed a set of maintainability tasks. The results show that maintaining Web applications with OOH4RIA clearly improves the performance of subjects. It also tips the satisfaction balance in favor of OOH4RIA, although not significantly. Model-driven development methods seem to improve both the developers’ objective performance and subjective opinions on ease of use of the method. This notwithstanding, further experimentation is needed to be able to generalize the results to different populations, methods, languages and tools, different domains and different application sizes.
Resumo:
There are a large number of image processing applications that work with different performance requirements and available resources. Recent advances in image compression focus on reducing image size and processing time, but offer no real-time solutions for providing time/quality flexibility of the resulting image, such as using them to transmit the image contents of web pages. In this paper we propose a method for encoding still images based on the JPEG standard that allows the compression/decompression time cost and image quality to be adjusted to the needs of each application and to the bandwidth conditions of the network. The real-time control is based on a collection of adjustable parameters relating both to aspects of implementation and to the hardware with which the algorithm is processed. The proposed encoding system is evaluated in terms of compression ratio, processing delay and quality of the compressed image when compared with the standard method.