847 resultados para Combined Web crippling and Flange Crushing
Resumo:
This dataset contains the results of granulometric and bulk geochemical analyses of Van Veen surface samples obtained by the Alfred Wegener Institute (AWI) in the course of the 2012 and 2013 summer field seasons. The sampling was performed along transects in depths generally <13 m, to a distance of about <5 km off Herschel Island. In 2012, 75 samples in Pauline Cove and in the vicinity of Simpson Point were obtained. Sample collection was expanded in 2013, on transects established the previous year, with additional locations in Tetris Bay and Workboat Passage. Samples consisted of approximately 100 g of the top 3-6 cm of sediment, and were frozen in the field and freeze dried at the AWI before undergoing analytical procedures. Sample locations were recorded with the onboard global positioning system (GPS) unit. Grain size distributions in our study were obtained using laser diffractometry at the AWI (Beckman Coulter LS200) on the <1 mm fraction of samples oxidized with 30% H2O2 until effervescence ceased to remove organics. Some samples were also sieved using a sieve stack with 1 phi intervals. GRADISTAT (Blott and Pye, 2001) was used to calculate graphical grain size statistics (Folk and Ward, 1957). Grain diameters were logarithmically transformed to phi values, calculated as phi=-log2d, where d is the grain diameter in millimeters (Blott and Pye, 2001; Krumbein, 1934). Freeze dried samples were ground and ground using an Elemetar Vario EL III carbon-nitrogen-sulphur analyzer at the AWI to measure total carbon (TC) and total nitrogen (TN). Tungsten oxide was added to the samples as a catalyst to the pyrolysis. Following this analysis, total organic carbon (TOC) was determined using an Elementar VarioMax. Stable carbon isotope ratios of 13C/12C of 118 samples were determined on a DELTAplusXL mass spectrometer (ThermoFisher Scientific, Bremen) at the German Research Centre for Geosciences (GFZ) in Potsdam, Germany . An additional analysis on 69 samples was carried out at the University of Hamburg with an isotope ratio mass spectrometer (Delta V, Thermo Scientific, Germany) coupled to an elemental analyzer (Flash 2000, Thermo Scientific, Germany). Prior to analysis, soil samples were treated with phosphoric acid (43%) to release inorganic carbon. Values are expressed relative to Vienna Peedee belemnite (VPDB) using external standards (USGS40, -26.4 per mil VPDB and IVA soil 33802153, -27.5 per mil VPDB).
Resumo:
Rising temperatures and ocean acidification driven by anthropogenic carbon emissions threaten both tropical and temperate corals. However, the synergistic effect of these stressors on coral physiology is still poorly understood, in particular for cold-water corals. This study assessed changes in key physiological parameters (calcification, respiration and ammonium excretion) of the widespread cold-water coral Desmophyllum dianthus maintained for 8 months at two temperatures (ambient 12 °C and elevated 15 °C) and two pCO2 conditions (ambient 390 ppm and elevated 750 ppm). At ambient temperatures no change in instantaneous calcification, respiration or ammonium excretion rates was observed at either pCO2 levels. Conversely, elevated temperature (15 °C) significantly reduced calcification rates, and combined elevated temperature and pCO2 significantly reduced respiration rates. Changes in the ratio of respired oxygen to excreted nitrogen (O:N), which provides information on the main sources of energy being metabolized, indicated a shift from mixed use of protein and carbohydrate/lipid as metabolic substrates under control conditions, to less efficient protein-dominated catabolism under both stressors. Overall, this study shows that the physiology of D. dianthus is more sensitive to thermal than pCO2 stress, and that the predicted combination of rising temperatures and ocean acidification in the coming decades may severely impact this cold-water coral species.
Resumo:
With global climate change, ocean warming and acidification occur concomitantly. In this study, we tested the hypothesis that increasing CO2 levels affect the acid-base balance and reduce the activity capacity of the Arctic spider crab Hyas araneus, especially at the limits of thermal tolerance. Crabs were acclimated to projected oceanic CO2 levels for 12 days (today: 380, towards the year 2100: 750 and 1,120 and beyond: 3,000 ?atm) and at two temperatures (1 and 4 °C). Effects of these treatments on the righting response (RR) were determined (1) at acclimation temperatures followed by (2) righting when exposed to an additional acute (15 min) heat stress at 12 °C. Prior to (resting) and after the consecutive stresses of combined righting activity and heat exposure, acid-base status and lactate contents were measured in the haemolymph. Under resting conditions, CO2 caused a decrease in haemolymph pH and an increase in oxygen partial pressure. Despite some buffering via an accumulation of bicarbonate, the extracellular acidosis remained uncompensated at 1 °C, a trend exacerbated when animals were acclimated to 4 °C. The additional combined exposure to activity and heat had only a slight effect on blood gas and acid-base status. Righting activity in all crabs incubated at 1 and 4 °C was unaffected by elevated CO2 levels or acute heat stress but was significantly reduced when both stressors acted synergistically. This impact was much stronger in the group acclimated at 1 °C where some individuals acclimated to high CO2 levels stopped responding. Lactate only accumulated in the haemolymph after combined righting and heat stress. In the group acclimated to 1 °C, lactate content was highest under normocapnia and lowest at the highest CO2 level in line with the finding that RR was largely reduced. In crabs acclimated to 4 °C, the RR was less affected by CO2 such that activity caused lactate to increase with rising CO2 levels. In line with the concept of oxygen and capacity limited thermal tolerance, all animals exposed to temperature extremes displayed a reduction in scope for performance, a trend exacerbated by increasing CO2 levels. Additionally, the differences seen between cold- and warm-acclimated H. araneus after heat stress indicate that a small shift to higher acclimation temperatures also alleviates the response to temperature extremes, indicating a shift in the thermal tolerance window which reduces susceptibility to additional CO2 exposure.
Resumo:
Nowadays video and web conferencing systems have become effective tools for communication and collaboration inside organizations. However, although these systems have evolved and now provide very nice features (e.g. sharing multimedia and documents), they are still too focused on the moment the meeting takes place. The existing systems provide very few facilities to organize the meeting and they do not take advantage of the possibilities the generated content offers once the meeting is finished. In this paper, we analyze the life cycle of a web conference and how existing systems monitor these conferences. Finally we present our solution, based on our know-how in videoconference management and our experience with these existing systems.
Resumo:
This article proposes a MAS architecture for network diagnosis under uncertainty. Network diagnosis is divided into two inference processes: hypothesis generation and hypothesis confirmation. The first process is distributed among several agents based on a MSBN, while the second one is carried out by agents using semantic reasoning. A diagnosis ontology has been defined in order to combine both inference processes. To drive the deliberation process, dynamic data about the influence of observations are taken during diagnosis process. In order to achieve quick and reliable diagnoses, this influence is used to choose the best action to perform. This approach has been evaluated in a P2P video streaming scenario. Computational and time improvements are highlight as conclusions.
Resumo:
In the paper we report on the results of our experiments on the construction of the opinion ontology. Our aim is to show the benefits of publishing in the open, on the Web, the results of the opinion mining process in a structured form. On the road to achieving this, we attempt to answer the research question to what extent opinion information can be formalized in a unified way. Furthermore, as part of the evaluation, we experiment with the usage of Semantic Web technologies and show particular use cases that support our claims.
Resumo:
Biomedical ontologies are key elements for building up the Life Sciences Semantic Web. Reusing and building biomedical ontologies requires flexible and versatile tools to manipulate them efficiently, in particular for enriching their axiomatic content. The Ontology Pre Processor Language (OPPL) is an OWL-based language for automating the changes to be performed in an ontology. OPPL augments the ontologists’ toolbox by providing a more efficient, and less error-prone, mechanism for enriching a biomedical ontology than that obtained by a manual treatment. Results We present OPPL-Galaxy, a wrapper for using OPPL within Galaxy. The functionality delivered by OPPL (i.e. automated ontology manipulation) can be combined with the tools and workflows devised within the Galaxy framework, resulting in an enhancement of OPPL. Use cases are provided in order to demonstrate OPPL-Galaxy’s capability for enriching, modifying and querying biomedical ontologies. Conclusions Coupling OPPL-Galaxy with other bioinformatics tools of the Galaxy framework results in a system that is more than the sum of its parts. OPPL-Galaxy opens a new dimension of analyses and exploitation of biomedical ontologies, including automated reasoning, paving the way towards advanced biological data analyses.
Resumo:
The Networks of Evolutionary Processors (NEPs) are computing mechanisms directly inspired from the behavior of cell populations more specifically the point mutations in DNA strands. These mechanisms are been used for solving NP-complete problems by means of a parallel computation postulation. This paper describes an implementation of the basic model of NEP using Web technologies and includes the possibility of designing some of the most common variants of it by means the use of the web page design which eases the configuration of a given problem. It is a system intended to be used in a multicore processor in order to benefit from the multi thread use.
Resumo:
In this paper, calculus of variations and combined blade element and momentum theory (BEMT) are used to demonstrate that, in hover, when neither root nor tip losses are considered; the rotor, which minimizes the total power (MPR), generates an induced velocity that varies linearly along the blade span. The angle of attack of every blade element is constant and equal to its optimum value. The traditional ideal twist (ITR) and optimum (OR) rotors are revisited in the context of this variational framework. Two more optimum rotors are obtained considering root and tip losses, the ORL, and the MPRL. A comparison between these five rotors is presented and discussed. The MPR and MPRL present a remarkable saving of power for low values of both thrust coefficient and maximum aerodynamic efficiency. The result obtained can be exploited to improve the aerodynamic behaviour of rotary wing micro air vehicles (MAV). A comparison with experimental results obtained from the literature is presented.
Resumo:
En este trabajo se han analizado varios problemas en el contexto de la elasticidad no lineal basándose en modelos constitutivos representativos. En particular, se han analizado problemas relacionados con el fenómeno de perdida de estabilidad asociada con condiciones de contorno en el caso de material reforzados con fibras. Cada problema se ha formulado y se ha analizado por separado en diferentes capítulos. En primer lugar se ha mostrado el análisis del gradiente de deformación discontinuo para un material transversalmente isótropo, en particular, el modelo del material considerado consiste de una base neo-Hookeana isótropa incrustada con fibras de refuerzo direccional caracterizadas con un solo parámetro. La solución de este problema se vincula con instabilidades que dan lugar al mecanismo de fallo conocido como banda de cortante. La perdida de elipticidad de las ecuaciones diferenciales de equilibrio es una condición necesaria para que aparezca este tipo de soluciones y por tanto las inestabilidades asociadas. En segundo lugar se ha analizado una deformación combinada de extensión, inación y torsión de un tubo cilíndrico grueso donde se ha encontrado que la deformación citada anteriormente puede ser controlada solo para determinadas direcciones de las fibras refuerzo. Para entender el comportamiento elástico del tubo considerado se ha ilustrado numéricamente los resultados obtenidos para las direcciones admisibles de las fibras de refuerzo bajo la deformación considerada. En tercer lugar se ha estudiado el caso de un tubo cilíndrico grueso reforzado con dos familias de fibras sometido a cortante en la dirección azimutal para un modelo de refuerzo especial. En este problema se ha encontrado que las inestabilidades que aparecen en el material considerado están asociadas con lo que se llama soluciones múltiples de la ecuación diferencial de equilibrio. Se ha encontrado que el fenómeno de instabilidad ocurre en un estado de deformación previo al estado de deformación donde se pierde la elipticidad de la ecuación diferencial de equilibrio. También se ha demostrado que la condición de perdida de elipticidad y ^W=2 = 0 (la segunda derivada de la función de energía con respecto a la deformación) son dos condiciones necesarias para la existencia de soluciones múltiples. Finalmente, se ha analizado detalladamente en el contexto de elipticidad un problema de un tubo cilíndrico grueso sometido a una deformación combinada en las direcciones helicoidal, axial y radial para distintas geotermias de las fibras de refuerzo . In the present work four main problems have been addressed within the framework of non-linear elasticity based on representative constitutive models. Namely, problems related to the loss of stability phenomena associated with boundary value problems for fibre-reinforced materials. Each of the considered problems is formulated and analysed separately in different chapters. We first start with the analysis of discontinuous deformation gradients for a transversely isotropic material under plane deformation. In particular, the material model is an augmented neo-Hookean base with a simple unidirectional reinforcement characterised by a single parameter. The solution of this problem is related to material instabilities and it is associated with a shear band-type failure mode. The loss of ellipticity of the governing differential equations is a necessary condition for the existence of these material instabilities. The second problem involves a detailed analysis of the combined non-linear extension, inflation and torsion of a thick-walled circular cylindrical tube where it has been found that the aforementioned deformation is controllable only for certain preferred directions of transverse isotropy. Numerical results have been illustrated to understand the elastic behaviour of the tube for the admissible preferred directions under the considered deformation. The third problem deals with the analysis of a doubly fibre-reinforced thickwalled circular cylindrical tube undergoing pure azimuthal shear for a special class of the reinforcing model where multiple non-smooth solutions emerge. The associated instability phenomena are found to occur prior to the point where the nominal stress tensor changes monotonicity in a particular direction. It has been also shown that the loss of ellipticity condition that arises from the equilibrium equation and ^W=2 = 0 (the second derivative of the strain-energy function with respect to the deformation) are equivalent necessary conditions for the emergence of multiple solutions for the considered material. Finally, a detailed analysis in the basis of the loss of ellipticity of the governing differential equations for a combined helical, axial and radial elastic deformations of a fibre-reinforced circular cylindrical tube is carried out.
Resumo:
La Ciencia Ciudadana nace del resultado de involucrar en las investigaciones científicas a todo tipo de personas, las cuales pueden participar en un determinado experimento analizando o recopilando datos. No hace falta que tengan una formación científica para poder participar, es decir cualquiera puede contribuir con su granito de arena. La ciencia ciudadana se ha convertido en un elemento a tener en cuenta a la hora de realizar tareas científicas que requieren mucha dedicación, o que simplemente por el volumen de trabajo que estas implican, resulta casi imposible que puedan ser realizadas por una sola persona o un pequeño grupo de trabajo. El proyecto GLORIA (GLObal Robotic-telescopes Intelligent Array) es la primera red de telescopios robóticos del mundo de acceso libre que permite a los usuarios participar en la investigación astronómica mediante la observación con telescopios robóticos, y/o analizando los datos que otros usuarios han adquirido con GLORIA, o desde otras bases de datos de libre acceso. Con el objetivo de contribuir a esta iniciativa se ha propuesto crear una plataforma web que pasará a formar parte del Proyecto GLORIA, en la que se puedan realizar experimentos astronómicos. Con el objetivo de fomentar la ciencia y el aprendizaje colaborativo se propone construir una aplicación web que se ejecute en la plataforma Facebook. Los experimentos los proporciona la red de telescopios del proyecto GLORIA mediante servicios web y están definidos mediante XML. La aplicación web recibe el XML con la descripción del experimento, lo interpreta y lo representa en la plataforma Facebook para que los usuarios potenciales puedan realizar los experimentos. Los resultados de los experimentos realizados se envían a una base de datos de libre acceso que será gestionada por el proyecto GLORIA, para su posterior análisis por parte de expertos. ---ABSTRACT---The citizen’s science is born out of the result of involving all type of people in scientific investigations, in which, they can participate in a determined experiment analyzing or compiling data. There is no need to have a scientific training in order to participate, but, anyone could contribute doing one’s bit. The citizen’s science has become an element to take into account when carrying out scientific tasks that require a lot dedication, or that, for the volume of work that these involve, are nearly impossible to be carried out by one person or a small working group. The GLORIA Project (Global Robotic-Telescopes Intelligent Array) is the first network of free access robotic telescopes in the world that permits the users to participate in the astronomic investigation by means of observation with robotic telescopes, and/or analyzing data from other users that have obtained through GLORIA, or from other free-access databases. With the aim of contributing to this initiative, a web platform has been created and will be part of the GLORIA Project, in which astronomic experiments can be carried out. With the objective of promoting science and collaborative apprenticeship, a web application carried out in the FACEBOOK platform is to be built. The experiments are founded by the telescopes network of the GLORIA project by means of web services and are defined through XML. The web application receives the XML with the description of the experiment, interprets it and represents it in the FACEBOOK platform in order for potential users may perform the experiments. The results of the experiments carried out are sent to a free-access database that will be managed by the GLORIA Project for its analysis on the part of experts.
Resumo:
This document is the result of a process of web development to create a tool that will allow to Cracow University of Technology consult, create and manage timetables. The technologies chosen for this purpose are Apache Tomcat Server, My SQL Community Server, JDBC driver, Java Servlets and JSPs for the server side. The client part counts on Javascript, jQuery, AJAX and CSS technologies to perform the dynamism. The document will justify the choice of these technologies and will explain some development tools that help in the integration and development of all this elements: specifically, NetBeans IDE and MySQL workbench have been used as helpful tools. After explaining all the elements involved in the development of the web application, the architecture and the code developed are explained through UML diagrams. Some implementation details related to security are also deeper explained through sequence diagrams. As the source code of the application is provided, an installation manual has been developed to run the project. In addition, as the platform is intended to be a beta that will be grown, some unimplemented ideas for future development are also exposed. Finally, some annexes with important files and scripts related to the initiation of the platform are attached. This project started through an existing tool that needed to be expanded. The main purpose of the project along its development has focused on setting the roots for a whole new platform that will replace the existing one. For this goal, it has been needed to make a deep inspection on the existing web technologies: a web server and a SQL database had to be chosen. Although the alternatives were a lot, Java technology for the server was finally selected because of the big community backwards, the easiness of modelling the language through UML diagrams and the fact of being free license software. Apache Tomcat is the open source server that can use Java Servlet and JSP technology. Related to the SQL database, MySQL Community Server is the most popular open-source SQL Server, with a big community after and quite a lot of tools to manage the server. JDBC is the driver needed to put in contact Java and MySQL. Once we chose the technologies that would be part of the platform, the development process started. After a detailed explanation of the development environment installation, we used UML use case diagrams to set the main tasks of the platform; UML class diagrams served to establish the existing relations between the classes generated; the architecture of the platform was represented through UML deployment diagrams; and Enhanced entity–relationship (EER) model were used to define the tables of the database and their relationships. Apart from the previous diagrams, some implementation issues were explained to make a better understanding of the developed code - UML sequence diagrams helped to explain this. Once the whole platform was properly defined and developed, the performance of the application has been shown: it has been proved that with the current state of the code, the platform covers the use cases that were set as the main target. Nevertheless, some requisites needed for the proper working of the platform have been specified. As the project is aimed to be grown, some ideas that could not be added to this beta have been explained in order not to be missed for future development. Finally, some annexes containing important configuration issues for the platform have been added after proper explanation, as well as an installation guide that will let a new developer get the project ready. In addition to this document some other files related to the project are provided: - Javadoc. The Javadoc containing the information of every Java class created is necessary for a better understanding of the source code. - database_model.mwb. This file contains the model of the database for MySQL Workbench. This model allows, among other things, generate the MySQL script for the creation of the tables. - ScheduleManager.war. The WAR file that will allow loading the developed application into Tomcat Server without using NetBeans. - ScheduleManager.zip. The source code exported from NetBeans project containing all Java packages, JSPs, Javascript files and CSS files that are part of the platform. - config.properties. The configuration file to properly get the names and credentials to use the database, also explained in Annex II. Example of config.properties file. - db_init_script.sql. The SQL query to initiate the database explained in Annex III. SQL statements for MySQL initialization. RESUMEN. Este proyecto tiene como punto de partida la necesidad de evolución de una herramienta web existente. El propósito principal del proyecto durante su desarrollo se ha centrado en establecer las bases de una completamente nueva plataforma que reemplazará a la existente. Para lograr esto, ha sido necesario realizar una profunda inspección en las tecnologías web existentes: un servidor web y una base de datos SQL debían ser elegidos. Aunque existen muchas alternativas, la tecnología Java ha resultado ser elegida debido a la gran comunidad de desarrolladores que tiene detrás, además de la facilidad que proporciona este lenguaje a la hora de modelarlo usando diagramas UML. Tampoco hay que olvidar que es una tecnología de uso libre de licencia. Apache Tomcat es el servidor de código libre que permite emplear Java Servlets y JSPs para hacer uso de la tecnología de Java. Respecto a la base de datos SQL, el servidor más popular de código libre es MySQL, y cuenta también con una gran comunidad detrás y buenas herramientas de modelado, creación y gestión de la bases de datos. JDBC es el driver que va a permitir comunicar las aplicaciones Java con MySQL. Tras elegir las tecnologías que formarían parte de esta nueva plataforma, el proceso de desarrollo tiene comienzo. Tras una extensa explicación de la instalación del entorno de desarrollo, se han usado diagramas de caso de UML para establecer cuáles son los objetivos principales de la plataforma; los diagramas de clases nos permiten realizar una organización del código java desarrollado de modo que sean fácilmente entendibles las relaciones entre las diferentes clases. La arquitectura de la plataforma queda definida a través de diagramas de despliegue. Por último, diagramas EER van a definir las relaciones entre las tablas creadas en la base de datos. Aparte de estos diagramas, algunos detalles de implementación se van a justificar para tener una mejor comprensión del código desarrollado. Diagramas de secuencia ayudarán en estas explicaciones. Una vez que toda la plataforma haya quedad debidamente definida y desarrollada, se va a realizar una demostración de la misma: se demostrará cómo los objetivos generales han sido alcanzados con el desarrollo actual del proyecto. No obstante, algunos requisitos han sido aclarados para que la plataforma trabaje adecuadamente. Como la intención del proyecto es crecer (no es una versión final), algunas ideas que se han podido llevar acabo han quedado descritas de manera que no se pierdan. Por último, algunos anexos que contienen información importante acerca de la plataforma se han añadido tras la correspondiente explicación de su utilidad, así como una guía de instalación que va a permitir a un nuevo desarrollador tener el proyecto preparado. Junto a este documento, ficheros conteniendo el proyecto desarrollado quedan adjuntos. Estos ficheros son: - Documentación Javadoc. Contiene la información de las clases Java que han sido creadas. - database_model.mwb. Este fichero contiene el modelo de la base de datos para MySQL Workbench. Esto permite, entre otras cosas, generar el script de iniciación de la base de datos para la creación de las tablas. - ScheduleManager.war. El fichero WAR que permite desplegar la plataforma en un servidor Apache Tomcat. - ScheduleManager.zip. El código fuente exportado directamente del proyecto de Netbeans. Contiene todos los paquetes de Java generados, ficheros JSPs, Javascript y CSS que forman parte de la plataforma. - config.properties. Ejemplo del fichero de configuración que permite obtener los nombres de la base de datos - db_init_script.sql. Las consultas SQL necesarias para la creación de la base de datos.
Resumo:
En la realización de este proyecto se ha tratado principalmente la temática del web scraping sobre documentos HTML en Android. Como resultado del mismo, se ha propuesto una metodología para poder realizar web scraping en aplicaciones implementadas para este sistema operativo y se desarrollará una aplicación basada en esta metodología que resulte útil a los alumnos de la escuela. Web scraping se puede definir como una técnica basada en una serie de algoritmos de búsqueda de contenido con el fin de obtener una determinada información de páginas web, descartando aquella que no sea relevante. Como parte central, se ha dedicado bastante tiempo al estudio de los navegadores y servidores Web, y del lenguaje HTML presente en casi todas las páginas web en la actualidad así como de los mecanismos utilizados para la comunicación entre cliente y servidor ya que son los pilares en los que se basa esta técnica. Se ha realizado un estudio de las técnicas y herramientas necesarias, aportándose todos los conceptos teóricos necesarios, así como la proposición de una posible metodología para su implementación. Finalmente se ha codificado la aplicación UPMdroid, desarrollada con el fin de ejemplificar la implementación de la metodología propuesta anteriormente y a la vez desarrollar una aplicación cuya finalidad es brindar al estudiante de la ETSIST un soporte móvil en Android que le facilite el acceso y la visualización de aquellos datos más importantes del curso académico como son: el horario de clases y las calificaciones de las asignaturas en las que se matricule. Esta aplicación, además de implementar la metodología propuesta, es una herramienta muy interesante para el alumno, ya que le permite utilizar de una forma sencilla e intuitiva gran número de funcionalidades de la escuela solucionando así los problemas de visualización de contenido web en los dispositivos. ABSTRACT. The main topic of this project is about the web scraping over HTML documents on Android OS. As a result thereof, it is proposed a methodology to perform web scraping in deployed applications for this operating system and based on this methodology that is useful to the ETSIST school students. Web scraping can be defined as a technique based on a number of content search algorithms in order to obtain certain information from web pages, discarding those that are not relevant. As a main part, has spent considerable time studying browsers and Web servers, and the HTML language that is present today in almost all websites as well as the mechanisms used for communication between client and server because they are the pillars which this technique is based. We performed a study of the techniques and tools needed, providing all the necessary theoretical concepts, as well as the proposal of a possible methodology for implementation. Finally it has codified UPMdroid application, developed in order to illustrate the implementation of the previously proposed methodology and also to give the student a mobile ETSIST Android support to facilitate access and display those most important data of the current academic year such as: class schedules and scores for the subjects in which you are enrolled. This application, in addition to implement the proposed methodology is also a very interesting tool for the student, as it allows a simple and intuitive way of use these school functionalities thus fixing the viewing web content on devices.