916 resultados para web tool
Resumo:
Parmodel is a web server for automated comparative modeling and evaluation of protein structures. The aim of this tool is to help inexperienced users to perform modeling, assessment, visualization, and optimization of protein models as well as crystallographers to evaluate structures solved experimentally. It is subdivided in four modules: Parmodel Modeling, Parmodel Assessment, Parmodel Visualization, and Parmodel Optimization. The main module is the Parmodel Modeling that allows the building of several models ford a same protein in a reduced time, through the distribution of modeling processes on a Beowulf cluster. Parmodel automates and integrates the main softwares used in comparative modeling as MODELLER, Whatcheck, Procheck, Raster3D, Molscript, and Gromacs. This web server is freely accessible at http://www.biocristalografia.df.ibilce.unesp.br/tools/parmodel. (C) 2004 Elsevier B.V. All rights reserved.
Resumo:
This work describes a new web system to aid project management that was created to correct the principal deficiencies identified in systems having a common purpose which are at present available, as well as to follow the guidelines that are proposed in the Project Management Body of Knowledge (PMBoK) and the quality characteristics described in the ISO/IEC 9126 norm. As from the adopted methodology, the system was structured to attend the real necessities of project managers and also to contribute towards obtaining quality results from the projects. The validation of the proposed solution was done with the collaboration of professionals that used the functions available in it for a period of 15 days. Results attested to the quality and adequacy of the developed system.
Usabilidade na web: o usuário como agente-facilitador no desenvolvimento de interfaces de home pages
Resumo:
Pós-graduação em Design - FAAC
Resumo:
Pós-graduação em Ciência da Informação - FFC
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
Food webs have been used in order to understand the trophic relationship among organisms within an ecosystem, however the extension by which sampling efficiency could affect food web responses remain poorly understood. Still, there is a lack of long-term sampling data for many insect groups, mainly related to the interactions between herbivores and their host plants. In the first chapter, I describe a source food web based on the Senegalia tenuifolia plant by identifying the associated insect species and the interactions among them and with this host plant. Furthermore, I check for the data robustness from each trophic level and propose a cost-efficiently methodology. The results from this chapter show that the collected dataset and the methodology presented are a good tool for sample most insect richness of a source food web. In total the food web comprises 27 species belonging to four trophic levels. In the second chapter, I demonstrate the temporal variation in the species richness and abundance from each trophic level, as well as the relationship among distinct trophic levels. Moreover, I investigate the diversity patterns of the second and third trophic level by assessing the contribution of alfa and beta-diversity components along the years. This chapter shows that in our system the parasitoid abundance is regulated by the herbivore abundances. Besides, the species richness and abundances of the trophic levels vary temporally. It also shows that alfa-diversity was the diversity component that most contribute to the herbivore species diversity (2nd trophic level), while the contribution of alfa- and beta-diversity changed along the years for parasitoid diversity (3rd level). Overall, this dissertation describes a source food web and bring insights into some food web challenges related to the sampling effort to gather enough species from all trophic levels. It also discuss the relation among communities associated with distinct trophic levels and their temporal variation and diversity patterns. Finally, this dissertation contributes for the world food web database and in understanding the interactions among its trophic levels and each trophic level pattern along time and space
Resumo:
A web tem sido bastante utilizada ultimamente por empresas e instituições de ensino como forma de oferecer aprendizado. A criação de conteúdos para o aprendizado via web é uma atividade que tem demandado recursos humanos especializados em tecnologias e conceitos, de forma que os conteúdos criados não sejam restritos a ferramentas, plataformas e formatos de apresentação. Este trabalho apresenta o desenvolvimento de um protótipo de uma ferramenta de autoria de conteúdo para aprendizado via web, através da qual os conteúdos criados são independentes da ferramenta de autoria, da plataforma operacional e dos formatos de apresentação que sejam utilizados. Por ser um trabalho de desenvolvimento, um processo é definido e suas fases e atividades são utilizadas para organizar trabalho e documentar todas as decisões tomadas.
Resumo:
Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)
Resumo:
The tool proposed, known as WSPControl, enables remote monitoring of computers across the Internet using distributed applications. Through a Web Services architecture is possible the communication between these distributed applications across heterogeneous platforms, also eliminates the need for additional settings in computer networks, such as release of ports or proxy. The tool is divided into three modules, namely: • Client Interface: developed in C Sharp, is responsible for capturing data on performance of the monitored computer also connects to the Web Services to report this data. • Web Services Interface: developed in PHP using the PHP SOAP library, is responsible for facilitating the communication between internet applications and client. • Internet Interface: developed in PHP, is responsible for reading and interpreting the information captured these available on the Internet
Resumo:
Over the last decades changes have occurred in communication within and between enterprises, made easier by technologies suchas E-commerce, Internet, ERP systems and remote meetings and there was a rapid progress in network technology, which has changed the way business is done. A standardized way to offer services over the internet is using web services. Web services are a kind of remote procedure call and are generally used to integrate systems, independent of language, both client and server. It is common to use several web services run in sequence to perform a business process. To this type of process, gives the name of workflow. Thus, Web services are the primary components of workflows. A tool that provides a way of visualizing the behavior of a workflow can assist the administrator and is required. The present work presents the development of a tool that allows the administrator to classify visually services components and evaluate their importance in the final performance of a workflow. As proof of concept we used several virtual servers and computers where each computer has received a set of web services. A proxy was added between each call of workflows collecting relevant information and storing them in a database for later analysis. The analysis was based on Quality of Service parameters
Resumo:
The University of São Paulo has been experiencing the increase in contents in electronic and digital formats, distributed by different suppliers and hosted remotely or in clouds, and is faced with the also increasing difficulties related to facilitating access to this digital collection by its users besides coexisting with the traditional world of physical collections. A possible solution was identified in the new generation of systems called Web Scale Discovery, which allow better management, data integration and agility of search. Aiming to identify if and how such a system would meet the USP demand and expectation and, in case it does, to identify what the analysis criteria of such a tool would be, an analytical study with an essentially documental base was structured, as from a revision of the literature and from data available in official websites and of libraries using this kind of resources. The conceptual base of the study was defined after the identification of software assessment methods already available, generating a standard with 40 analysis criteria, from details on the unique access interface to information contents, web 2.0 characteristics, intuitive interface, facet navigation, among others. The details of the studies conducted into four of the major systems currently available in this software category are presented, providing subsidies for the decision-making of other libraries interested in such systems.
Resumo:
[ES] Esta aplicación surgió de la necesidad real del equipo de desarrolladores encargados de llevar a cabo el proceso de despliegue de aplicaciones web en empresas que siguen modelos de desarrollo iterativos, con grandes procesos de despliegue que deben ser aplicados con bastante frecuencia. El objetivo fundamental del proyecto es proporcionar a estos usuarios una aplicación que sirva de ayuda en el proceso de despliegue y centralice toda la información importante y útil que debe ser comprobada de manera constante y automatice todas aquellas tareas triviales que constituyen una pérdida de tiempo para los encargados del proceso de despliegue y disminuyen la concentración en el propio proceso y los detalles importantes y por lo tanto la calidad final del proceso de despliegue. La aplicación desarrollada será útil para la organización y control del proceso y además ofrecerá beneficios en cuanto a comunicación entre aquellos interesados en la nueva versión a desplegar. De modo general, la aplicación ofrecerá una forma de monitorizar los distintos estados de la nueva versión y mantendrá un sistema de rastreo de todos los eventos sucedidos durante la etapa de despliegue e información relacionada, además de automatizar tantas tareas como sea posible.
Resumo:
[ES] El objetivo de este Trabajo Final de Grado (TFG) es la creación de un prototipo de aplicación web para la gestión de recursos geoespaciales. Esta propuesta surgió a partir de la necesidad de disponer de una herramienta que no tuviera que ser instalada en un dispositivo, sino servida por un servidor web, permitiendo su acceso desde cualquier parte y dispositivo. El resultado fue el Gestor Web de Recursos Geoespaciales con Tecnología OpenLayers, una aplicación que combina diversas herramientas (OpenLayers, GeoServer, PostgreSQL, jQuery…) – todas ellas basadas en Software Libre – para cumplir funcionalidades como la creación de primitivas vectoriales sobre un mapa, gestión y visualización de la información asociada, edición de estilos, modificación de coordenadas, etc. siendo todas éstas funcionalidades características de un Sistema de Información Geográfica (SIG) y ofreciendo una interfaz de uso cómoda y eficaz, que abstraiga al usuario de detalles internos y complejos. El material desarrollado dispone del potencial necesario para convertirse en una solución a las necesidades de gestión de información geoespacial de la ULPGC, especialmente en el campus de Tafira, sobre el que se ha ejemplificado su uso. Además, a diferencia de las herramientas ofertadas por empresas como Google o Microsoft, esta aplicación está por completo bajo una licencia GNU GPL v3, lo que permite que se pueda indagar dentro de su código, mejorarlo y añadir funcionalidades a cualquier persona interesada.
Resumo:
[ES] El Trabajo de Fin de Grado, Monitor Web de Expresiones Regulares (MWRegEx), es una herramienta basada en tecnologías web, desarrollada usando el entorno Visual Studio. El objetivo principal de la aplicación es dar apoyo a la docencia de expresiones regulares, en el marco de la enseñanza del manejo de ristras de caracteres en las asignaturas de programación del Grado en Ingeniería Informática. La aplicación permite obtener el dibujo de un autómata de una expresión regular, facilitando su comprensión; además, permite aplicar la expresión a diferentes ristras de caracteres, mostrando las coincidencias encontradas, y ofrece una versión de la expresión adaptada a su uso en literales string de lenguajes como Java y otros. La herramienta se ha implementado en dos partes: un servicio web, escrito en C#, donde se realizan todos los análisis de las expresiones regulares y las ristras a contrastar; y un cliente web, implementado usando tecnología asp.net, con JavaScript y JQuery, que gestiona la interfaz de usuario y muestra los resultados. Esta separación permite que el servicio web pueda ser reutilizado con otras aplicaciones cliente. El autómata que representa una expresión regular esta dibujado usando la librería Raphaël JavaScript que permite manejar los elementos SVG. Cada elemento de la expresión regular tiene un dibujo diferente y único para así diferenciarlo. Toda la interfaz gráfica de usuario está internacionalizada de manera tal que pueda adaptarse a diferentes idiomas y regiones sin la necesidad de realizar cambios de ingeniería ni en el código. Tanto el servicio web como la parte cliente están estructurados para que se puedan agregar nuevas modificaciones sin que esto genere una onda expansiva a lo largo de las diversas clases existentes.
Resumo:
As distributed collaborative applications and architectures are adopting policy based management for tasks such as access control, network security and data privacy, the management and consolidation of a large number of policies is becoming a crucial component of such policy based systems. In large-scale distributed collaborative applications like web services, there is the need of analyzing policy interactions and integrating policies. In this thesis, we propose and implement EXAM-S, a comprehensive environment for policy analysis and management, which can be used to perform a variety of functions such as policy property analyses, policy similarity analysis, policy integration etc. As part of this environment, we have proposed and implemented new techniques for the analysis of policies that rely on a deep study of state of the art techniques. Moreover, we propose an approach for solving heterogeneity problems that usually arise when considering the analysis of policies belonging to different domains. Our work focuses on analysis of access control policies written in the dialect of XACML (Extensible Access Control Markup Language). We consider XACML policies because XACML is a rich language which can represent many policies of interest to real world applications and is gaining widespread adoption in the industry.