914 resultados para GUI legacy Windows Form web-application


Relevância:

100.00% 100.00%

Publicador:

Resumo:

El tema de investigación de esta monografía de grado, es un estudio nunca antes realizado, original, y que permitirá analizar por primera vez, en el ámbito universitario, el proceso que se está llevando a cabo en este país, y dará cuenta de cómo un Estado con una tradición de conflictos con sus vecinos, es capaz de neutralizar esta situación, en favor de una cooperación que beneficie sus intereses nacionales. Es por eso, que para este estudio, será necesario precisar las consecuencias económicas y políticas, positivas y negativas que ha desencadenado esta alianza. También, será necesario analizar la relación de Pakistán con algunos países islámicos como Irán, Arabia Saudita y Afganistán, para evaluar como ha sido tomada esta alianza por el mundo musulmán, y finalmente, establecer a manera de conclusión, un paralelo entre Costos y Beneficios, que nos permita establecer si Pakistán se ha beneficiado más de lo que se ha perjudicado de esta alianza, y si esta nueva sociedad formada desde 2001, es el motor generador de estabilidad en la región del sur de Asia. 2 Esta monografía, como ya se ha mencionado, trata un tema poco abordado que no ha sido estudiado antes, por lo que se hace necesaria la utilización de una investigación exploratoria, que lleve a familiarizarnos con fenómenos relativamente desconocidos. Por ello, para el desarrollo de este trabajo de grado se tomarán como ayuda bibliográfica revistas electrónicas, artículos de periódicos, internet y documentos oficiales de varios gobiernos disponibles en la web.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

La principal contribución del trabajo es estudiar como representar la información turística de manera que sea fácilmente interpretable por programas de cálculo de rutas. Esto permitirá implementar asistentes que creen rutas turísticas en función de los gustos del usuario y que proporcionen información sobre sus puntos de interés de forma transparente con independencia de su formato o ubicación en la web

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Extracte de les notícies sobre tesis llegides que al llarg del 2011 s'han publicat a la web.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

"Student’s Watcher” is a small Web application which wants to show in a visual, simple and fast way, the evolution of the students. The main project table displays such things as marks and comments about students. We can add a comment for each mark to explain why this mark. The objective is to be able to know if some student has a problem, how is going his year, marks in other courses, or even, to know if he has a bad week in a different subjects. We can see the evolution of students in past years to do an objective comparison. It also allows inserting global comments of student, we have a list of these, and all professors can add new ones, where we can see more general valuations. “Student’s Watcher” was begun in ASP.net, but finally my project would be developed in PHP, HTML and CSS. This project wants to be a comparison between two of most important languages used nowadays, ASPX and PHP

Relevância:

100.00% 100.00%

Publicador:

Resumo:

G-Rex is light-weight Java middleware that allows scientific applications deployed on remote computer systems to be launched and controlled as if they are running on the user's own computer. G-Rex is particularly suited to ocean and climate modelling applications because output from the model is transferred back to the user while the run is in progress, which prevents the accumulation of large amounts of data on the remote cluster. The G-Rex server is a RESTful Web application that runs inside a servlet container on the remote system, and the client component is a Java command line program that can easily be incorporated into existing scientific work-flow scripts. The NEMO and POLCOMS ocean models have been deployed as G-Rex services in the NERC Cluster Grid, and G-Rex is the core grid middleware in the GCEP and GCOMS e-science projects.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Compute grids are used widely in many areas of environmental science, but there has been limited uptake of grid computing by the climate modelling community, partly because the characteristics of many climate models make them difficult to use with popular grid middleware systems. In particular, climate models usually produce large volumes of output data, and running them usually involves complicated workflows implemented as shell scripts. For example, NEMO (Smith et al. 2008) is a state-of-the-art ocean model that is used currently for operational ocean forecasting in France, and will soon be used in the UK for both ocean forecasting and climate modelling. On a typical modern cluster, a particular one year global ocean simulation at 1-degree resolution takes about three hours when running on 40 processors, and produces roughly 20 GB of output as 50000 separate files. 50-year simulations are common, during which the model is resubmitted as a new job after each year. Running NEMO relies on a set of complicated shell scripts and command utilities for data pre-processing and post-processing prior to job resubmission. Grid Remote Execution (G-Rex) is a pure Java grid middleware system that allows scientific applications to be deployed as Web services on remote computer systems, and then launched and controlled as if they are running on the user's own computer. Although G-Rex is general purpose middleware it has two key features that make it particularly suitable for remote execution of climate models: (1) Output from the model is transferred back to the user while the run is in progress to prevent it from accumulating on the remote system and to allow the user to monitor the model; (2) The client component is a command-line program that can easily be incorporated into existing model work-flow scripts. G-Rex has a REST (Fielding, 2000) architectural style, which allows client programs to be very simple and lightweight and allows users to interact with model runs using only a basic HTTP client (such as a Web browser or the curl utility) if they wish. This design also allows for new client interfaces to be developed in other programming languages with relatively little effort. The G-Rex server is a standard Web application that runs inside a servlet container such as Apache Tomcat and is therefore easy to install and maintain by system administrators. G-Rex is employed as the middleware for the NERC1 Cluster Grid, a small grid of HPC2 clusters belonging to collaborating NERC research institutes. Currently the NEMO (Smith et al. 2008) and POLCOMS (Holt et al, 2008) ocean models are installed, and there are plans to install the Hadley Centre’s HadCM3 model for use in the decadal climate prediction project GCEP (Haines et al., 2008). The science projects involving NEMO on the Grid have a particular focus on data assimilation (Smith et al. 2008), a technique that involves constraining model simulations with observations. The POLCOMS model will play an important part in the GCOMS project (Holt et al, 2008), which aims to simulate the world’s coastal oceans. A typical use of G-Rex by a scientist to run a climate model on the NERC Cluster Grid proceeds as follows :(1) The scientist prepares input files on his or her local machine. (2) Using information provided by the Grid’s Ganglia3 monitoring system, the scientist selects an appropriate compute resource. (3) The scientist runs the relevant workflow script on his or her local machine. This is unmodified except that calls to run the model (e.g. with “mpirun”) are simply replaced with calls to "GRexRun" (4) The G-Rex middleware automatically handles the uploading of input files to the remote resource, and the downloading of output files back to the user, including their deletion from the remote system, during the run. (5) The scientist monitors the output files, using familiar analysis and visualization tools on his or her own local machine. G-Rex is well suited to climate modelling because it addresses many of the middleware usability issues that have led to limited uptake of grid computing by climate scientists. It is a lightweight, low-impact and easy-to-install solution that is currently designed for use in relatively small grids such as the NERC Cluster Grid. A current topic of research is the use of G-Rex as an easy-to-use front-end to larger-scale Grid resources such as the UK National Grid service.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Web Services for Remote Portlets (WSRP) is gaining attention among portal developers and vendors to enable easy development, increased richness in functionality, pluggability, and flexibility of deployment. Whilst currently not supporting all WSRP functionalities, open-source portal frameworks could in future use WSRP Consumers to access remote portlets found from a WSRP Producer registry service. This implies that we need a central registry for the remote portlets and a more expressive WSRP Consumer interface to implement the remote portlet functions. This paper reports on an investigation into a new system architecture, which includes a Web Services repository, registry, and client interface. The Web Services repository holds portlets as remote resource producers. A new data structure for expressing remote portlets is found and published by populating a Universal Description, Discovery and Integration (UDDI) registry. A remote portlet publish and search engine for UDDI has also been developed. Finally, a remote portlet client interface was developed as a Web application. The client interface supports remote portlet features, as well as window status and mode functions. Copyright (c) 2007 John Wiley & Sons, Ltd.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Parkinson’s disease is a clinical syndrome manifesting with slowness and instability. As it is a progressive disease with varying symptoms, repeated assessments are necessary to determine the outcome of treatment changes in the patient. In the recent past, a computer-based method was developed to rate impairment in spiral drawings. The downside of this method is that it cannot separate the bradykinetic and dyskinetic spiral drawings. This work intends to construct the computer method which can overcome this weakness by using the Hilbert-Huang Transform (HHT) of tangential velocity. The work is done under supervised learning, so a target class is used which is acquired from a neurologist using a web interface. After reducing the dimension of HHT features by using PCA, classification is performed. C4.5 classifier is used to perform the classification. Results of the classification are close to random guessing which shows that the computer method is unsuccessful in assessing the cause of drawing impairment in spirals when evaluated against human ratings. One promising reason is that there is no difference between the two classes of spiral drawings. Displaying patients self ratings along with the spirals in the web application is another possible reason for this, as the neurologist may have relied too much on this in his own ratings.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper examines the value of real-time traffic information gathered through Geographic Information Systems for achieving an optimal vehicle routing within a dynamically stochastic transportation network. We present a systematic approach in determining the dynamically varying parameters and implementation attributes that were used for the development of a Web-based transportation routing application integrated with real-time GIS services. We propose and implement an optimal routing algorithm by modifying Dijkstra’s algorithm in order to incorporate stochastically changing traffic flows. We describe the significant features of our Web application in making use of the real-time dynamic traffic flow information from GIS services towards achieving total costs savings and vehicle usage reduction. These features help users and vehicle drivers in improving their service levels and productivity as the Web application enables them to interactively find the optimal path and in identifying destinations effectively.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Disulfide torsional energy, a good predictor of disulfide redox potential in proteins, may be estimated by interpolation on a potential energy surface (PES) describing the twisting of diethyl disulfide through its three central dihedral angles. Here we update PES calculations at the M05-2X level of theory with the 6-31G(d) basis set. Although the surface shows no qualitative differences from an earlier MP2(full) PES, energy differences greater than 1 kJ mol–1 were seen for conformations with χ2 between –60° and 30°, or with χ3 below 60° or above 130°. This is particularly significant for highly strained disulfides that are likely to be spontaneously reduced by mechanical means. In benchmarking against the high-level G3X method, M05-2X showed significantly reduced root mean squared deviation compared with MP2(full) (1.0 versus 2.0 kJ mol–1 respectively). Results are incorporated into a web application that calculates relative torsional energies from disulfide dihedral angles (http://www.sbinf.org/applications/pes.html).

Relevância:

100.00% 100.00%

Publicador:

Resumo:

For years, opinion polls rely on data collected through telephone or person-to-person surveys. The process is costly, inconvenient, and slow. Recently online search data has emerged as potential proxies for the survey data. However considerable human involvement is still needed for the selection of search indices, a task that requires knowledge of both the target issue and how search terms are used by the online community. The robustness of such manually selected search indices can be questionable. In this paper, we propose an automatic polling system through a novel application of machine learning. In this system, the needs for examining, comparing, and selecting search indices have been eliminated through automatic generation of candidate search indices and intelligent combination of the indices. The results include a publicly accessible web application that provides real-time, robust, and accurate measurements of public opinions on several subjects of general interest.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Nos últimos anos, um grande esforço tem sido despendido no estudo de formas de representar documentos textuais, chamados semi-estruturados, objetivando extrair informações destes documentos com a mesma eficiência com que essas são extraídas de bancos de dados relacionais e orientados a objetos. A pesquisa, em dados semi-estruturados, tornou-se fundamental com o crescimento da Web como fonte e repositório de dados, uma vez que os métodos de pesquisa existentes, baseados em navegação e busca por palavra-chave, mostraram-se insuficientes para satisfazer as necessidades de consulta em escala cada vez maior. Com o surgimento da XML, e a tendência de se tornar a linguagem padrão na Web, fez com que a representação de informações fosse dirigida para este novo padrão, porque disponibiliza um bom intercâmbio de informações e a produção de documentos eletrônicos. Existe a necessidade de se disponibilizar os documentos e as informações na rede em formato Web, HTML ou XML. Sendo assim, consultar documentos XML representa um desafio para a comunidade de pesquisa em banco de dados, pois implica em disponibilizar os grandes volumes de dados já existentes em formato XML, surgindo a necessidade de ferramentas de consulta que sejam ao mesmo tempo, flexíveis o suficiente para compreender a heterogeneidade dos documentos e, poderosas ao ponto de extraírem informações com rapidez e correção. Este trabalho apresenta uma abordagem sobre a linguagem XML, sua importância, vantagens, principais aplicações e as linguagens de consulta para dados XML. Após, é detalhada uma aplicação para Web utilizando a tecnologia XML, baseado em Ontologias, e a sua disponibilização na Web. A aplicação desenvolvida utilizou XML e linguagens de consulta para XML e com suporte a XML, baseando-se em Ontologias, com o objetivo de permitir consultas e o armazenamento de informações referentes aos alunos concluintes de determinados cursos da Universidade da Região da Campanha - Urcamp/Bagé, colocando à disposição da Universidade uma nova ferramenta, que disponibiliza informações referentes aos cursos em questão, utilizando uma nova tecnologia, que tende a tornar-se padrão na Web.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

No presente trabalho é proposto um conjunto de atividades de aquisição de dados para um laboratório didático de Física, no contexto de ensino tecnológico de escolas técnicas. São construídas rotinas de programação com a ferramenta Visual Basic for Applications (VBA), incorporadas em planilhas Excel, sendo a aquisição de dados através da porta de joystick da placa de som do microcomputador. As atividades propostas envolvem especialmente experimentos no campo da mecânica, no qual são investigados o sistema massa-mola, pêndulo simples e movimento de rotação de um pião. A metodologia proposta para estas atividades é embasada na teórica construtivista de Vygotsky. O material foi aplicado e avaliado em uma oficina oferecida para alunos de cursos superiores de tecnologia (Automação Industrial e Manutenção Industrial e Sistemas de Informação) do CEFET/PR, unidade sudoeste do Paraná em 2005. O produto educacional deste trabalho consiste em um texto de apoio sobre medidas físicas e erros e sobre o uso da linguagem VBA em sistemas de aquisição de dados com a placa de som do microcomputador, e rotinas incorporadas a planilhas Excel que permitem a aquisição automática dos dados trabalhados nas atividades experimentais. No texto de apoio, também são sugeridas atividades para os alunos. As planilhas Excel (incluindo as macros VBA) constam o CD-ROM anexo a este trabalho, e estão disponibilizadas na web.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This study aims to identify, through the application of webometric indicators, which Post-Graduate Courses in Engineering recommended by the Coordination of Improvement of Higher Personnel Education (CAPES) in Brazil stand out in the web space, in relation to the communication process and dissemination of scientific information in the academic environment. For this, we analyzed the structures content of the sites, the use, through the conduct of investigations and searches, the quality of information available, as well as the structure of existent hypertexts in the sites of this universe of search. The tools and methodologies adopted for this study are: search engines (Google, Yahoo), Mapper software (Xenu Link Sleuth) and analysis software and visualization of networks (and Ucinet6 NetDraw). Webometric indicators are also used, such as size of the web sites, visibility, web impact factor, brightness and density of the network. These instruments provide a brief analysis and evaluation for this webometric study. Therefore, from the incursion of the literature used, it appears that there are many advantages of using this type of metric study in the so called Information Society. The obtained results could identify which postgraduate courses in engineering has a better availability of their information on the Web, as well to define which of these courses stands out in relation to the use of their information, which has been outstanding in respect to its impact factor and which offers a greater number of links that serve as a source of information for its users, contributing, in its turn, with the navigability of the same network. In summary, it is asserted that the webometric study presents promising results, which are able to achieve the proposed objectives, as well as identify the factors that contribute significantly to the good visualization of these sites in the network, thus helping the spread of information and scientific communication through the use of the Web.