958 resultados para WEB (Computer program language)


Relevância:

40.00% 40.00%

Publicador:

Resumo:

El projecte té per objecte explorar les possibilitats dels sistemes d'informació geogràfica en els estudis de vigilància epidemiològica (VE).

Relevância:

40.00% 40.00%

Publicador:

Resumo:

L'objectiu d'aquest projecte és millorar la situació actual del servei tècnic de reparació i manteniment d'equipament mèdic de l'Hospital Verge de la Cinta i dels centres d'atenció primària de Tortosa i crear unes interfícies clares, ordenades, fàcils i intuïtives per als usuaris, ja siguin els clients o els propis empleats, i que serviran de pont de comunicació amb el servidor central de l'organització on estarà tota la seva informació. Per crear les interfícies s'utilitzarà la metodologia del disseny centrat en l'usuari (DCU) i, a través d'un prototip, s'avaluaran les tasques que han de realitzar els usuaris (clients i empleats).

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Across Latin America 420 indigenous languages are spoken. Spanish is considered a second language in indigenous communities and is progressively introduced in education. However, most of the tools to support teaching processes of a second language have been developed for the most common languages such as English, French, German, Italian, etc. As a result, only a small amount of learning objects and authoring tools have been developed for indigenous people considering the specific needs of their population. This paper introduces Multilingual–Tiny as a web authoring tool to support the virtual experience of indigenous students and teachers when they are creating learning objects in indigenous languages or in Spanish language, in particular, when they have to deal with the grammatical structures of Spanish. Multilingual–Tiny has a module based on the Case-based Reasoning technique to provide recommendations in real time when teachers and students write texts in Spanish. An experiment was performed in order to compare some local similarity functions to retrieve cases from the case library taking into account the grammatical structures. As a result we found the similarity function with the best performance

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Communication, the flow of ideas and information between individuals in a social context, is the heart of educational experience. Constructivism and constructivist theories form the foundation for the collaborative learning processes of creating and sharing meaning in online educational contexts. The Learning and Collaboration in Technology-enhanced Contexts (LeCoTec) course comprised of 66 participants drawn from four European universities (Oulu, Turku, Ghent and Ramon Llull). These participants were split into 15 groups with the express aim of learning about computer-supported collaborative learning (CSCL). The Community of Inquiry model (social, cognitive and teaching presences) provided the content and tools for learning and researching the collaborative interactions in this environment. The sampled comments from the collaborative phase were collected and analyzed at chain-level and group-level, with the aim of identifying the various message types that sustained high learning outcomes. Furthermore, the Social Network Analysis helped to view the density of whole group interactions, as well as the popular and active members within the highly collaborating groups. It was observed that long chains occur in groups having high quality outcomes. These chains were also characterized by Social, Interactivity, Administrative and Content comment-types. In addition, high outcomes were realized from the high interactive cases and high-density groups. In low interactive groups, commenting patterned around the one or two central group members. In conclusion, future online environments should support high-order learning and develop greater metacognition and self-regulation. Moreover, such an environment, with a wide variety of problem solving tools, would enhance interactivity.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

A web service is a software system that provides a machine-processable interface to the other machines over the network using different Internet protocols. They are being increasingly used in the industry in order to automate different tasks and offer services to a wider audience. The REST architectural style aims at producing scalable and extensible web services using technologies that play well with the existing tools and infrastructure of the web. It provides a uniform set of operation that can be used to invoke a CRUD interface (create, retrieve, update and delete) of a web service. The stateless behavior of the service interface requires that every request to a resource is independent of the previous ones facilitating scalability. Automated systems, e.g., hotel reservation systems, provide advanced scenarios for stateful services that require a certain sequence of requests that must be followed in order to fulfill the service goals. Designing and developing such services for advanced scenarios with REST constraints require rigorous approaches that are capable of creating web services that can be trusted for their behavior. Systems that can be trusted for their behavior can be termed as dependable systems. This thesis presents an integrated design, analysis and validation approach that facilitates the service developer to create dependable and stateful REST web services. The main contribution of this thesis is that we provide a novel model-driven methodology to design behavioral REST web service interfaces and their compositions. The behavioral interfaces provide information on what methods can be invoked on a service and the pre- and post-conditions of these methods. The methodology uses Unified Modeling Language (UML), as the modeling language, which has a wide user base and has mature tools that are continuously evolving. We have used UML class diagram and UML state machine diagram with additional design constraints to provide resource and behavioral models, respectively, for designing REST web service interfaces. These service design models serve as a specification document and the information presented in them have manifold applications. The service design models also contain information about the time and domain requirements of the service that can help in requirement traceability which is an important part of our approach. Requirement traceability helps in capturing faults in the design models and other elements of software development environment by tracing back and forth the unfulfilled requirements of the service. The information about service actors is also included in the design models which is required for authenticating the service requests by authorized actors since not all types of users have access to all the resources. In addition, following our design approach, the service developer can ensure that the designed web service interfaces will be REST compliant. The second contribution of this thesis is consistency analysis of the behavioral REST interfaces. To overcome the inconsistency problem and design errors in our service models, we have used semantic technologies. The REST interfaces are represented in web ontology language, OWL2, that can be part of the semantic web. These interfaces are used with OWL 2 reasoners to check unsatisfiable concepts which result in implementations that fail. This work is fully automated thanks to the implemented translation tool and the existing OWL 2 reasoners. The third contribution of this thesis is the verification and validation of REST web services. We have used model checking techniques with UPPAAL model checker for this purpose. The timed automata of UML based service design models are generated with our transformation tool that are verified for their basic characteristics like deadlock freedom, liveness, reachability and safety. The implementation of a web service is tested using a black-box testing approach. Test cases are generated from the UPPAAL timed automata and using the online testing tool, UPPAAL TRON, the service implementation is validated at runtime against its specifications. Requirement traceability is also addressed in our validation approach with which we can see what service goals are met and trace back the unfulfilled service goals to detect the faults in the design models. A final contribution of the thesis is an implementation of behavioral REST interfaces and service monitors from the service design models. The partial code generation tool creates code skeletons of REST web services with method pre and post-conditions. The preconditions of methods constrain the user to invoke the stateful REST service under the right conditions and the post condition constraint the service developer to implement the right functionality. The details of the methods can be manually inserted by the developer as required. We do not target complete automation because we focus only on the interface aspects of the web service. The applicability of the approach is demonstrated with a pedagogical example of a hotel room booking service and a relatively complex worked example of holiday booking service taken from the industrial context. The former example presents a simple explanation of the approach and the later worked example shows how stateful and timed web services offering complex scenarios and involving other web services can be constructed using our approach.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

With the growth in new technologies, using online tools have become an everyday lifestyle. It has a greater impact on researchers as the data obtained from various experiments needs to be analyzed and knowledge of programming has become mandatory even for pure biologists. Hence, VTT came up with a new tool, R Executables (REX) which is a web application designed to provide a graphical interface for biological data functions like Image analysis, Gene expression data analysis, plotting, disease and control studies etc., which employs R functions to provide results. REX provides a user interactive application for the biologists to directly enter the values and run the required analysis with a single click. The program processes the given data in the background and prints results rapidly. Due to growth of data and load on server, the interface has gained problems concerning time consumption, poor GUI, data storage issues, security, minimal user interactive experience and crashes with large amount of data. This thesis handles the methods by which these problems were resolved and made REX a better application for the future. The old REX was developed using Python Django and now, a new programming language, Vaadin has been implemented. Vaadin is a Java framework for developing web applications and the programming language is extremely similar to Java with new rich components. Vaadin provides better security, better speed, good and interactive interface. In this thesis, subset functionalities of REX was selected which includes IST bulk plotting and image segmentation and implemented those using Vaadin. A code of 662 lines was programmed by me which included Vaadin as the front-end handler while R language was used for back-end data retrieval, computing and plotting. The application is optimized to allow further functionalities to be migrated with ease from old REX. Future development is focused on including Hight throughput screening functions along with gene expression database handling

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The current study investigated the effects that barriers (both real and perceived) had on participation and completion of speech and language programs for preschool children with communication delays. I compared 36 families of preschool children with an identified communication delay that have completed services (completers) to 13 families that have not completed services (non-completers) prescribed by Speech and Language professionals. Data findings reported were drawn from an interview with the mother, a speech and language assessment of the child, and an extensive package of measures completed by the mother. Children ranged in age from 32 to 71 mos. These data were collected as part of a project funded by the Canadian Language and Literacy Research Networks of Centres of Excellence. Findings suggest that completers and non-completers shared commonalities in a number of parenting characteristics but differed significantly in two areas. Mothers in the noncompleting group were more permissive and had lower maternal education than mothers in the completing families. From a systemic standpoint, families also differed in the number of perceived barriers to treatment experienced during their time with Speech Services Niagara. Mothers in the non-completing group experienced more perceived barriers to treatment than completing mothers. Specifically, these mothers perceived more stressors and obstacles that competed with treatment, perceived more treatment demands and they perceived the relevance of treatment as less important than the completing group. Despite this, the findings suggest that non-completing families were 100% satisfied with services. Contrary to predictions, there were no significant differences in child characterisfics and economic characteristics between completers and non-completers. The findings in this study are considered exploratory and tentative due to the small sample size.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Although much research has explored computer mediated communication for its application in second language instruction, there still exists a need for empirical results from research to guide practitioners who wish to introduce web-based activities into their instruction. This study was undertaken to explore collaborative online task-based activities for the instruction of ESL academic writing. Nine ESL students in their midtwenties, enrolled at a community college in Ontario, engaged in two separate online prewriting activities in both a synchronous and an asynchronous environment. The students were interviewed in order to explore their perceptions of how the activities affected the generation and organization of ideas for academic essays. These interviews were triangulated with examples of the students' online writing, nonparticipatory observations of the students' interactions, and a discussion with the course instructor. The results of the study reveal that a small majority of students felt that brainstorming in writing with their peers in an asynchronous online discussion created a grammatical and lexical framework that supported idea generation and organization. The students did not feel that the synchronous chat activity was as successful. Although they felt that this activity also contributed to the generation of ideas, synchronous chat introduced a level of difficulty in communication that hindered the students' engagement in the task and failed to assist them with the organization of their ideas. The students also noted positive aspects of the web-based activities that were not related to prewriting tasks, for example, improved typing and word processing skills. Directions for future research could explore whether online prewriting activities can assist students in the creation of essays that are syntactically or lexically complex.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Over the past decade, research has suggested that phonological and word awareness skills (i.e., the ability to reflect on and manipulate the components of language) are important for early reading acquisition. This study examined the phonological and word awareness skills of language-delayed kindergarten children at the beginning and end of a language intervention program using five tasks. The results were compared to the performances of average kindergarten children who did not participate in the language intervention program. There were significant performance differences for all tasks, favouring the average children, at the beginning of the intervention program. However, at the end of the training interval, the languagedelayed children performed as well as the average children.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This study investigated the needs of adult ESL learners intending to pursue higher education in Canada. Its chief purpose was to enable educators and administrators to design ESL programs that would prepare students to function at optimal levels in academic and social settings during their university studies. The study adopted a mixed research method that was predominantly qualitative in its orientation and narrative in its implementation. It focused on an Intensive English Language Program (IELP) offered at an Ontario university. Using a holistic approach, the study sought to represent the various perspectives of all the participants in the program: the students, the instructors, and the administrators. Analysis of both qualitative and quantitative data gathered from 17 students, 6 instructors, and 1 administrator in the IELP showed that to a large extent the academic needs ofESL learners in the IELP were generally not being met. Most notably, the study found that learners were not receiving sufficient training in speaking and listening skills, a factor that contributed to their sense of insecurity and lack of confidence in their ability to communicate successfully in academic and social settings. The study also revealed that the solutions to many of the problems it identified lay not in the classroom but in the way the ESL program was structured administratively. One major recommendation to come out of the study is that programs like the IELP should be restructured so as to give them greater flexibility in meeting individual needs. While the study labored under certain limitations and did not achieve all of its goals, it did succeed in creating awareness ofthe problems and in establishing a methodological approach that can serve as a framework within which future research may be conducted in this somewhat neglected area.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

L'utilisation des méthodes formelles est de plus en plus courante dans le développement logiciel, et les systèmes de types sont la méthode formelle qui a le plus de succès. L'avancement des méthodes formelles présente de nouveaux défis, ainsi que de nouvelles opportunités. L'un des défis est d'assurer qu'un compilateur préserve la sémantique des programmes, de sorte que les propriétés que l'on garantit à propos de son code source s'appliquent également au code exécutable. Cette thèse présente un compilateur qui traduit un langage fonctionnel d'ordre supérieur avec polymorphisme vers un langage assembleur typé, dont la propriété principale est que la préservation des types est vérifiée de manière automatisée, à l'aide d'annotations de types sur le code du compilateur. Notre compilateur implante les transformations de code essentielles pour un langage fonctionnel d'ordre supérieur, nommément une conversion CPS, une conversion des fermetures et une génération de code. Nous présentons les détails des représentation fortement typées des langages intermédiaires, et les contraintes qu'elles imposent sur l'implantation des transformations de code. Notre objectif est de garantir la préservation des types avec un minimum d'annotations, et sans compromettre les qualités générales de modularité et de lisibilité du code du compilateur. Cet objectif est atteint en grande partie dans le traitement des fonctionnalités de base du langage (les «types simples»), contrairement au traitement du polymorphisme qui demande encore un travail substantiel pour satisfaire la vérification de type.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Les structures avec des lieurs sont très communes en informatique. Les langages de programmation et les systèmes logiques sont des exemples de structures avec des lieurs. La manipulation de lieurs est délicate, de sorte que l’écriture de programmes qui ma- nipulent ces structures tirerait profit d’un soutien spécifique pour les lieurs. L’environ- nement de programmation Beluga est un exemple d’un tel système. Nous développons et présentons ici un compilateur pour ce système. Parmi les programmes pour lesquels Beluga est spécialement bien adapté, plusieurs peuvent bénéficier d’un compilateur. Par exemple, les programmes pour valider les types (les "type-checkers"), les compilateurs et les interpréteurs tirent profit du soutien spécifique des lieurs et des types dépendants présents dans le langage. Ils nécessitent tous également une exécution efficace, que l’on propose d’obtenir par le biais d’un compilateur. Le but de ce travail est de présenter un nouveau compilateur pour Beluga, qui emploie une représentation interne polyvalente et permet de partager du code entre plusieurs back-ends. Une contribution notable est la compilation du filtrage de Beluga, qui est particulièrement puissante dans ce langage.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Modern computer systems are plagued with stability and security problems: applications lose data, web servers are hacked, and systems crash under heavy load. Many of these problems or anomalies arise from rare program behavior caused by attacks or errors. A substantial percentage of the web-based attacks are due to buffer overflows. Many methods have been devised to detect and prevent anomalous situations that arise from buffer overflows. The current state-of-art of anomaly detection systems is relatively primitive and mainly depend on static code checking to take care of buffer overflow attacks. For protection, Stack Guards and I-leap Guards are also used in wide varieties.This dissertation proposes an anomaly detection system, based on frequencies of system calls in the system call trace. System call traces represented as frequency sequences are profiled using sequence sets. A sequence set is identified by the starting sequence and frequencies of specific system calls. The deviations of the current input sequence from the corresponding normal profile in the frequency pattern of system calls is computed and expressed as an anomaly score. A simple Bayesian model is used for an accurate detection.Experimental results are reported which show that frequency of system calls represented using sequence sets, captures the normal behavior of programs under normal conditions of usage. This captured behavior allows the system to detect anomalies with a low rate of false positives. Data are presented which show that Bayesian Network on frequency variations responds effectively to induced buffer overflows. It can also help administrators to detect deviations in program flow introduced due to errors.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Web services from different partners can be combined to applications that realize a more complex business goal. Such applications built as Web service compositions define how interactions between Web services take place in order to implement the business logic. Web service compositions not only have to provide the desired functionality but also have to comply with certain Quality of Service (QoS) levels. Maximizing the users' satisfaction, also reflected as Quality of Experience (QoE), is a primary goal to be achieved in a Service-Oriented Architecture (SOA). Unfortunately, in a dynamic environment like SOA unforeseen situations might appear like services not being available or not responding in the desired time frame. In such situations, appropriate actions need to be triggered in order to avoid the violation of QoS and QoE constraints. In this thesis, proper solutions are developed to manage Web services and Web service compositions with regard to QoS and QoE requirements. The Business Process Rules Language (BPRules) was developed to manage Web service compositions when undesired QoS or QoE values are detected. BPRules provides a rich set of management actions that may be triggered for controlling the service composition and for improving its quality behavior. Regarding the quality properties, BPRules allows to distinguish between the QoS values as they are promised by the service providers, QoE values that were assigned by end-users, the monitored QoS as measured by our BPR framework, and the predicted QoS and QoE values. BPRules facilitates the specification of certain user groups characterized by different context properties and allows triggering a personalized, context-aware service selection tailored for the specified user groups. In a service market where a multitude of services with the same functionality and different quality values are available, the right services need to be selected for realizing the service composition. We developed new and efficient heuristic algorithms that are applied to choose high quality services for the composition. BPRules offers the possibility to integrate multiple service selection algorithms. The selection algorithms are applicable also for non-linear objective functions and constraints. The BPR framework includes new approaches for context-aware service selection and quality property predictions. We consider the location information of users and services as context dimension for the prediction of response time and throughput. The BPR framework combines all new features and contributions to a comprehensive management solution. Furthermore, it facilitates flexible monitoring of QoS properties without having to modify the description of the service composition. We show how the different modules of the BPR framework work together in order to execute the management rules. We evaluate how our selection algorithms outperform a genetic algorithm from related research. The evaluation reveals how context data can be used for a personalized prediction of response time and throughput.