852 resultados para SERVER
Resumo:
Com Sona l'ESO és un projecte en què s’integren música i dansa, literatura, llengua i educació amb una doble finalitat: traure fora de les quatre parets de l’aula el treball que es fa dia a dia amb l’alumnat i oferir-los l’oportunitat d’actuar en directe damunt d’un escenari. En aquest article es descriu el treball dins i fora de l’aula que culmina en aquest espectacle durant 16 edicions està recorrent diferents poblacions i centres del territori lingüístic i es posa en valor el caràcter multidisciplinari, la implicació social, el protagonisme de l’alumnat i l’èxit d’un treball que esdevé plenament aprenentatge significatiu.
Resumo:
In today's internet world, web browsers are an integral part of our day-to-day activities. Therefore, web browser security is a serious concern for all of us. Browsers can be breached in different ways. Because of the over privileged access, extensions are responsible for many security issues. Browser vendors try to keep safe extensions in their official extension galleries. However, their security control measures are not always effective and adequate. The distribution of unsafe extensions through different social engineering techniques is also a very common practice. Therefore, before installation, users should thoroughly analyze the security of browser extensions. Extensions are not only available for desktop browsers, but many mobile browsers, for example, Firefox for Android and UC browser for Android, are also furnished with extension features. Mobile devices have various resource constraints in terms of computational capabilities, power, network bandwidth, etc. Hence, conventional extension security analysis techniques cannot be efficiently used by end users to examine mobile browser extension security issues. To overcome the inadequacies of the existing approaches, we propose CLOUBEX, a CLOUd-based security analysis framework for both desktop and mobile Browser EXtensions. This framework uses a client-server architecture model. In this framework, compute-intensive security analysis tasks are generally executed in a high-speed computing server hosted in a cloud environment. CLOUBEX is also enriched with a number of essential features, such as client-side analysis, requirements-driven analysis, high performance, and dynamic decision making. At present, the Firefox extension ecosystem is most susceptible to different security attacks. Hence, the framework is implemented for the security analysis of the Firefox desktop and Firefox for Android mobile browser extensions. A static taint analysis is used to identify malicious information flows in the Firefox extensions. In CLOUBEX, there are three analysis modes. A dynamic decision making algorithm assists us to select the best option based on some important parameters, such as the processing speed of a client device and network connection speed. Using the best analysis mode, performance and power consumption are improved significantly. In the future, this framework can be leveraged for the security analysis of other desktop and mobile browser extensions, too.
Resumo:
Introduction – Based on a previous project of University of Lisbon (UL) – a Bibliometric Benchmarking Analysis of University of Lisbon, for the period of 2000-2009 – a database was created to support research information (ULSR). However this system was not integrated with other existing systems at University, as the UL Libraries Integrated System (SIBUL) and the Repository of University of Lisbon (Repositório.UL). Since libraries were called to be part of the process, the Faculty of Pharmacy Library’ team felt that it was very important to get all systems connected or, at least, to use that data in the library systems. Objectives – The main goals were to centralize all the scientific research produced at Faculty of Pharmacy, made it available to the entire Faculty, involve researchers and library team, capitalize and reinforce team work with the integration of several distinct projects and reducing tasks’ redundancy. Methods – Our basis was the imported data collection from the ISI Web of Science (WoS), for the period of 2000-2009, into ULSR. All the researchers and indexed publications at WoS, were identified. A first validation to identify all the researchers and their affiliation (university, faculty, department and unit) was done. The final validation was done by each researcher. In a second round, concerning the same period, all Pharmacy Faculty researchers identified their published scientific work in other databases/resources (NOT WoS). To our strategy, it was important to get all the references and essential/critical to relate them with the correspondent digital objects. To each researcher previously identified, was requested to register all their references of the ‘NOT WoS’ published works, at ULSR. At the same time, they should submit all PDF files (for both WoS and NOT WoS works) in a personal area of the Web server. This effort enabled us to do a more reliable validation and prepare the data and metadata to be imported to Repository and to Library Catalogue. Results – 558 documents related with 122 researchers, were added into ULSR. 1378 bibliographic records (WoS + NOT WoS) were converted into UNIMARC and Dublin Core formats. All records were integrated in the catalogue and repository. Conclusions – Although different strategies could be adopted, according to each library team, we intend to share this experience and give some tips of what could be done and how Faculty of Pharmacy created and implemented her strategy.
Resumo:
Relatório de estágio apresentada para cumprimento dos requisitos necessários à obtenção do grau de Mestre em Sistemas de Informação Organizacionais
Resumo:
Mode of access: Internet.
Resumo:
Cybercrime and related malicious activity in our increasingly digital world has become more prevalent and sophisticated, evading traditional security mechanisms. Digital forensics has been proposed to help investigate, understand and eventually mitigate such attacks. The practice of digital forensics, however, is still fraught with various challenges. Some of the most prominent of these challenges include the increasing amounts of data and the diversity of digital evidence sources appearing in digital investigations. Mobile devices and cloud infrastructures are an interesting specimen, as they inherently exhibit these challenging circumstances and are becoming more prevalent in digital investigations today. Additionally they embody further characteristics such as large volumes of data from multiple sources, dynamic sharing of resources, limited individual device capabilities and the presence of sensitive data. These combined set of circumstances make digital investigations in mobile and cloud environments particularly challenging. This is not aided by the fact that digital forensics today still involves manual, time consuming tasks within the processes of identifying evidence, performing evidence acquisition and correlating multiple diverse sources of evidence in the analysis phase. Furthermore, industry standard tools developed are largely evidence-oriented, have limited support for evidence integration and only automate certain precursory tasks, such as indexing and text searching. In this study, efficiency, in the form of reducing the time and human labour effort expended, is sought after in digital investigations in highly networked environments through the automation of certain activities in the digital forensic process. To this end requirements are outlined and an architecture designed for an automated system that performs digital forensics in highly networked mobile and cloud environments. Part of the remote evidence acquisition activity of this architecture is built and tested on several mobile devices in terms of speed and reliability. A method for integrating multiple diverse evidence sources in an automated manner, supporting correlation and automated reasoning is developed and tested. Finally the proposed architecture is reviewed and enhancements proposed in order to further automate the architecture by introducing decentralization particularly within the storage and processing functionality. This decentralization also improves machine to machine communication supporting several digital investigation processes enabled by the architecture through harnessing the properties of various peer-to-peer overlays. Remote evidence acquisition helps to improve the efficiency (time and effort involved) in digital investigations by removing the need for proximity to the evidence. Experiments show that a single TCP connection client-server paradigm does not offer the required scalability and reliability for remote evidence acquisition and that a multi-TCP connection paradigm is required. The automated integration, correlation and reasoning on multiple diverse evidence sources demonstrated in the experiments improves speed and reduces the human effort needed in the analysis phase by removing the need for time-consuming manual correlation. Finally, informed by published scientific literature, the proposed enhancements for further decentralizing the Live Evidence Information Aggregator (LEIA) architecture offer a platform for increased machine-to-machine communication thereby enabling automation and reducing the need for manual human intervention.
Resumo:
Thesis (Ph.D.)--University of Washington, 2016-06
Resumo:
The literature contains a number of reports of early work involving telemedicine and chronic disease; however, there are comparatively few studies in asthma. Most of the telemedicine studies in asthma have investigated the use of remote monitoring of patients in the home, e.g. transmitting spirometry data via a telephone modem to a central server. The primary objective of these studies was to improve management. A secondary benefit was that patient adherence to prescribed treatment is also likely to be improved. Early results are encouraging; home monitoring in a randomized controlled trial in Japan significantly reduced the number of emergency room visits by patients with poorly controlled asthma. Other studies have described the cost-benefits of a specialist asthma nurse who can manage patients by telephone contact, as well as deliver asthma education. Many web-based systems are available for the general public or healthcare professionals to improve education in asthma, although their quality is highly variable. The work on telemedicine in asthma clearly shows that the technique holds promise in a number of areas. Unfortunately - as in telemedicine generally - most of the literature in patients with asthma refers to pilot trials and feasibility studies, with short-term outcomes. Large-scale, formal research trials are required to establish the cost effectiveness of telemedicine in asthma.
Resumo:
Three different, well established systems for e-referral were examined. They ranged from a system in a single country handling a large number of cases (60,000 per year) to a global system covering many countries which handled fewer cases (150 per year). Nonetheless, there appeared to be a number of common features. Whether the purpose is e-transfer or e-consultation, the underlying model of the e-referral process is: the referrer initiates an e-request; the organization managing the process receives it, the organization allocates it for reply; the responder replies to the initiator. Various things can go wrong and the organization managing the e-referral process needs to be able to track requests through the system; this requires various performance metrics. E-referral can be conducted using email, or as messages passed either directly between computer systems or via a Web-link to a server. The experience of the three systems studied shows that significant changes in work practice are needed to launch an e-referral service successfully. The use of e-referral between primary and secondary care improves access to services and can be shown to be cost-effective.
Resumo:
The Swinfen Charitable Trust (SCT) provided two kinds of telemedical support to Iraq during 2004. Starting in January 2004, the Al-Yarmouk Teaching Hospital in Baghdad was able to refer cases into the well established global e-health network that the SCT has operated for the last five years. (In the first quarter of 2004, the SCT dealt with a total of 57 referrals from 15 hospitals in eight countries.) Two cases were referred from Baghdad in March 2004, both gynaecological, which were dealt with by consultants from the UK and Australia. The SCT administrators visited Basrah during April 2004 and met Iraqi doctors at the Shaibah Hospital as part of the international initiatives to improve health care there. Following this visit, the SCT network expanded to include another four hospitals in Iraq (Table 1). In addition, the SCT provided an electronic health records (EHR) system to support the rebuilding of maternity services, which has been led by the British Royal Colleges. The maternity records system is a Web-based EHR system, running on a secure server, which allows integrated access from antenatal clinics, from hospitals and from postnatal clinics in Iraq. Patients can view their own notes, thus promoting ownership of medical information, and doctors can view the notes of their own patients, from any Internet-connected PC. No special software is required by the user.
Resumo:
An ontology is increasingly becoming an essential tool for solving problems in many research areas. The ontology is a complex information object. It can contain millions of concepts in complex relationships. When we want to manage complex information objects, we generally turn to information systems technology. An information system intended to manage ontology is called an ontology server. The ontology server technology is at the time of writing quite immature. Therefore, this paper reviews and compares the main ontology servers that have been reported in the literatures. As a result, we point out several research questions related to server technology