552 resultados para Open mobile-guarding


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Specialist scholarly books, including monographs, allow researchers to present their work, pose questions and to test and extend areas of theory through long-form writing. In spite of the fact that research communities all over the world value monographs and depend heavily on them as a requirement of tenure and promotion in many disciplines, sales of this kind of book are in free fall, with some estimates suggesting declines of as much as 90% over twenty years (Willinsky 2006). Cashstrapped monograph publishers have found themselves caught in a negative cycle of increasing prices and falling sales, with few resources left to support experimentation, business model innovation or engagement with digital technology and Open Access (OA). This chapter considers an important attempt to tackle failing markets for scholarly monographs, and to enable the wider adoption of OA licenses for book-length works: the 2012 – 2014 Knowledge Unlatched pilot. Knowledge Unlatched is a bold attempt to reconfigure the market for specialist scholarly books: moving it beyond the sale of ‘content’ towards a model that supports the services valued by scholarly and wider communities in the context of digital possibility. Its success has powerful implications for the way we understand copyright’s role in the creative industries, and the potential for established institutions and infrastructure to support the open and networked dynamics of a digital age.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Networked digital technologies and Open Access (OA) are transforming the processes and institutions of research, knowledge creation and dissemination globally: enabling new forms of collaboration, allowing researchers to be seen and heard in new ways and reshaping relationships between stakeholders across the global academic publishing system. This article draws on Joseph Nye’s concept of ‘Soft Power’ to explore the role that OA is playing in helping to reshape academic publishing in China. It focusses on two important areas of OA development: OA journals and national-level repositories. OA is being supported at the highest levels, and there is potential for it to play an important role in increasing the status and impact of Chinese scholarship. Investments in OA also have the potential to help China to re-position itself within international copyright discourses: moving beyond criticism for failure to enforce the rights of foreign copyright owners and progressing an agenda that places greater emphasis on equality of access to the resources needed to foster innovation. However, the potential for OA to help China to build and project its soft power is being limited by the legacies of the print era, as well as the challenges of efficiently governing the national research and innovation systems.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A Delay Tolerant Network (DTN) is a dynamic, fragmented, and ephemeral network formed by a large number of highly mobile nodes. DTNs are ephemeral networks with highly mobile autonomous nodes. This requires distributed and self-organised approaches to trust management. Revocation and replacement of security credentials under adversarial influence by preserving the trust on the entity is still an open problem. Existing methods are mostly limited to detection and removal of malicious nodes. This paper makes use of the mobility property to provide a distributed, self-organising, and scalable revocation and replacement scheme. The proposed scheme effectively utilises the Leverage of Common Friends (LCF) trust system concepts to revoke compromised security credentials, replace them with new ones, whilst preserving the trust on them. The level of achieved entity confidence is thereby preserved. Security and performance of the proposed scheme is evaluated using an experimental data set in comparison with other schemes based around the LCF concept. Our extensive experimental results show that the proposed scheme distributes replacement credentials up to 35% faster and spreads spoofed credentials of strong collaborating adversaries up to 50% slower without causing any significant increase on the communication and storage overheads, when compared to other LCF based schemes.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Mobile applications are being increasingly deployed on a massive scale in various mobile sensor grid database systems. With limited resources from the mobile devices, how to process the huge number of queries from mobile users with distributed sensor grid databases becomes a critical problem for such mobile systems. While the fundamental semantic cache technique has been investigated for query optimization in sensor grid database systems, the problem is still difficult due to the fact that more realistic multi-dimensional constraints have not been considered in existing methods. To solve the problem, a new semantic cache scheme is presented in this paper for location-dependent data queries in distributed sensor grid database systems. It considers multi-dimensional constraints or factors in a unified cost model architecture, determines the parameters of the cost model in the scheme by using the concept of Nash equilibrium from game theory, and makes semantic cache decisions from the established cost model. The scenarios of three factors of semantic, time and locations are investigated as special cases, which improve existing methods. Experiments are conducted to demonstrate the semantic cache scheme presented in this paper for distributed sensor grid database systems.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This thesis investigates factors that impact the energy efficiency of a mining operation. An innovative mathematical framework and solution approach are developed to model, solve and analyse an open-pit coal mine. A case study in South East Queensland is investigated to validate the approach and explore the opportunities for using it to aid long, medium and short term decision makers.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The use of social networking has exploded, with millions of people using various web- and mobile-based services around the world. This increase in social networking use has led to user anxiety related to privacy and the unauthorised exposure of personal information. Large-scale sharing in virtual spaces means that researchers, designers and developers now need to re-consider the issues and challenges of maintaining privacy when using social networking services. This paper provides a comprehensive survey of the current state-of-the-art privacy in social networks for both desktop and mobile uses and devices from various architectural vantage points. The survey will assist researchers and analysts in academia and industry to move towards mitigating many of the privacy issues in social networks.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Open biomass burning from wildfires and the prescribed burning of forests and farmland is a frequent occurrence in South-East Queensland (SEQ), Australia. This work reports on data collected from 10-30 September 2011, which covers the days before (10-14 September), during (15-20 September) and after (21-30 September) a period of biomass burning in SEQ. The aim of this project was to comprehensively quantify the impact of the biomass burning on air quality in Brisbane, the capital city of Queensland. A multi-parameter field measurement campaign was conducted and ambient air quality data from 13 monitoring stations across SEQ were analysed. During the burning period, the average concentrations of all measured pollutants increased (from 20% to 430%) compared to the non-burning period (both before and after burning), except for total xylenes. The average concentration of O3, NO2, SO2, benzene, formaldehyde, PM10, PM2.5 and visibility-reducing particles reached their highest levels for the year, which were up to 10 times higher than annual average levels, while PM10, PM2.5 and SO2 concentrations exceeded the WHO 24-hour guidelines and O3 concentration exceeded the WHO maximum 8-hour average threshold during the burning period. Overall spatial variations showed that all measured pollutants, with the exception of O3, were closer to spatial homogeneity during the burning compared to the non-burning period. In addition to the above, elevated concentrations of three biomass burning organic tracers (levoglucosan, mannosan and galactosan), together with the amount of non-refractory organic particles (PM1) and the average value of f60 (attributed to levoglucosan), reinforce that elevated pollutant concentration levels were due to emissions from open biomass burning events, 70% of which were prescribed burning events. This study, which is the first and most comprehensive of its kind in Australia, provides quantitative evidence of the significant impact of open biomass burning events, especially prescribed burning, on urban air quality. The current results provide a solid platform for more detailed health and modelling investigations in the future.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper we focus on the challenging problem of place categorization and semantic mapping on a robot with-out environment-specific training. Motivated by their ongoing success in various visual recognition tasks, we build our system upon a state-of-the-art convolutional network. We overcome its closed-set limitations by complementing the network with a series of one-vs-all classifiers that can learn to recognize new semantic classes online. Prior domain knowledge is incorporated by embedding the classification system into a Bayesian filter framework that also ensures temporal coherence. We evaluate the classification accuracy of the system on a robot that maps a variety of places on our campus in real-time. We show how semantic information can boost robotic object detection performance and how the semantic map can be used to modulate the robot’s behaviour during navigation tasks. The system is made available to the community as a ROS module.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Background The irreversible ErbB family blocker afatinib and the reversible EGFR tyrosine kinase inhibitor gefitinib are approved for first-line treatment of EGFR mutation-positive non-small-cell lung cancer (NSCLC). We aimed to compare the efficacy and safety of afatinib and gefitinib in this setting. Methods This multicentre, international, open-label, exploratory, randomised controlled phase 2B trial (LUX-Lung 7) was done at 64 centres in 13 countries. Treatment-naive patients with stage IIIB or IV NSCLC and a common EGFR mutation (exon 19 deletion or Leu858Arg) were randomly assigned (1:1) to receive afatinib (40 mg per day) or gefitinib (250 mg per day) until disease progression, or beyond if deemed beneficial by the investigator. Randomisation, stratified by EGFR mutation type and status of brain metastases, was done centrally using a validated number generating system implemented via an interactive voice or web-based response system with a block size of four. Clinicians and patients were not masked to treatment allocation; independent review of tumour response was done in a blinded manner. Coprimary endpoints were progression-free survival by independent central review, time-to-treatment failure, and overall survival. Efficacy analyses were done in the intention-to-treat population and safety analyses were done in patients who received at least one dose of study drug. This ongoing study is registered with ClinicalTrials.gov, number NCT01466660. Findings Between Dec 13, 2011, and Aug 8, 2013, 319 patients were randomly assigned (160 to afatinib and 159 to gefitinib). Median follow-up was 27·3 months (IQR 15·3–33·9). Progression-free survival (median 11·0 months [95% CI 10·6–12·9] with afatinib vs 10·9 months [9·1–11·5] with gefitinib; hazard ratio [HR] 0·73 [95% CI 0·57–0·95], p=0·017) and time-to-treatment failure (median 13·7 months [95% CI 11·9–15·0] with afatinib vs 11·5 months [10·1–13·1] with gefitinib; HR 0·73 [95% CI 0·58–0·92], p=0·0073) were significantly longer with afatinib than with gefitinib. Overall survival data are not mature. The most common treatment-related grade 3 or 4 adverse events were diarrhoea (20 [13%] of 160 patients given afatinib vs two [1%] of 159 given gefitinib) and rash or acne (15 [9%] patients given afatinib vs five [3%] of those given gefitinib) and liver enzyme elevations (no patients given afatinib vs 14 [9%] of those given gefitinib). Serious treatment-related adverse events occurred in 17 (11%) patients in the afatinib group and seven (4%) in the gefitinib group. Ten (6%) patients in each group discontinued treatment due to drug-related adverse events. 15 (9%) fatal adverse events occurred in the afatinib group and ten (6%) in the gefitinib group. All but one of these deaths were considered unrelated to treatment; one patient in the gefitinib group died from drug-related hepatic and renal failure. Interpretation Afatinib significantly improved outcomes in treatment-naive patients with EGFR-mutated NSCLC compared with gefitinib, with a manageable tolerability profile. These data are potentially important for clinical decision making in this patient population.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Background: This multicentre, open-label, randomized, controlled phase II study evaluated cilengitide in combination with cetuximab and platinum-based chemotherapy, compared with cetuximab and chemotherapy alone, as first-line treatment of patients with advanced non-small-cell lung cancer (NSCLC). Patients and methods: Patients were randomized 1:1:1 to receive cetuximab plus platinum-based chemotherapy alone (control), or combined with cilengitide 2000 mg 1×/week i.v. (CIL-once) or 2×/week i.v. (CIL-twice). A protocol amendment limited enrolment to patients with epidermal growth factor receptor (EGFR) histoscore ≥200 and closed the CIL-twice arm for practical feasibility issues. Primary end point was progression-free survival (PFS; independent read); secondary end points included overall survival (OS), safety, and biomarker analyses. A comparison between the CIL-once and control arms is reported, both for the total cohorts, as well as for patients with EGFR histoscore ≥200. Results: There were 85 patients in the CIL-once group and 84 in the control group. The PFS (independent read) was 6.2 versus 5.0 months for CIL-once versus control [hazard ratio (HR) 0.72; P = 0.085]; for patients with EGFR histoscore ≥200, PFS was 6.8 versus 5.6 months, respectively (HR 0.57; P = 0.0446). Median OS was 13.6 for CIL-once versus 9.7 months for control (HR 0.81; P = 0.265). In patients with EGFR ≥200, OS was 13.2 versus 11.8 months, respectively (HR 0.95; P = 0.855). No major differences in adverse events between CIL-once and control were reported; nausea (59% versus 56%, respectively) and neutropenia (54% versus 46%, respectively) were the most frequent. There was no increased incidence of thromboembolic events or haemorrhage in cilengitide-treated patients. αvβ3 and αvβ5 expression was neither a predictive nor a prognostic indicator. Conclusions: The addition of cilengitide to cetuximab/chemotherapy indicated potential clinical activity, with a trend for PFS difference in the independent-read analysis. However, the observed inconsistencies across end points suggest additional investigations are required to substantiate a potential role of other integrin inhibitors in NSCLC treatment.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The Body Area Network (BAN) is an emerging technology that focuses on monitoring physiological data in, on and around the human body. BAN technology permits wearable and implanted sensors to collect vital data about the human body and transmit it to other nodes via low-energy communication. In this paper, we investigate interactions in terms of data flows between parties involved in BANs under four different scenarios targeting outdoor and indoor medical environments: hospital, home, emergency and open areas. Based on these scenarios, we identify data flow requirements between BAN elements such as sensors and control units (CUs) and parties involved in BANs such as the patient, doctors, nurses and relatives. Identified requirements are used to generate BAN data flow models. Petri Nets (PNs) are used as the formal modelling language. We check the validity of the models and compare them with the existing related work. Finally, using the models, we identify communication and security requirements based on the most common active and passive attack scenarios.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The future of civic engagement is characterised by both technological innovation as well as new technological user practices that are fuelled by trends towards mobile, personal devices; broadband connectivity; open data; urban interfaces; and cloud computing. These technology trends are progressing at a rapid pace, and have led global technology vendors to package and sell the “Smart City” as a centralised service delivery platform predicted to optimise and enhance cities’ key performance indicators – and generate a profitable market. The top-down deployment of these large and proprietary technology platforms have helped sectors such as energy, transport, and healthcare to increase efficiencies. However, an increasing number of scholars and commentators warn of another “IT bubble” emerging. Along with some city leaders, they argue that the top-down approach does not fit the governance dynamics and values of a liberal democracy when applied across sectors. A thorough understanding is required, of the socio-cultural nuances of how people work, live, play across different environments, and how they employ social media and mobile devices to interact with, engage in, and constitute public realms. Although the term “slacktivism” is sometimes used to denote a watered down version of civic engagement and activism that is reduced to clicking a “Like” button and signing online petitions, we believe that we are far from witnessing another Biedermeier period that saw people focus on the domestic and the non-political. There is plenty of evidence to the contrary, such as post-election violence in Kenya in 2008, the Occupy movements in New York, Hong Kong and elsewhere, the Arab Spring, Stuttgart 21, Fukushima, the Taksim Gezi Park in Istanbul, and the Vinegar Movement in Brazil in 2013. These examples of civic action shape the dynamics of governments, and in turn, call for new processes to be incorporated into governance structures. Participatory research into these new processes across the triad of people, place and technology is a significant and timely investment to foster productive, sustainable, and liveable human habitats. With this article, we want to reframe the current debates in academia and priorities in industry and government to allow citizens and civic actors to take their rightful centrepiece place in civic movements. This calls for new participatory approaches for co-inquiry and co-design. It is an evolving process with an explicit agenda to facilitate change, and we propose participatory action research (PAR) as an indispensable component in the journey to develop new governance infrastructures and practices for civic engagement. We do not limit our definition of civic technologies to tools specifically designed to simply enhance government and governance, such as renewing your car registration online or casting your vote electronically on election day. Rather, we are interested in civic media and technologies that foster citizen engagement in the widest sense, and particularly the participatory design of such civic technologies that strive to involve citizens in political debate and action as well as question conventional approaches to political issues. The rationale for this approach is an alternative to smart cities in a “perpetual tomorrow,” based on many weak and strong signals of civic actions revolving around technology seen today. It seeks to emphasise and direct attention to active citizenry over passive consumerism, human actors over human factors, culture over infrastructure, and prosperity over efficiency. First, we will have a look at some fundamental issues arising from applying simplistic smart city visions to the kind of a problem a city poses. We focus on the touch points between “the city” and its civic body, the citizens. In order to provide for meaningful civic engagement, the city must provide appropriate interfaces.