875 resultados para big data storage


Relevância:

80.00% 80.00%

Publicador:

Resumo:

Conducting research in the rapidly evolving fields constituting the digital social sciences raises challenging ethical and technical issues, especially when the subject matter includes activities of stigmatised populations. Our study of a dark-web drug-use community provides a case example of ‘how to’ conduct studies in digital environments where sensitive and illicit activities are discussed. In this paper we present the workflow from our digital ethnography and consider the consequences of particular choices of action upon knowledge production. Key considerations that our workflow responded to include adapting to volatile field-sites, researcher safety in digital environments, data security and encryption, and ethical-legal challenges. We anticipate that this workflow may assist other researchers to emulate, test and adapt our approach to the diverse range of illicit studies online. In this paper we argue that active engagement with stigmatised communities through multi-sited digital ethnography can complement and augment the findings of digital trace analyses.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Cet essai est présenté en tant que mémoire de maîtrise dans le cadre du programme de droit des technologies de l’information. Ce mémoire traite de différents modèles d’affaires qui ont pour caractéristique commune de commercialiser les données dans le contexte des technologies de l’information. Les pratiques commerciales observées sont peu connues et l’un des objectifs est d’informer le lecteur quant au fonctionnement de ces pratiques. Dans le but de bien situer les enjeux, cet essai discutera d’abord des concepts théoriques de vie privée et de protection des renseignements personnels. Une fois ce survol tracé, les pratiques de « data brokerage », de « cloud computing » et des solutions « analytics » seront décortiquées. Au cours de cette description, les enjeux juridiques soulevés par chaque aspect de la pratique en question seront étudiés. Enfin, le dernier chapitre de cet essai sera réservé à deux enjeux, soit le rôle du consentement et la sécurité des données, qui ne relèvent pas d’une pratique commerciale spécifique, mais qui sont avant tout des conséquences directes de l’évolution des technologies de l’information.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Virtually every sector of business and industry that uses computing, including financial analysis, search engines, and electronic commerce, incorporate Big Data analysis into their business model. Sophisticated clustering algorithms are popular for deducing the nature of data by assigning labels to unlabeled data. We address two main challenges in Big Data. First, by definition, the volume of Big Data is too large to be loaded into a computer’s memory (this volume changes based on the computer used or available, but there is always a data set that is too large for any computer). Second, in real-time applications, the velocity of new incoming data prevents historical data from being stored and future data from being accessed. Therefore, we propose our Streaming Kernel Fuzzy c-Means (stKFCM) algorithm, which reduces both computational complexity and space complexity significantly. The proposed stKFCM only requires O(n2) memory where n is the (predetermined) size of a data subset (or data chunk) at each time step, which makes this algorithm truly scalable (as n can be chosen based on the available memory). Furthermore, only 2n2 elements of the full N × N (where N >> n) kernel matrix need to be calculated at each time-step, thus reducing both the computation time in producing the kernel elements and also the complexity of the FCM algorithm. Empirical results show that stKFCM, even with relatively very small n, can provide clustering performance as accurately as kernel fuzzy c-means run on the entire data set while achieving a significant speedup.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

I Big Data stanno guidando una rivoluzione globale. In tutti i settori, pubblici o privati, e le industrie quali Vendita al dettaglio, Sanità, Media e Trasporti, i Big Data stanno influenzando la vita di miliardi di persone. L’impatto dei Big Data è sostanziale, ma così discreto da passare inosservato alla maggior parte delle persone. Le applicazioni di Business Intelligence e Advanced Analytics vogliono studiare e trarre informazioni dai Big Data. Si studia il passaggio dalla prima alla seconda, mettendo in evidenza aspetti simili e differenze.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The Everglades R-EMAP project for year 2005 produced large quantities of data collected at 232 sampling sites. Data collection and analysis is an on-going long-term activity conducted by scientists of different disciplines at irregular intervals of several years. The data sets collected for 2005 include bio-geo-chemical (including mercury and hydro period), fish, invertebrate, periphyton, and plant data. Each sampling site is associated with a location, a description of the site to provide a general overview and photographs to provide a pictorial impression. The Geographic Information Systems and Remote Sensing Center(GISRSC) at Florida International University (FIU) has designed and implemented an enterprise database for long-term storage of the project�s data in a central repository, providing the framework of data storage for the continuity of future sampling campaigns and allowing integration of new sample data as it becomes available. In addition GISRSC provides this interactive web application for easy, quick and effective retrieval and visualization of that data.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Cet essai est présenté en tant que mémoire de maîtrise dans le cadre du programme de droit des technologies de l’information. Ce mémoire traite de différents modèles d’affaires qui ont pour caractéristique commune de commercialiser les données dans le contexte des technologies de l’information. Les pratiques commerciales observées sont peu connues et l’un des objectifs est d’informer le lecteur quant au fonctionnement de ces pratiques. Dans le but de bien situer les enjeux, cet essai discutera d’abord des concepts théoriques de vie privée et de protection des renseignements personnels. Une fois ce survol tracé, les pratiques de « data brokerage », de « cloud computing » et des solutions « analytics » seront décortiquées. Au cours de cette description, les enjeux juridiques soulevés par chaque aspect de la pratique en question seront étudiés. Enfin, le dernier chapitre de cet essai sera réservé à deux enjeux, soit le rôle du consentement et la sécurité des données, qui ne relèvent pas d’une pratique commerciale spécifique, mais qui sont avant tout des conséquences directes de l’évolution des technologies de l’information.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

As economies, societies, and environments change, official statistics evolve and develop to reflect those changes. In reaction to disruptive innovations arising from globalisation, technological advances, and cultural changes, the pace of change of official statistics will accelerate in the future. The motivation for change may also be more existential than that of the past as official statisticians consider the survival of their discipline. This article examines some of the emerging developments and questions whether they present threats or offer opportunities.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Multiferroic materials displaying coupled ferroelectric and ferromagnetic order parameters could provide a means for data storage whereby bits could be written electrically and read magnetically, or vice versa. Thin films of Aurivillius phase Bi6Ti2.8Fe1.52Mn0.68O18, previously prepared by a chemical solution deposition (CSD) technique, are multiferroics demonstrating magnetoelectric coupling at room temperature. Here, we demonstrate the growth of a similar composition, Bi6Ti2.99Fe1.46Mn0.55O18, via the liquid injection chemical vapor deposition technique. High-resolution magnetic measurements reveal a considerably higher in-plane ferromagnetic signature than CSD grown films (MS = 24.25 emu/g (215 emu/cm3), MR = 9.916 emu/g (81.5 emu/cm3), HC = 170 Oe). A statistical analysis of the results from a thorough microstructural examination of the samples, allows us to conclude that the ferromagnetic signature can be attributed to the Aurivillius phase, with a confidence level of 99.95%. In addition, we report the direct piezoresponse force microscopy visualization of ferroelectric switching while going through a full in-plane magnetic field cycle, where increased volumes (8.6 to 14% compared with 4 to 7% for the CSD-grown films) of the film engage in magnetoelectric coupling and demonstrate both irreversible and reversible magnetoelectric domain switching.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

As mechatronic devices and components become increasingly integrated with and within wider systems concepts such as Cyber-Physical Systems and the Internet of Things, designer engineers are faced with new sets of challenges in areas such as privacy. The paper looks at the current, and potential future, of privacy legislation, regulations and standards and considers how these are likely to impact on the way in which mechatronics is perceived and viewed. The emphasis is not therefore on technical issues, though these are brought into consideration where relevant, but on the soft, or human centred, issues associated with achieving user privacy.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The proliferation of cloud computing allows users to flexibly store, re-compute or transfer large generated datasets with multiple cloud service providers. However, due to the pay-As-you-go model, the total cost of using cloud services depends on the consumption of storage, computation and bandwidth resources which are three key factors for the cost of IaaS-based cloud resources. In order to reduce the total cost for data, given cloud service providers with different pricing models on their resources, users can flexibly choose a cloud service to store a generated dataset, or delete it and choose a cloud service to regenerate it whenever reused. However, finding the minimum cost is a complicated yet unsolved problem. In this paper, we propose a novel algorithm that can calculate the minimum cost for storing and regenerating datasets in clouds, i.e. whether datasets should be stored or deleted, and furthermore where to store or to regenerate whenever they are reused. This minimum cost also achieves the best trade-off among computation, storage and bandwidth costs in multiple clouds. Comprehensive analysis and rigid theorems guarantee the theoretical soundness of the paper, and general (random) simulations conducted with popular cloud service providers' pricing models demonstrate the excellent performance of our approach.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

With multimedia dominating the digital contents, Device-To-Device (D2D) communication has been proposed as a promising data offloading solution in the big data area. As the quality of experience (QoE) is a major determining factor in the success of new multimedia applications, we propose a QoEdriven cooperative content dissemination (QeCS) scheme in this work. Specifically, all users predict the QoE of the potential connections characterized by the mean opinion score (MOS), and send the results to the content provider (CP). Then CP formulates a weighted directed graph according to the network topology and MOS of each potential connection. In order to stimulate cooperation among the users, the content dissemination mechanism is designed through seeking 1-factor of the weighted directed graph with the maximum weight thus achieving maximum total user MOS. Additionally, a debt mechanism is adopted to combat the cheat attacks. Furthermore, we extend the proposed QeCS scheme by considering a constrained condition to the optimization problem for fairness improvement. Extensive simulation results demonstrate that the proposed QeCS scheme achieves both efficiency and fairness especially in large scale and density networks.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Scale-free networks are often used to model a wide range of real-world networks, such as social, technological, and biological networks. Understanding the structure of scale-free networks evolves into a big data problem for business, management, and protein function prediction. In the past decade, there has been a surge of interest in exploring the properties of scale-free networks. Two interesting properties have attracted much attention: the assortative mixing and community structure. However, these two properties have been studied separately in either theoretical models or real-world networks. In this paper, we show that the structural features of communities are highly related with the assortative mixing in scale-free networks. According to the value of assortativity coefficient, scale-free networks can be categorized into assortative, disassortative, and neutral networks, respectively. We systematically analyze the community structure in these three types of scale-free networks through six metrics: node embeddedness, link density, hub dominance, community compactness, the distribution of community sizes, and the presence of hierarchical communities. We find that the three types of scale-free networks exhibit significant differences in these six metrics of community structures. First, assortative networks present high embeddedness, meaning that many links lying within communities but few links lying between communities. This leads to the high link density of communities. Second, disassortative networks exhibit great hubs in communities, which results in the high compactness of communities that nodes can reach each other via short paths. Third, in neutral networks, a big portion of links act as community bridges, so they display sparse and less compact communities. In addition, we find that (dis)assortative networks show hierarchical community structure with power-law-distributed community sizes, while neutral networks present no hierarchy. Understanding the structure of communities from the angle of assortative mixing patterns of nodes can provide insights into the network structure and guide us in modeling information propagation in different categories of scale-free networks.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Cloud computing systems and services have become major targets for cyberattackers. To provide strong protection of cloud platforms, infrastructure, hosted applications, and data stored in the cloud, we need to address the security issue from a range of perspectives-from secure data and application outsourcing, to anonymous communication, to secure multiparty computation. This special issue on cloud security aims to address the importance of protecting and securing cloud platforms, infrastructures, hosted applications, and data storage.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

BACKGROUND: As more and more researchers are turning to big data for new opportunities of biomedical discoveries, machine learning models, as the backbone of big data analysis, are mentioned more often in biomedical journals. However, owing to the inherent complexity of machine learning methods, they are prone to misuse. Because of the flexibility in specifying machine learning models, the results are often insufficiently reported in research articles, hindering reliable assessment of model validity and consistent interpretation of model outputs.

OBJECTIVE: To attain a set of guidelines on the use of machine learning predictive models within clinical settings to make sure the models are correctly applied and sufficiently reported so that true discoveries can be distinguished from random coincidence.

METHODS: A multidisciplinary panel of machine learning experts, clinicians, and traditional statisticians were interviewed, using an iterative process in accordance with the Delphi method.

RESULTS: The process produced a set of guidelines that consists of (1) a list of reporting items to be included in a research article and (2) a set of practical sequential steps for developing predictive models.

CONCLUSIONS: A set of guidelines was generated to enable correct application of machine learning models and consistent reporting of model specifications and results in biomedical research. We believe that such guidelines will accelerate the adoption of big data analysis, particularly with machine learning methods, in the biomedical research community.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Objetivo: Identificar las barreras para la unificación de una Historia Clínica Electrónica –HCE- en Colombia. Materiales y Métodos: Se realizó un estudio cualitativo. Se realizaron entrevistas semiestructuradas a profesionales y expertos de 22 instituciones del sector salud, de Bogotá y de los departamentos de Cundinamarca, Santander, Antioquia, Caldas, Huila, Valle del Cauca. Resultados: Colombia se encuentra en una estructuración para la implementación de la Historia Clínica Electrónica Unificada -HCEU-. Actualmente, se encuentra en unificación en 42 IPSs públicas en el departamento de Cundinamarca, el desarrollo de la HCEU en el país es privado y de desarrollo propio debido a las necesidades particulares de cada IPS. Conclusiones: Se identificaron barreras humanas, financieras, legales, organizacionales, técnicas y profesionales en los departamentos entrevistados. Se identificó que la unificación de la HCE depende del acuerdo de voluntades entre las IPSs del sector público, privado, EPSs, y el Gobierno Nacional.