818 resultados para big data
Resumo:
Title: Let’s SoFWIReD! Time: Wed, 21 May 2014 11:00-11:50 Location: Building 32, Room 3077 Speaker: Dr Sepi Chakaveh Abstract The information age as we know it has its roots in several enabling technologies – most of all the World Wide Web – for the provision of truly global connectivity. The emergence of a Web of Big Data in terms of the publication and analysis of Open Data provides new insights about the impact of the Web in our society. The second most important technology in this regard has been the emergence of streaming processes based on new and innovative compression methods such as MP3 so that audio and video content becomes accessible to everyone on the Web. The SoFWIReD team is developing comprehensive, interoperable platforms for data and knowledge driven processing of Open Data and will investigate aspects of collective intelligence. Insights generated in the project will form the basis for supporting companies through consulting, organisational development, and software solutions so that they can master the collective intelligence transition. The seminar will present how the project addresses the research topics of web observatory, dynamic media objects, crowd-sourced open data and Internet services. At the end of a talk a number of demos will be shown in the context of SoFWIReD’s Dynamic Media Object.
Resumo:
Los aportes teóricos y aplicados de la complejidad en economía han tomado tantas direcciones y han sido tan frenéticos en las últimas décadas, que no existe un trabajo reciente, hasta donde conocemos, que los compile y los analice de forma integrada. El objetivo de este proyecto, por tanto, es desarrollar un estado situacional de las diferentes aplicaciones conceptuales, teóricas, metodológicas y tecnológicas de las ciencias de la complejidad en la economía. Asimismo, se pretende analizar las tendencias recientes en el estudio de la complejidad de los sistemas económicos y los horizontes que las ciencias de la complejidad ofrecen de cara al abordaje de los fenómenos económicos del mundo globalizado contemporáneo.
Resumo:
El presente trabajo investigativo tiene como objetivo principal dar respuesta a la interrogante acerca de en qué casos y condiciones las interacciones virtuales contribuyen a conformar movilizaciones reales a partir de la información que circula en redes sociales. Para cumplirlo, en primer lugar, se partió de una revisión teórica de autores trascendentales como Howard Rheingold, Manuel Castells, Antonio Damasio, Pierre Levy o Antonio Negri, lo que dio como resultado un primer hallazgo: el vínculo entre la comunicación, las nuevas redes sociales, los medios tradicionales y las emociones que se gestan en ellos y que pueden hacer eco en el individuo hasta promover su movilización y la acción social. Sobre esta base teórica, el siguiente paso fue determinar cómo aquello se presentaba en un grupo definido de usuarios de redes sociales, concretamente el colectivo que hizo que el caso de Karina del Pozo fuera tendencia. Para ello, con la aplicación de las herramientas que arrojaron datos cuantificables como el Big Data y el Important Data, se procedió al trabajo de campo que constó de dos momentos. El primero de ellos, la fase de recopilación de datos; y el segundo, de análisis e interpretación sobre los resultados obtenidos. Como deducciones del estudio a partir del planteamiento teórico y de la investigación, la movilización hacia un determinado objetivo más allá de las redes sociales, es el discurso y del relato periodístico en medios tradicionales que generaron una empatía narrativa, situando al espectador en un lugar virtualmente cercano al hecho. Estos elementos además de ser una respuesta al asesinato de Karina del Pozo fueron un cuestionamiento a la sociedad y a sus prácticas, al machismo, a la violencia de género, pero además significaron la manifestación de estas mismas condiciones cuando se empezó a culpabilizar a la víctima, atribuyéndole la responsabilidad de los hechos que acabaron con su vida. Estas emociónes, valores, pensamientos o sentimientos similares, reflejados en la adopción de una determinada posición frente a un mismo hecho. En el caso de quienes se movilizaron, se estaría cumpliendo la presencia de emociones que van desde el miedo, la ira, la indignación y en contraposición, la solidaridad y la esperanza por cambiar una realidad a través de la acción.
Resumo:
SOA (Service Oriented Architecture), workflow, the Semantic Web, and Grid computing are key enabling information technologies in the development of increasingly sophisticated e-Science infrastructures and application platforms. While the emergence of Cloud computing as a new computing paradigm has provided new directions and opportunities for e-Science infrastructure development, it also presents some challenges. Scientific research is increasingly finding that it is difficult to handle “big data” using traditional data processing techniques. Such challenges demonstrate the need for a comprehensive analysis on using the above mentioned informatics techniques to develop appropriate e-Science infrastructure and platforms in the context of Cloud computing. This survey paper describes recent research advances in applying informatics techniques to facilitate scientific research particularly from the Cloud computing perspective. Our particular contributions include identifying associated research challenges and opportunities, presenting lessons learned, and describing our future vision for applying Cloud computing to e-Science. We believe our research findings can help indicate the future trend of e-Science, and can inform funding and research directions in how to more appropriately employ computing technologies in scientific research. We point out the open research issues hoping to spark new development and innovation in the e-Science field.
Resumo:
Automatic generation of classification rules has been an increasingly popular technique in commercial applications such as Big Data analytics, rule based expert systems and decision making systems. However, a principal problem that arises with most methods for generation of classification rules is the overfit-ting of training data. When Big Data is dealt with, this may result in the generation of a large number of complex rules. This may not only increase computational cost but also lower the accuracy in predicting further unseen instances. This has led to the necessity of developing pruning methods for the simplification of rules. In addition, classification rules are used further to make predictions after the completion of their generation. As efficiency is concerned, it is expected to find the first rule that fires as soon as possible by searching through a rule set. Thus a suit-able structure is required to represent the rule set effectively. In this chapter, the authors introduce a unified framework for construction of rule based classification systems consisting of three operations on Big Data: rule generation, rule simplification and rule representation. The authors also review some existing methods and techniques used for each of the three operations and highlight their limitations. They introduce some novel methods and techniques developed by them recently. These methods and techniques are also discussed in comparison to existing ones with respect to efficient processing of Big Data.
Resumo:
The induction of classification rules from previously unseen examples is one of the most important data mining tasks in science as well as commercial applications. In order to reduce the influence of noise in the data, ensemble learners are often applied. However, most ensemble learners are based on decision tree classifiers which are affected by noise. The Random Prism classifier has recently been proposed as an alternative to the popular Random Forests classifier, which is based on decision trees. Random Prism is based on the Prism family of algorithms, which is more robust to noise. However, like most ensemble classification approaches, Random Prism also does not scale well on large training data. This paper presents a thorough discussion of Random Prism and a recently proposed parallel version of it called Parallel Random Prism. Parallel Random Prism is based on the MapReduce programming paradigm. The paper provides, for the first time, novel theoretical analysis of the proposed technique and in-depth experimental study that show that Parallel Random Prism scales well on a large number of training examples, a large number of data features and a large number of processors. Expressiveness of decision rules that our technique produces makes it a natural choice for Big Data applications where informed decision making increases the user’s trust in the system.
Resumo:
This study assesses Autism-Spectrum Quotient (AQ) scores in a ‘big data’ sample collected through the UK Channel 4 television website, following the broadcasting of a medical education program. We examine correlations between the AQ and age, sex, occupation, and UK geographic region in 450,394 individuals. We predicted that age and geography would not be correlated with AQ, whilst sex and occupation would have a correlation. Mean AQ for the total sample score was m = 19.83 (SD = 8.71), slightly higher than a previous systematic review of 6,900 individuals in a non-clinical sample (mean of means = 16.94) This likely reflects that this big-data sample includes individuals with autism who in the systematic review score much higher (mean of means = 35.19). As predicted, sex and occupation differences were observed: on average, males (m = 21.55, SD = 8.82) scored higher than females (m = 18.95; SD = 8.52), and individuals working in a STEM career (m = 21.92, SD = 8.92) scored higher than individuals non-STEM careers (m = 18.92, SD = 8.48). Also as predicted, age and geographic region were not meaningfully correlated with AQ. These results support previous findings relating to sex and STEM careers in the largest set of individuals for which AQ scores have been reported and suggest the AQ is a useful self-report measure of autistic traits
Resumo:
This introduction to the Virtual Special Issue surveys the development of spatial housing economics from its roots in neo-classical theory, through more recent developments in social interactions modelling, and touching on the role of institutions, path dependence and economic history. The survey also points to some of the more promising future directions for the subject that are beginning to appear in the literature. The survey covers elements hedonic models, spatial econometrics, neighbourhood models, housing market areas, housing supply, models of segregation, migration, housing tenure, sub-national house price modelling including the so-called ripple effect, and agent-based models. Possible future directions are set in the context of a selection of recent papers that have appeared in Urban Studies. Nevertheless, there are still important gaps in the literature that merit further attention, arising at least partly from emerging policy problems. These include more research on housing and biodiversity, the relationship between housing and civil unrest, the effects of changing age distributions - notably housing for the elderly - and the impact of different international institutional structures. Methodologically, developments in Big Data provide an exciting framework for future work.
Resumo:
I consider the case for genuinely anonymous web searching. Big data seems to have it in for privacy. The story is well known, particularly since the dawn of the web. Vastly more personal information, monumental and quotidian, is gathered than in the pre-digital days. Once gathered it can be aggregated and analyzed to produce rich portraits, which in turn permit unnerving prediction of our future behavior. The new information can then be shared widely, limiting prospects and threatening autonomy. How should we respond? Following Nissenbaum (2011) and Brunton and Nissenbaum (2011 and 2013), I will argue that the proposed solutions—consent, anonymity as conventionally practiced, corporate best practices, and law—fail to protect us against routine surveillance of our online behavior. Brunton and Nissenbaum rightly maintain that, given the power imbalance between data holders and data subjects, obfuscation of one’s online activities is justified. Obfuscation works by generating “misleading, false, or ambiguous data with the intention of confusing an adversary or simply adding to the time or cost of separating good data from bad,” thus decreasing the value of the data collected (Brunton and Nissenbaum, 2011). The phenomenon is as old as the hills. Natural selection evidently blundered upon the tactic long ago. Take a savory butterfly whose markings mimic those of a toxic cousin. From the point of view of a would-be predator the data conveyed by the pattern is ambiguous. Is the bug lunch or potential last meal? In the light of the steep costs of a mistake, the savvy predator goes hungry. Online obfuscation works similarly, attempting for instance to disguise the surfer’s identity (Tor) or the nature of her queries (Howe and Nissenbaum 2009). Yet online obfuscation comes with significant social costs. First, it implies free riding. If I’ve installed an effective obfuscating program, I’m enjoying the benefits of an apparently free internet without paying the costs of surveillance, which are shifted entirely onto non-obfuscators. Second, it permits sketchy actors, from child pornographers to fraudsters, to operate with near impunity. Third, online merchants could plausibly claim that, when we shop online, surveillance is the price we pay for convenience. If we don’t like it, we should take our business to the local brick-and-mortar and pay with cash. Brunton and Nissenbaum have not fully addressed the last two costs. Nevertheless, I think the strict defender of online anonymity can meet these objections. Regarding the third, the future doesn’t bode well for offline shopping. Consider music and books. Intrepid shoppers can still find most of what they want in a book or record store. Soon, though, this will probably not be the case. And then there are those who, for perfectly good reasons, are sensitive about doing some of their shopping in person, perhaps because of their weight or sexual tastes. I argue that consumers should not have to pay the price of surveillance every time they want to buy that catchy new hit, that New York Times bestseller, or a sex toy.