982 resultados para Knowledge maps


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Web-scale knowledge retrieval can be enabled by distributed information retrieval, clustering Web clients to a large-scale computing infrastructure for knowledge discovery from Web documents. Based on this infrastructure, we propose to apply semiotic (i.e., sub-syntactical) and inductive (i.e., probabilistic) methods for inferring concept associations in human knowledge. These associations can be combined to form a fuzzy (i.e.,gradual) semantic net representing a map of the knowledge in the Web. Thus, we propose to provide interactive visualizations of these cognitive concept maps to end users, who can browse and search the Web in a human-oriented, visual, and associative interface.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper gives an insight into cognitive computing for smart cities, resulting in cognitive cities. Cognitive cities and cognitive computing research with the underlying concepts of knowledge graphs and fuzzy cognitive maps are presented and supported by existing tools (i.e., IBM Watson and Google Now) and intended tools (meta-app). The paper illustrates FCM as a suiting instrument to represent information/knowledge in a city environment driven by human-technology interaction, enforcing the concept of cognitive cities. A proposed paper prototype combines the findings of the paper and shows the next step in the implementation of the proposed meta-app.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The new computing paradigm known as cognitive computing attempts to imitate the human capabilities of learning, problem solving, and considering things in context. To do so, an application (a cognitive system) must learn from its environment (e.g., by interacting with various interfaces). These interfaces can run the gamut from sensors to humans to databases. Accessing data through such interfaces allows the system to conduct cognitive tasks that can support humans in decision-making or problem-solving processes. Cognitive systems can be integrated into various domains (e.g., medicine or insurance). For example, a cognitive system in cities can collect data, can learn from various data sources and can then attempt to connect these sources to provide real time optimizations of subsystems within the city (e.g., the transportation system). In this study, we provide a methodology for integrating a cognitive system that allows data to be verbalized, making the causalities and hypotheses generated from the cognitive system more understandable to humans. We abstract a city subsystem—passenger flow for a taxi company—by applying fuzzy cognitive maps (FCMs). FCMs can be used as a mathematical tool for modeling complex systems built by directed graphs with concepts (e.g., policies, events, and/or domains) as nodes and causalities as edges. As a verbalization technique we introduce the restriction-centered theory of reasoning (RCT). RCT addresses the imprecision inherent in language by introducing restrictions. Using this underlying combinatorial design, our approach can handle large data sets from complex systems and make the output understandable to humans.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Synchronizing mind maps with fuzzy cognitive maps can help to handle complex problems with many involved stakeholders by taking advantage of human creativity. The proposed approach has the capacity to instantiate cognitive cities by including cognitive computing. A use case in the context of decision-finding (concerning a transportation system) is presented to illustrate the approach.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

State of the art methods for disparity estimation achieve good results for single stereo frames, but temporal coherence in stereo videos is often neglected. In this paper we present a method to compute temporally coherent disparity maps. We define an energy over whole stereo sequences and optimize their Conditional Random Field (CRF) distributions using mean-field approximation. We introduce novel terms for smoothness and consistency between the left and right views, and perform CRF optimization by fast, iterative spatio-temporal filtering with linear complexity in the total number of pixels. Our results rank among the state of the art while having significantly less flickering artifacts in stereo sequences.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In 2014, UniDive (The University of Queensland Underwater Club) conducted an ecological assessment of the Point Lookout Dive sites for comparison with similar surveys conducted in 2001 - the PLEA project. Involvement in the project was voluntary. Members of UniDive who were marine experts conducted training for other club members who had no, or limited, experience in identifying marine organisms and mapping habitats. Since the 2001 detailed baseline study, no similar seasonal survey has been conducted. The 2014 data is particularly important given that numerous changes have taken place in relation to the management of, and potential impacts on, these reef sites. In 2009, Moreton Bay Marine Park was re-zoned, and Flat Rock was converted to a marine national park zone (Green zone) with no fishing or anchoring. In 2012, four permanent moorings were installed at Flat Rock. Additionally, the entire area was exposed to the potential effects of the 2011 and 2013 Queensland floods, including flood plumes which carried large quantities of sediment into Moreton Bay and surrounding waters. The population of South East Queensland has increased from 2.49 million in 2001 to 3.18 million in 2011 (BITRE, 2013). This rapidly expanding coastal population has increased the frequency and intensity of both commercial and recreational activities around Point Lookout dive sites (EPA 2008). Habitats were mapped using a combination of towed GPS photo transects, aerial photography and expert knowledge. This data provides georeferenced information regarding the major features of each of the Point Lookout Dive Sites.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Long term global archives of high-moderate spatial resolution, multi-spectral satellite imagery are now readily accessible, but are not being fully utilised by management agencies due to the lack of appropriate methods to consistently produce accurate and timely management ready information. This work developed an object-based remote sensing approach to map land cover and seagrass distribution in an Australian coastal environment for a 38 year Landsat image time-series archive (1972-2010). Landsat Multi-Spectral Scanner (MSS), Thematic Mapper (TM) and Enhanced Thematic Mapper (ETM+) imagery were used without in situ field data input (but still using field knowledge) to produce land and seagrass cover maps every year data were available, resulting in over 60 map products over the 38 year archive. Land cover was mapped annually using vegetation, bare ground, urban and agricultural classes. Seagrass distribution was also mapped annually, and in some years monthly, via horizontal projected foliage cover classes, sand and deep water. Land cover products were validated using aerial photography and seagrass maps were validated with field survey data, producing several measures of accuracy. An average overall accuracy of 65% and 80% was reported for seagrass and land cover products respectively, which is consistent with other studies in the area. This study is the first to show moderate spatial resolution, long term annual changes in land cover and seagrass in an Australian environment, created without the use of in situ data; and only one of a few similar studies globally. The land cover products identify several long term trends; such as significant increases in South East Queensland's urban density and extent, vegetation clearing in rural and rural-residential areas, and inter-annual variation in dry vegetation types in western South East Queensland. The seagrass cover products show that there has been a minimal overall change in seagrass extent, but that seagrass cover level distribution is extremely dynamic; evidenced by large scale migrations of higher seagrass cover levels and several sudden and significant changes in cover level. These mapping products will allow management agencies to build a baseline assessment of their resources, understand past changes and help inform implementation and planning of management policy to address potential future changes.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

To deliver sample estimates provided with the necessary probability foundation to permit generalization from the sample data subset to the whole target population being sampled, probability sampling strategies are required to satisfy three necessary not sufficient conditions: (i) All inclusion probabilities be greater than zero in the target population to be sampled. If some sampling units have an inclusion probability of zero, then a map accuracy assessment does not represent the entire target region depicted in the map to be assessed. (ii) The inclusion probabilities must be: (a) knowable for nonsampled units and (b) known for those units selected in the sample: since the inclusion probability determines the weight attached to each sampling unit in the accuracy estimation formulas, if the inclusion probabilities are unknown, so are the estimation weights. This original work presents a novel (to the best of these authors' knowledge, the first) probability sampling protocol for quality assessment and comparison of thematic maps generated from spaceborne/airborne Very High Resolution (VHR) images, where: (I) an original Categorical Variable Pair Similarity Index (CVPSI, proposed in two different formulations) is estimated as a fuzzy degree of match between a reference and a test semantic vocabulary, which may not coincide, and (II) both symbolic pixel-based thematic quality indicators (TQIs) and sub-symbolic object-based spatial quality indicators (SQIs) are estimated with a degree of uncertainty in measurement in compliance with the well-known Quality Assurance Framework for Earth Observation (QA4EO) guidelines. Like a decision-tree, any protocol (guidelines for best practice) comprises a set of rules, equivalent to structural knowledge, and an order of presentation of the rule set, known as procedural knowledge. The combination of these two levels of knowledge makes an original protocol worth more than the sum of its parts. The several degrees of novelty of the proposed probability sampling protocol are highlighted in this paper, at the levels of understanding of both structural and procedural knowledge, in comparison with related multi-disciplinary works selected from the existing literature. In the experimental session the proposed protocol is tested for accuracy validation of preliminary classification maps automatically generated by the Satellite Image Automatic MapperT (SIAMT) software product from two WorldView-2 images and one QuickBird-2 image provided by DigitalGlobe for testing purposes. In these experiments, collected TQIs and SQIs are statistically valid, statistically significant, consistent across maps and in agreement with theoretical expectations, visual (qualitative) evidence and quantitative quality indexes of operativeness (OQIs) claimed for SIAMT by related papers. As a subsidiary conclusion, the statistically consistent and statistically significant accuracy validation of the SIAMT pre-classification maps proposed in this contribution, together with OQIs claimed for SIAMT by related works, make the operational (automatic, accurate, near real-time, robust, scalable) SIAMT software product eligible for opening up new inter-disciplinary research and market opportunities in accordance with the visionary goal of the Global Earth Observation System of Systems (GEOSS) initiative and the QA4EO international guidelines.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The adaptation to the European Higher Education Area (EHEA) is becoming a great challenge for the University Community, especially for its teaching and research staff, which is involved actively in the teaching-learning process. It is also inducing a paradigm change for lecturers and students. Among the methodologies used for processes of teaching innovation, system thinking plays an important role when working mainly with mind maps, and is focused to highlighting the essence of the knowledge, allowing its visual representation. In this paper, a method for using these mind maps for organizing a particular subject is explained. This organization is completed with the definition of duration, precedence relationships and resources for each of these activities, as well as with their corresponding monitoring. Mind maps are generated by means of the MINDMANAGER package whilst Ms-PROJECT is used for establishing tasks relationships, durations, resources, and monitoring. Summarizing, a procedure and the necessary set of applications for self organizing and managing (timed) scheduled teaching tasks has been described in this paper.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The adaptation to the European Higher Education Area (EHEA) is becoming a great challenge for the University Community, especially for its teaching and research staff, which is involved actively in the teaching-learning process. It is also inducing a paradigm change for lecturers and students. Among the methodologies used for processes of teaching innovation, system thinking plays an important role when working mainly with mind maps, and is focused to highlighting the essence of the knowledge, allowing its visual representation. In this paper, a method for using these mind maps for organizing a particular subject is explained. This organization is completed with the definition of duration, precedence relationships and resources for each of these activities, as well as with their corresponding monitoring. Mind maps are generated by means of the MINDMANAGER package whilst Ms-PROJECT is used for establishing tasks relationships, durations, resources, and monitoring. Summarizing, a procedure and the necessary set of applications for self organizing and managing (timed) scheduled teaching tasks has been described in this paper

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We introduce a method of functionally classifying genes by using gene expression data from DNA microarray hybridization experiments. The method is based on the theory of support vector machines (SVMs). SVMs are considered a supervised computer learning method because they exploit prior knowledge of gene function to identify unknown genes of similar function from expression data. SVMs avoid several problems associated with unsupervised clustering methods, such as hierarchical clustering and self-organizing maps. SVMs have many mathematical features that make them attractive for gene expression analysis, including their flexibility in choosing a similarity function, sparseness of solution when dealing with large data sets, the ability to handle large feature spaces, and the ability to identify outliers. We test several SVMs that use different similarity metrics, as well as some other supervised learning methods, and find that the SVMs best identify sets of genes with a common function using expression data. Finally, we use SVMs to predict functional roles for uncharacterized yeast ORFs based on their expression data.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Feature selection is an important and active issue in clustering and classification problems. By choosing an adequate feature subset, a dataset dimensionality reduction is allowed, thus contributing to decreasing the classification computational complexity, and to improving the classifier performance by avoiding redundant or irrelevant features. Although feature selection can be formally defined as an optimisation problem with only one objective, that is, the classification accuracy obtained by using the selected feature subset, in recent years, some multi-objective approaches to this problem have been proposed. These either select features that not only improve the classification accuracy, but also the generalisation capability in case of supervised classifiers, or counterbalance the bias toward lower or higher numbers of features that present some methods used to validate the clustering/classification in case of unsupervised classifiers. The main contribution of this paper is a multi-objective approach for feature selection and its application to an unsupervised clustering procedure based on Growing Hierarchical Self-Organising Maps (GHSOMs) that includes a new method for unit labelling and efficient determination of the winning unit. In the network anomaly detection problem here considered, this multi-objective approach makes it possible not only to differentiate between normal and anomalous traffic but also among different anomalies. The efficiency of our proposals has been evaluated by using the well-known DARPA/NSL-KDD datasets that contain extracted features and labelled attacks from around 2 million connections. The selected feature sets computed in our experiments provide detection rates up to 99.8% with normal traffic and up to 99.6% with anomalous traffic, as well as accuracy values up to 99.12%.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

O contexto tecnológico em que vivemos é uma realidade. E a tendência é para ser assim também no futuro. Cada vez mais. É o caso das representações de locais e entidades em mapas digitais na web. Na visão de Crocker (2014), esta tendência é ainda mais acentuada, no âmbito das aplicações móveis, como mostram as mais diversas location-based applications. No setor do desporto e da respetiva gestão nem sempre foi fácil desenvolver aplicações, recorrendo a este tipo de representações espaciais. A tecnologia não era fácil e o know-how não era adequadamente qualificado. Mas, as empresas fornecedoras de tecnologia geoespacial simplificaram o desenvolvimento de aplicações web nesta área, através da utilização de application programming interfaces (API). Como refere Svennerberg (2010), estas API’s servem de interface entre um serviço proporcionado por uma empresa, caso da Google Maps (2013) e uma aplicação web ou móvel que utiliza esses serviços. Foi com este objetivo que desenvolvemos uma aplicação web, utilizando as metodologias próprias neste domínio, como a framework de Zachman (2009), tal como foi originalmente adaptada por Whitten e Bentley (2005), onde um dos módulos é precisamente a representação de espaços desportivos, recorrendo à utilização dos serviços da Google Maps. Para além disso, toda a aplicação é suportada numa abordagem Model-View-Control (MVC). Para conseguir representar as instalações desportivas num mapa, criámos uma base de dados MySQL, com dados de longitude e latitude, de cada instalação desportiva. Através de JavaScript criou-se o mapa propriamente dito, indicando o tipo (mapa de estradas, satélite ou street view) e as respetivas opções (nível de zoom, alinhamento, controlo de interface e posicionamente, entre muitas outras opções). O passo seguinte consistiu em passar os dados para o frontend da aplicação web. Para isso, recorreu-se à integração do PHP com as livrarias externas de código JavaSrcipt, criadas especificamente para o efeito (caso da MarkerManager). A implementação destas funcionalidades permite georeferenciar todos os tipos e géneros de espaços desportivos de um concelho, região ou País. Obteve-se ainda know-how, background e massa crítica, para o desenvolvimento de novas funcionalidades. A sua utilização em dispositivos móveis é outra das possibilidades atualmente já em desenvolvimento.