999 resultados para K-arc


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Random Indexing K-tree is the combination of two algorithms suited for large scale document clustering.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper describes the approach taken to the clustering task at INEX 2009 by a group at the Queensland University of Technology. The Random Indexing (RI) K-tree has been used with a representation that is based on the semantic markup available in the INEX 2009 Wikipedia collection. The RI K-tree is a scalable approach to clustering large document collections. This approach has produced quality clustering when evaluated using two different methodologies.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Multi-disciplinary approaches to complex problems are becoming more common – they enable criteria manifested in distinct (and potentially conflicting) domains to be jointly balanced and satisfied. In this paper we present airport terminals as a case study which requires multi-disciplinary knowledge in order to balance conflicting security, economic and passenger-driven needs and correspondingly enhance the design, management and operation of airport terminals. The need for a truly multi-disciplinary scientific approach which integrates information, process, people, technology and space domains is highlighted through a brief discussion of two challenges currently faced by airport operators. The paper outlines the approach taken by this project, detailing the aims and objectives of each of seven diverse research programs.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Digital collections are growing exponentially in size as the information age takes a firm grip on all aspects of society. As a result Information Retrieval (IR) has become an increasingly important area of research. It promises to provide new and more effective ways for users to find information relevant to their search intentions. Document clustering is one of the many tools in the IR toolbox and is far from being perfected. It groups documents that share common features. This grouping allows a user to quickly identify relevant information. If these groups are misleading then valuable information can accidentally be ignored. There- fore, the study and analysis of the quality of document clustering is important. With more and more digital information available, the performance of these algorithms is also of interest. An algorithm with a time complexity of O(n2) can quickly become impractical when clustering a corpus containing millions of documents. Therefore, the investigation of algorithms and data structures to perform clustering in an efficient manner is vital to its success as an IR tool. Document classification is another tool frequently used in the IR field. It predicts categories of new documents based on an existing database of (doc- ument, category) pairs. Support Vector Machines (SVM) have been found to be effective when classifying text documents. As the algorithms for classifica- tion are both efficient and of high quality, the largest gains can be made from improvements to representation. Document representations are vital for both clustering and classification. Representations exploit the content and structure of documents. Dimensionality reduction can improve the effectiveness of existing representations in terms of quality and run-time performance. Research into these areas is another way to improve the efficiency and quality of clustering and classification results. Evaluating document clustering is a difficult task. Intrinsic measures of quality such as distortion only indicate how well an algorithm minimised a sim- ilarity function in a particular vector space. Intrinsic comparisons are inherently limited by the given representation and are not comparable between different representations. Extrinsic measures of quality compare a clustering solution to a “ground truth” solution. This allows comparison between different approaches. As the “ground truth” is created by humans it can suffer from the fact that not every human interprets a topic in the same manner. Whether a document belongs to a particular topic or not can be subjective.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Numerous challenges remain in the successful clinical translation of cell-based therapies for musculoskeletal tissue repair, including the identification of an appropriate cell source and a viable cell delivery system. The aim of this study was to investigate the attachment, colonization, and osteogenic differentiation of two stem cell types, human mesenchymal stem cells (hMSCs) and human amniotic fluid stem (hAFS) cells, on electrospun nanofiber meshes. We demonstrate that nanofiber meshes are able to support these cell functions robustly, with both cell types demonstrating strong osteogenic potential. Differences in the kinetics of osteogenic differentiation were observed between hMSCs and hAFS cells, with the hAFS cells displaying a delayed alkaline phosphatase peak, but elevated mineral deposition, compared to hMSCs. We also compared the cell behavior on nanofiber meshes to that on tissue culture plastic, and observed that there is delayed initial attachment and proliferation on meshes, but enhanced mineralization at a later time point. Finally, cell-seeded nanofiber meshes were found to be effective in colonizing three-dimensional scaffolds in an in vitro system. This study provides support for the use of the nanofiber mesh as a model surface for cell culture in vitro, and a cell delivery vehicle for the repair of bone defects in vivo.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Separability is a concept that is very difficult to define, and yet much of our scientific method is implicitly based upon the assumption that systems can sensibly be reduced to a set of interacting components. This paper examines the notion of separability in the creation of bi-ambiguous compounds that is based upon the CHSH and CH inequalities. It reports results of an experiment showing that violations of the CHSH and CH inequality can occur in human conceptual combination.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Background: Poor feeding practices in early childhood contribute to the burden of childhood malnutrition and morbidity. Objective: To estimate the key indicators of breastfeeding and complementary feeding and the determinants of selected feeding practices in Sri Lanka. Methods: The sample consisted of 1,127 children aged 0 to 23 months from the Sri Lanka Demographic and Health Survey 2000. The key infant feeding indicators were estimated and selected indicators were examined against a set of individual-, household-, and community- level variables using univariate and multivariate analyses. Results: Breastfeeding was initiated within the first hour after birth in 56.3% of infants, 99.7% had ever been breastfed, 85.0% were currently being breastfed, and 27.2% were being bottle-fed. Of infants under 6 months of age, 60.6% were fully breastfed, and of those aged 6 to 9 months, 93.4% received complementary foods. The likelihood of not initiating breastfeeding within the first hour after birth was higher for mothers who underwent cesarean delivery (OR = 3.23) and those who were not visited by a Public Health Midwife at home during pregnancy (OR = 1.81). The rate of full breastfeeding was significantly lower among mothers who did not receive postnatal home visits by a Public Health Midwife. Bottlefeeding rates were higher among infants whose mothers had ever been employed (OR = 1.86), lived in a metropolitan area (OR = 3.99), or lived in the South-Central Hill country (OR = 3.11) and were lower among infants of mothers with secondary education (OR = 0.27). Infants from the urban (OR = 8.06) and tea estate (OR = 12.63) sectors were less likely to receive timely complementary feeding than rural infants. Conclusions: Antenatal and postnatal contacts with Public Health Midwives were associated with improved breastfeeding practices. Breastfeeding promotion strategies should specifically focus on the estate and urban or metropolitan communities.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

China’s Creative Industries explores the role of new technologies, globalization and higher levels of connectivity in re-defining relationships between ‘producers’ and ‘consumers’ in 21st century China. The evolution of new business models, the impact of state regulation, the rise of entrepreneurial consumers and the role of intellectual property rights are traced through China’s film, music and fashion industries. The book argues that social network markets, consumer entrepreneurship and business model evolution are driving forces in the production and commercialization of cultural commodities. In doing so it raises important questions about copyright’s role in the business of culture, particularly in a digital age.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Type unions, pointer variables and function pointers are a long standing source of subtle security bugs in C program code. Their use can lead to hard-to-diagnose crashes or exploitable vulnerabilities that allow an attacker to attain privileged access over classified data. This paper describes an automatable framework for detecting such weaknesses in C programs statically, where possible, and for generating assertions that will detect them dynamically, in other cases. Exclusively based on analysis of the source code, it identifies required assertions using a type inference system supported by a custom made symbol table. In our preliminary findings, our type system was able to infer the correct type of unions in different scopes, without manual code annotations or rewriting. Whenever an evaluation is not possible or is difficult to resolve, appropriate runtime assertions are formed and inserted into the source code. The approach is demonstrated via a prototype C analysis tool.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This article introduces a “pseudo classical” notion of modelling non-separability. This form of non-separability can be viewed as lying between separability and quantum-like non-separability. Non-separability is formalized in terms of the non-factorizabilty of the underlying joint probability distribution. A decision criterium for determining the non-factorizability of the joint distribution is related to determining the rank of a matrix as well as another approach based on the chi-square-goodness-of-fit test. This pseudo-classical notion of non-separability is discussed in terms of quantum games and concept combinations in human cognition.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

There are at least four key challenges in the online news environment that computational journalism may address. Firstly, news providers operate in a rapidly evolving environment and larger businesses are typically slower to adapt to market innovations. News consumption patterns have changed and news providers need to find new ways to capture and retain digital users. Meanwhile, declining financial performance has led to cost cuts in mass market newspapers. Finally investigative reporting is typically slow, high cost and may be tedious, and yet is valuable to the reputation of a news provider. Computational journalism involves the application of software and technologies to the activities of journalism, and it draws from the fields of computer science, social science and communications. New technologies may enhance the traditional aims of journalism, or may require “a new breed of people who are midway between technologists and journalists” (Irfan Essa in Mecklin 2009: 3). Historically referred to as ‘computer assisted reporting’, the use of software in online reportage is increasingly valuable due to three factors: larger datasets are becoming publicly available; software is becoming sophisticated and ubiquitous; and the developing Australian digital economy. This paper introduces key elements of computational journalism – it describes why it is needed; what it involves; benefits and challenges; and provides a case study and examples. Computational techniques can quickly provide a solid factual basis for original investigative journalism and may increase interaction with readers, when correctly used. It is a major opportunity to enhance the delivery of original investigative journalism, which ultimately may attract and retain readers online.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

As organizations reach to higher levels of business process management maturity, they often find themselves maintaining repositories of hundreds or even thousands of process models, representing valuable knowledge about their operations. Over time, process model repositories tend to accumulate duplicate fragments (also called clones) as new process models are created or extended by copying and merging fragments from other models. This calls for methods to detect clones in process models, so that these clones can be refactored as separate subprocesses in order to improve maintainability. This paper presents an indexing structure to support the fast detection of clones in large process model repositories. The proposed index is based on a novel combination of a method for process model decomposition (specifically the Refined Process Structure Tree), with established graph canonization and string matching techniques. Experiments show that the algorithm scales to repositories with hundreds of models. The experimental results also show that a significant number of non-trivial clones can be found in process model repositories taken from industrial practice.