590 resultados para Automatized Indexing
Resumo:
"The indexing has been under the immediate charge of Herbert W. Denio ... assisted by Miss Helen Merrill ... [et al.]."
Resumo:
Description based on: 1980 ed.
Resumo:
Cybercrime and related malicious activity in our increasingly digital world has become more prevalent and sophisticated, evading traditional security mechanisms. Digital forensics has been proposed to help investigate, understand and eventually mitigate such attacks. The practice of digital forensics, however, is still fraught with various challenges. Some of the most prominent of these challenges include the increasing amounts of data and the diversity of digital evidence sources appearing in digital investigations. Mobile devices and cloud infrastructures are an interesting specimen, as they inherently exhibit these challenging circumstances and are becoming more prevalent in digital investigations today. Additionally they embody further characteristics such as large volumes of data from multiple sources, dynamic sharing of resources, limited individual device capabilities and the presence of sensitive data. These combined set of circumstances make digital investigations in mobile and cloud environments particularly challenging. This is not aided by the fact that digital forensics today still involves manual, time consuming tasks within the processes of identifying evidence, performing evidence acquisition and correlating multiple diverse sources of evidence in the analysis phase. Furthermore, industry standard tools developed are largely evidence-oriented, have limited support for evidence integration and only automate certain precursory tasks, such as indexing and text searching. In this study, efficiency, in the form of reducing the time and human labour effort expended, is sought after in digital investigations in highly networked environments through the automation of certain activities in the digital forensic process. To this end requirements are outlined and an architecture designed for an automated system that performs digital forensics in highly networked mobile and cloud environments. Part of the remote evidence acquisition activity of this architecture is built and tested on several mobile devices in terms of speed and reliability. A method for integrating multiple diverse evidence sources in an automated manner, supporting correlation and automated reasoning is developed and tested. Finally the proposed architecture is reviewed and enhancements proposed in order to further automate the architecture by introducing decentralization particularly within the storage and processing functionality. This decentralization also improves machine to machine communication supporting several digital investigation processes enabled by the architecture through harnessing the properties of various peer-to-peer overlays. Remote evidence acquisition helps to improve the efficiency (time and effort involved) in digital investigations by removing the need for proximity to the evidence. Experiments show that a single TCP connection client-server paradigm does not offer the required scalability and reliability for remote evidence acquisition and that a multi-TCP connection paradigm is required. The automated integration, correlation and reasoning on multiple diverse evidence sources demonstrated in the experiments improves speed and reduces the human effort needed in the analysis phase by removing the need for time-consuming manual correlation. Finally, informed by published scientific literature, the proposed enhancements for further decentralizing the Live Evidence Information Aggregator (LEIA) architecture offer a platform for increased machine-to-machine communication thereby enabling automation and reducing the need for manual human intervention.
Resumo:
Online geographic information systems provide the means to extract a subset of desired spatial information from a larger remote repository. Data retrieved representing real-world geographic phenomena are then manipulated to suit the specific needs of an end-user. Often this extraction requires the derivation of representations of objects specific to a particular resolution or scale from a single original stored version. Currently standard spatial data handling techniques cannot support the multi-resolution representation of such features in a database. In this paper a methodology to store and retrieve versions of spatial objects at, different resolutions with respect to scale using standard database primitives and SQL is presented. The technique involves heavy fragmentation of spatial features that allows dynamic simplification into scale-specific object representations customised to the display resolution of the end-user's device. Experimental results comparing the new approach to traditional R-Tree indexing and external object simplification reveal the former performs notably better for mobile and WWW applications where client-side resources are limited and retrieved data loads are kept relatively small.
Resumo:
Background The aims of this study were threefold. First, to ascertain whether personality disorder (PD) was a significant predictor of disability (as measured in a variety of ways) over and above that contributed by Axis I mental disorders and physical conditions. Second, whether the number of PD diagnoses given to an individual resulted in increasing severity of disability, and third, whether PD was a significant predictor of health and mental health consultations with GPs, psychiatrists, and psychologists, respectively, over the last 12 months. Method Data were obtained from the National Survey of Mental Health and Wellbeing, conducted between May and August 1997. A stratified random sample of households was generated, from which all those aged 18 and over were considered potential interviewees. There were 10 641 respondents to the survey, and this represented a response rate of 78%. Each interviewee was asked questions indexing specific ICD-10 PD criteria. Results Five measures of disability were examined. It was found that PD was a significant predictor of disability once Axis I and physical conditions were taken into account for four of the five disability measures. For three of the dichotomously-scored disability measures, odds ratios ranged from 1.88 to 6.32 for PD, whilst for the dimensionally-scored Mental Summary Subscale of the SF-12, a beta weight of -0.17 was recorded for PD. As regards number of PDs having a quasi-linear relationship to disability, there was some indication of this on the SF-12 Mental Summary Subscale and the two role functioning measures, and less so on the other two measures. As regards mental consultations, PD was a predictor of visits to GPs, psychiatrists and psychologists, over and above Axis I disorders and physical conditions. Conclusion The study reports findings from a nationwide survey conducted within Australia and as such the data are less influenced by the selection and setting bias inherent in other germane studies. However, it does support previous findings that PD is a significant predictor of disability and mental health consultations independent of Axis I disorders and physical conditions.
Resumo:
Multiresolution Triangular Mesh (MTM) models are widely used to improve the performance of large terrain visualization by replacing the original model with a simplified one. MTM models, which consist of both original and simplified data, are commonly stored in spatial database systems due to their size. The relatively slow access speed of disks makes data retrieval the bottleneck of such terrain visualization systems. Existing spatial access methods proposed to address this problem rely on main-memory MTM models, which leads to significant overhead during query processing. In this paper, we approach the problem from a new perspective and propose a novel MTM called direct mesh that is designed specifically for secondary storage. It supports available indexing methods natively and requires no modification to MTM structure. Experiment results, which are based on two real-world data sets, show an average performance improvement of 5-10 times over the existing methods.
Resumo:
In many advanced applications, data are described by multiple high-dimensional features. Moreover, different queries may weight these features differently; some may not even specify all the features. In this paper, we propose our solution to support efficient query processing in these applications. We devise a novel representation that compactly captures f features into two components: The first component is a 2D vector that reflects a distance range ( minimum and maximum values) of the f features with respect to a reference point ( the center of the space) in a metric space and the second component is a bit signature, with two bits per dimension, obtained by analyzing each feature's descending energy histogram. This representation enables two levels of filtering: The first component prunes away points that do not share similar distance ranges, while the bit signature filters away points based on the dimensions of the relevant features. Moreover, the representation facilitates the use of a single index structure to further speed up processing. We employ the classical B+-tree for this purpose. We also propose a KNN search algorithm that exploits the access orders of critical dimensions of highly selective features and partial distances to prune the search space more effectively. Our extensive experiments on both real-life and synthetic data sets show that the proposed solution offers significant performance advantages over sequential scan and retrieval methods using single and multiple VA-files.
Resumo:
In this paper, we compare a well-known semantic spacemodel, Latent Semantic Analysis (LSA) with another model, Hyperspace Analogue to Language (HAL) which is widely used in different area, especially in automatic query refinement. We conduct this comparative analysis to prove our hypothesis that with respect to ability of extracting the lexical information from a corpus of text, LSA is quite similar to HAL. We regard HAL and LSA as black boxes. Through a Pearsonrsquos correlation analysis to the outputs of these two black boxes, we conclude that LSA highly co-relates with HAL and thus there is a justification that LSA and HAL can potentially play a similar role in the area of facilitating automatic query refinement. This paper evaluates LSA in a new application area and contributes an effective way to compare different semantic space models.
Resumo:
Terrain can be approximated by a triangular mesh consisting millions of 3D points. Multiresolution triangular mesh (MTM) structures are designed to support applications that use terrain data at variable levels of detail (LOD). Typically, an MTM adopts a tree structure where a parent node represents a lower-resolution approximation of its descendants. Given a region of interest (ROI) and a LOD, the process of retrieving the required terrain data from the database is to traverse the MTM tree from the root to reach all the nodes satisfying the ROI and LOD conditions. This process, while being commonly used for multiresolution terrain visualization, is inefficient as either a large number of sequential I/O operations or fetching a large amount of extraneous data is incurred. Various spatial indexes have been proposed in the past to address this problem, however level-by-level tree traversal remains a common practice in order to obtain topological information among the retrieved terrain data. A new MTM data structure called direct mesh is proposed. We demonstrate that with direct mesh the amount of data retrieval can be substantially reduced. Comparing with existing MTM indexing methods, a significant performance improvement has been observed for real-life terrain data.
Resumo:
Few works address methodological issues of how to conduct strategy-as-practice research and even fewer focus on how to analyse the subsequent data in ways that illuminate strategy as an everyday, social practice. We address this gap by proposing a quantitative method for analysing observational data, which can complement more traditional qualitative methodologies. We propose that rigorous but context-sensitive coding of transcripts can render everyday practice analysable statistically. Such statistical analysis provides a means for analytically representing patterns and shifts within the mundane, repetitive elements through which practice is accomplished. We call this approach the Event Database (EDB) and it consists of five basic coding categories that help us capture the stream of practice. Indexing codes help to index or categorise the data, in order to give context and offer some basic information about the event under discussion. Indexing codes are descriptive codes, which allow us to catalogue and classify events according to their assigned characteristics. Content codes are to do with the qualitative nature of the event; this is the essence of the event. It is a description that helps to inform judgements about the phenomenon. Nature codes help us distinguish between discursive and tangible events. We include this code to acknowledge that some events differ qualitatively from other events. Type events are codes abstracted from the data in order to help us classify events based on their description or nature. This involves significantly more judgement than the index codes but consequently is also more meaningful. Dynamics codes help us capture some of the movement or fluidity of events. This category has been included to let us capture the flow of activity over time.
Resumo:
This paper summarizes the scientific work presented at the 32nd European Conference on Information Retrieval. It demonstrates that information retrieval (IR) as a research area continues to thrive with progress being made in three complementary sub-fields, namely IR theory and formal methods together with indexing and query representation issues, furthermore Web IR as a primary application area and finally research into evaluation methods and metrics. It is the combination of these areas that gives IR its solid scientific foundations. The paper also illustrates that significant progress has been made in other areas of IR. The keynote speakers addressed three such subject fields, social search engines using personalization and recommendation technologies, the renewed interest in applying natural language processing to IR, and multimedia IR as another fast-growing area.
Resumo:
Few works address methodological issues of how to conduct strategy-as-practice research and even fewer focus on how to analyse the subsequent data in ways that illuminate strategy as an everyday, social practice. We address this gap by proposing a quantitative method for analysing observational data, which can complement more traditional qualitative methodologies. We propose that rigorous but context-sensitive coding of transcripts can render everyday practice analysable statistically. Such statistical analysis provides a means for analytically representing patterns and shifts within the mundane, repetitive elements through which practice is accomplished. We call this approach the Event Database (EDB) and it consists of five basic coding categories that help us capture the stream of practice. Indexing codes help to index or categorise the data, in order to give context and offer some basic information about the event under discussion. Indexing codes are descriptive codes, which allow us to catalogue and classify events according to their assigned characteristics. Content codes are to do with the qualitative nature of the event; this is the essence of the event. It is a description that helps to inform judgements about the phenomenon. Nature codes help us distinguish between discursive and tangible events. We include this code to acknowledge that some events differ qualitatively from other events. Type events are codes abstracted from the data in order to help us classify events based on their description or nature. This involves significantly more judgement than the index codes but consequently is also more meaningful. Dynamics codes help us capture some of the movement or fluidity of events. This category has been included to let us capture the flow of activity over time.
Resumo:
The essential first step for a beginning reader is to learn to match printed forms to phonological representations. For a new word, this is an effortful process where each grapheme must be translated individually (serial decoding). The role of phonological awareness in developing a decoding strategy is well known. We examined whether beginner readers recruit different skills depending on the nature of the words being read (familiar words vs. nonwords). Print knowledge, phoneme and rhyme awareness, rapid automatized naming (RAN), phonological short term memory (STM), nonverbal reasoning, vocabulary, auditory skills and visual attention were measured in 392 pre-readers aged 4 to 5 years. Word and nonword reading were measured 9 months later. We used structural equation modeling to examine the skills-reading relationship and modeled correlations between our two reading outcomes and among all pre-reading skills. We found that a broad range of skills were associated with reading outcomes: early print knowledge, phonological STM, phoneme awareness and RAN. Whereas all these skills were directly predictive of nonword reading, early print knowledge was the only direct predictor of word reading. Our findings suggest that beginner readers draw most heavily on their existing print knowledge to read familiar words.
Resumo:
Readers may have noted that a short but very important announcement was made in the last issue of CLAE, at the top of the contents page. CLAE has been accepted by Thomson Reuters for abstracting and indexing in its SciSearch, Journal Citation Reports, and Current Contents services. This will ensure a greater visibility to the international research community. In addition, in June 2012 CLAE will receive its very first official Impact Factor – a measure of journal influence of importance to authors and readers alike. The impact factor value has not yet been decided but internal estimates by Elsevier estimate it will be around 1, and it will be applied to all CLAE issue back to January 2009 (volume 32). I would guess readers at this stage would have one of two responses – either ‘that's good news’ or perhaps ‘what's an impact factor?’ If you are in the latter camp then allow me to try and explain. Basically the impact factor or citation index of a journal is based on how many times in the previous year papers published in that journal in the previous two years were cited by authors publishing in other journals. So the 2012 impact factor for CLAE is calculated on how many times in 2011 papers that were published in CLAE in 2010 and 2009 were cited in other journals in 2011, divided by the number of papers published in CLAE 2010 and 2009. Essentially authors will try and get their work published in journals with a higher impact factor as it is thought that the paper will be cited more by other authors or the paper will have higher visibility in the arena. For universities having its published output in higher journals is one of the markers used to judge esteem. For individual authors publishing in journals with a higher impact factor or the number of times one of their papers is published is something that they are likely to add to their CVs or demonstrate the importance of their work. Journals with higher impact factors tend to be more review journals or journals with a wider spectrum so for a relatively small journal with a specialised field like CLAE it is great to be listed with a citation index. The awarding of a citation index crowns many changes that CLAE has undergone since the current Editor took the reins in 2005. CLAE has increased from four issues (in 2004) to six issues per year with at least one review article per issue and one article with continuing education per issue. The rejection rate has gone up significantly meaning that only best papers are published (currently it stands at 37%). CLAE has been Medline/Pubmed indexed for a few years now which is also a very important factor in improving visibility of the journal. The submission and reviewing process for CLAE in now entirely online and finally the editorial board has changed from being merely a list of keynote people to being an active group of keynote people who are enthusiastically involved with the journal. From the editorial board one person is appointed as a Reviews Editor plus we have two additional editors who work as Regional Editors. As ever, on behalf of CLAE I would like to thank the BCLA Council for their continued support (especially Vivien Freeman) and Elsevier for their continuing guidance (in particular Andrew Miller and Rosie Davey) and the excellent Editorial Board (Christopher Snyder, Pauline Cho, Eric Papas, Jan Bergmanson, Roger Buckley, Patrick Caroline, Dwight Cavanagh, Robin Chalmers, Michael Doughty, Nathan Efron, Michel Guillon, Nizar Hirji, Meng Lin, Florence Malet, Philip Morgan, Deborah Sweeney, Brian Tighe, Eef van Der Worp, Barry Weissman, Mark Willcox, James Wolffsohn and Craig Woods). And finally, a big thanks to the authors and reviewers who work tirelessly putting manuscripts together for publication in CLAE. Copyright © 2012 Published by Elsevier Ltd.