895 resultados para Geographical computer applications
Resumo:
"Period of performance: September, l968- December, 1973."
Resumo:
Includes index.
Resumo:
This work attempts to create a systemic design framework for man-machine interfaces which is self consistent, compatible with other concepts, and applicable to real situations. This is tackled by examining the current architecture of computer applications packages. The treatment in the main is philosophical and theoretical and analyses the origins, assumptions and current practice of the design of applications packages. It proposes that the present form of packages is fundamentally contradictory to the notion of packaging itself. This is because as an indivisible ready-to-implement solution, current package architecture displays the following major disadvantages. First, it creates problems as a result of user-package interactions, in which the designer tries to mould all potential individual users, no matter how diverse they are, into one model. This is worsened by the minute provision, if any, of important properties such as flexibility, independence and impartiality. Second, it displays rigid structure that reduces the variety and/or multi-use of the component parts of such a package. Third, it dictates specific hardware and software configurations which probably results in reducing the number of degrees of freedom of its user. Fourth, it increases the dependence of its user upon its supplier through inadequate documentation and understanding of the package. Fifth, it tends to cause a degeneration of the expertise of design of the data processing practitioners. In view of this understanding an alternative methodological design framework which is both consistent with systems approach and the role of a package in its likely context is proposed. The proposition is based upon an extension of the identified concept of the hierarchy of holons* which facilitates the examination of the complex relationships of a package with its two principal environments. First, the user characteristics and his decision making practice and procedures; implying an examination of the user's M.I.S. network. Second, the software environment and its influence upon a package regarding support, control and operation of the package. The framework is built gradually as discussion advances around the central theme of a compatible M.I.S., software and model design. This leads to the formation of the alternative package architecture that is based upon the design of a number of independent, self-contained small parts. Such is believed to constitute the nucleus around which not only packages can be more effectively designed, but is also applicable to many man-machine systems design.
Resumo:
Las tecnologías de la información y la comunicación están consiguiendo que la información geográfica sea asequible a un mayor número de profesionales a través de las Tecnologías de la Información Geográfica. La intervención multidisciplinar en el territorio enriquece la investigación y las formas de aplicación de este tipo de recursos tecnológicos. Pero esta facilidad tecnológica puede suponer el riesgo de un uso inadecuado, por falta de conocimientos técnicos adecuados a la complejidad de la información geográfica o por el mal uso de las aplicaciones informáticas. El trabajo catastral puede beneficiarse mucho del empleo de estas tecnologías de información geográfica, al facilitar el uso, la comunicación y su administración electrónica, pero el desconocimiento de las propiedades geométricas y topológicas de la información geográfica puede llevar a cometer errores de graves consecuencias a profesionales no especializados. En este artículo ofrecemos el resultado de la investigación del trabajo de diversos juristas y técnicos, con el objetivo de desarrollar métodos automatizados y aplicaciones informáticas que permitan a los especialistas no expertos en Cartografía usar este tipo de información con garantías de exactitud al más alto nivel, como una solución eficaz para que la información geográfica con calidad topológica enriquezca la seguridad jurídica en el tráfico inmobiliario.
Resumo:
As computer applications become more available—both technically and economically—construction project managers are increasingly able to access advanced computer tools capable of transforming the role that project managers have typically performed. Competence at using these tools requires a dual commitment in training—from the individual and the firm. Improving the computer skills of project managers can provide construction firms with a competitive advantage to differentiate from others in an increasingly competitive international market. Yet, few published studies have quantified what existing level of competence construction project managers have. Identification of project managers’ existing computer application skills is a necessary first step to developing more directed training to better capture the benefits of computer applications. This paper discusses the yet to be released results of a series of surveys undertaken in Malaysia, Singapore, Indonesia, Australia and the United States through QUT’s School of Construction Management and Property and the M.E. Rinker, Sr. School of Building Construction at the University of Florida. This international survey reviews the use and reported competence in using a series of commercially-available computer applications by construction project managers. The five different country locations of the survey allow cross-national comparisons to be made between project managers undertaking continuing professional development programs. The results highlight a shortfall in the ability of construction project managers to capture potential benefits provided by advanced computer applications and provide directions for targeted industry training programs. This international survey also provides a unique insight to the cross-national usage of advanced computer applications and forms an important step in this ongoing joint review of technology and the construction project manager.
Resumo:
Department of Computer Applications, Cochin University of Science and Technology
Resumo:
The present work deals with the A study of morphological opertors with applications. Morphology is now a.necessary tool for engineers involved with imaging applications. Morphological operations have been viewed as filters the properties of which have been well studied (Heijmans, 1994). Another well-known class of non-linear filters is the class of rank order filters (Pitas and Venetsanopoulos, 1990). Soft morphological filters are a combination of morphological and weighted rank order filters (Koskinen, et al., 1991, Kuosmanen and Astola, 1995). They have been introduced to improve the behaviour of traditional morphological filters in noisy environments. The idea was to slightly relax the typical morphological definitions in such a way that a degree of robustness is achieved, while most of the desirable properties of typical morphological operations are maintained. Soft morphological filters are less sensitive to additive noise and to small variations in object shape than typical morphological filters. They can remove positive and negative impulse noise, preserving at the same time small details in images. Currently, Mathematical Morphology allows processing images to enhance fuzzy areas, segment objects, detect edges and analyze structures. The techniques developed for binary images are a major step forward in the application of this theory to gray level images. One of these techniques is based on fuzzy logic and on the theory of fuzzy sets.Fuzzy sets have proved to be strongly advantageous when representing in accuracies, not only regarding the spatial localization of objects in an image but also the membership of a certain pixel to a given class. Such inaccuracies are inherent to real images either because of the presence of indefinite limits between the structures or objects to be segmented within the image due to noisy acquisitions or directly because they are inherent to the image formation methods.
Resumo:
This paper describes an innovative approach to develop the understanding about the relevance of mathematics to computer science. The mathematical subjects are introduced through an application-to-model scheme that lead computer science students to a better understanding of why they have to learn math and learn it effectively. Our approach consists of a single one semester course, taught at the first semester of the program, where the students are initially exposed to some typical computer applications. When they recognize the applications' complexity, the instructor gives the mathematical models supporting such applications, even before a formal introduction to the model in a math course. We applied this approach at Unesp (Brazil) and the results include a large reduction in the rate of students that abandon the college and better students in the final years of our program.
Resumo:
In order to simplify computer management, several system administrators are adopting advanced techniques to manage software configuration of enterprise computer networks, but the tight coupling between hardware and software makes every PC an individual managed entity, lowering the scalability and increasing the costs to manage hundreds or thousands of PCs. Virtualization is an established technology, however its use is been more focused on server consolidation and virtual desktop infrastructure, not for managing distributed computers over a network. This paper discusses the feasibility of the Distributed Virtual Machine Environment, a new approach for enterprise computer management that combines virtualization and distributed system architecture as the basis of the management architecture. © 2008 IEEE.
Resumo:
Epilepsy is a very complex disease which can have a variety of etiologies, co-morbidities, and a long list of psychosocial factors4. Clinical management of epilepsy patients typically includes serological tests, EEG's, and imaging studies to determine the single best antiepileptic drug (AED). Self-management is a vital component of achieving optimal health when living with a chronic disease. For patients with epilepsy self-management includes any necessary actions to control seizures and cope with any subsequent effects of the condition9; including aspects of treatment, seizure, and lifestyle. The use of computer-based applications can allow for more effective use of clinic visits and ultimately enhance the patient-provider relationship through focused discussion of determinants affecting self-management. ^ The purpose of this study is to conduct a systematic literature review on informatics application in epilepsy self-management in an effort to describe current evidence for informatics applications and decision support as an adjunct to successful clinical management of epilepsy. Each publication was analyzed for the type of study design utilized. ^ A total of 68 publications were included and categorized by the study design used, development stage, and clinical domain. Descriptive study designs comprised of three-fourths of the publications and indicate an underwhelming use of prospective studies. The vast majority of prospective studies also focused on clinician use to increase knowledge in treating patients with epilepsy. ^ Due to the chronic nature of epilepsy and the difficulty that both clinicians and patients can experience in managing epilepsy, more prospective studies are needed to evaluate applications that can effectively increase management activities. Within the last two decades of epilepsy research, management studies have employed the use of biomedical informatics applications. While the use of computer applications to manage epilepsy has increased, more progress is needed.^
Resumo:
In this article, we present a new framework oriented to teach Computer Vision related subjects called JavaVis. It is a computer vision library divided in three main areas: 2D package is featured for classical computer vision processing; 3D package, which includes a complete 3D geometric toolset, is used for 3D vision computing; Desktop package comprises a tool for graphic designing and testing of new algorithms. JavaVis is designed to be easy to use, both for launching and testing existing algorithms and for developing new ones.
Resumo:
Context The School of Information Technology at QUT has recently undertaken a major restructuring of their Bachelor of Information Technology (BIT) course. Some of the aims of this restructuring include a reduction in first year attrition and to provide an attractive degree course that meets both student and industry expectations. Emphasis has been placed on the first semester in the context of retaining students by introducing a set of four units that complement one another and provide introductory material on technology, programming and related skills, and generic skills that will aid the students throughout their undergraduate course and in their careers. This discussion relates to one of these four fist semester units, namely Building IT Systems. The aim of this unit is to create small Information Technology (IT) systems that use programming or scripting, databases as either standalone applications or web applications. In the prior history of teaching introductory computer programming at QUT, programming has been taught as a stand alone subject and integration of computer applications with other systems such as databases and networks was not undertaken until students had been given a thorough grounding in those topics as well. Feedback has indicated that students do not believe that working with a database requires programming skills. In fact, the teaching of the building blocks of computer applications have been compartmentalized and taught in isolation from each other. The teaching of introductory computer programming has been an industry requirement of IT degree courses as many jobs require at least some knowledge of the topic. Yet, computer programming is not a skill that all students have equal capabilities of learning (Bruce et al., 2004) and this is clearly shown by the volume of publications dedicated to this topic in the literature over a broad period of time (Eckerdal & Berglund, 2005; Mayer, 1981; Winslow, 1996). The teaching of this introductory material has been done pretty much the same way over the past thirty years. During this period of time that introductory computer programming courses have been taught at QUT, a number of different programming languages and programming paradigms have been used and different approaches to teaching and learning have been attempted in an effort to find the golden thread that would allow students to learn this complex topic. Unfortunately, computer programming is not a skill that can be learnt in one semester. Some basics can be learnt but it can take many years to master (Norvig, 2001). Faculty data typically has shown a bimodal distribution of results for students undertaking introductory programming courses with a high proportion of students receiving a high mark and a high proportion of students receiving a low or failing mark. This indicates that there are students who understand and excel with the introductory material while there is another group who struggle to understand the concepts and practices required to be able to translate a specification or problem statement into a computer program that achieves what is being requested. The consequence of a large group of students failing the introductory programming course has been a high level of attrition amongst first year students. This attrition level does not provide good continuity in student numbers in later years of the degree program and the current approach is not seen as sustainable.