989 resultados para Computer Oriented Statistics


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Na década de 80, o surgimento de programas de computadores mais amigáveis para usuários e produtores de informação e a evolução tecnológica fizeram com que as instituições, públicas e privadas, se aperfeiçoassem em estudos sobre sistemas de produção cartográfica apoiados por computador, visando a implementação de Sistemas de Informação Geográfica (SIG). A pouca simultaneidade de forças entre órgãos interessados, resultou em uma grande quantidade de arquivos digitais com a necessidade de padronização. Em 2007, a Comissão Nacional de Cartografia (CONCAR) homologou a Estrutura de Dados Geoespaciais Vetoriais (EDGV) a fim de minimizar o problema da falta de padronização de bases cartográficas. A presente dissertação tem como foco elaborar uma metodologia de trabalho para o processo de conversão de bases cartográficas digitais existentes no padrão da Mapoteca Topográfica Digital (MTD), do Instituto Brasileiro de Geografia e Estatística (IBGE), para o padrão da EDGV, bem como suas potencialidades e limitações para integração e padronização de bases cartográficas digitais. Será feita uma aplicação da metodologia utilizando a carta topográfica de Saquarema, na escala de 1:50.000, vetorizada na Coordenação de Cartografia (CCAR) do IBGE e disponível na Internet. Como a EDGV foi elaborada segundo técnicas de modelagem orientada a objetos, foi necessário um mapeamento para banco de dados relacional, já que este ainda é utilizado pela maioria dos usuários e produtores de informação geográfica. Um dos objetivos específicos é elaborar um esquema de banco de dados, ou seja, um banco de dados vazio contendo todas as classes de objetos, atributos e seus respectivos domínios existentes na EDGV para que possa ser utilizado no processo de produção cartográfica do IBGE. Este esquema conterá todas as descrições dos objetos e de seus respectivos atributos, além de já permitir que o usuário selecione o domínio de um determinado atributo em uma lista pré definida, evitando que ocorra erro no preenchimento dados. Esta metodologia de trabalho será de grande importância para o processo de conversão das bases cartográficas existentes no IBGE e, com isso, gerar e disponibilizar bases cartográficas no padrão da EDGV.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The personal computer has become commonplace on the desk of most scientists. As hardware costs have plummeted, software capabilities have expanded enormously, permitting the scientist to examine extremely large datasets in novel ways. Advances in networking now permit rapid transfer of large datasets, which can often be used unchanged from one machine to the next. In spite of these significant advances, many scientists still use their personal computers only for word processing or e-mail, or as "dumb terminals". Many are simply unaware of the richness of software now available to statistically analyze and display scientific data in highly innovative ways. This paper presents several examples drawn from actual climate data analysis that illustrate some novel and practical features of several widely-used software packages for Macintosh computers.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

CAD software can be structured as a set of modular 'software tools' only if there is some agreement on the data structures which are to be passed between tools. Beyond this basic requirement, it is desirable to give the agreed structures the status of 'data types' in the language used for interactive design. The ultimate refinement is to have a data management capability which 'understands' how to manipulate such data types. In this paper the requirements of CACSD are formulated from the point of view of Database Management Systems. Progress towards meeting these requirements in both the DBMS and the CACSD community is reviewed. The conclusion reached is that there has been considerable movement towards the realisation of software tools for CACSD, but that this owes more to modern ideas about programming languages, than to DBMS developments. The DBMS field has identified some useful concepts, but further significant progress is expected to come from the exploitation of concepts such as object-oriented programming, logic programming, or functional programming.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Computer modelling approaches have significant potential to enable decision-making about various aspects of responsive manufacturing. In order to understand the system prior to the selection of any responsiveness strategy, multiple process segments of organisations need to be modelled. The article presents a novel systematic approach for creating coherent sets of unified enterprise, simulation and other supporting models that collectively facilitate responsiveness. In this approach, enterprise models are used to explicitly define relatively enduring relationships between (i) production planning and control (PPC) processes, that implement a particular strategy and (ii) process-oriented elements of production systems, that are work loaded by the PPC processes. Coherent simulation models, can in part be derived from the enterprise models, so that they computer execute production system behaviours. In this way, time-based performance outcomes can be simulated; so that the impacts of alternative PPC strategies on the planning and controlling historical or forecasted patterns of workflow, through (current and possible future) production system models, can be analysed. The article describes the unified modelling approach conceived and its application in a furniture industry case study small and medium enterprise (SME). Copyright © 2010 Inderscience Enterprises Ltd.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Humans recognize optical reflectance properties of surfaces such as metal, plastic, or paper from a single image without knowledge of illumination. We develop a machine vision system to perform similar recognition tasks automatically. Reflectance estimation under unknown, arbitrary illumination proves highly underconstrained due to the variety of potential illumination distributions and surface reflectance properties. We have found that the spatial structure of real-world illumination possesses some of the statistical regularities observed in the natural image statistics literature. A human or computer vision system may be able to exploit this prior information to determine the most likely surface reflectance given an observed image. We develop an algorithm for reflectance classification under unknown real-world illumination, which learns relationships between surface reflectance and certain features (statistics) computed from a single observed image. We also develop an automatic feature selection method.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This report describes a computer system that creates simple computer animation in response to high-level, vague, and incomplete descriptions of films. It makes its films by collecting and evaluating suggestions from several different bodies of knowledge. The order in which it makes its choices is influenced by the focus of the film. Difficult choices are postponed to be resumed when more of the film has been determined. The system was implemented in an object-oriented language based upon computational entities called "actors". The goal behind the construction of the system is that, whenever faced with a choice, it should sensibly choose between alternatives based upon the description of the film and as much general knowledge as possible. The system is presented as a computational model of creativity and aesthetics.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The proposed research will focus on developing a novel approach to solve Software Service Evolution problems in Computing Clouds. The approach will support dynamic evolution of the software service in clouds via a set of discovered evolution patterns. An initial survey informed us that such an approach does not exist yet and is in urgent need. Evolution Requirement can be classified into evolution features; researchers can describe the whole requirement by using evolution feature typology, the typology will define the relation and dependency between each features. After the evolution feature typology has been constructed, evolution model will be created to make the evolution more specific. Aspect oriented approach can be used for enhance evolution feature-model modularity. Aspect template code generation technique will be used for model transformation in the end. Product Line Engineering contains all the essential components for driving the whole evolution process.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Janet Taylor, Ross D King, Thomas Altmann and Oliver Fiehn (2002). Application of metabolomics to plant genotype discrimination using statistics and machine learning. 1st European Conference on Computational Biology (ECCB). (published as a journal supplement in Bioinformatics 18: S241-S248).

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Murphy, L. and Thomas, L. 2008. Dangers of a fixed mindset: implications of self-theories research for computer science education. In Proceedings of the 13th Annual Conference on innovation and Technology in Computer Science Education (Madrid, Spain, June 30 - July 02, 2008). ITiCSE '08. ACM, New York, NY, 271-275.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Sanders, K. and Thomas, L. 2007. Checklists for grading object-oriented CS1 programs: concepts and misconceptions. In Proceedings of the 12th Annual SIGCSE Conference on innovation and Technology in Computer Science Education (Dundee, Scotland, June 25 - 27, 2007). ITiCSE '07. ACM, New York, NY, 166-170

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Wireless sensor networks have recently emerged as enablers of important applications such as environmental, chemical and nuclear sensing systems. Such applications have sophisticated spatial-temporal semantics that set them aside from traditional wireless networks. For example, the computation of temperature averaged over the sensor field must take into account local densities. This is crucial since otherwise the estimated average temperature can be biased by over-sampling areas where a lot more sensors exist. Thus, we envision that a fundamental service that a wireless sensor network should provide is that of estimating local densities. In this paper, we propose a lightweight probabilistic density inference protocol, we call DIP, which allows each sensor node to implicitly estimate its neighborhood size without the explicit exchange of node identifiers as in existing density discovery schemes. The theoretical basis of DIP is a probabilistic analysis which gives the relationship between the number of sensor nodes contending in the neighborhood of a node and the level of contention measured by that node. Extensive simulations confirm the premise of DIP: it can provide statistically reliable and accurate estimates of local density at a very low energy cost and constant running time. We demonstrate how applications could be built on top of our DIP-based service by computing density-unbiased statistics from estimated local densities.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This report describes our attempt to add animation as another data type to be used on the World Wide Web. Our current network infrastructure, the Internet, is incapable of carrying video and audio streams for them to be used on the web for presentation purposes. In contrast, object-oriented animation proves to be efficient in terms of network resource requirements. We defined an animation model to support drawing-based and frame-based animation. We also extended the HyperText Markup Language in order to include this animation mode. BU-NCSA Mosanim, a modified version of the NCSA Mosaic for X(v2.5), is available to demonstrate the concept and potentials of animation in presentations an interactive game playing over the web.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Computer based mathematical models describing the aircraft evacuation process have a vital role to play in aviation safety. However such models have a heavy dependency on real evacuation data in order to (a) identify the key processes and factors associated with evacuation, (b) quantify variables and parameters associated with the identified factors/processes and finally (c) validate the models. The Fire Safety Engineering Group of the University of Greenwich is undertaking a large data extraction exercise from three major data sources in order to address these issues. This paper describes the extraction and application of data from one of these sources - aviation accident reports. To aid in the storage and analysis of the raw data, a computer database known as AASK (aircraft accident statistics and knowledge) is under development. AASK is being developed to store human observational and anecdotal data contained in accident reports and interview transcripts. AASK comprises four component sub-databases. These consist of the ACCIDENT (crash details), FLIGHT ATTENDANT (observations and actions of the flight attendants), FATALS (details concerning passenger fatalities) and PAX (observations and accounts from individual passengers) databases. AASK currently contains information from 25 survivable aviation accidents covering the period 4 April 1977 to 6 August 1995, involving some 2415 passengers, 2210 survivors, 205 fatalities and accounts from 669 people. In addition to aiding the development of aircraft evacuation models, AASK is also being used to challenge some of the myths which proliferate in the aviation safety industry such as, passenger exit selection during evacuation, nature and frequency of seat jumping, speed of passenger response and group dynamics. AASK can also be used to aid in the development of a more comprehensive approach to conducting post accident interviews, and will eventually be used to store the data directly.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The last few years have seen a substantial increase in the geometric complexity for 3D flow simulation. In this paper we describe the challenges in generating computation grids for 3D aerospace configuations and demonstrate the progress made to eventually achieve a push button technology for CAD to visualized flow. Special emphasis is given to the interfacing from the grid generator to the flow solver by semi-automatic generation of boundary conditions during the grid generation process. In this regard, once a grid has been generated, push button technology of most commercial flow solvers has been achieved. This will be demonstrated by the ad hoc simulation for the Hopper configuration.