24 resultados para Information science|Computer science
Resumo:
The shuttle radar topography mission (SRTM), was flow on the space shuttle Endeavour in February 2000, with the objective of acquiring a digital elevation model of all land between 60 degrees north latitude and 56 degrees south latitude, using interferometric synthetic aperture radar (InSAR) techniques. The SRTM data are distributed at horizontal resolution of 1 arc-second (similar to 30m) for areas within the USA and at 3 arc-second (similar to 90m) resolution for the rest of the world. A resolution of 90m can be considered suitable for the small or medium-scale analysis, but it is too coarse for more detailed purposes. One alternative is to interpolate the SRTM data at a finer resolution; it will not increase the level of detail of the original digital elevation model (DEM), but it will lead to a surface where there is the coherence of angular properties (i.e. slope, aspect) between neighbouring pixels, which is an important characteristic when dealing with terrain analysis. This work intents to show how the proper adjustment of variogram and kriging parameters, namely the nugget effect and the maximum distance within which values are used in interpolation, can be set to achieve quality results on resampling SRTM data from 3"" to 1"". We present for a test area in western USA, which includes different adjustment schemes (changes in nugget effect value and in the interpolation radius) and comparisons with the original 1"" model of the area, with the national elevation dataset (NED) DEMs, and with other interpolation methods (splines and inverse distance weighted (IDW)). The basic concepts for using kriging to resample terrain data are: (i) working only with the immediate neighbourhood of the predicted point, due to the high spatial correlation of the topographic surface and omnidirectional behaviour of variogram in short distances; (ii) adding a very small random variation to the coordinates of the points prior to interpolation, to avoid punctual artifacts generated by predicted points with the same location than original data points and; (iii) using a small value of nugget effect, to avoid smoothing that can obliterate terrain features. Drainages derived from the surfaces interpolated by kriging and by splines have a good agreement with streams derived from the 1"" NED, with correct identification of watersheds, even though a few differences occur in the positions of some rivers in flat areas. Although the 1"" surfaces resampled by kriging and splines are very similar, we consider the results produced by kriging as superior, since the spline-interpolated surface still presented some noise and linear artifacts, which were removed by kriging.
Resumo:
This paper presents a new framework for generating triangular meshes from textured color images. The proposed framework combines a texture classification technique, called W-operator, with Imesh, a method originally conceived to generate simplicial meshes from gray scale images. An extension of W-operators to handle textured color images is proposed, which employs a combination of RGB and HSV channels and Sequential Floating Forward Search guided by mean conditional entropy criterion to extract features from the training data. The W-operator is built into the local error estimation used by Imesh to choose the mesh vertices. Furthermore, the W-operator also enables to assign a label to the triangles during the mesh construction, thus allowing to obtain a segmented mesh at the end of the process. The presented results show that the combination of W-operators with Imesh gives rise to a texture classification-based triangle mesh generation framework that outperforms pixel based methods. Crown Copyright (C) 2009 Published by Elsevier Inc. All rights reserved.
Resumo:
We review some issues related to the implications of different missing data mechanisms on statistical inference for contingency tables and consider simulation studies to compare the results obtained under such models to those where the units with missing data are disregarded. We confirm that although, in general, analyses under the correct missing at random and missing completely at random models are more efficient even for small sample sizes, there are exceptions where they may not improve the results obtained by ignoring the partially classified data. We show that under the missing not at random (MNAR) model, estimates on the boundary of the parameter space as well as lack of identifiability of the parameters of saturated models may be associated with undesirable asymptotic properties of maximum likelihood estimators and likelihood ratio tests; even in standard cases the bias of the estimators may be low only for very large samples. We also show that the probability of a boundary solution obtained under the correct MNAR model may be large even for large samples and that, consequently, we may not always conclude that a MNAR model is misspecified because the estimate is on the boundary of the parameter space.
Resumo:
Cytochrome P450 (CYP450) is a class of enzymes where the substrate identification is particularly important to know. It would help medicinal chemists to design drugs with lower side effects due to drug-drug interactions and to extensive genetic polymorphism. Herein, we discuss the application of the 2D and 3D-similarity searches in identifying reference Structures with higher capacity to retrieve Substrates of three important CYP enzymes (CYP2C9, CYP2D6, and CYP3A4). On the basis of the complementarities of multiple reference structures selected by different similarity search methods, we proposed the fusion of their individual Tanimoto scores into a consensus Tanimoto score (T(consensus)). Using this new score, true positive rates of 63% (CYP2C9) and 81% (CYP2D6) were achieved with false positive rates of 4% for the CYP2C9-CYP2D6 data Set. Extended similarity searches were carried out oil a validation data set, and the results showed that by using the T(consensus) score, not only the area of a ROC graph increased, but also more substrates were recovered at the beginning of a ranked list.
Resumo:
This paper analyzes the changes the ways of organizing memory have undergone since ancient times, turning them into the current artificial memory systems. It aims to draw a parallel between the art of memory (which associates images to specific texts) and the hypertext (which also uses associations, but in a non-linear way). Our methodology consisted of a qualitative approach, involving the collection of texts about the art of memory and hypertext; this enables us to salvage the historical-cultural changes which have modified form and use of the art of memory and allowed the creation of hypertext. It also analyzes the similarities among artificial memory systems created by different cultures in order to prevent loss of knowledge produced by society.
Resumo:
This paper, first result of a larger research, proposes a query about some aspects of social representation of libraries and librarians, as they appear in literary and cinematographic productions. Little by little, this query, which arose from purposes of organizing catalogues, revealed elements that established different series, in which the narrative genre (literary or cinematographic) has no relevance to either libraries or librarians` representations. The presence of these elements seems to show some expectations and utopias in relation to the common knowledge, independently from narratives being located in the past, in the present or in the future, stimulating reflection on some medieval and baroque traditions about the library universe and its main characters, the librarians. The cinematographic material selected for research was The time machine, Farenheit 451, The day after tomorrow, Star Wars - episode II and the novels Martin Eden, The man without qualities, The time machine and La sombra del viento.
Resumo:
This work maps and analyses cross-citations in the areas of Biology, Mathematics, Physics and Medicine in the English version of Wikipedia, which are represented as an undirected complex network where the entries correspond to nodes and the citations among the entries are mapped as edges. We found a high value of clustering coefficient for the areas of Biology and Medicine, and a small value for Mathematics and Physics. The topological organization is also different for each network, including a modular structure for Biology and Medicine, a sparse structure for Mathematics and a dense core for Physics. The networks have degree distributions that can be approximated by a power-law with a cut-off. The assortativity of the isolated networks has also been investigated and the results indicate distinct patterns for each subject. We estimated the betweenness centrality of each node considering the full Wikipedia network, which contains the nodes of the four subjects and the edges between them. In addition, the average shortest path length between the subjects revealed a close relationship between the subjects of Biology and Physics, and also between Medicine and Physics. Our results indicate that the analysis of the full Wikipedia network cannot predict the behavior of the isolated categories since their properties can be very different from those observed in the full network. (C) 2011 Elsevier Ltd. All rights reserved.
Resumo:
In this paper, a simple relation between the Leimkuhler curve and the mean residual life is established. The result is illustrated with several models commonly used in informetrics, such as exponential, Pareto and lognormal. Finally, relationships with some other reliability concepts are also presented. (C) 2010 Elsevier Ltd. All rights reserved.
THE EXTENT OF MULTIDISCIPLINARY AUTHORSHIP OF ARTICLES ON SCIENTOMETRICS AND BIBLIOMETRICS IN BRAZIL
Resumo:
The publications in scientometrics and bibliometrics with Brazilian authorship expanded exponentially in the 1990-2006 period, reaching 13 times in the Web of Science database and 19.5 times in the Google Scholar database. This increase is rather superior to that of the total Brazilian scientific production in the same time period (5.6 times in the Web of Science). Some characteristics to be noticed in this rise are: 1) The total number of articles during this period was 197; in that, 78% were published in 57 Brazilian journals and 22% in 13 international journals. 2) The national and international articles averaged 4.3 and 5.9 citations/article, respectively; two journals stood out among these, the national Ciencia da Informacao (44 articles averaging 6.7 citations/article) and the international Scientometrics (32 articles averaging 6.2 citations/article). 3) The articles encompass an impressive participation of authors from areas other than information science; only one-fourth of the authors are bound to the information science field, the remaining ones being distributed among the areas of humanities/business administration, biology/biomedicine, health and hard sciences. The occurrence of adventitious authors at this level of multidisciplinarity is uncommon in science. However, the possible benefits of such patterns are not clear in view of a fragmented intercommunication among the authors, as noticed through the citations. The advantages of changing this trend and of using other scientometric and bibliometric databases, such as SciELO, to avoid an almost exclusive use of the Web of Science database, are discussed.