10 resultados para Processing wikipedia data
em Biblioteca Digital da Produção Intelectual da Universidade de São Paulo
Resumo:
The web services (WS) technology provides a comprehensive solution for representing, discovering, and invoking services in a wide variety of environments, including Service Oriented Architectures (SOA) and grid computing systems. At the core of WS technology lie a number of XML-based standards, such as the Simple Object Access Protocol (SOAP), that have successfully ensured WS extensibility, transparency, and interoperability. Nonetheless, there is an increasing demand to enhance WS performance, which is severely impaired by XML's verbosity. SOAP communications produce considerable network traffic, making them unfit for distributed, loosely coupled, and heterogeneous computing environments such as the open Internet. Also, they introduce higher latency and processing delays than other technologies, like Java RMI and CORBA. WS research has recently focused on SOAP performance enhancement. Many approaches build on the observation that SOAP message exchange usually involves highly similar messages (those created by the same implementation usually have the same structure, and those sent from a server to multiple clients tend to show similarities in structure and content). Similarity evaluation and differential encoding have thus emerged as SOAP performance enhancement techniques. The main idea is to identify the common parts of SOAP messages, to be processed only once, avoiding a large amount of overhead. Other approaches investigate nontraditional processor architectures, including micro-and macrolevel parallel processing solutions, so as to further increase the processing rates of SOAP/XML software toolkits. This survey paper provides a concise, yet comprehensive review of the research efforts aimed at SOAP performance enhancement. A unified view of the problem is provided, covering almost every phase of SOAP processing, ranging over message parsing, serialization, deserialization, compression, multicasting, security evaluation, and data/instruction-level processing.
Resumo:
CHARACTERIZATION STUDY OF CAMBUCI FRUIT [Campomanesia phaea (O. Berg.) Landrum] AND ITS APPLICATION IN JELLY PROCESSING The objective of this paper was to study possible differences of varieties of cambuci fruit reported by producers by means of fruit characterization and elaboration of jelly. The fruits were divided in four possible varieties groups, named as A, B, C and D, and submitted to analysis regarding weight, ash, moisture, pH, soluble solids, acidity, ratio, ascorbic acid and water activity. To the preparation of the jelly was chosen the variety A with two formulations, 50%50% and 40%60%, fruit and sugar respectively. Concerning the possible varieties of cambuci only water activity was not significantly different. The acidity and ratio of B variety are noteworthy, because it did differ from the others presenting a more satisfactory result for in natura consumption. All other tests presented statistic alteration of at least one variety, but since these data can be influenced by edaphoclimatic conditions they must be carefully observed. Lower results for pH, acidity, luminosity and degrees hue, were observed for the 60% sugar jelly sample, which contributed for its better results in the preference test for the attributes appearance, color, flavor and texture. The evaluation of attribute aroma of both samples (50% and 60%) did not differ.
Resumo:
This study evaluated whether processing non-timber forest products (NTFPs) and establishing trade partnerships between forest communities and companies enhance the outcomes of NTFP commercialization. In particular, we evaluated whether product processing, partnerships, or their combination was associated with a number of outcomes related to the well-being of forest inhabitants and forest conservation. We based our analyses on ethnographic and quantitative data (i.e., survey and systematic observations) gathered at seven communities from five societies of the Brazilian and Bolivian Amazon. Our results indicated that product processing and partnerships do not represent a silver bullet able to improve the results of NTFP commercialization in terms of well-being and conservation indicators. Compared with cases without interventions, households adopting partnerships but not product processing were most often associated with improved economic proxies of well-being (total income, NTFP income, food consumption and gender equality in income). In comparison, the combination of product processing and partnerships was associated with similar outcomes. Unexpectedly, product processing alone was associated with negative outcomes in the economic indicators of well-being. All of the investigated strategies were associated with less time spent in social and cultural activities. With respect to forest conservation, the strategies that included a partnership with or without processing produced similar results: while household deforestation tended to decrease, the hunting impact increased. Processing alone was also associated with higher levels of hunting, though it did not reduce deforestation. Our results indicate that establishing partnerships may enhance the outcomes of NTFP trade in terms of the financial outcomes of local communities, but practitioners need to use caution when adopting the processing strategy and they need to evaluate potential negative results for indicators of social and cultural activities. With respect to conservation, the three strategies are promising for reducing deforestation, but more pervasive impacts, such as hunting, might increase.
Resumo:
This study investigated whether there are differences in the Speech-Evoked Auditory Brainstem Response among children with Typical Development (TD), (Central) Auditory Processing Disorder (C) APD, and Language Impairment (LI). The speech-evoked Auditory Brainstem Response was tested in 57 children (ages 6-12). The children were placed into three groups: TD (n = 18), (C)APD (n = 18) and LI (n = 21). Speech-evoked ABR were elicited using the five-formant syllable/da/. Three dimensions were defined for analysis, including timing, harmonics, and pitch. A comparative analysis of the responses between the typical development children and children with (C)APD and LI revealed abnormal encoding of the speech acoustic features that are characteristics of speech perception in children with (C)APD and LI, although the two groups differed in their abnormalities. While the children with (C)APD might had a greater difficulty distinguishing stimuli based on timing cues, the children with LI had the additional difficulty of distinguishing speech harmonics, which are important to the identification of speech sounds. These data suggested that an inefficient representation of crucial components of speech sounds may contribute to the difficulties with language processing found in children with LI. Furthermore, these findings may indicate that the neural processes mediated by the auditory brainstem differ among children with auditory processing and speech-language disorders. (C) 2012 Elsevier B.V. All rights reserved.
Resumo:
Spatial data warehouses (SDWs) allow for spatial analysis together with analytical multidimensional queries over huge volumes of data. The challenge is to retrieve data related to ad hoc spatial query windows according to spatial predicates, avoiding the high cost of joining large tables. Therefore, mechanisms to provide efficient query processing over SDWs are essential. In this paper, we propose two efficient indices for SDW: the SB-index and the HSB-index. The proposed indices share the following characteristics. They enable multidimensional queries with spatial predicate for SDW and also support predefined spatial hierarchies. Furthermore, they compute the spatial predicate and transform it into a conventional one, which can be evaluated together with other conventional predicates by accessing a star-join Bitmap index. While the SB-index has a sequential data structure, the HSB-index uses a hierarchical data structure to enable spatial objects clustering and a specialized buffer-pool to decrease the number of disk accesses. The advantages of the SB-index and the HSB-index over the DBMS resources for SDW indexing (i.e. star-join computation and materialized views) were investigated through performance tests, which issued roll-up operations extended with containment and intersection range queries. The performance results showed that improvements ranged from 68% up to 99% over both the star-join computation and the materialized view. Furthermore, the proposed indices proved to be very compact, adding only less than 1% to the storage requirements. Therefore, both the SB-index and the HSB-index are excellent choices for SDW indexing. Choosing between the SB-index and the HSB-index mainly depends on the query selectivity of spatial predicates. While low query selectivity benefits the HSB-index, the SB-index provides better performance for higher query selectivity.
Resumo:
The autoregressive (AR) estimator, a non-parametric method, is used to analyze functional magnetic resonance imaging (fMRI) data. The same method has been used, with success, in several other time series data analysis. It uses exclusively the available experimental data points to estimate the most plausible power spectra compatible with the experimental data and there is no need to make any assumption about non-measured points. The time series, obtained from fMRI block paradigm data, is analyzed by the AR method to determine the brain active regions involved in the processing of a given stimulus. This method is considerably more reliable than the fast Fourier transform or the parametric methods. The time series corresponding to each image pixel is analyzed using the AR estimator and the corresponding poles are obtained. The pole distribution gives the shape of power spectra, and the pixels with poles at the stimulation frequency are considered as the active regions. The method was applied in simulated and real data, its superiority is shown by the receiver operating characteristic curves which were obtained using the simulated data.
Resumo:
Magnetic nanoparticles are promising for a variety of applications, such as biomedical devices, spin electronics, magnetic data storage media, to name a few. However, these goals may only be reached if stable and organized structures are fabricated. In this article, we report on a single-step synthetic route with the coprecipitation method, in which iron oxide magnetic nanoparticles (Fe3O4 NPs) were stabilized in aqueous media using the poly(diallyldimethylammonium chloride) (PDAC) polyelectrolyte. The Fe3O4 NPs had a diameter of ca. 5 nm, according to transmission electron microscopy (TEM) images, being arranged in an inverse spinel structure typical of magnetite. An investigation with infrared spectroscopy indicated that the mechanisms of stabilization in the polymer matrix were based on the interaction between quaternary amide groups from PDAC and the nanoparticle surface. The Fe3O4-PDAC NPs exhibited considerable magnetic susceptibility, with a monotonic increase in the magnetization with decreasing temperature. These Fe3O4-PDAC NPs were immobilized in layer-by-layer (LbL) films, being alternated with layers of poly(vinylsulfonic acid) (PVS). The LbL films were much rougher than typical films made with polyelectrolytes, and Fe3O4-PDAC NPs have been responsible for the high electrocatalytic activity toward H2O2 reduction, with an overpotential shift of 0.69 V. Overall, the stability, magnetic properties and film-forming ability indicate that the Fe3O4-PDAC NPs may be used for nanoelectronics and bioelectrochemical devices requiring reversible and magnetic redox materials.
Resumo:
Current commercial and academic OLAP tools do not process XML data that contains XLink. Aiming at overcoming this issue, this paper proposes an analytical system composed by LMDQL, an analytical query language. Also, the XLDM metamodel is given to model cubes of XML documents with XLink and to deal with syntactic, semantic and structural heterogeneities commonly found in XML documents. As current W3C query languages for navigating in XML documents do not support XLink, XLPath is discussed in this article to provide features for the LMDQL query processing. A prototype system enabling the analytical processing of XML documents that use XLink is also detailed. This prototype includes a driver, named sql2xquery, which performs the mapping of SQL queries into XQuery. To validate the proposed system, a case study and its performance evaluation are presented to analyze the impact of analytical processing over XML/XLink documents.
Resumo:
Current scientific applications have been producing large amounts of data. The processing, handling and analysis of such data require large-scale computing infrastructures such as clusters and grids. In this area, studies aim at improving the performance of data-intensive applications by optimizing data accesses. In order to achieve this goal, distributed storage systems have been considering techniques of data replication, migration, distribution, and access parallelism. However, the main drawback of those studies is that they do not take into account application behavior to perform data access optimization. This limitation motivated this paper which applies strategies to support the online prediction of application behavior in order to optimize data access operations on distributed systems, without requiring any information on past executions. In order to accomplish such a goal, this approach organizes application behaviors as time series and, then, analyzes and classifies those series according to their properties. By knowing properties, the approach selects modeling techniques to represent series and perform predictions, which are, later on, used to optimize data access operations. This new approach was implemented and evaluated using the OptorSim simulator, sponsored by the LHC-CERN project and widely employed by the scientific community. Experiments confirm this new approach reduces application execution time in about 50 percent, specially when handling large amounts of data.
Resumo:
Abstract Background The study and analysis of gene expression measurements is the primary focus of functional genomics. Once expression data is available, biologists are faced with the task of extracting (new) knowledge associated to the underlying biological phenomenon. Most often, in order to perform this task, biologists execute a number of analysis activities on the available gene expression dataset rather than a single analysis activity. The integration of heteregeneous tools and data sources to create an integrated analysis environment represents a challenging and error-prone task. Semantic integration enables the assignment of unambiguous meanings to data shared among different applications in an integrated environment, allowing the exchange of data in a semantically consistent and meaningful way. This work aims at developing an ontology-based methodology for the semantic integration of gene expression analysis tools and data sources. The proposed methodology relies on software connectors to support not only the access to heterogeneous data sources but also the definition of transformation rules on exchanged data. Results We have studied the different challenges involved in the integration of computer systems and the role software connectors play in this task. We have also studied a number of gene expression technologies, analysis tools and related ontologies in order to devise basic integration scenarios and propose a reference ontology for the gene expression domain. Then, we have defined a number of activities and associated guidelines to prescribe how the development of connectors should be carried out. Finally, we have applied the proposed methodology in the construction of three different integration scenarios involving the use of different tools for the analysis of different types of gene expression data. Conclusions The proposed methodology facilitates the development of connectors capable of semantically integrating different gene expression analysis tools and data sources. The methodology can be used in the development of connectors supporting both simple and nontrivial processing requirements, thus assuring accurate data exchange and information interpretation from exchanged data.