1000 resultados para 280599 Data Format not elsewhere classified


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Significant advances have been made in the last decade to quantify the process of wet granulation. The attributes of product granules from the granulation process are controlled by a combination of three groups of processes occurring in the granulator: (1) wetting and nucleation, (2) growth and consolidation and (3) breakage and attrition. For the first two of these processes, the key controlling dimensionless groups are defined and regime maps are presented and validated with data from tumbling and mixer granulators. Granulation is an example of particle design. For quantitative analysis, both careful characterisation of the feed formulation and knowledge of operating parameters are required. A key thesis of this paper is that the design, scaleup and operation of granulation processes can now be considered as quantitative engineering rather than a black art. Résumé

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Retrieving large amounts of information over wide area networks, including the Internet, is problematic due to issues arising from latency of response, lack of direct memory access to data serving resources, and fault tolerance. This paper describes a design pattern for solving the issues of handling results from queries that return large amounts of data. Typically these queries would be made by a client process across a wide area network (or Internet), with one or more middle-tiers, to a relational database residing on a remote server. The solution involves implementing a combination of data retrieval strategies, including the use of iterators for traversing data sets and providing an appropriate level of abstraction to the client, double-buffering of data subsets, multi-threaded data retrieval, and query slicing. This design has recently been implemented and incorporated into the framework of a commercial software product developed at Oracle Corporation.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Large amounts of information can be overwhelming and costly to process, especially when transmitting data over a network. A typical modern Geographical Information System (GIS) brings all types of data together based on the geographic component of the data and provides simple point-and-click query capabilities as well as complex analysis tools. Querying a Geographical Information System, however, can be prohibitively expensive due to the large amounts of data which may need to be processed. Since the use of GIS technology has grown dramatically in the past few years, there is now a need more than ever, to provide users with the fastest and least expensive query capabilities, especially since an approximated 80 % of data stored in corporate databases has a geographical component. However, not every application requires the same, high quality data for its processing. In this paper we address the issues of reducing the cost and response time of GIS queries by preaggregating data by compromising the data accuracy and precision. We present computational issues in generation of multi-level resolutions of spatial data and show that the problem of finding the best approximation for the given region and a real value function on this region, under a predictable error, in general is "NP-complete.