939 resultados para Bancos de dados


Relevância:

20.00% 20.00%

Publicador:

Resumo:

The last years have presented an increase in the acceptance and adoption of the parallel processing, as much for scientific computation of high performance as for applications of general intention. This acceptance has been favored mainly for the development of environments with massive parallel processing (MPP - Massively Parallel Processing) and of the distributed computation. A common point between distributed systems and MPPs architectures is the notion of message exchange, that allows the communication between processes. An environment of message exchange consists basically of a communication library that, acting as an extension of the programming languages that allow to the elaboration of applications parallel, such as C, C++ and Fortran. In the development of applications parallel, a basic aspect is on to the analysis of performance of the same ones. Several can be the metric ones used in this analysis: time of execution, efficiency in the use of the processing elements, scalability of the application with respect to the increase in the number of processors or to the increase of the instance of the treat problem. The establishment of models or mechanisms that allow this analysis can be a task sufficiently complicated considering parameters and involved degrees of freedom in the implementation of the parallel application. An joined alternative has been the use of collection tools and visualization of performance data, that allow the user to identify to points of strangulation and sources of inefficiency in an application. For an efficient visualization one becomes necessary to identify and to collect given relative to the execution of the application, stage this called instrumentation. In this work it is presented, initially, a study of the main techniques used in the collection of the performance data, and after that a detailed analysis of the main available tools is made that can be used in architectures parallel of the type to cluster Beowulf with Linux on X86 platform being used libraries of communication based in applications MPI - Message Passing Interface, such as LAM and MPICH. This analysis is validated on applications parallel bars that deal with the problems of the training of neural nets of the type perceptrons using retro-propagation. The gotten conclusions show to the potentiality and easinesses of the analyzed tools.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Self-organizing maps (SOM) are artificial neural networks widely used in the data mining field, mainly because they constitute a dimensionality reduction technique given the fixed grid of neurons associated with the network. In order to properly the partition and visualize the SOM network, the various methods available in the literature must be applied in a post-processing stage, that consists of inferring, through its neurons, relevant characteristics of the data set. In general, such processing applied to the network neurons, instead of the entire database, reduces the computational costs due to vector quantization. This work proposes a post-processing of the SOM neurons in the input and output spaces, combining visualization techniques with algorithms based on gravitational forces and the search for the shortest path with the greatest reward. Such methods take into account the connection strength between neighbouring neurons and characteristics of pattern density and distances among neurons, both associated with the position that the neurons occupy in the data space after training the network. Thus, the goal consists of defining more clearly the arrangement of the clusters present in the data. Experiments were carried out so as to evaluate the proposed methods using various artificially generated data sets, as well as real world data sets. The results obtained were compared with those from a number of well-known methods existent in the literature

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The control of industrial processes has become increasingly complex due to variety of factory devices, quality requirement and market competition. Such complexity requires a large amount of data to be treated by the three levels of process control: field devices, control systems and management softwares. To use data effectively in each one of these levels is extremely important to industry. Many of today s industrial computer systems consist of distributed software systems written in a wide variety of programming languages and developed for specific platforms, so, even more companies apply a significant investment to maintain or even re-write their systems for different platforms. Furthermore, it is rare that a software system works in complete isolation. In industrial automation is common that, software had to interact with other systems on different machines and even written in different languages. Thus, interoperability is not just a long-term challenge, but also a current context requirement of industrial software production. This work aims to propose a middleware solution for communication over web service and presents an user case applying the solution developed to an integrated system for industrial data capture , allowing such data to be available simplified and platformindependent across the network

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Crude oil is a complex liquid mixture of organic and inorganic compounds that are dominated by hydrocarbons. It is a mixture of alkanes from the simplest to more complex aromatic compounds that are present derivatives such as gasoline, diesel, alcohol, kerosene, naphtha, etc.. These derivatives are extracted from any oil, however, only with a very high quality, in other words, when the content of hydrocarbons of low molecular weight is high means that production of these compounds is feasible. The American Petroleum Institute (API) developed a classification system for the various types of oil. In Brazil, the quality of most of the oil taken from wells is very low, so it is necessary to generate new technology to develop best practices for refining in order to produce petroleum products of higher commercial value. Therefore, it is necessary to study the thermodynamic equilibrium properties of its derivative compounds of interest. This dissertation aims to determine vapor-liquid equilibrium (VLE) data for the systems Phenilcyclohexane - CO2, and Cyclohexane - Phenilcyclohexane - CO2 at high pressure and temperatures between 30 to 70oC. Furthermore, comparisons between measured VLE experimental data from this work and from the literature in relation to the Peng- Robinson molecular thermodynamic model, using a simulation program SPECS IVCSEP v5.60 and two adjustable interaction parameters, have been performed for modeling and simulation purposes. Finally, the developed apparatus for determination of phase equilibrium data at high pressures is presented

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The objective of this work was the development and improvement of the mathematical models based on mass and heat balances, representing the drying transient process fruit pulp in spouted bed dryer with intermittent feeding. Mass and energy balance for drying, represented by a system of differential equations, were developed in Fortran language and adapted to the condition of intermittent feeding and mass accumulation. Were used the DASSL routine (Differential Algebraic System Solver) for solving the differential equation system and used a heuristic optimization algorithm in parameter estimation, the Particle Swarm algorithm. From the experimental data food drying, the differential models were used to determine the quantity of water and the drying air temperature at the exit of a spouted bed and accumulated mass of powder in the dryer. The models were validated using the experimental data of drying whose operating conditions, air temperature, flow rate and time intermittency, varied within the limits studied. In reviewing the results predicted, it was found that these models represent the experimental data of the kinetics of production and accumulation of powder and humidity and air temperature at the outlet of the dryer

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In the present work are established initially the fundamental relationships of thermodynamics that govern the equilibrium between phases, the models used for the description of the behavior non ideal of the liquid and vapor phases in conditions of low pressures. This work seeks the determination of vapor-liquid equilibrium (VLE) data for a series of multicomponents mixtures of saturated aliphatic hydrocarbons, prepared synthetically starting from substances with analytical degree and the development of a new dynamic cell with circulation of the vapor phase. The apparatus and experimental procedures developed are described and applied for the determination of VLE data. VLE isobarics data were obtained through a Fischer s ebulliometer of circulation of both phases, for the systems pentane + dodecane, heptane + dodecane and decane + dodecane. Using the two new dynamic cells especially projected, of easy operation and low cost, with circulation of the vapor phase, data for the systems heptane + decane + dodecane, acetone + water, tween 20 + dodecane, phenol + water and distillation curves of a gasoline without addictive were measured. Compositions of the equilibrium phases were found by densimetry, chromatography, and total organic carbon analyzer. Calibration curves of density versus composition were prepared from synthetic mixtures and the behavior excess volumes were evaluated. The VLE data obtained experimentally for the hydrocarbon and aqueous systems were submitted to the test of thermodynamic consistency, as well as the obtained from the literature data for another binary systems, mainly in the bank DDB (Dortmund Data Bank), where the Gibbs-Duhem equation is used obtaining a satisfactory data base. The results of the thermodynamic consistency tests for the binary and ternary systems were evaluated in terms of deviations for applications such as model development. Later, those groups of data (tested and approved) were used in the KijPoly program for the determination of the binary kij parameters of the cubic equations of state original Peng-Robinson and with the expanded alpha function. These obtained parameters can be applied for simulation of the reservoirs petroleum conditions and of the several distillation processes found in the petrochemistry industry, through simulators. The two designed dynamic cells used equipments of national technology for the determination of VLE data were well succeed, demonstrating efficiency and low cost. Multicomponents systems, mixtures of components of different molecular weights and also diluted solutions may be studied in these developed VLE cells

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The nonionic surfactants when in aqueous solution, have the property of separating into two phases, one called diluted phase, with low concentration of surfactant, and the other one rich in surfactants called coacervate. The application of this kind of surfactant in extraction processes from aqueous solutions has been increasing over time, which implies the need for knowledge of the thermodynamic properties of these surfactants. In this study were determined the cloud point of polyethoxylated surfactants from nonilphenolpolietoxylated family (9,5 , 10 , 11, 12 and 13), the family from octilphenolpolietoxylated (10 e 11) and polyethoxylated lauryl alcohol (6 , 7, 8 and 9) varying the degree of ethoxylation. The method used to determine the cloud point was the observation of the turbidity of the solution heating to a ramp of 0.1 ° C / minute and for the pressure studies was used a cell high-pressure maximum ( 300 bar). Through the experimental data of the studied surfactants were used to the Flory - Huggins models, UNIQUAC and NRTL to describe the curves of cloud point, and it was studied the influence of NaCl concentration and pressure of the systems in the cloud point. This last parameter is important for the processes of oil recovery in which surfactant in solution are used in high pressures. While the effect of NaCl allows obtaining cloud points for temperatures closer to the room temperature, it is possible to use in processes without temperature control. The numerical method used to adjust the parameters was the Levenberg - Marquardt. For the model Flory- Huggins parameter settings were determined as enthalpy of the mixing, mixing entropy and the number of aggregations. For the UNIQUAC and NRTL models were adjusted interaction parameters aij using a quadratic dependence with temperature. The parameters obtained had good adjust to the experimental data RSMD < 0.3 %. The results showed that both, ethoxylation degree and pressure increase the cloudy points, whereas the NaCl decrease

Relevância:

20.00% 20.00%

Publicador:

Resumo:

O artigo relata parte de estudo desenvolvido junto ao Núcleo de Ensino de Araraquara para caracterizar o professorado que atua em cursos de Magistério (antigo Normal). Por meio de questionários foram obtidas e analisadas informações relativas a sexo, idade, estado civil, experiências anteriores e responsabilidade pelo sustento da família. A análise focaliza relações com o processo educativo, com outros estudos e com aspectos extra-escolares.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In the present work are established initially the fundamental relationships of thermodynamics that govern the equilibrium between phases, the models used for the description of the behavior non ideal of the liquid and vapor phases in conditions of low pressures. This work seeks the determination of vapor-liquid equilibrium (VLE) data for a series of multicomponents mixtures of saturated aliphatic hydrocarbons, prepared synthetically starting from substances with analytical degree and the development of a new dynamic cell with circulation of the vapor phase. The apparatus and experimental procedures developed are described and applied for the determination of VLE data. VLE isobarics data were obtained through a Fischer's ebulliometer of circulation of both phases, for the systems pentane + dodecane, heptane + dodecane and decane + dodecane. Using the two new dynamic cells especially projected, of easy operation and low cost, with circulation of the vapor phase, data for the systems heptane + decane + dodecane, acetone + water, tween 20 + dodecane, phenol + water and distillation curves of a gasoline without addictive were measured. Compositions of the equilibrium phases were found by densimetry, chromatography, and total organic carbon analyzer. Calibration curves of density versus composition were prepared from synthetic mixtures and the behavior excess volumes were evaluated. The VLE data obtained experimentally for the hydrocarbon and aqueous systems were submitted to the test of thermodynamic consistency, as well as the obtained from the literature data for another binary systems, mainly in the bank DDB (Dortmund Data Bank), where the Gibbs-Duhem equation is used obtaining a satisfactory data base. The results of the thermodynamic consistency tests for the binary and ternary systems were evaluated in terms of deviations for applications such as model development. Later, those groups of data (tested and approved) were used in the KijPoly program for the determination of the binary kij parameters of the cubic equations of state original Peng-Robinson and with the expanded alpha function. These obtained parameters can be applied for simulation of the reservoirs petroleum conditions and of the several distillation processes found in the petrochemistry industry, through simulators. The two designed dynamic cells used equipments of national technology for the determination Humberto Neves Maia de Oliveira Tese de Doutorado PPGEQ/PRH-ANP 14/UFRN of VLE data were well succeed, demonstrating efficiency and low cost. Multicomponents systems, mixtures of components of different molecular weights and also diluted solutions may be studied in these developed VLE cells

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)

Relevância:

20.00% 20.00%

Publicador:

Resumo:

As plantas aquáticas têm papel fundamental no equilíbrio dos ecossistemas, porém seu crescimento desequilibrado pode obstruir canais, represas e reservatórios e afetar múltiplos usos da água. em relação a plantas aquáticas submersas, a utilização de medidas de controle torna-se mais complexa, em face da dificuldade em mapear e quantificar volumetricamente as áreas colonizadas. Nessas situações, considera-se que o uso de dados hidroacústicos possibilite o mapeamento e a mensuração dessas áreas, auxiliando na elaboração de propostas de manejo sustentáveis desse tipo de vegetação aquática. Assim, o presente trabalho utilizou dados acústicos e a técnica de krigagem para realizar a inferência espacial do biovolume de plantas aquáticas submersas. Os dados foram obtidos em três levantamentos ecobatimétricos realizados em uma área de estudos localizada no rio Paraná, caracterizada por condições favoráveis para proliferação de vegetação aquática submersa e dificuldade de navegação. Para delimitar as áreas caracterizadas pela presença de plantas aquáticas submersas, utilizou-se uma imagem multiespectral de alta resolução espacial World View-2. O mapeamento do biovolume das plantas aquáticas submersas nas áreas de ocorrência do fenômeno foi realizado a partir da inferência do biovolume por krigagem e do fatiamento dos valores inferidos em intervalos de 15%. A partir do mapa gerado, foi possível identificar os locais de maior concentração de macrófitas submersas, com predominância de valores de biovolume entre 15-30% e 30-45%, confirmando a viabilidade da utilização da krigagem na inferência espacial do biovolume, a partir de medidas ecobatimétricas georreferenciadas e com o suporte de imagem de alta resolução espacial.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper proposes a methodology for automatic extraction of building roof contours from a Digital Elevation Model (DEM), which is generated through the regularization of an available laser point cloud. The methodology is based on two steps. First, in order to detect high objects (buildings, trees etc.), the DEM is segmented through a recursive splitting technique and a Bayesian merging technique. The recursive splitting technique uses the quadtree structure for subdividing the DEM into homogeneous regions. In order to minimize the fragmentation, which is commonly observed in the results of the recursive splitting segmentation, a region merging technique based on the Bayesian framework is applied to the previously segmented data. The high object polygons are extracted by using vectorization and polygonization techniques. Second, the building roof contours are identified among all high objects extracted previously. Taking into account some roof properties and some feature measurements (e. g., area, rectangularity, and angles between principal axes of the roofs), an energy function was developed based on the Markov Random Field (MRF) model. The solution of this function is a polygon set corresponding to building roof contours and is found by using a minimization technique, like the Simulated Annealing (SA) algorithm. Experiments carried out with laser scanning DEM's showed that the methodology works properly, as it delivered roof contours with approximately 90% shape accuracy and no false positive was verified.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This research presents a methodology for prediction of building shadows cast on urban roads existing on high-resolution aerial imagery. Shadow elements can be used in the modeling of contextual information, whose use has become more and more common in image analysis complex processes. The proposed methodology consists in three sequential steps. First, the building roof contours are manually extracted from an intensity image generated by the transformation of a digital elevation model (DEM) obtained from airborne laser scanning data. In similarly, the roadside contours are extracted, now from the radiometric information of the laser scanning data. Second, the roof contour polygons are projected onto the adjacent roads by using the parallel projection straight lines, whose directions are computed from the solar ephemeris, which depends on the aerial image acquisition time. Finally, parts of shadow polygons that are free from building perspective obstructions are determined, given rise to new shadow polygons. The results obtained in the experimental evaluation of the methodology showed that the method works properly, since it allowed the prediction of shadow in high-resolution imagery with high accuracy and reliability.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper, a methodology is proposed for the geometric refinement of laser scanning building roof contours using high-resolution aerial images and Markov Random Field (MRF) models. The proposed methodology takes for granted that the 3D description of each building roof reconstructed from the laser scanning data (i.e., a polyhedron) is topologically correct and that it is only necessary to improve its accuracy. Since roof ridges are accurately extracted from laser scanning data, our main objective is to use high-resolution aerial images to improve the accuracy of roof outlines. In order to meet this goal, the available roof contours are first projected onto the image-space. After that, the projected polygons and the straight lines extracted from the image are used to establish an MRF description, which is based on relations ( relative length, proximity, and orientation) between the two sets of straight lines. The energy function associated with the MRF is minimized by using a modified version of the brute force algorithm, resulting in the grouping of straight lines for each roof object. Finally, each grouping of straight lines is topologically reconstructed based on the topology of the corresponding laser scanning polygon projected onto the image-space. The preliminary results showed that the proposed methodology is promising, since most sides of the refined polygons are geometrically better than corresponding projected laser scanning straight lines.