998 resultados para Modelos de dados de contagem


Relevância:

30.00% 30.00%

Publicador:

Resumo:

The last years have presented an increase in the acceptance and adoption of the parallel processing, as much for scientific computation of high performance as for applications of general intention. This acceptance has been favored mainly for the development of environments with massive parallel processing (MPP - Massively Parallel Processing) and of the distributed computation. A common point between distributed systems and MPPs architectures is the notion of message exchange, that allows the communication between processes. An environment of message exchange consists basically of a communication library that, acting as an extension of the programming languages that allow to the elaboration of applications parallel, such as C, C++ and Fortran. In the development of applications parallel, a basic aspect is on to the analysis of performance of the same ones. Several can be the metric ones used in this analysis: time of execution, efficiency in the use of the processing elements, scalability of the application with respect to the increase in the number of processors or to the increase of the instance of the treat problem. The establishment of models or mechanisms that allow this analysis can be a task sufficiently complicated considering parameters and involved degrees of freedom in the implementation of the parallel application. An joined alternative has been the use of collection tools and visualization of performance data, that allow the user to identify to points of strangulation and sources of inefficiency in an application. For an efficient visualization one becomes necessary to identify and to collect given relative to the execution of the application, stage this called instrumentation. In this work it is presented, initially, a study of the main techniques used in the collection of the performance data, and after that a detailed analysis of the main available tools is made that can be used in architectures parallel of the type to cluster Beowulf with Linux on X86 platform being used libraries of communication based in applications MPI - Message Passing Interface, such as LAM and MPICH. This analysis is validated on applications parallel bars that deal with the problems of the training of neural nets of the type perceptrons using retro-propagation. The gotten conclusions show to the potentiality and easinesses of the analyzed tools.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The objective of this work was the development and improvement of the mathematical models based on mass and heat balances, representing the drying transient process fruit pulp in spouted bed dryer with intermittent feeding. Mass and energy balance for drying, represented by a system of differential equations, were developed in Fortran language and adapted to the condition of intermittent feeding and mass accumulation. Were used the DASSL routine (Differential Algebraic System Solver) for solving the differential equation system and used a heuristic optimization algorithm in parameter estimation, the Particle Swarm algorithm. From the experimental data food drying, the differential models were used to determine the quantity of water and the drying air temperature at the exit of a spouted bed and accumulated mass of powder in the dryer. The models were validated using the experimental data of drying whose operating conditions, air temperature, flow rate and time intermittency, varied within the limits studied. In reviewing the results predicted, it was found that these models represent the experimental data of the kinetics of production and accumulation of powder and humidity and air temperature at the outlet of the dryer

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This work studies the development, implementation and improvement of a macroscopic model to describe the behavior of the spouted bed dryer with continuous feeding for pastes and suspensions drying. This model is based on the CST model (Freire et al., 2009) and the model of Fernandes (2005), whose theoretical foundation is based on macroscopic mass and heat balances for the three phases involved in the process: gas, liquid and solid. Because this technique is quite relevant, the studies of modeling and simulation of spouted bed drying are essential in the analysis of the process as a whole, because through them it is possible to predict and understand the behavior of the process, which contributes significantly to more efficient project and operation. The development and understanding of the phenomena involved in the drying process can be obtained by comparing the experimental data with those from computer simulations. Such knowledge is critical for choosing properly the process conditions in order to obtain a good drying efficiency. Over the past few years, researches and development of works in the field of pastes and suspensions drying in spouted bed has been gaining ground in Brazil. The Particulate Systems Laboratory at Universidade Federal do Rio Grande do Norte, has been developing several researches and generating a huge collection of experimental data concerning the drying of fruit pulps, vegetables pastes, goat milk and suspensions of agro-industrial residues. From this collection, some data of goat milk and residue from acerola (Malpighia glabra L.) drying were collected. For the first time, these data were used for the development and validation of a model that can describe the behavior of spouted bed dryer. Thus, it was possible to model the dryer and to evaluate the influence of process variables (paste feeding, temperature and flow rate of the drying air) in the drying dynamics. We also performed water evaporation experiments in order to understand and to study the behavior of the dryer wall temperature and the evaporation rate. All these analysis will contribute to future works involving the implementation of control strategies in the pastes and suspensions drying. The results obtained in transient analysis were compared with experimental data indicating that this model well represents the process

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the present work are established initially the fundamental relationships of thermodynamics that govern the equilibrium between phases, the models used for the description of the behavior non ideal of the liquid and vapor phases in conditions of low pressures. This work seeks the determination of vapor-liquid equilibrium (VLE) data for a series of multicomponents mixtures of saturated aliphatic hydrocarbons, prepared synthetically starting from substances with analytical degree and the development of a new dynamic cell with circulation of the vapor phase. The apparatus and experimental procedures developed are described and applied for the determination of VLE data. VLE isobarics data were obtained through a Fischer s ebulliometer of circulation of both phases, for the systems pentane + dodecane, heptane + dodecane and decane + dodecane. Using the two new dynamic cells especially projected, of easy operation and low cost, with circulation of the vapor phase, data for the systems heptane + decane + dodecane, acetone + water, tween 20 + dodecane, phenol + water and distillation curves of a gasoline without addictive were measured. Compositions of the equilibrium phases were found by densimetry, chromatography, and total organic carbon analyzer. Calibration curves of density versus composition were prepared from synthetic mixtures and the behavior excess volumes were evaluated. The VLE data obtained experimentally for the hydrocarbon and aqueous systems were submitted to the test of thermodynamic consistency, as well as the obtained from the literature data for another binary systems, mainly in the bank DDB (Dortmund Data Bank), where the Gibbs-Duhem equation is used obtaining a satisfactory data base. The results of the thermodynamic consistency tests for the binary and ternary systems were evaluated in terms of deviations for applications such as model development. Later, those groups of data (tested and approved) were used in the KijPoly program for the determination of the binary kij parameters of the cubic equations of state original Peng-Robinson and with the expanded alpha function. These obtained parameters can be applied for simulation of the reservoirs petroleum conditions and of the several distillation processes found in the petrochemistry industry, through simulators. The two designed dynamic cells used equipments of national technology for the determination of VLE data were well succeed, demonstrating efficiency and low cost. Multicomponents systems, mixtures of components of different molecular weights and also diluted solutions may be studied in these developed VLE cells

Relevância:

30.00% 30.00%

Publicador:

Resumo:

O objetivo deste trabalho foi avaliar o desempenho de modelos isotrópicos de estimativa do total de radiação incidente em superfícies inclinadas e propor estimativas com base nas correlações entre os índices de claridade horizontais e inclinados, em diferentes condições de cobertura de céu, em Botucatu, SP. Foram avaliadas superfícies com inclinação de 12,85º, 22,85º e 32,85º, pelos modelos isotrópicos propostos por Liu & Jordan, Revfeim, Jimenez & Castro, Koronakis, a teoria Circunsolar, e a correlação entre os índices de claridade horizontais e inclinados, para diferentes condições de cobertura de céu. O banco de dados de radiação global utilizado corresponde ao período de 1998 a 2007, com intervalos de 4/1998 a 8/2001 para a inclinação de 22,85º, de 9/2001 a 2/2003 para 12,85º e de 1/2004 a 12/2007 para 32,85º. O desempenho dos modelos foi avaliado pelos indicadores estatísticos erro absoluto médio, raiz quadrada do quadrado médio do erro e índice d de Wilmott. Os modelos de Liu & Jordan, Koronakis e de Revfeim apresentaram os melhores desempenhos em dias nublados, em todas as inclinações. As coberturas de céu parcialmente difuso e parcialmente aberto, nos maiores ângulos de inclinação, apresentaram as maiores dispersões entre valores estimados e medidos, independentemente do modelo. As equações estatísticas apresentaram bons resultados em aplicações com agrupamentos de dados mensais.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The nonionic surfactants when in aqueous solution, have the property of separating into two phases, one called diluted phase, with low concentration of surfactant, and the other one rich in surfactants called coacervate. The application of this kind of surfactant in extraction processes from aqueous solutions has been increasing over time, which implies the need for knowledge of the thermodynamic properties of these surfactants. In this study were determined the cloud point of polyethoxylated surfactants from nonilphenolpolietoxylated family (9,5 , 10 , 11, 12 and 13), the family from octilphenolpolietoxylated (10 e 11) and polyethoxylated lauryl alcohol (6 , 7, 8 and 9) varying the degree of ethoxylation. The method used to determine the cloud point was the observation of the turbidity of the solution heating to a ramp of 0.1 ° C / minute and for the pressure studies was used a cell high-pressure maximum ( 300 bar). Through the experimental data of the studied surfactants were used to the Flory - Huggins models, UNIQUAC and NRTL to describe the curves of cloud point, and it was studied the influence of NaCl concentration and pressure of the systems in the cloud point. This last parameter is important for the processes of oil recovery in which surfactant in solution are used in high pressures. While the effect of NaCl allows obtaining cloud points for temperatures closer to the room temperature, it is possible to use in processes without temperature control. The numerical method used to adjust the parameters was the Levenberg - Marquardt. For the model Flory- Huggins parameter settings were determined as enthalpy of the mixing, mixing entropy and the number of aggregations. For the UNIQUAC and NRTL models were adjusted interaction parameters aij using a quadratic dependence with temperature. The parameters obtained had good adjust to the experimental data RSMD < 0.3 %. The results showed that both, ethoxylation degree and pressure increase the cloudy points, whereas the NaCl decrease

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the present work are established initially the fundamental relationships of thermodynamics that govern the equilibrium between phases, the models used for the description of the behavior non ideal of the liquid and vapor phases in conditions of low pressures. This work seeks the determination of vapor-liquid equilibrium (VLE) data for a series of multicomponents mixtures of saturated aliphatic hydrocarbons, prepared synthetically starting from substances with analytical degree and the development of a new dynamic cell with circulation of the vapor phase. The apparatus and experimental procedures developed are described and applied for the determination of VLE data. VLE isobarics data were obtained through a Fischer's ebulliometer of circulation of both phases, for the systems pentane + dodecane, heptane + dodecane and decane + dodecane. Using the two new dynamic cells especially projected, of easy operation and low cost, with circulation of the vapor phase, data for the systems heptane + decane + dodecane, acetone + water, tween 20 + dodecane, phenol + water and distillation curves of a gasoline without addictive were measured. Compositions of the equilibrium phases were found by densimetry, chromatography, and total organic carbon analyzer. Calibration curves of density versus composition were prepared from synthetic mixtures and the behavior excess volumes were evaluated. The VLE data obtained experimentally for the hydrocarbon and aqueous systems were submitted to the test of thermodynamic consistency, as well as the obtained from the literature data for another binary systems, mainly in the bank DDB (Dortmund Data Bank), where the Gibbs-Duhem equation is used obtaining a satisfactory data base. The results of the thermodynamic consistency tests for the binary and ternary systems were evaluated in terms of deviations for applications such as model development. Later, those groups of data (tested and approved) were used in the KijPoly program for the determination of the binary kij parameters of the cubic equations of state original Peng-Robinson and with the expanded alpha function. These obtained parameters can be applied for simulation of the reservoirs petroleum conditions and of the several distillation processes found in the petrochemistry industry, through simulators. The two designed dynamic cells used equipments of national technology for the determination Humberto Neves Maia de Oliveira Tese de Doutorado PPGEQ/PRH-ANP 14/UFRN of VLE data were well succeed, demonstrating efficiency and low cost. Multicomponents systems, mixtures of components of different molecular weights and also diluted solutions may be studied in these developed VLE cells

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Digital Elevation Models (DEM) are numerical representations of a portion of the earth surface. Among several factors which affect the quality of a DEM, it should be emphasized the attention on the input data and the choice of the interpolating algorithm. On the other hand, several numerical models are used nowadays to characterize nearshore hydrodynamics and morphological changes in coastal areas, whose validation is based on field data collection. Independent on the complexity of the physical processes which are modeled, little attention has been given to the intrinsic bathymetric interpolation built within the numerical models of the specific application. Therefore, this study aims to investigate and to quantify the influence of the bathymetry, as obtained by a DEM, on the hydrodynamic circulation model at a coastal stretch, off the coast of the State of Rio Grande do Norte, Northeast Brazil. This coastal region is characterized by strong hydrodynamic and littoral processes, resulting in a very dynamic morphology with shallow coastal bathymetry. Important economic activities, such as oil exploitation and production, fisheries, salt ponds, shrimp farms and tourism, also bring impacts upon the local ecosystems and influence themselves the local hydrodynamics. This fact makes the region one of the most important for the development of the State, but also enhances the possibility of serious environmental accidents. As a hydrodynamic model, SisBaHiA® - Environmental Hydrodynamics System ( Sistema Básico de Hidrodinâmica Ambiental ) was chosen, for it has been successfully employed at several locations along the Brazilian coast. This model was developed at the Coastal and Oceanographical Engineering Group of the Ocean Engineering Program at the Federal University of Rio de Janeiro. Several interpolating methods were tested for the construction of the DEM, namely Natural Neighbor, Kriging, Triangulation with Linear Interpolation, Inverse Distance to a Power, Nearest Neighbor, and Minimum Curvature, all implemented within the software Surfer®. The bathymetry which was used as reference for the DEM was obtained from nautical charts provided by the Brazilian Hydrographic Service of the Brazilian Navy and from a field survey conducted in 2005. Changes in flow velocity and free surface elevation were evaluated under three aspects: a spatial vision along three profiles perpendicular to the coast and one profile longitudinal to the coast as shown; a temporal vision from three central nodes of the grid during 30 days; a hodograph analysis of components of speed in U and V, by different tidal cycles. Small, but negligible, variations in sea surface elevation were identified. However, the differences in flow and direction of velocities were significant, depending on the DEM

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The object of this study is the construction of situation models in the discourse pattern comic book narrative, and sits in the field of Cognitive Linguistics. Its main foundations are the notions of embodied mind (LAKOFF; JOHNSON, 1999), mental simulation (BARSALOU, 1999), discourse pattern (DUQUE; COSTA, 2012) and situation models (ZWAAN, 1999). I stem from the hypothesis that the process of meaning construction in narratives is attached to the simulation of space, of time and of the characters goals and actions within the story world, dimensions which make up the situation models elaborated by the reader. The simulation of these experiences during discourse processing originates from the fact of it having an embodied and cultural basis, i.e., upon being confronted with the clues found in the narrative, the cognitive structures that make up the reader s personal and social memories are triggered and make it possible to activate information which in turn refer to his/her physical and social experiences, built up in the environment in which he/she lives. As regards comic book narratives, the construction of situation models is closely related to the recurring activation of certain cognitive structures originating from graphic resources that are typical of that discourse pattern. These conclusions were drawn from the data analysis taken from the work Palestina (SACCO, 2003; 2004; 2011)

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This work is a detailed study of self-similar models for the expansion of extragalactic radio sources. A review is made of the definitions of AGN, the unified model is discussed and the main characteristics of double radio sources are examined. Three classification schemes are outlined and the self-similar models found in the literature are studied in detail. A self-similar model is proposed that represents a generalization of the models found in the literature. In this model, the area of the head of the jet varies with the size of the jet with a power law with an exponent γ. The atmosphere has a variable density that may or may not be spherically symmetric and it is taken into account the time variation of the cinematic luminosity of the jet according to a power law with an exponent h. It is possible to show that models Type I, II and III are particular cases of the general model and one also discusses the evolution of the sources radio luminosity. One compares the evolutionary curves of the general model with the particular cases and with the observational data in a P-D diagram. The results show that the model allows a better agreement with the observations depending on the appropriate choice of the model parameters.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Recent astronomical observations (involving supernovae type Ia, cosmic background radiation anisotropy and galaxy clusters probes) have provided strong evidence that the observed universe is described by an accelerating, flat model whose space-time properties can be represented by the FriedmannRobertsonWalker (FRW) metric. However, the nature of the substance or mechanism behind the current cosmic acceleration remains unknown and its determination constitutes a challenging problem for modern cosmology. In the general relativistic description, an accelerat ing regime is usually obtained by assuming the existence of an exotic energy component endowed with negative pressure, called dark energy, which is usually represented by a cosmological constant ¤ associated to the vacuum energy density. All observational data available so far are in good agreement with the concordance cosmic ¤CDM model. Nevertheless, such models are plagued with several problems thereby inspiring many authors to propose alternative candidates in the relativistic context. In this thesis, a new kind of accelerating flat model with no dark energy and fully dominated by cold dark matter (CDM) is proposed. The number of CDM particles is not conserved and the present accelerating stage is a consequence of the negative pressure describing the irreversible process of gravitational particle creation. In order to have a transition from a decelerating to an accelerating regime at low redshifts, the matter creation rate proposed here depends on 2 parameters (y and ߯): the first one identifies a constant term of the order of H0 and the second one describes a time variation proportional to he Hubble parameter H(t). In this scenario, H0 does not need to be small in order to solve the age problem and the transition happens even if there is no matter creation during the radiation and part of the matter dominated phase (when the ß term is negligible). Like in flat ACDM scenarios, the dimming of distant type Ia supernovae can be fitted with just one free parameter, and the coincidence problem plaguing the models driven by the cosmological constant. ACDM is absent. The limits endowed with with the existence of the quasar APM 08279+5255, located at z = 3:91 and with an estimated ages between 2 and 3 Gyr are also investigated. In the simplest case (ß = 0), the model is compatible with the existence of the quasar for y > 0:56 whether the age of the quasar is 2.0 Gyr. For 3 Gyr the limit derived is y > 0:72. New limits for the formation redshift of the quasar are also established

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Coordenação de Aperfeiçoamento de Pessoal de Nível Superior

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this work, we study the survival cure rate model proposed by Yakovlev et al. (1993), based on a competing risks structure concurring to cause the event of interest, and the approach proposed by Chen et al. (1999), where covariates are introduced to model the risk amount. We focus the measurement error covariates topics, considering the use of corrected score method in order to obtain consistent estimators. A simulation study is done to evaluate the behavior of the estimators obtained by this method for finite samples. The simulation aims to identify not only the impact on the regression coefficients of the covariates measured with error (Mizoi et al. 2007) but also on the coefficients of covariates measured without error. We also verify the adequacy of the piecewise exponential distribution to the cure rate model with measurement error. At the end, model applications involving real data are made

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this work we study the survival cure rate model proposed by Yakovlev (1993) that are considered in a competing risk setting. Covariates are introduced for modeling the cure rate and we allow some covariates to have missing values. We consider only the cases by which the missing covariates are categorical and implement the EM algorithm via the method of weights for maximum likelihood estimation. We present a Monte Carlo simulation experiment to compare the properties of the estimators based on this method with those estimators under the complete case scenario. We also evaluate, in this experiment, the impact in the parameter estimates when we increase the proportion of immune and censored individuals among the not immune one. We demonstrate the proposed methodology with a real data set involving the time until the graduation for the undergraduate course of Statistics of the Universidade Federal do Rio Grande do Norte

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In Survival Analysis, long duration models allow for the estimation of the healing fraction, which represents a portion of the population immune to the event of interest. Here we address classical and Bayesian estimation based on mixture models and promotion time models, using different distributions (exponential, Weibull and Pareto) to model failure time. The database used to illustrate the implementations is described in Kersey et al. (1987) and it consists of a group of leukemia patients who underwent a certain type of transplant. The specific implementations used were numeric optimization by BFGS as implemented in R (base::optim), Laplace approximation (own implementation) and Gibbs sampling as implemented in Winbugs. We describe the main features of the models used, the estimation methods and the computational aspects. We also discuss how different prior information can affect the Bayesian estimates