903 resultados para mathematical modeling of PTO
Resumo:
Mathematics Subject Classification: 26A33
Resumo:
Mathematical modeling may have different purposes in chemical and biochemical engineering sciences. One of them is to confirm or to reject kinetic models for certain processes, or to evaluate the importance of some transport phenomena on the net chemical or biochemical reaction rate. In the present paper different microbial processes are considered and modeled for evaluation of kinetic constants for batch and continuous processes accomplished by free and immobilized microbial cells. The practical examples are from the field of wastewater treatment and biosynthesis of products, like enzymes, lactic acid, gluconic acid, etc. By the aid of mathematical modeling the kinetics and the type of inhibition are specified for microbial wastewater denitrification and biodegradation of halogenated hydrocarbons. The importance of free and immobilized cells and their separate contribution to the overall microbial process is also evaluated for some fermentation processes: gluconic acid production, dichloroethane biodegradation, lactic acid fermentation and monochloroacetic acid biodegradation.
Resumo:
We propose a new approach to the mathematical modelling of microbial growth. Our approach differs from familiar Monod type models by considering two phases in the physiological states of the microorganisms and makes use of basic relations from enzyme kinetics. Such an approach may be useful in the modelling and control of biotechnological processes, where microorganisms are used for various biodegradation purposes and are often put under extreme inhibitory conditions. Some computational experiments are performed in support of our modelling approach.
Resumo:
Report published in the Proceedings of the National Conference on "Education and Research in the Information Society", Plovdiv, May, 2014
Resumo:
Inverse analysis for reactive transport of chlorides through concrete in the presence of electric field is presented. The model is solved using MATLAB’s built-in solvers “pdepe.m” and “ode15s.m”. The results from the model are compared with experimental measurements from accelerated migration test and a function representing the lack of fit is formed. This function is optimised with respect to varying amount of key parameters defining the model. Levenberg-Marquardt trust-region optimisation approach is employed. The paper presents a method by which the degree of inter-dependency between parameters and sensitivity (significance) of each parameter towards model predictions can be studied on models with or without clearly defined governing equations. Eigen value analysis of the Hessian matrix was employed to investigate and avoid over-parametrisation in inverse analysis. We investigated simultaneous fitting of parameters for diffusivity, chloride binding as defined by Freundlich isotherm (thermodynamic) and binding rate (kinetic parameter). Fitting of more than 2 parameters, simultaneously, demonstrates a high degree of parameter inter-dependency. This finding is significant as mathematical models for representing chloride transport rely on several parameters for each mode of transport (i.e., diffusivity, binding, etc.), which combined may lead to unreliable simultaneous estimation of parameters.
Resumo:
Thesis (Ph.D.)--University of Washington, 2016-08
Resumo:
The blast furnace is the main ironmaking production unit in the world which converts iron ore with coke and hot blast into liquid iron, hot metal, which is used for steelmaking. The furnace acts as a counter-current reactor charged with layers of raw material of very different gas permeability. The arrangement of these layers, or burden distribution, is the most important factor influencing the gas flow conditions inside the furnace, which dictate the efficiency of the heat transfer and reduction processes. For proper control the furnace operators should know the overall conditions in the furnace and be able to predict how control actions affect the state of the furnace. However, due to high temperatures and pressure, hostile atmosphere and mechanical wear it is very difficult to measure internal variables. Instead, the operators have to rely extensively on measurements obtained at the boundaries of the furnace and make their decisions on the basis of heuristic rules and results from mathematical models. It is particularly difficult to understand the distribution of the burden materials because of the complex behavior of the particulate materials during charging. The aim of this doctoral thesis is to clarify some aspects of burden distribution and to develop tools that can aid the decision-making process in the control of the burden and gas distribution in the blast furnace. A relatively simple mathematical model was created for simulation of the distribution of the burden material with a bell-less top charging system. The model developed is fast and it can therefore be used by the operators to gain understanding of the formation of layers for different charging programs. The results were verified by findings from charging experiments using a small-scale charging rig at the laboratory. A basic gas flow model was developed which utilized the results of the burden distribution model to estimate the gas permeability of the upper part of the blast furnace. This combined formulation for gas and burden distribution made it possible to implement a search for the best combination of charging parameters to achieve a target gas temperature distribution. As this mathematical task is discontinuous and non-differentiable, a genetic algorithm was applied to solve the optimization problem. It was demonstrated that the method was able to evolve optimal charging programs that fulfilled the target conditions. Even though the burden distribution model provides information about the layer structure, it neglects some effects which influence the results, such as mixed layer formation and coke collapse. A more accurate numerical method for studying particle mechanics, the Discrete Element Method (DEM), was used to study some aspects of the charging process more closely. Model charging programs were simulated using DEM and compared with the results from small-scale experiments. The mixed layer was defined and the voidage of mixed layers was estimated. The mixed layer was found to have about 12% less voidage than layers of the individual burden components. Finally, a model for predicting the extent of coke collapse when heavier pellets are charged over a layer of lighter coke particles was formulated based on slope stability theory, and was used to update the coke layer distribution after charging in the mathematical model. In designing this revision, results from DEM simulations and charging experiments for some charging programs were used. The findings from the coke collapse analysis can be used to design charging programs with more stable coke layers.
Resumo:
Background Plant-soil interaction is central to human food production and ecosystem function. Thus, it is essential to not only understand, but also to develop predictive mathematical models which can be used to assess how climate and soil management practices will affect these interactions. Scope In this paper we review the current developments in structural and chemical imaging of rhizosphere processes within the context of multiscale mathematical image based modeling. We outline areas that need more research and areas which would benefit from more detailed understanding. Conclusions We conclude that the combination of structural and chemical imaging with modeling is an incredibly powerful tool which is fundamental for understanding how plant roots interact with soil. We emphasize the need for more researchers to be attracted to this area that is so fertile for future discoveries. Finally, model building must go hand in hand with experiments. In particular, there is a real need to integrate rhizosphere structural and chemical imaging with modeling for better understanding of the rhizosphere processes leading to models which explicitly account for pore scale processes.
Resumo:
This work aims to study the application of Genetic Algorithms in anaerobic digestion modeling, in particular when using dynamical models. Along the work, different types of bioreactors are shown, such as batch, semi-batch and continuous, as well as their mathematical modeling. The work intendeds to estimate the parameter values of two biological reaction model. For that, simulated results, where only one output variable, the produced biogas, is known, are fitted to the model results. For this reason, the problems associated with reverse optimization are studied, using some graphics that provide clues to the sensitivity and identifiability associated with the problem. Particular solutions obtained by the identifiability analysis using GENSSI and DAISY softwares are also presented. Finally, the optimization is performed using genetic algorithms. During this optimization the need to improve the convergence of genetic algorithms was felt. This need has led to the development of an adaptation of the genetic algorithms, which we called Neighbored Genetic Algorithms (NGA1 and NGA2). In order to understand if this new approach overcomes the Basic Genetic Algorithms (BGA) and achieves the proposed goals, a study of 100 full optimization runs for each situation was further developed. Results show that NGA1 and NGA2 are statistically better than BGA. However, because it was not possible to obtain consistent results, the Nealder-Mead method was used, where the initial guesses were the estimated results from GA; Algoritmos Evolucionários para a Modelação de Bioreactores Resumo: Neste trabalho procura-se estudar os algoritmos genéticos com aplicação na modelação da digestão anaeróbia e, em particular, quando se utilizam modelos dinâmicos. Ao longo do mesmo, são apresentados diferentes tipos de bioreactores, como os batch, semi-batch e contínuos, bem como a modelação matemática dos mesmos. Neste trabalho procurou-se estimar o valor dos parâmetros que constam num modelo de digestão anaeróbia para o ajustar a uma situação simulada onde apenas se conhece uma variável de output, o biogas produzido. São ainda estudados os problemas associados à optimização inversa com recurso a alguns gráficos que fornecem pistas sobre a sensibilidade e identifiacabilidade associadas ao problema da modelação da digestão anaeróbia. São ainda apresentadas soluções particulares de idenficabilidade obtidas através dos softwares GENSSI e DAISY. Finalmente é realizada a optimização do modelo com recurso aos algoritmos genéticos. No decorrer dessa optimização sentiu-se a necessidade de melhorar a convergência e, portanto, desenvolveu-se ainda uma adaptação dos algoritmos genéticos a que se deu o nome de Neighboured Genetic Algorithms (NGA1 e NGA2). No sentido de se compreender se as adaptações permitiam superar os algoritmos genéticos básicos e atingir as metas propostas, foi ainda desenvolvido um estudo em que o processo de optimização foi realizado 100 vezes para cada um dos métodos, o que permitiu concluir, estatisticamente, que os BGA foram superados pelos NGA1 e NGA2. Ainda assim, porque não foi possivel obter consistência nos resultados, foi usado o método de Nealder-Mead utilizado como estimativa inicial os resultados obtidos pelos algoritmos genéticos.
Resumo:
American tegumentary leishmaniasis (ATL) is a disease transmitted to humans by the female sandflies of the genus Lutzomyia. Several factors are involved in the disease transmission cycle. In this work only rainfall and deforestation were considered to assess the variability in the incidence of ATL. In order to reach this goal, monthly recorded data of the incidence of ATL in Orán, Salta, Argentina, were used, in the period 1985-2007. The square root of the relative incidence of ATL and the corresponding variance were formulated as time series, and these data were smoothed by moving averages of 12 and 24 months, respectively. The same procedure was applied to the rainfall data. Typical months, which are April, August, and December, were found and allowed us to describe the dynamical behavior of ATL outbreaks. These results were tested at 95% confidence level. We concluded that the variability of rainfall would not be enough to justify the epidemic outbreaks of ATL in the period 1997-2000, but it consistently explains the situation observed in the years 2002 and 2004. Deforestation activities occurred in this region could explain epidemic peaks observed in both years and also during the entire time of observation except in 2005-2007.
Resumo:
Onion (Allium cepa) is one of the most cultivated and consumed vegetables in Brazil and its importance is due to the large laborforce involved. One of the main pests that affect this crop is the Onion Thrips (Thrips tabaci), but the spatial distribution of this insect, although important, has not been considered in crop management recommendations, experimental planning or sampling procedures. Our purpose here is to consider statistical tools to detect and model spatial patterns of the occurrence of the onion thrips. In order to characterize the spatial distribution pattern of the Onion Thrips a survey was carried out to record the number of insects in each development phase on onion plant leaves, on different dates and sample locations, in four rural properties with neighboring farms under different infestation levels and planting methods. The Mantel randomization test proved to be a useful tool to test for spatial correlation which, when detected, was described by a mixed spatial Poisson model with a geostatistical random component and parameters allowing for a characterization of the spatial pattern, as well as the production of prediction maps of susceptibility to levels of infestation throughout the area.
Resumo:
Below cloud scavenging processes have been investigated considering a numerical simulation, local atmospheric conditions and particulate matter (PM) concentrations, at different sites in Germany. The below cloud scavenging model has been coupled with bulk particulate matter counter TSI (Trust Portacounter dataset, consisting of the variability prediction of the particulate air concentrations during chosen rain events. The TSI samples and meteorological parameters were obtained during three winter Campaigns: at Deuselbach, March 1994, consisting in three different events; Sylt, April 1994 and; Freiburg, March 1995. The results show a good agreement between modeled and observed air concentrations, emphasizing the quality of the conceptual model used in the below cloud scavenging numerical modeling. The results between modeled and observed data have also presented high square Pearson coefficient correlations over 0.7 and significant, except the Freiburg Campaign event. The differences between numerical simulations and observed dataset are explained by the wind direction changes and, perhaps, the absence of advection mass terms inside the modeling. These results validate previous works based on the same conceptual model.
Resumo:
This work presents a model for the magnetic Barkhausen jump in low carbon content steels. The outcomes of the model evidence that the Barkhausen jump height depends on the coercive field of the pinning site and on the mean free path of the domain wall between pinning sites. These results are used to deduce the influence of the microstructural features and of the magnetizing parameters on the amplitude and duration of the Barkhausen jumps. In particular, a theoretical expression, establishing the dependence of the Barkbausen jump height on the carbon content and grain size, is obtained. The model also reveals the dependence of the Barkhausen jump on the applied frequency and amplitude. Theoretical and experimental results are presented and compared, being in good agreement. (C) 2008 American Institute of Physics.
Resumo:
Atmospheric aerosol particles serving as cloud condensation nuclei (CCN) are key elements of the hydrological cycle and climate. We have measured and characterized CCN at water vapor supersaturations in the range of S=0.10-0.82% in pristine tropical rainforest air during the AMAZE-08 campaign in central Amazonia. The effective hygroscopicity parameters describing the influence of chemical composition on the CCN activity of aerosol particles varied in the range of kappa approximate to 0.1-0.4 (0.16+/-0.06 arithmetic mean and standard deviation). The overall median value of kappa approximate to 0.15 was by a factor of two lower than the values typically observed for continental aerosols in other regions of the world. Aitken mode particles were less hygroscopic than accumulation mode particles (kappa approximate to 0.1 at D approximate to 50 nm; kappa approximate to 0.2 at D approximate to 200 nm), which is in agreement with earlier hygroscopicity tandem differential mobility analyzer (H-TDMA) studies. The CCN measurement results are consistent with aerosol mass spectrometry (AMS) data, showing that the organic mass fraction (f(org)) was on average as high as similar to 90% in the Aitken mode (D <= 100 nm) and decreased with increasing particle diameter in the accumulation mode (similar to 80% at D approximate to 200 nm). The kappa values exhibited a negative linear correlation with f(org) (R(2)=0.81), and extrapolation yielded the following effective hygroscopicity parameters for organic and inorganic particle components: kappa(org)approximate to 0.1 which can be regarded as the effective hygroscopicity of biogenic secondary organic aerosol (SOA) and kappa(inorg)approximate to 0.6 which is characteristic for ammonium sulfate and related salts. Both the size dependence and the temporal variability of effective particle hygroscopicity could be parameterized as a function of AMS-based organic and inorganic mass fractions (kappa(p)=kappa(org) x f(org)+kappa(inorg) x f(inorg)). The CCN number concentrations predicted with kappa(p) were in fair agreement with the measurement results (similar to 20% average deviation). The median CCN number concentrations at S=0.1-0.82% ranged from N(CCN,0.10)approximate to 35 cm(-3) to N(CCN,0.82)approximate to 160 cm(-3), the median concentration of aerosol particles larger than 30 nm was N(CN,30)approximate to 200 cm(-3), and the corresponding integral CCN efficiencies were in the range of N(CCN,0.10/NCN,30)approximate to 0.1 to N(CCN,0.82/NCN,30)approximate to 0.8. Although the number concentrations and hygroscopicity parameters were much lower in pristine rainforest air, the integral CCN efficiencies observed were similar to those in highly polluted megacity air. Moreover, model calculations of N(CCN,S) assuming an approximate global average value of kappa approximate to 0.3 for continental aerosols led to systematic overpredictions, but the average deviations exceeded similar to 50% only at low water vapor supersaturation (0.1%) and low particle number concentrations (<= 100 cm(-3)). Model calculations assuming aconstant aerosol size distribution led to higher average deviations at all investigated levels of supersaturation: similar to 60% for the campaign average distribution and similar to 1600% for a generic remote continental size distribution. These findings confirm earlier studies suggesting that aerosol particle number and size are the major predictors for the variability of the CCN concentration in continental boundary layer air, followed by particle composition and hygroscopicity as relatively minor modulators. Depending on the required and applicable level of detail, the information and parameterizations presented in this paper should enable efficient description of the CCN properties of pristine tropical rainforest aerosols of Amazonia in detailed process models as well as in large-scale atmospheric and climate models.
Resumo:
A thermodynamic approach is presented to model devices manufactured with cellular polymers. They are heterogeneous nonpolar space-charge electrets that exhibit much higher piezoelectricity than the well-known ferroelectric polymers. Their pyroelectric and piezoelectric properties are characterized by adequate coefficients which quantify the performance of devices manufactured with those materials. The method presented in this contribution to calculate those coefficients is exact and consistent avoiding ad hoc simplifications introduced in other approaches. The results obtained by this method allow drawing conclusions regarding device optimization.