108 resultados para NUMERICAL METHODS
Resumo:
An investigation was undertaken to test the effectiveness of two procedures for recording boundaries and plot positions for scientific studies on farms on Leyte Island, the Philippines. The accuracy of a Garmin 76 Global Positioning System (GPS) unit and a compass and chain was checked under the same conditions. Tree canopies interfered with the ability of the satellite signal to reach the GPS and therefore the GPS survey was less accurate than the compass and chain survey. Where a high degree of accuracy is required, a compass and chain survey remains the most effective method of surveying land underneath tree canopies, providing operator error is minimised. For a large number of surveys and thus large amounts of data, a GPS is more appropriate than a compass and chain survey because data are easily up-loaded into a Geographic Information System (GIS). However, under dense canopies where satellite signals cannot reach the GPS, it may be necessary to revert to a compass survey or a combination of both methods.
Resumo:
The second edition of An Introduction to Efficiency and Productivity Analysis is designed to be a general introduction for those who wish to study efficiency and productivity analysis. The book provides an accessible, well-written introduction to the four principal methods involved: econometric estimation of average response models; index numbers, data envelopment analysis (DEA); and stochastic frontier analysis (SFA). For each method, a detailed introduction to the basic concepts is presented, numerical examples are provided, and some of the more important extensions to the basic methods are discussed. Of special interest is the systematic use of detailed empirical applications using real-world data throughout the book. In recent years, there have been a number of excellent advance-level books published on performance measurement. This book, however, is the first systematic survey of performance measurement with the express purpose of introducing the field to a wide audience of students, researchers, and practitioners. Indeed, the 2nd Edition maintains its uniqueness: (1) It is a well-written introduction to the field. (2) It outlines, discusses and compares the four principal methods for efficiency and productivity analysis in a well-motivated presentation. (3) It provides detailed advice on computer programs that can be used to implement these performance measurement methods. The book contains computer instructions and output listings for the SHAZAM, LIMDEP, TFPIP, DEAP and FRONTIER computer programs. More extensive listings of data and computer instruction files are available on the book's website: (www.uq.edu.au/economics/cepa/crob2005).
Resumo:
In this paper, a progressive asymptotic approach procedure is presented for solving the steady-state Horton-Rogers-Lapwood problem in a fluid-saturated porous medium. The Horton-Rogers-Lapwood problem possesses a bifurcation and, therefore, makes the direct use of conventional finite element methods difficult. Even if the Rayleigh number is high enough to drive the occurrence of natural convection in a fluid-saturated porous medium, the conventional methods will often produce a trivial non-convective solution. This difficulty can be overcome using the progressive asymptotic approach procedure associated with the finite element method. The method considers a series of modified Horton-Rogers-Lapwood problems in which gravity is assumed to tilt a small angle away from vertical. The main idea behind the progressive asymptotic approach procedure is that through solving a sequence of such modified problems with decreasing tilt, an accurate non-zero velocity solution to the Horton-Rogers-Lapwood problem can be obtained. This solution provides a very good initial prediction for the solution to the original Horton-Rogers-Lapwood problem so that the non-zero velocity solution can be successfully obtained when the tilted angle is set to zero. Comparison of numerical solutions with analytical ones to a benchmark problem of any rectangular geometry has demonstrated the usefulness of the present progressive asymptotic approach procedure. Finally, the procedure has been used to investigate the effect of basin shapes on natural convection of pore-fluid in a porous medium. (C) 1997 by John Wiley & Sons, Ltd.
Resumo:
The ability to predict leaf area and leaf area index is crucial in crop simulation models that predict crop growth and yield. Previous studies have shown existing methods of predicting leaf area to be inadequate when applied to a broad range of cultivars with different numbers of leaves. The objectives of the study were to (i) develop generalised methods of modelling individual and total plant leaf area, and leaf senescence, that do not require constants that are specific to environments and/or genotypes, (ii) re-examine the base, optimum, and maximum temperatures for calculation of thermal time for leaf senescence, and (iii) assess the method of calculation of individual leaf area from leaf length and leaf width in experimental work. Five cultivars of maize differing widely in maturity and adaptation were planted in October 1994 in south-eastern Queensland, and grown under non-limiting conditions of water and plant nutrient supplies. Additional data for maize plants with low total leaf number (12-17) grown at Katumani Research Centre, Kenya, were included to extend the range in the total leaf number per plant. The equation for the modified (slightly skewed) bell curve could be generalised for modelling individual leaf area, as all coefficients in it were related to total leaf number. Use of coefficients for individual genotypes can be avoided, and individual and total plant leaf area can be calculated from total leaf number. A single, logistic equation, relying on maximum plant leaf area and thermal time from emergence, was developed to predict leaf senescence. The base, optimum, and maximum temperatures for calculation of thermal time for leaf senescence were 8, 34, and 40 degrees C, and apply for the whole crop-cycle when used in modelling of leaf senescence. Thus, the modelling of leaf production and senescence is simplified, improved, and generalised. Consequently, the modelling of leaf area index (LAI) and variables that rely on LAI will be improved. For experimental purposes, we found that the calculation of leaf area from leaf length and leaf width remains appropriate, though the relationship differed slightly from previously published equations.
Resumo:
A new method of poly-beta-hydroxybutyrate (PHB) extraction from recombinant E. coli is proposed, using homogenization and centrifugation coupled with sodium hypochlorite treatment. The size of PHB granules and cell debris in homogenates was characterised as a function of the number of homogenization passes. Simulation was used to develop the PHB and cell debris fractionation system, enabling numerical examination of the effects of repeated homogenization and centrifuge-feedrate variation. The simulation provided a good prediction of experimental performance. Sodium hypochlorite treatment was necessary to optimise PHB fractionation. A PHB recovery of 80% at a purity of 96.5% was obtained with the final optimised process. Protein and DNA contained in the resultant product were negligible. The developed process holds promise for significantly reducing the recovery cost associated with PHB manufacture.
Resumo:
The moving finite element collocation method proposed by Kill et al. (1995) Chem. Engng Sci. 51 (4), 2793-2799 for solution of problems with steep gradients is further developed to solve transient problems arising in the field of adsorption. The technique is applied to a model of adsorption in solids with bidisperse pore structures. Numerical solutions were found to match the analytical solution when it exists (i.e. when the adsorption isotherm is linear). The method is simple yet sufficiently accurate for use in adsorption problems, where global collocation methods fail. (C) 1998 Elsevier Science Ltd. All rights reserved.
Resumo:
Conotoxins are valuable probes of receptors and ion channels because of their small size and highly selective activity. alpha-Conotoxin EpI, a 16-residue peptide from the mollusk-hunting Conus episcopatus, has the amino acid sequence GCCSDPRCNMNNPDY(SO3H)C-NH2 and appears to be an extremely potent and selective inhibitor of the alpha 3 beta 2 and alpha 3 beta 4 neuronal subtypes of the nicotinic acetylcholine receptor (nAChR). The desulfated form of EpI ([Tyr(15)]EpI) has a potency and selectivity for the nAChR receptor similar to those of EpI. Here we describe the crystal structure of [Tyr(15)]EpI solved at a resolution of 1.1 Angstrom using SnB. The asymmetric unit has a total of 284 non-hydrogen atoms, making this one of the largest structures solved de novo try direct methods. The [Tyr(15)]EpI structure brings to six the number of alpha-conotoxin structures that have been determined to date. Four of these, [Tyr(15)]EpI, PnIA, PnIB, and MII, have an alpha 4/7 cysteine framework and are selective for the neuronal subtype of the nAChR. The structure of [Tyr(15)]EpI has the same backbone fold as the other alpha 4/7-conotoxin structures, supporting the notion that this conotoxin cysteine framework and spacing give rise to a conserved fold. The surface charge distribution of [Tyr(15)]EpI is similar to that of PnIA and PnIB but is likely to be different from that of MII, suggesting that [Tyr(15)]EpI and MII may have different binding modes for the same receptor subtype.
Resumo:
Expokit provides a set of routines aimed at computing matrix exponentials. More precisely, it computes either a small matrix exponential in full, the action of a large sparse matrix exponential on an operand vector, or the solution of a system of linear ODEs with constant inhomogeneity. The backbone of the sparse routines consists of matrix-free Krylov subspace projection methods (Arnoldi and Lanczos processes), and that is why the toolkit is capable of coping with sparse matrices of large dimension. The software handles real and complex matrices and provides specific routines for symmetric and Hermitian matrices. The computation of matrix exponentials is a numerical issue of critical importance in the area of Markov chains and furthermore, the computed solution is subject to probabilistic constraints. In addition to addressing general matrix exponentials, a distinct attention is assigned to the computation of transient states of Markov chains.
Resumo:
This paper investigates the effective diagnostic technique(s) for assessing the condition of insulation in aged power transformers. A number of electrical, mechanical and chemical techniques were investigated. Many of these techniques are already used by the utility engineers and two comparatively new techniques are proposed in this paper. Results showing the effectiveness of these diagnostics are presented and correlation between the techniques are also presented. Finally, merits and suitability of different techniques are discussed in this paper.
Resumo:
A version of the Agricultural Production Systems Simulator (APSIM) capable of simulating the key agronomic aspects of intercropping maize between legume shrub hedgerows was described and parameterised in the first paper of this series (Nelson et al., this issue). In this paper, APSIM is used to simulate maize yields and soil erosion from traditional open-field farming and hedgerow intercropping in the Philippine uplands. Two variants of open-field farming were simulated using APSIM, continuous and fallow, for comparison with intercropping maize between leguminous shrub hedgerows. Continuous open-field maize farming was predicted to be unsustainable in the long term, while fallow open-field farming was predicted to slow productivity decline by spreading the effect of erosion over a larger cropping area. Hedgerow intercropping was predicted to reduce erosion by maintaining soil surface cover during periods of intense rainfall, contributing to sustainable production of maize in the long term. In the third paper in this series, Nelson et al. (this issue) use cost-benefit analysis to compare the economic viability of hedgerow intercropping relative to traditional open-field farming of maize in relatively inaccessible upland areas. (C) 1998 Elsevier Science Ltd. All rights reserved.
Resumo:
Two previous papers in this series (Nelson et al., this issue) described the use of the Agricultural Production Systems Simulator (APSIM) to simulate the effect of erosion on maize yields from open-field farming and hedgerow intercropping in the Philippine uplands. In this paper, maize yields simulated with APSIM are used to compare the economic viability of intercropping maize between leguminous shrub hedgerows with that of continuous and fallow open-field farming of maize. The analysis focuses on the economic incentives of upland farmers to adopt hedgerow intercropping, discussing farmers' planning horizons, access to credit and security of land tenure, as well as maize pricing in the Philippines. Insecure land tenure has limited the planning horizons of upland farmers, and high establishment costs reduce the economic viability of hedgerow intercropping relative to continuous and fallow open-field farming in the short term, In the long term, high discount rates and share-tenancy arrangements in which landlords do not contribute to establishment costs reduce the economic viability of hedgerow intercropping relative to fallow open-field farming, (C) 1998 Elsevier Science Ltd. All rights reserved.
Resumo:
We introduce the study of dynamical quantum noise in Bose-Einstein condensates through numerical simulation of stochastic partial differential equations obtained using phase-space representations. We derive evolution equations for a single trapped condensate in both the positive-P and Wigner representations and perform simulations to compare the predictions of the two methods. The positive-P approach is found to be highly susceptible to the stability problems that have been observed in other strongly nonlinear, weakly damped systems. Using the Wigner representation, we examine the evolution of several quantities of interest using from a variety of choices of initial stare for the condensate and compare results to those for single-mode models. [S1050-2947(98)06612-8].
Resumo:
The use of computational fluid dynamics simulations for calibrating a flush air data system is described, In particular, the flush air data system of the HYFLEX hypersonic vehicle is used as a case study. The HYFLEX air data system consists of nine pressure ports located flush with the vehicle nose surface, connected to onboard pressure transducers, After appropriate processing, surface pressure measurements can he converted into useful air data parameters. The processing algorithm requires an accurate pressure model, which relates air data parameters to the measured pressures. In the past, such pressure models have been calibrated using combinations of flight data, ground-based experimental results, and numerical simulation. We perform a calibration of the HYFLEX flush air data system using computational fluid dynamics simulations exclusively, The simulations are used to build an empirical pressure model that accurately describes the HYFLEX nose pressure distribution ol cr a range of flight conditions. We believe that computational fluid dynamics provides a quick and inexpensive way to calibrate the air data system and is applicable to a broad range of flight conditions, When tested with HYFLEX flight data, the calibrated system is found to work well. It predicts vehicle angle of attack and angle of sideslip to accuracy levels that generally satisfy flight control requirements. Dynamic pressure is predicted to within the resolution of the onboard inertial measurement unit. We find that wind-tunnel experiments and flight data are not necessary to accurately calibrate the HYFLEX flush air data system for hypersonic flight.