971 resultados para Spatially-explicit models


Relevância:

80.00% 80.00%

Publicador:

Resumo:

Perturbation of natural ecosystems, namely by increasing freshwater use and its degradative use, as well as topsoil erosion by water of land-use production systems, have been emerging as topics of high environmental concern. Freshwater use has become a focus of attention in the last few years for all stakeholders involved in the production of goods, mainly agro-industrial and forest-based products, which are freshwater-intensive consumers, requiring large inputs of green and blue water. This thesis presents a global review on the available Water Footprint Assessment and Life Cycle Assessment (LCA)-based methods for measuring and assessing the environmental relevance of freshwater resources use, based on a life cycle perspective. Using some of the available midpoint LCA-based methods, the freshwater use-related impacts of a Portuguese wine (white ‘vinho verde’) were assessed. However, the relevance of environmental green water has been neglected because of the absence of a comprehensive impact assessment method associated with green water flows. To overcome this constraint, this thesis helps to improve and enhance the LCA-based methods by providing a midpoint and spatially explicit Life Cycle Impact Assessment (LCIA) method for assessing impacts on terrestrial green water flow and addressing reductions in surface blue water production caused by reductions in surface runoff due to land-use production systems. The applicability of the proposed method is illustrated by a case study on Eucalyptus globulus conducted in Portugal, as the growth of short rotation forestry is largely dependent on local precipitation. Topsoil erosion by water has been characterised as one of the most upsetting problems for rivers. Because of this, this thesis also focuses on the ecosystem impacts caused by suspended solids (SS) from topsoil erosion that reach freshwater systems. A framework to conduct a spatially distributed SS delivery to freshwater streams and a fate and effect LCIA method to derive site-specific characterisation factors (CFs) for endpoint damage on aquatic ecosystem diversity, namely on algae, macrophyte, and macroinvertebrates organisms, were developed. The applicability of this framework, combined with the derived site-specific CFs, is shown by conducting a case study on E. globulus stands located in Portugal as an example of a land use based system. A spatially explicit LCA assessment was shown to be necessary, since the impacts associated with both green water flows and SS vary greatly as a function of spatial location.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Well-designed marine protected area (MPA) networks can deliver a range of ecological, economic and social benefits, and so a great deal of research has focused on developing spatial conservation prioritization tools to help identify important areas. However, whilst these software tools are designed to identify MPA networks that both represent biodiversity and minimize impacts on stakeholders, they do not consider complex ecological processes. Thus, it is difficult to determine the impacts that proposed MPAs could have on marine ecosystem health, fisheries and fisheries sustainability. Using the eastern English Channel as a case study, this paper explores an approach to address these issues by identifying a series of MPA networks using the Marxan and Marxan with Zones conservation planning software and linking them with a spatially explicit ecosystem model developed in Ecopath with Ecosim. We then use these to investigate potential trade-offs associated with adopting different MPA management strategies. Limited-take MPAs, which restrict the use of some fishing gears, could have positive benefits for conservation and fisheries in the eastern English Channel, even though they generally receive far less attention in research on MPA network design. Our findings, however, also clearly indicate that no-take MPAs should form an integral component of proposed MPA networks in the eastern English Channel, as they not only result in substantial increases in ecosystem biomass, fisheries catches and the biomass of commercially valuable target species, but are fundamental to maintaining the sustainability of the fisheries. Synthesis and applications. Using the existing software tools Marxan with Zones and Ecopath with Ecosim in combination provides a powerful policy-screening approach. This could help inform marine spatial planning by identifying potential conflicts and by designing new regulations that better balance conservation objectives and stakeholder interests. In addition, it highlights that appropriate combinations of no-take and limited-take marine protected areas might be the most effective when making trade-offs between long-term ecological benefits and short-term political acceptability.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

A spatially explicit coupled hydrodynamic-biogeochemical model was developed to study a coastal ecosystem under the combined effects of mussel aquaculture, nutrient loading and climate change. The model was applied to St Peter's Bay (SPB), Prince Edward Island, Eastern Canada. Approximately 40 % of the SPB area is dedicated to mussel (Mytilus edulis) longline culture. Results indicate that the two main food sources for mussels, phytoplankton and organic detritus, are most depleted in the central part of the embayment. Results also suggest that the system is near its ultimate capacity, a state where the energy cycle is restricted to nitrogen-phytoplankton-detritus-mussels with few resources left to be transferred to higher trophic levels. Annually, mussel meat harvesting extracts nitrogen (N) resources equivalent to 42 % of river inputs or 46.5 % of the net phytoplankton primary production. Under such extractive pressure, the phytoplankton biomass is being curtailed to 1980's levels when aquaculture was not yet developed and N loading was half the present level. Current mussel stocks also decrease bay-scale sedimentation rates by 14 %. Finally, a climate change scenario (year 2050) predicted a 30 % increase in mussel production, largely driven by more efficient utilization of the phytoplankton spring bloom. However, the predicted elevated summer temperatures (> 25 A degrees C) may also have deleterious physiological effects on mussels and possibly increase summer mortality levels. In conclusion, cultivated bivalves may play an important role in remediating the negative impacts of land-derived nutrient loading. Climate change may lead to increases in production and ecological carrying capacity as long as the cultivated species can tolerate warmer summer conditions.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Forests have a prominent role in carbon storage and sequestration. Anthropogenic forcing has the potential to accelerate climate change and alter the distribution of forests. How forests redistribute spatially and temporally in response to climate change can alter their carbon sequestration potential. The driving question for this research was: How does plant migration from climate change impact vegetation distribution and carbon sequestration potential over continental scales? Large-scale simulation of the equilibrium response of vegetation and carbon from future climate change has shown relatively modest net gains in sequestration potential, but studies of the transient response has been limited to the sub-continent or landscape scale. The transient response depends on fine scale processes such as competition, disturbance, landscape characteristics, dispersal, and other factors, which makes it computational prohibitive at large domain sizes. To address this, this research used an advanced mechanistic model (Ecosystem Demography Model, ED) that is individually based, but pseudo-spatial, that reduces computational intensity while maintaining the fine scale processes that drive the transient response. First, the model was validated against remote sensing data for current plant functional type distribution in northern North America with a current climatology, and then a future climatology was used to predict the potential equilibrium redistribution of vegetation and carbon from future climate change. Next, to enable transient calculations, a method was developed to simulate the spatially explicit process of dispersal in pseudo-spatial modeling frameworks. Finally, the new dispersal sub-model was implemented in the mechanistic ecosystem model, and a model experimental design was designed and completed to estimate the transient response of vegetation and carbon to climate change. The potential equilibrium forest response to future climate change was found to be large, with large gross changes in distribution of plant functional types and comparatively smaller changes in net carbon sequestration potential for the region. However, the transient response was found to be on the order of centuries, and to depend strongly on disturbance rates and dispersal distances. Future work should explore the impact of species-specific disturbance and dispersal rates, landscape fragmentation, and other processes that influence migration rates and have been simulated at the sub-continent scale, but now at continental scales, and explore a range of alternative future climate scenarios as they continue to be developed.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Live animal trade is considered a major mode of introduction of viruses from enzootic foci into disease-free areas. Due to societal and behavioural changes, some wild animal species may nowadays be considered as pet species. The species diversity of animals involved in international trade is thus increasing. This could benefit pathogens that have a broad host range such as arboviruses. The objective of this study was to analyze the risk posed by live animal imports for the introduction, in the European Union (EU), of four arboviruses that affect human and horses: Eastern and Western equine encephalomyelitis, Venezuelan equine encephalitis and Japanese encephalitis. Importation data for a five-years period (2005-2009, extracted from the EU TRACES database), environmental data (used as a proxy for the presence of vectors) and horses and human population density data (impacting the occurrence of clinical cases) were combined to derive spatially explicit risk indicators for virus introduction and for the potential consequences of such introductions. Results showed the existence of hotspots where the introduction risk was the highest in Belgium, in the Netherlands and in the north of Italy. This risk was higher for Eastern equine encephalomyelitis (EEE) than for the three other diseases. It was mainly attributed to exotic pet species such as rodents, reptiles or cage birds, imported in small-sized containments from a wide variety of geographic origins. The increasing species and origin diversity of these animals may have in the future a strong impact on the risk of introduction of arboviruses in the EU.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

In the protein folding problem, solvent-mediated forces are commonly represented by intra-chain pairwise contact energy. Although this approximation has proven to be useful in several circumstances, it is limited in some other aspects of the problem. Here we show that it is possible to achieve two models to represent the chain-solvent system. one of them with implicit and other with explicit solvent, such that both reproduce the same thermodynamic results. Firstly, lattice models treated by analytical methods, were used to show that the implicit and explicitly representation of solvent effects can be energetically equivalent only if local solvent properties are time and spatially invariant. Following, applying the same reasoning Used for the lattice models, two inter-consistent Monte Carlo off-lattice models for implicit and explicit solvent are constructed, being that now in the latter the solvent properties are allowed to fluctuate. Then, it is shown that the chain configurational evolution as well as the globule equilibrium conformation are significantly distinct for implicit and explicit solvent systems. Actually, strongly contrasting with the implicit solvent version, the explicit solvent model predicts: (i) a malleable globule, in agreement with the estimated large protein-volume fluctuations; (ii) thermal conformational stability, resembling the conformational hear resistance of globular proteins, in which radii of gyration are practically insensitive to thermal effects over a relatively wide range of temperatures; and (iii) smaller radii of gyration at higher temperatures, indicating that the chain conformational entropy in the unfolded state is significantly smaller than that estimated from random coil configurations. Finally, we comment on the meaning of these results with respect to the understanding of the folding process. (C) 2009 Elsevier B.V. All rights reserved.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

We show that a simple mixing idea allows one to establish a number of explicit formulas for ruin probabilities and related quantities in collective risk models with dependence among claim sizes and among claim inter-occurrence times. Examples include compound Poisson risk models with completely monotone marginal claim size distributions that are dependent according to Archimedean survival copulas as well as renewal risk models with dependent inter-occurrence times.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Application of semi-distributed hydrological models to large, heterogeneous watersheds deals with several problems. On one hand, the spatial and temporal variability in catchment features should be adequately represented in the model parameterization, while maintaining the model complexity in an acceptable level to take advantage of state-of-the-art calibration techniques. On the other hand, model complexity enhances uncertainty in adjusted model parameter values, therefore increasing uncertainty in the water routing across the watershed. This is critical for water quality applications, where not only streamflow, but also a reliable estimation of the surface versus subsurface contributions to the runoff is needed. In this study, we show how a regularized inversion procedure combined with a multiobjective function calibration strategy successfully solves the parameterization of a complex application of a water quality-oriented hydrological model. The final value of several optimized parameters showed significant and consistentdifferences across geological and landscape features. Although the number of optimized parameters was significantly increased by the spatial and temporal discretization of adjustable parameters, the uncertainty in water routing results remained at reasonable values. In addition, a stepwise numerical analysis showed that the effects on calibration performance due to inclusion of different data types in the objective function could be inextricably linked. Thus caution should be taken when adding or removing data from an aggregated objective function.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Several previous studies have attempted to assess the sublimation depth-scales of ice particles from clouds into clear air. Upon examining the sublimation depth-scales in the Met Office Unified Model (MetUM), it was found that the MetUM has evaporation depth-scales 2–3 times larger than radar observations. Similar results can be seen in the European Centre for Medium-Range Weather Forecasts (ECMWF), Regional Atmospheric Climate Model (RACMO) and Météo-France models. In this study, we use radar simulation (converting model variables into radar observations) and one-dimensional explicit microphysics numerical modelling to test and diagnose the cause of the deep sublimation depth-scales in the forecast model. The MetUM data and parametrization scheme are used to predict terminal velocity, which can be compared with the observed Doppler velocity. This can then be used to test the hypothesis as to why the sublimation depth-scale is too large within the MetUM. Turbulence could lead to dry air entrainment and higher evaporation rates; particle density may be wrong, particle capacitance may be too high and lead to incorrect evaporation rates or the humidity within the sublimating layer may be incorrectly represented. We show that the most likely cause of deep sublimation zones is an incorrect representation of model humidity in the layer. This is tested further by using a one-dimensional explicit microphysics model, which tests the sensitivity of ice sublimation to key atmospheric variables and is capable of including sonde and radar measurements to simulate real cases. Results suggest that the MetUM grid resolution at ice cloud altitudes is not sufficient enough to maintain the sharp drop in humidity that is observed in the sublimation zone.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

With the prospect of exascale computing, computational methods requiring only local data become especially attractive. Consequently, the typical domain decomposition of atmospheric models means horizontally-explicit vertically-implicit (HEVI) time-stepping schemes warrant further attention. In this analysis, Runge-Kutta implicit-explicit schemes from the literature are analysed for their stability and accuracy using a von Neumann stability analysis of two linear systems. Attention is paid to the numerical phase to indicate the behaviour of phase and group velocities. Where the analysis is tractable, analytically derived expressions are considered. For more complicated cases, amplification factors have been numerically generated and the associated amplitudes and phase diagnosed. Analysis of a system describing acoustic waves has necessitated attributing the three resultant eigenvalues to the three physical modes of the system. To do so, a series of algorithms has been devised to track the eigenvalues across the frequency space. The result enables analysis of whether the schemes exactly preserve the non-divergent mode; and whether there is evidence of spurious reversal in the direction of group velocities or asymmetry in the damping for the pair of acoustic modes. Frequency ranges that span next-generation high-resolution weather models to coarse-resolution climate models are considered; and a comparison is made of errors accumulated from multiple stability-constrained shorter time-steps from the HEVI scheme with a single integration from a fully implicit scheme over the same time interval. Two schemes, “Trap2(2,3,2)” and “UJ3(1,3,2)”, both already used in atmospheric models, are identified as offering consistently good stability and representation of phase across all the analyses. Furthermore, according to a simple measure of computational cost, “Trap2(2,3,2)” is the least expensive.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Generalized linear mixed models (GLMMs) provide an elegant framework for the analysis of correlated data. Due to the non-closed form of the likelihood, GLMMs are often fit by computational procedures like penalized quasi-likelihood (PQL). Special cases of these models are generalized linear models (GLMs), which are often fit using algorithms like iterative weighted least squares (IWLS). High computational costs and memory space constraints often make it difficult to apply these iterative procedures to data sets with very large number of cases. This paper proposes a computationally efficient strategy based on the Gauss-Seidel algorithm that iteratively fits sub-models of the GLMM to subsetted versions of the data. Additional gains in efficiency are achieved for Poisson models, commonly used in disease mapping problems, because of their special collapsibility property which allows data reduction through summaries. Convergence of the proposed iterative procedure is guaranteed for canonical link functions. The strategy is applied to investigate the relationship between ischemic heart disease, socioeconomic status and age/gender category in New South Wales, Australia, based on outcome data consisting of approximately 33 million records. A simulation study demonstrates the algorithm's reliability in analyzing a data set with 12 million records for a (non-collapsible) logistic regression model.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

There is an emerging interest in modeling spatially correlated survival data in biomedical and epidemiological studies. In this paper, we propose a new class of semiparametric normal transformation models for right censored spatially correlated survival data. This class of models assumes that survival outcomes marginally follow a Cox proportional hazard model with unspecified baseline hazard, and their joint distribution is obtained by transforming survival outcomes to normal random variables, whose joint distribution is assumed to be multivariate normal with a spatial correlation structure. A key feature of the class of semiparametric normal transformation models is that it provides a rich class of spatial survival models where regression coefficients have population average interpretation and the spatial dependence of survival times is conveniently modeled using the transformed variables by flexible normal random fields. We study the relationship of the spatial correlation structure of the transformed normal variables and the dependence measures of the original survival times. Direct nonparametric maximum likelihood estimation in such models is practically prohibited due to the high dimensional intractable integration of the likelihood function and the infinite dimensional nuisance baseline hazard parameter. We hence develop a class of spatial semiparametric estimating equations, which conveniently estimate the population-level regression coefficients and the dependence parameters simultaneously. We study the asymptotic properties of the proposed estimators, and show that they are consistent and asymptotically normal. The proposed method is illustrated with an analysis of data from the East Boston Ashma Study and its performance is evaluated using simulations.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

A novel GPU-based nonparametric moving object detection strategy for computer vision tools requiring real-time processing is proposed. An alternative and efficient Bayesian classifier to combine nonparametric background and foreground models allows increasing correct detections while avoiding false detections. Additionally, an efficient region of interest analysis significantly reduces the computational cost of the detections.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Existing theories of semantic cognition propose models of cognitive processing occurring in a conceptual space, where ‘meaning’ is derived from the spatial relationships between concepts’ mapped locations within the space. Information visualisation is a growing area of research within the field of information retrieval, and methods for presenting database contents visually in the form of spatial data management systems (SDMSs) are being developed. This thesis combined these two areas of research to investigate the benefits associated with employing spatial-semantic mapping (documents represented as objects in two- and three-dimensional virtual environments are proximally mapped dependent on the semantic similarity of their content) as a tool for improving retrieval performance and navigational efficiency when browsing for information within such systems. Positive effects associated with the quality of document mapping were observed; improved retrieval performance and browsing behaviour were witnessed when mapping was optimal. It was also shown using a third dimension for virtual environment (VE) presentation provides sufficient additional information regarding the semantic structure of the environment that performance is increased in comparison to using two-dimensions for mapping. A model that describes the relationship between retrieval performance and browsing behaviour was proposed on the basis of findings. Individual differences were not found to have any observable influence on retrieval performance or browsing behaviour when mapping quality was good. The findings from this work have implications for both cognitive modelling of semantic information, and for designing and testing information visualisation systems. These implications are discussed in the conclusions of this work.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Йордан Йорданов, Андрей Василев - В работата се изследват методи за решаването на задачи на оптималното управление в дискретно време с безкраен хоризонт и явни управления. Дадена е обосновка на една процедура за решаване на такива задачи, базирана на множители на Лагранж, коята често се употребява в икономическата литература. Извеждени са необходимите условия за оптималност на базата на уравнения на Белман и са приведени достатъчни условия за оптималност при допускания, които често се използват в икономиката.