860 resultados para conceptual data modelling
Resumo:
Doutoramento em Engenharia Florestal e dos Recursos Naturais - Instituto Superior de Agronomia - UL
Resumo:
Predicting the evolution of a coastal cell requires the identification of the key drivers of morphology. Soft coastlines are naturally dynamic but severe storm events and even human intervention can accelerate any changes that are occurring. However, when erosive events such as barrier breaching occur with no obvious contributory factors, a deeper understanding of the underlying coastal processes is required. Ideally conclusions on morphological drivers should be drawn from field data collection and remote sensing over a long period of time. Unfortunately, when the Rossbeigh barrier beach in Dingle Bay, County Kerry, began to erode rapidly in the early 2000’s, eventually leading to it breaching in 2008, no such baseline data existed. This thesis presents a study of the morphodynamic evolution of the Inner Dingle Bay coastal system. The study combines existing coastal zone analysis approaches with experimental field data collection techniques and a novel approach to long term morphodynamic modelling to predict the evolution of the barrier beach inlet system. A conceptual model describing the long term evolution of Inner Dingle Bay in 5 stages post breaching was developed. The dominant coastal processes driving the evolution of the coastal system were identified and quantified. A new methodology of long term process based numerical modelling approach to coastal evolution was developed. This method was used to predict over 20 years of coastal evolution in Inner Dingle Bay. On a broader context this thesis utilised several experimental coastal zone data collection and analysis methods such as ocean radar and grain size trend analysis. These were applied during the study and their suitability to a dynamic coastal system was assessed.
Resumo:
The phosphatidylinositide 3-kinases (PI3K) and mammalian target of rapamycin-1 (mTOR1) are two key targets for anti-cancer therapy. Predicting the response of the PI3K/AKT/mTOR1 signalling pathway to targeted therapy is made difficult because of network complexities. Systems biology models can help explore those complexities but the value of such models is dependent on accurate parameterisation. Motivated by a need to increase accuracy in kinetic parameter estimation, and therefore the predictive power of the model, we present a framework to integrate kinetic data from enzyme assays into a unified enzyme kinetic model. We present exemplar kinetic models of PI3K and mTOR1, calibrated on in vitro enzyme data and founded on Michaelis-Menten (MM) approximation. We describe the effects of an allosteric mTOR1 inhibitor (Rapamycin) and ATP-competitive inhibitors (BEZ2235 and LY294002) that show dual inhibition of mTOR1 and PI3K. We also model the kinetics of phosphatase and tensin homolog (PTEN), which modulates sensitivity of the PI3K/AKT/mTOR1 pathway to these drugs. Model validation with independent data sets allows investigation of enzyme function and drug dose dependencies in a wide range of experimental conditions. Modelling of the mTOR1 kinetics showed that Rapamycin has an IC50 independent of ATP concentration and that it is a selective inhibitor of mTOR1 substrates S6K1 and 4EBP1: it retains 40% of mTOR1 activity relative to 4EBP1 phosphorylation and inhibits completely S6K1 activity. For the dual ATP-competitive inhibitors of mTOR1 and PI3K, LY294002 and BEZ235, we derived the dependence of the IC50 on ATP concentration that allows prediction of the IC50 at different ATP concentrations in enzyme and cellular assays. Comparison of the drug effectiveness in enzyme and cellular assays showed that some features of these drugs arise from signalling modulation beyond the on-target action and MM approximation and require a systems-level consideration of the whole PI3K/PTEN/AKT/mTOR1 network in order to understand mechanisms of drug sensitivity and resistance in different cancer cell lines. We suggest that using these models in systems biology investigation of the PI3K/AKT/mTOR1 signalling in cancer cells can bridge the gap between direct drug target action and the therapeutic response to these drugs and their combinations.
Resumo:
The fisheries for mackerel scad, Decapterus macarellus, are particularly important in Cape Verde, constituting almost 40% of total catches at the peak of the fishery in 1997 and 1998 ( 3700 tonnes). Catches have been stable at a much lower level of about 2 100 tonnes in recent years. Given the importance of mackerel scad in terms of catch weight and local food security, there is an urgent need for updated assessment. Stock assessment was carried out using a Bayesian approach to biomass dynamic modelling. In order to tackle the problem of a non-informative CPUE series, the intrinsic rate of increase, r, was estimated separately, and the ratio B-0/X, initial biomass relative to carrying capacity, was assumed based on available information. The results indicated that the current level of fishing is sustainable. The probability of collapse is low, particularly in the short-term, and it is likely that biomass may increase further above B-msy, indicating a healthy stock level. It would appear that it is relatively safe to increase catches even up to 4000 tonnes. However, the marginal posterior of r was almost identical to the prior, indicating that there is relatively low information content in CPUE. This was also the case in relation to B-0/X There have been substantial increases in fishing efficiency, which have not been adequately captured by the measure used for effort (days or trips), implying that the results may be overly optimistic and should be considered preliminary. (c) 2006 Elsevier B.V. All rights reserved.
Resumo:
There are many natural events that can negatively affect the urban ecosystem, but weather-climate variations are certainly among the most significant. The history of settlements has been characterized by extreme events like earthquakes and floods, which repeat themselves at different times, causing extensive damage to the built heritage on a structural and urban scale. Changes in climate also alter various climatic subsystems, changing rainfall regimes and hydrological cycles, increasing the frequency and intensity of extreme precipitation events (heavy rainfall). From an hydrological risk perspective, it is crucial to understand future events that could occur and their magnitude in order to design safer infrastructures. Unfortunately, it is not easy to understand future scenarios as the complexity of climate is enormous. For this thesis, precipitation and discharge extremes were primarily used as data sources. It is important to underline that the two data sets are not separated: changes in rainfall regime, due to climate change, could significantly affect overflows into receiving water bodies. It is imperative that we understand and model climate change effects on water structures to support the development of adaptation strategies. The main purpose of this thesis is to search for suitable water structures for a road located along the Tione River. Therefore, through the analysis of the area from a hydrological point of view, we aim to guarantee the safety of the infrastructure over time. The observations made have the purpose to underline how models such as a stochastic one can improve the quality of an analysis for design purposes, and influence choices.
Resumo:
Geographic Data Warehouses (GDW) are one of the main technologies used in decision-making processes and spatial analysis, and the literature proposes several conceptual and logical data models for GDW. However, little effort has been focused on studying how spatial data redundancy affects SOLAP (Spatial On-Line Analytical Processing) query performance over GDW. In this paper, we investigate this issue. Firstly, we compare redundant and non-redundant GDW schemas and conclude that redundancy is related to high performance losses. We also analyze the issue of indexing, aiming at improving SOLAP query performance on a redundant GDW. Comparisons of the SB-index approach, the star-join aided by R-tree and the star-join aided by GiST indicate that the SB-index significantly improves the elapsed time in query processing from 25% up to 99% with regard to SOLAP queries defined over the spatial predicates of intersection, enclosure and containment and applied to roll-up and drill-down operations. We also investigate the impact of the increase in data volume on the performance. The increase did not impair the performance of the SB-index, which highly improved the elapsed time in query processing. Performance tests also show that the SB-index is far more compact than the star-join, requiring only a small fraction of at most 0.20% of the volume. Moreover, we propose a specific enhancement of the SB-index to deal with spatial data redundancy. This enhancement improved performance from 80 to 91% for redundant GDW schemas.
Resumo:
Context. Be stars undergo outbursts producing a circumstellar disk from the ejected material. The beating of non-radial pulsations has been put forward as a possible mechanism of ejection. Aims. We analyze the pulsational behavior of the early B0.5IVe star HD 49330 observed during the first CoRoT long run towards the Galactical anticenter (LRA1). This Be star is located close to the lower edge of the beta Cephei instability strip in the HR diagram and showed a 0.03 mag outburst during the CoRoT observations. It is thus an ideal case for testing the aforementioned hypothesis. Methods. We analyze the CoRoT light curve of HD 49330 using Fourier methods and non-linear least square fitting. Results. In this star, we find pulsation modes typical of beta Cep stars (p modes) and SPB stars (g modes) with amplitude variations along the run directly correlated with the outburst. These results provide new clues about the origin of the Be phenomenon as well as strong constraints on the seismic modelling of Be stars.
Resumo:
Ecological niche modelling combines species occurrence points with environmental raster layers in order to obtain models for describing the probabilistic distribution of species. The process to generate an ecological niche model is complex. It requires dealing with a large amount of data, use of different software packages for data conversion, for model generation and for different types of processing and analyses, among other functionalities. A software platform that integrates all requirements under a single and seamless interface would be very helpful for users. Furthermore, since biodiversity modelling is constantly evolving, new requirements are constantly being added in terms of functions, algorithms and data formats. This evolution must be accompanied by any software intended to be used in this area. In this scenario, a Service-Oriented Architecture (SOA) is an appropriate choice for designing such systems. According to SOA best practices and methodologies, the design of a reference business process must be performed prior to the architecture definition. The purpose is to understand the complexities of the process (business process in this context refers to the ecological niche modelling problem) and to design an architecture able to offer a comprehensive solution, called a reference architecture, that can be further detailed when implementing specific systems. This paper presents a reference business process for ecological niche modelling, as part of a major work focused on the definition of a reference architecture based on SOA concepts that will be used to evolve the openModeller software package for species modelling. The basic steps that are performed while developing a model are described, highlighting important aspects, based on the knowledge of modelling experts. In order to illustrate the steps defined for the process, an experiment was developed, modelling the distribution of Ouratea spectabilis (Mart.) Engl. (Ochnaceae) using openModeller. As a consequence of the knowledge gained with this work, many desirable improvements on the modelling software packages have been identified and are presented. Also, a discussion on the potential for large-scale experimentation in ecological niche modelling is provided, highlighting opportunities for research. The results obtained are very important for those involved in the development of modelling tools and systems, for requirement analysis and to provide insight on new features and trends for this category of systems. They can also be very helpful for beginners in modelling research, who can use the process and the experiment example as a guide to this complex activity. (c) 2008 Elsevier B.V. All rights reserved.
Resumo:
The thermodynamic assessment of an Al(2)O(3)-MnO pseudo-binary system has been carried out with the use of an ionic model. The use of the electro-neutrality principles in addition to the constitutive relations, between site fractions of the species on each sub-lattice, the thermodynamics descriptions of each solid phase has been determined to make possible the solubility description. Based on the thermodynamics descriptions of each phase in addition to thermo-chemical data obtained from the literature, the Gibbs energy functions were optimized for each phase of the Al(2)O(3)-MnO system with the support of PARROT(R) module from ThemoCalc(R) package. A thermodynamic database was obtained, in agreement with the thermo-chemical data extracted from the literature, to describe the Al(2)O(3)-MnO system including the solubility description of solid phases. (C) 2009 Elsevier Ltd. All rights reserved.
Resumo:
A four-parameter extension of the generalized gamma distribution capable of modelling a bathtub-shaped hazard rate function is defined and studied. The beauty and importance of this distribution lies in its ability to model monotone and non-monotone failure rate functions, which are quite common in lifetime data analysis and reliability. The new distribution has a number of well-known lifetime special sub-models, such as the exponentiated Weibull, exponentiated generalized half-normal, exponentiated gamma and generalized Rayleigh, among others. We derive two infinite sum representations for its moments. We calculate the density of the order statistics and two expansions for their moments. The method of maximum likelihood is used for estimating the model parameters and the observed information matrix is obtained. Finally, a real data set from the medical area is analysed.
Resumo:
This article considers alternative methods to calculate the fair premium rate of crop insurance contracts based on county yields. The premium rate was calculated using parametric and nonparametric approaches to estimate the conditional agricultural yield density. These methods were applied to a data set of county yield provided by the Statistical and Geography Brazilian Institute (IBGE), for the period of 1990 through 2002, for soybean, corn and wheat, in the State of Paran. In this article, we propose methodological alternatives to pricing crop insurance contracts resulting in more accurate premium rates in a situation of limited data.
Resumo:
A large number of models have been derived from the two-parameter Weibull distribution and are referred to as Weibull models. They exhibit a wide range of shapes for the density and hazard functions, which makes them suitable for modelling complex failure data sets. The WPP and IWPP plot allows one to determine in a systematic manner if one or more of these models are suitable for modelling a given data set. This paper deals with this topic.
Resumo:
Data mining is the process to identify valid, implicit, previously unknown, potentially useful and understandable information from large databases. It is an important step in the process of knowledge discovery in databases, (Olaru & Wehenkel, 1999). In a data mining process, input data can be structured, seme-structured, or unstructured. Data can be in text, categorical or numerical values. One of the important characteristics of data mining is its ability to deal data with large volume, distributed, time variant, noisy, and high dimensionality. A large number of data mining algorithms have been developed for different applications. For example, association rules mining can be useful for market basket problems, clustering algorithms can be used to discover trends in unsupervised learning problems, classification algorithms can be applied in decision-making problems, and sequential and time series mining algorithms can be used in predicting events, fault detection, and other supervised learning problems (Vapnik, 1999). Classification is among the most important tasks in the data mining, particularly for data mining applications into engineering fields. Together with regression, classification is mainly for predictive modelling. So far, there have been a number of classification algorithms in practice. According to (Sebastiani, 2002), the main classification algorithms can be categorized as: decision tree and rule based approach such as C4.5 (Quinlan, 1996); probability methods such as Bayesian classifier (Lewis, 1998); on-line methods such as Winnow (Littlestone, 1988) and CVFDT (Hulten 2001), neural networks methods (Rumelhart, Hinton & Wiliams, 1986); example-based methods such as k-nearest neighbors (Duda & Hart, 1973), and SVM (Cortes & Vapnik, 1995). Other important techniques for classification tasks include Associative Classification (Liu et al, 1998) and Ensemble Classification (Tumer, 1996).
Resumo:
The performance of three analytical methods for multiple-frequency bioelectrical impedance analysis (MFBIA) data was assessed. The methods were the established method of Cole and Cole, the newly proposed method of Siconolfi and co-workers and a modification of this procedure. Method performance was assessed from the adequacy of the curve fitting techniques, as judged by the correlation coefficient and standard error of the estimate, and the accuracy of the different methods in determining the theoretical values of impedance parameters describing a set of model electrical circuits. The experimental data were well fitted by all curve-fitting procedures (r = 0.9 with SEE 0.3 to 3.5% or better for most circuit-procedure combinations). Cole-Cole modelling provided the most accurate estimates of circuit impedance values, generally within 1-2% of the theoretical values, followed by the Siconolfi procedure using a sixth-order polynomial regression (1-6% variation). None of the methods, however, accurately estimated circuit parameters when the measured impedances were low (<20 Omega) reflecting the electronic limits of the impedance meter used. These data suggest that Cole-Cole modelling remains the preferred method for the analysis of MFBIA data.
Resumo:
Normal mixture models are being increasingly used to model the distributions of a wide variety of random phenomena and to cluster sets of continuous multivariate data. However, for a set of data containing a group or groups of observations with longer than normal tails or atypical observations, the use of normal components may unduly affect the fit of the mixture model. In this paper, we consider a more robust approach by modelling the data by a mixture of t distributions. The use of the ECM algorithm to fit this t mixture model is described and examples of its use are given in the context of clustering multivariate data in the presence of atypical observations in the form of background noise.