945 resultados para Data modeling
Resumo:
Mine simulation depends on data that is both coherent and representative of the mining operation. This paper describes a methodology for modeling operational data which has been developed for mine simulation. The methodology has been applied to a case study of an open-pit mine, where the cycle times of the truck fleet have been modeled for mine simulation purposes. The results obtained have shown that once the operational data has been treated using the proposed methodology, the system variables have proven to be adherent to theoretical distributions. The research indicated the need jar tracking the origin of data inconsistencies through the development of a process to manage inconsistent data from the mining operation.
Resumo:
One of the most important advantages of database systems is that the underlying mathematics is rich enough to specify very complex operations with a small number of statements in the database language. This research covers an aspect of biological informatics that is the marriage of information technology and biology, involving the study of real-world phenomena using virtual plants derived from L-systems simulation. L-systems were introduced by Aristid Lindenmayer as a mathematical model of multicellular organisms. Not much consideration has been given to the problem of persistent storage for these simulations. Current procedures for querying data generated by L-systems for scientific experiments, simulations and measurements are also inadequate. To address these problems the research in this paper presents a generic process for data-modeling tools (L-DBM) between L-systems and database systems. This paper shows how L-system productions can be generically and automatically represented in database schemas and how a database can be populated from the L-system strings. This paper further describes the idea of pre-computing recursive structures in the data into derived attributes using compiler generation. A method to allow a correspondence between biologists' terms and compiler-generated terms in a biologist computing environment is supplied. Once the L-DBM gets any specific L-systems productions and its declarations, it can generate the specific schema for both simple correspondence terminology and also complex recursive structure data attributes and relationships.
Resumo:
INTRODUCTION: There are several risk scores for stratification of patients with ST-segment elevation myocardial infarction (STEMI), the most widely used of which are the TIMI and GRACE scores. However, these are complex and require several variables. The aim of this study was to obtain a reduced model with fewer variables and similar predictive and discriminative ability. METHODS: We studied 607 patients (age 62 years, SD=13; 76% male) who were admitted with STEMI and underwent successful primary angioplasty. Our endpoints were all-cause in-hospital and 30-day mortality. Considering all variables from the TIMI and GRACE risk scores, multivariate logistic regression models were fitted to the data to identify the variables that best predicted death. RESULTS: Compared to the TIMI score, the GRACE score had better predictive and discriminative performance for in-hospital mortality, with similar results for 30-day mortality. After data modeling, the variables with highest predictive ability were age, serum creatinine, heart failure and the occurrence of cardiac arrest. The new predictive model was compared with the GRACE risk score, after internal validation using 10-fold cross validation. A similar discriminative performance was obtained and some improvement was achieved in estimates of probabilities of death (increased for patients who died and decreased for those who did not). CONCLUSION: It is possible to simplify risk stratification scores for STEMI and primary angioplasty using only four variables (age, serum creatinine, heart failure and cardiac arrest). This simplified model maintained a good predictive and discriminative performance for short-term mortality.
Resumo:
Roughly fifteen years ago, the Church of Jesus Christ of Latter-day Saints published a new proposed standard file format. They call it GEDCOM. It was designed to allow different genealogy programs to exchange data.Five years later, in may 2000, appeared the GENTECH Data Modeling Project, with the support of the Federation of Genealogical Societies (FGS) and other American genealogical societies. They attempted to define a genealogical logic data model to facilitate data exchange between different genealogical programs. Although genealogists deal with an enormous variety of data sources, one of the central concepts of this data model was that all genealogical data could be broken down into a series of short, formal genealogical statements. It was something more versatile than only export/import data records on a predefined fields. This project was finally absorbed in 2004 by the National Genealogical Society (NGS).Despite being a genealogical reference in many applications, these models have serious drawbacks to adapt to different cultural and social environments. At the present time we have no formal proposal for a recognized standard to represent the family domain.Here we propose an alternative conceptual model, largely inherited from aforementioned models. The design is intended to overcome their limitations. However, its major innovation lies in applying the ontological paradigm when modeling statements and entities.
Resumo:
Diffusion MRI has evolved towards an important clinical diagnostic and research tool. Though clinical routine is using mainly diffusion weighted and tensor imaging approaches, Q-ball imaging and diffusion spectrum imaging techniques have become more widely available. They are frequently used in research-oriented investigations in particular those aiming at measuring brain network connectivity. In this work, we aim at assessing the dependency of connectivity measurements on various diffusion encoding schemes in combination with appropriate data modeling. We process and compare the structural connection matrices computed from several diffusion encoding schemes, including diffusion tensor imaging, q-ball imaging and high angular resolution schemes, such as diffusion spectrum imaging with a publically available processing pipeline for data reconstruction, tracking and visualization of diffusion MR imaging. The results indicate that the high angular resolution schemes maximize the number of obtained connections when applying identical processing strategies to the different diffusion schemes. Compared to the conventional diffusion tensor imaging, the added connectivity is mainly found for pathways in the 50-100mm range, corresponding to neighboring association fibers and long-range associative, striatal and commissural fiber pathways. The analysis of the major associative fiber tracts of the brain reveals striking differences between the applied diffusion schemes. More complex data modeling techniques (beyond tensor model) are recommended 1) if the tracts of interest run through large fiber crossings such as the centrum semi-ovale, or 2) if non-dominant fiber populations, e.g. the neighboring association fibers are the subject of investigation. An important finding of the study is that since the ground truth sensitivity and specificity is not known, the comparability between results arising from different strategies in data reconstruction and/or tracking becomes implausible to understand.
Resumo:
Data management consists of collecting, storing, and processing the data into the format which provides value-adding information for decision-making process. The development of data management has enabled of designing increasingly effective database management systems to support business needs. Therefore as well as advanced systems are designed for reporting purposes, also operational systems allow reporting and data analyzing. The used research method in the theory part is qualitative research and the research type in the empirical part is case study. Objective of this paper is to examine database management system requirements from reporting managements and data managements perspectives. In the theory part these requirements are identified and the appropriateness of the relational data model is evaluated. In addition key performance indicators applied to the operational monitoring of production are studied. The study has revealed that the appropriate operational key performance indicators of production takes into account time, quality, flexibility and cost aspects. Especially manufacturing efficiency has been highlighted. In this paper, reporting management is defined as a continuous monitoring of given performance measures. According to the literature review, the data management tool should cover performance, usability, reliability, scalability, and data privacy aspects in order to fulfill reporting managements demands. A framework is created for the system development phase based on requirements, and is used in the empirical part of the thesis where such a system is designed and created for reporting management purposes for a company which operates in the manufacturing industry. Relational data modeling and database architectures are utilized when the system is built for relational database platform.
Resumo:
This paper analyzes the measure of systemic importance ∆CoV aR proposed by Adrian and Brunnermeier (2009, 2010) within the context of a similar class of risk measures used in the risk management literature. In addition, we develop a series of testing procedures, based on ∆CoV aR, to identify and rank the systemically important institutions. We stress the importance of statistical testing in interpreting the measure of systemic importance. An empirical application illustrates the testing procedures, using equity data for three European banks.
Resumo:
A fundamental principle in practical nonlinear data modeling is the parsimonious principle of constructing the minimal model that explains the training data well. Leave-one-out (LOO) cross validation is often used to estimate generalization errors by choosing amongst different network architectures (M. Stone, "Cross validatory choice and assessment of statistical predictions", J. R. Stast. Soc., Ser. B, 36, pp. 117-147, 1974). Based upon the minimization of LOO criteria of either the mean squares of LOO errors or the LOO misclassification rate respectively, we present two backward elimination algorithms as model post-processing procedures for regression and classification problems. The proposed backward elimination procedures exploit an orthogonalization procedure to enable the orthogonality between the subspace as spanned by the pruned model and the deleted regressor. Subsequently, it is shown that the LOO criteria used in both algorithms can be calculated via some analytic recursive formula, as derived in this contribution, without actually splitting the estimation data set so as to reduce computational expense. Compared to most other model construction methods, the proposed algorithms are advantageous in several aspects; (i) There are no tuning parameters to be optimized through an extra validation data set; (ii) The procedure is fully automatic without an additional stopping criteria; and (iii) The model structure selection is directly based on model generalization performance. The illustrative examples on regression and classification are used to demonstrate that the proposed algorithms are viable post-processing methods to prune a model to gain extra sparsity and improved generalization.
Resumo:
Neste estudo são analisados, através de técnicas de dados em painel, os fatores determinantes dos níveis de ativos líquidos de empresas abertas do Brasil, Argentina, Chile, México e Peru no período de 1995 a 2009. O índice utilizado nas modelagens é denominado de ativo líquido (ou simplesmente caixa), o qual inclui os recursos disponíveis em caixa e as aplicações financeiras de curto prazo, divididos pelo total de ativos da firma. É possível identificar uma tendência crescente de acúmulo de ativos líquidos como proporção do total de ativos ao longo dos anos em praticamente todos os países. São encontradas evidências de que empresas com maiores oportunidades de crescimento, maior tamanho (medido pelo total de ativos), maior nível de pagamento de dividendos e maior nível de lucratividade, acumulam mais caixa na maior parte dos países analisados. Da mesma forma, empresas com maiores níveis de investimento em ativo imobilizado, maior geração de caixa, maior volatilidade do fluxo de caixa, maior alavancagem e maior nível de capital de giro, apresentam menor nível de acúmulo de ativos líquidos. São identificadas semelhanças de fatores determinantes de liquidez em relação a estudos empíricos com empresas de países desenvolvidos, bem como diferenças devido a fenômenos particulares de países emergentes, como por exemplo elevadas taxas de juros internas, diferentes graus de acessibilidade ao mercado de crédito internacional e a linhas de crédito de agências de fomento, equity kicking, entre outros. Em teste para a base de dados das maiores firmas do Brasil, é identificada a presença de níveis-alvo de caixa através de modelo auto-regressivo de primeira ordem (AR1). Variáveis presentes em estudos mais recentes com empresas de países desenvolvidos como aquisições, abertura recente de capital e nível de governança corporativa também são testadas para a base de dados do Brasil.
Resumo:
The oil industry, experiencing a great economic and environmental impact, has increasingly invested in researches aiming a more satisfactory treatment of its largest effluent, i.e., produced water. These are mostly discarded at sea, without reuse and after a basic treatment. Such effluent contains a range of organic compounds with high toxicity and are difficult to remove, such as polycyclic aromatic hydrocarbons, salts, heavy metals, etc.. The main objective of this work was to study the solar distillation of produced water pre-treated to remove salts and other contaminants trough of a hybrid system with a pre-heater. This developed apparatus was called solar system, which consists of a solar heater and a conventional distillation solar still. The first device consisted of a water tank, a solar flat plate collector and a thermal reservoir. The solar distillator is of simple effect, with 1m2 of flat area and 20° of inclination. This dissertation was divided in five steps: measurements in the solar system, i.e. temperatures and distillate flow rate and weather data; modeling and simulation of the system; study of vapor-liquid equilibrium of the synthetic wastewater by the aqueous solution of p-xylene; physical and chemical analyses of samples of the feed, distillate and residue, as well as climatology pertinent variables of Natal-RN. The solar system was tested separately, with the supply water, aqueous NaCl and synthetic oil produced water. Temperature measurements were taken every minute of the thermal reservoir, water tank and distillator (liquid and vapor phases). Data of solar radiation and rainfall were obtained from INPE (National Institute for Space Research). The solar pre-heater demonstrated to be effective for the liquid systems tested. The reservoir fluid had an average temperature of 58°C, which enabled the feed to be pre-heated in the distillator. The temperature profile in the solar distillator showed a similar behavior to daily solar radiation, with temperatures near 70°C. The distillation had an average yield of 2.4 L /day, i.e., an efficiency of 27.2%. Mathematical modeling aided the identification of the most important variables and parameters in the solar system. The study of the vapor-liquid equilibrium from Total Organic Carbon (TOC) analysis indicated heteroazeotropia and the vapor phase resulted more concentrated in p-xylene. The physical-chemical analysis of pH, conductivity, Total Dissolved Solids (TDS), chlorides, cations (including heavy metals) and anions, the effluent distillate showed satisfactory results, which presents a potential for reuse. The climatological study indicates the region of Natal-RN as favorable to the operation of solar systems, but the use of auxiliary heating during periods of higher rainfall and cloud cover is also recommended
Resumo:
This work focuses the geomorphological characterization and spatial data modeling in the shallow continental shelf within the Folha Touros limits (SB-25-CV-II), based on bathymetric data analysis and remote sensing products interpretation. The Rio Grande do Norte state is located in northeastern Brazil and the work area is located at the transition region between the eastern and northern portions of their coast. The bathymetric surveys were conduced between march and may 2009, using a 10 meters long vessel and 0.70 meters draught, equipped with global positioning system and echo sounder (dual beam, 200KHz , 14°). The fieldwork resulted in 44 bathymetric profiles espaced 1.5 km and 30 km average length. The bathymetric data amount were 111,200 points and were navigated 1395.7 km within na area about 1,850 km2. The bathymetric data were corrected for the tide level, vessel draught and were subsequently entered into a geographic information system for further processing. Analysis of remote sensing products was carried out using Landsat 7/ETM + band 1, from november 1999. The image was used for visualization and mapping submerged features. The results showed the presence of geomorphological features within the study area. Were observed, from the analysis of local bathymetry and satellite image, seven types of geomorphological features. The channels, with two longitudinals channels (e. g. San Roque and Cioba channels) and other perpendicular to the coast (e. g. Touros, Pititinga and Barretas). Coastal reef formations (Maracajaú, Rio do Fogo and Cioba). Longitudinal waves, described in the literature as longitudinal dunes. The occurrence of a transverse dune field. Another feature observed was the oceanic reefs, an rock alignment parallel to the coast. Were identified four riscas , from north to south: risca do Liso, Gameleira, Zumbi, Pititinga (the latter being described for the first time). Finally, an oceanic terrace was observed in the deepest area of study. Image interpretation corroborated with the in situ results, enabling visualization and description for all features in the region. The results were analysed in an integrating method (using the diferent methodologies applied in this work) and it was essential to describe all features in the area. This method allowed us to evaluate which methods generated better results to describe certain features. From these results was possible to prove the existence of submerged features in the eastern shallow continental shelf of Rio Grande do Norte. In this way, the conclusions was (1) this study contributed to the provision of new information about the area in question, particularly with regard to data collection in situ depths, (2) the method of data collection and interpretation proves to be effective because, through this, it was possible to visualize and interpret the features present in the study area and (3) the interpretation and discussion of results in an integrated method, using different methodologies, can provide better results
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
Pós-graduação em Ciências Cartográficas - FCT
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)