983 resultados para Incomplete Data


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Project justification is regarded as one of the major methodological deficits in Data Warehousing practice. As reasons for applying inappropriate methods, performing incomplete evaluations, or even entirely omitting justifications, the special nature of Data Warehousing benefits and the large portion of infrastructure-related activities are stated. In this paper, the economic justification of Data Warehousing projects is analyzed, and first results from a large academiaindustry collaboration project in the field of non-technical issues of Data Warehousing are presented. As conceptual foundations, the role of the Data Warehouse system in corporate application architectures is analyzed, and the specific properties of Data Warehousing projects are discussed. Based on an applicability analysis of traditional approaches to economic IT project justification, basic steps and responsibilities for the justification of Data Warehousing projects are derived.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Academic and industrial research in the late 90s have brought about an exponential explosion of DNA sequence data. Automated expert systems are being created to help biologists to extract patterns, trends and links from this ever-deepening ocean of information. Two such systems aimed on retrieving and subsequently utilizing phylogenetically relevant information have been developed in this dissertation, the major objective of which was to automate the often difficult and confusing phylogenetic reconstruction process. ^ Popular phylogenetic reconstruction methods, such as distance-based methods, attempt to find an optimal tree topology (that reflects the relationships among related sequences and their evolutionary history) by searching through the topology space. Various compromises between the fast (but incomplete) and exhaustive (but computationally prohibitive) search heuristics have been suggested. An intelligent compromise algorithm that relies on a flexible “beam” search principle from the Artificial Intelligence domain and uses the pre-computed local topology reliability information to adjust the beam search space continuously is described in the second chapter of this dissertation. ^ However, sometimes even a (virtually) complete distance-based method is inferior to the significantly more elaborate (and computationally expensive) maximum likelihood (ML) method. In fact, depending on the nature of the sequence data in question either method might prove to be superior. Therefore, it is difficult (even for an expert) to tell a priori which phylogenetic reconstruction method—distance-based, ML or maybe maximum parsimony (MP)—should be chosen for any particular data set. ^ A number of factors, often hidden, influence the performance of a method. For example, it is generally understood that for a phylogenetically “difficult” data set more sophisticated methods (e.g., ML) tend to be more effective and thus should be chosen. However, it is the interplay of many factors that one needs to consider in order to avoid choosing an inferior method (potentially a costly mistake, both in terms of computational expenses and in terms of reconstruction accuracy.) ^ Chapter III of this dissertation details a phylogenetic reconstruction expert system that selects a superior proper method automatically. It uses a classifier (a Decision Tree-inducing algorithm) to map a new data set to the proper phylogenetic reconstruction method. ^

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The consumption capital asset pricing model is the standard economic model used to capture stock market behavior. However, empirical tests have pointed out to its inability to account quantitatively for the high average rate of return and volatility of stocks over time for plausible parameter values. Recent research has suggested that the consumption of stockholders is more strongly correlated with the performance of the stock market than the consumption of non-stockholders. We model two types of agents, non-stockholders with standard preferences and stock holders with preferences that incorporate elements of the prospect theory developed by Kahneman and Tversky (1979). In addition to consumption, stockholders consider fluctuations in their financial wealth explicitly when making decisions. Data from the Panel Study of Income Dynamics are used to calibrate the labor income processes of the two types of agents. Each agent faces idiosyncratic shocks to his labor income as well as aggregate shocks to the per-share dividend but markets are incomplete and agents cannot hedge consumption risks completely. In addition, consumers face both borrowing and short-sale constraints. Our results show that in equilibrium, agents hold different portfolios. Our model is able to generate a time-varying risk premium of about 5.5% while maintaining a low risk free rate, thus suggesting a plausible explanation for the equity premium puzzle reported by Mehra and Prescott (1985).

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The discrete-time Markov chain is commonly used in describing changes of health states for chronic diseases in a longitudinal study. Statistical inferences on comparing treatment effects or on finding determinants of disease progression usually require estimation of transition probabilities. In many situations when the outcome data have some missing observations or the variable of interest (called a latent variable) can not be measured directly, the estimation of transition probabilities becomes more complicated. In the latter case, a surrogate variable that is easier to access and can gauge the characteristics of the latent one is usually used for data analysis. ^ This dissertation research proposes methods to analyze longitudinal data (1) that have categorical outcome with missing observations or (2) that use complete or incomplete surrogate observations to analyze the categorical latent outcome. For (1), different missing mechanisms were considered for empirical studies using methods that include EM algorithm, Monte Carlo EM and a procedure that is not a data augmentation method. For (2), the hidden Markov model with the forward-backward procedure was applied for parameter estimation. This method was also extended to cover the computation of standard errors. The proposed methods were demonstrated by the Schizophrenia example. The relevance of public health, the strength and limitations, and possible future research were also discussed. ^

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The coccolithophore Emiliania huxleyi (Lohmann) W. W. Hay et H. Mohler was cultured in natural seawater with the addition of either the microtubule-inhibitor colchicine, the actin-inhibitor cytochalasin B, or the photosynthesis inhibitor 3-(3,4 dichlorophenyl)-1,1-dimethyl-urea (DCMU). Additionally, E. huxleyi was cultured at different light intensities and temperatures. Growth rate was monitored, and coccolith morphology analyzed. While every treatment affected growth rate, the percentage of malformed coccoliths increased with colchicine, cytochalasin B, and at higher than optimal temperature. These results represent the first experimental evidence for the role of microtubules and actin microfilaments in coccolith morphogenesis.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A 160 m mostly turbiditic late Pleistocene sediment sequence (IODP Expedition 308, Hole U1319A) from the Brazos-Trinity intraslope basin system off Texas was investigated with paleo- and rock magnetic methods. Numerous layers depleted in iron oxides and enriched by the ferrimagnetic iron-sulfide mineral greigite (Fe3S4) were detected by diagnostic magnetic properties. From the distribution of these layers, their stratigraphic context and the present geochemical zonation, we develop two conceptual reaction models of greigite formation in non-steady depositional environments. The "sulfidization model" predicts single or twin greigite layers by incomplete transformation of iron monosulfides with polysulfides around the sulfate methane transition (SMT). The "oxidation model" explains greigite formation by partial oxidation of iron monosulfides near the iron redox boundary during periods of downward shifting oxidation fronts. The stratigraphic record provides evidence that both these greigite formation processes act here at typical depths of about 12-14 mbsf and 3-4 mbsf. Numerous "fossil" greigite layers most likely preserved by rapid upward shifts of the redox zonation denote past SMT and sea floor positions characterized by stagnant hemipelagic sedimentation conditions. Six diagenetic stages from a pristine magnetite-dominated to a fully greigite-dominated magnetic mineralogy were differentiated by combination of various hysteresis and remanence parameters.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Studies on the impact of historical, current and future global change require very high-resolution climate data (less or equal 1km) as a basis for modelled responses, meaning that data from digital climate models generally require substantial rescaling. Another shortcoming of available datasets on past climate is that the effects of sea level rise and fall are not considered. Without such information, the study of glacial refugia or early Holocene plant and animal migration are incomplete if not impossible. Sea level at the last glacial maximum (LGM) was approximately 125m lower, creating substantial additional terrestrial area for which no current baseline data exist. Here, we introduce the development of a novel, gridded climate dataset for LGM that is both very high resolution (1km) and extends to the LGM sea and land mask. We developed two methods to extend current terrestrial precipitation and temperature data to areas between the current and LGM coastlines. The absolute interpolation error is less than 1°C and 0.5 °C for 98.9% and 87.8% of all pixels for the first two 1 arc degree distance zones. We use the change factor method with these newly assembled baseline data to downscale five global circulation models of LGM climate to a resolution of 1km for Europe. As additional variables we calculate 19 'bioclimatic' variables, which are often used in climate change impact studies on biological diversity. The new LGM climate maps are well suited for analysing refugia and migration during Holocene warming following the LGM.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

There are many situations where input feature vectors are incomplete and methods to tackle the problem have been studied for a long time. A commonly used procedure is to replace each missing value with an imputation. This paper presents a method to perform categorical missing data imputation from numerical and categorical variables. The imputations are based on Simpson’s fuzzy min-max neural networks where the input variables for learning and classification are just numerical. The proposed method extends the input to categorical variables by introducing new fuzzy sets, a new operation and a new architecture. The procedure is tested and compared with others using opinion poll data.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Fission product yields are fundamental parameters for several nuclear engineering calculations and in particular for burn-up/activation problems. The impact of their uncertainties was widely studied in the past and valuations were released, although still incomplete. Recently, the nuclear community expressed the need for full fission yield covariance matrices to produce inventory calculation results that take into account the complete uncertainty data. In this work, we studied and applied a Bayesian/generalised least-squares method for covariance generation, and compared the generated uncertainties to the original data stored in the JEFF-3.1.2 library. Then, we focused on the effect of fission yield covariance information on fission pulse decay heat results for thermal fission of 235U. Calculations were carried out using different codes (ACAB and ALEPH-2) after introducing the new covariance values. Results were compared with those obtained with the uncertainty data currently provided by the library. The uncertainty quantification was performed with the Monte Carlo sampling technique. Indeed, correlations between fission yields strongly affect the statistics of decay heat. Introduction Nowadays, any engineering calculation performed in the nuclear field should be accompanied by an uncertainty analysis. In such an analysis, different sources of uncertainties are taken into account. Works such as those performed under the UAM project (Ivanov, et al., 2013) treat nuclear data as a source of uncertainty, in particular cross-section data for which uncertainties given in the form of covariance matrices are already provided in the major nuclear data libraries. Meanwhile, fission yield uncertainties were often neglected or treated shallowly, because their effects were considered of second order compared to cross-sections (Garcia-Herranz, et al., 2010). However, the Working Party on International Nuclear Data Evaluation Co-operation (WPEC)

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: The pupillary light reflex characterizes the direct and consensual response of the eye to the perceived brightness of a stimulus. It has been used as indicator of both neurological and optic nerve pathologies. As with other eye reflexes, this reflex constitutes an almost instantaneous movement and is linked to activation of the same midbrain area. The latency of the pupillary light reflex is around 200 ms, although the literature also indicates that the fastest eye reflexes last 20 ms. Therefore, a system with sufficiently high spatial and temporal resolutions is required for accurate assessment. In this study, we analyzed the pupillary light reflex to determine whether any small discrepancy exists between the direct and consensual responses, and to ascertain whether any other eye reflex occurs before the pupillary light reflex. Methods: We constructed a binocular video-oculography system two high-speed cameras that simultaneously focused on both eyes. This was then employed to assess the direct and consensual responses of each eye using our own algorithm based on Circular Hough Transform to detect and track the pupil. Time parameters describing the pupillary light reflex were obtained from the radius time-variation. Eight healthy subjects (4 women, 4 men, aged 24–45) participated in this experiment. Results: Our system, which has a resolution of 15 microns and 4 ms, obtained time parameters describing the pupillary light reflex that were similar to those reported in previous studies, with no significant differences between direct and consensual reflexes. Moreover, it revealed an incomplete reflex blink and an upward eye movement at around 100 ms that may correspond to Bell’s phenomenon. Conclusions: Direct and consensual pupillary responses do not any significant temporal differences. The system and method described here could prove useful for further assessment of pupillary and blink reflexes. The resolution obtained revealed the existence reported here of an early incomplete blink and an upward eye movement.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this study, a methodology based in a dynamical framework is proposed to incorporate additional sources of information to normalized difference vegetation index (NDVI) time series of agricultural observations for a phenological state estimation application. The proposed implementation is based on the particle filter (PF) scheme that is able to integrate multiple sources of data. Moreover, the dynamics-led design is able to conduct real-time (online) estimations, i.e., without requiring to wait until the end of the campaign. The evaluation of the algorithm is performed by estimating the phenological states over a set of rice fields in Seville (SW, Spain). A Landsat-5/7 NDVI series of images is complemented with two distinct sources of information: SAR images from the TerraSAR-X satellite and air temperature information from a ground-based station. An improvement in the overall estimation accuracy is obtained, especially when the time series of NDVI data is incomplete. Evaluations on the sensitivity to different development intervals and on the mitigation of discontinuities of the time series are also addressed in this work, demonstrating the benefits of this data fusion approach based on the dynamic systems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The Middle Valley segment at the northern end of the Juan de Fuca Ridge is a deep extensional rift blanketed with 200-500 m of Pleistocene turbiditic sediment. Sites 857 and 858 were drilled during Ocean Drilling Program Leg 139 to determine whether these two sites were hydrologically linked end members of an active hydrothermal circulation system. Site 858 was placed in an area of active hydrothermal discharge with fluids up to 270°C venting through anhydrite-bearing mounds on top of altered sediment. The shallow basement of fine-grained basalt that underlies the vents at Site 858 is interpreted as a seamount that was subsequently buried by turbidites. Site 857 was placed 1.6 km south of the Site 858 vents in a zone of high heat flow and numerous seismically imaged ridge-parallel faults. Drilling at Site 857 encountered sediments that are increasingly altered with depth and that overlie a series of mafic sills at depths of 460-940 m below sea floor. Sill margins and adjacent baked sediment are highly altered to magnesian chlorite and crosscut with veins filled with quartz, chlorite, sulfides, epidote, and wairakite. The sill interiors vary from slightly altered, with unaltered plagioclase and clinopyroxene in a mesostasis replaced by chlorite, to local zones of intense alteration and brecciation. In these latter zones, the sill interiors are pervasively replaced by chlorite, epidote, quartz, pyrite, titanite, and rare actinolite. The most complete replacement is associated with brecciated horizons with low recovery and slickensides on fracture surfaces, which we interpret as intersections between faults and the sills. Geochemically, the alteration of the sill complex is reflected in significant whole-rock depletions in Ca, Sr, and Na with corresponding enrichments in Mg, Al, and most metals. The latter results from the formation of conspicuous sulfide poikiloblasts. In contrast, metamorphism of the Site 858 seamount includes incomplete albitization of plagioclase phenocrysts and replacement of sparse mafic phenocrysts. Much of the basement alteration at Site 858 is confined to crosscutting veins except for a highly altered and veined horizon at the contact between basaltic basement and the overlying sediment. The sill complex at Site 857 is more highly depleted in 18O (d18O = 2.4 per mil - 4.7 per mil) and more pervasively replaced by secondary minerals relative to the extrusives at Site 858 (d18O = 4.5 per mil - 5.5 per mil). There is no evidence of significant albitization of the plagioclase at Site 857, suggesting high Ca/Na in the pore fluids. Fluid-inclusion data from hydrothermal minerals in altered mafic rocks and veins at Sites 857 and 858 show a consistency of homogenization temperatures, varying from 245 to 270°C, which is within the range of temperatures observed for the fluids venting at Site 858. The consistency of the fluid inclusion temperatures, the lack of albitization within the Site 857 sills, and the apparently low water/rock ratio collectively suggest that the sill complex at Site 857 is in thermal equilibrium and being altered by a highly evolved Ca-rich fluid similar to the fluids now venting at Site 858. The alteration evident in these two deep crustal drillsites is a result of the ongoing hydrothermal circulation and is consistent with downhole logging results, instrumented borehole results, and hydrothermal fluid chemistry. The pervasive alteration of the laterally extensive sill-sediment complex at Site 857 determines the chemistry of the fluids that are venting at Site 858. The limited alteration of the Site 858 lavas suggests that this basement edifice acts as a penetrator or ventilator for the regional hydrothermal reservoir with much of the flow focussed at the highly altered and veined sediment-basalt contact.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the wake of findings from the Bundaberg Hospital and Forster inquiries in Queensland, periodic public release of hospital performance reports has been recommended. A process for developing and releasing such reports is being established by Queensland Health, overseen by an independent expert panel. This recommendation presupposes that public reports based on routinely collected administrative data are accurate; that the public can access, correctly interpret and act upon report contents; that reports motivate hospital clinicians and managers to improve quality of care; and that there are no unintended adverse effects of public reporting. Available research suggests that primary data sources are often inaccurate and incomplete, that reports have low predictive value in detecting outlier hospitals, and that users experience difficulty in accessing and interpreting reports and tend to distrust their findings.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This special issue is a collection of the selected papers published on the proceedings of the First International Conference on Advanced Data Mining and Applications (ADMA) held in Wuhan, China in 2005. The articles focus on the innovative applications of data mining approaches to the problems that involve large data sets, incomplete and noise data, or demand optimal solutions.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

With an increased emphasis on outsourcing and shortening business cycles, contracts between firms have become more important. Carefully written contracts contribute to the efficiency and longevity of inter-firm relationships as they may constrain opportunism and are often a less costly governance mechanism than maintaining complex social relationships (Larson 1992). This exploratory examination adds to our understanding of how incomplete contracts affect interorganizational exchange. First, we consider the multiple dimensions of contract constraints (safeguards). We also investigate the extent that constraints affect decisions to enforce the relationship by delaying payments, and whether the decision is efficient. Finally, we examine the extent the constraints are effective (and ineffective) at reducing transaction problems associated with enforcement. Based on 971 observations of transactions using explicit, written terms and other secondary data in the context of IT transaction in The Netherlands we test our research propositions.