933 resultados para random oracle model
Resumo:
BACKGROUND: Physicians need a specific risk-stratification tool to facilitate safe and cost-effective approaches to the management of patients with cancer and acute pulmonary embolism (PE). The objective of this study was to develop a simple risk score for predicting 30-day mortality in patients with PE and cancer by using measures readily obtained at the time of PE diagnosis. METHODS: Investigators randomly allocated 1,556 consecutive patients with cancer and acute PE from the international multicenter Registro Informatizado de la Enfermedad TromboEmbólica to derivation (67%) and internal validation (33%) samples. The external validation cohort for this study consisted of 261 patients with cancer and acute PE. Investigators compared 30-day all-cause mortality and nonfatal adverse medical outcomes across the derivation and two validation samples. RESULTS: In the derivation sample, multivariable analyses produced the risk score, which contained six variables: age > 80 years, heart rate ≥ 110/min, systolic BP < 100 mm Hg, body weight < 60 kg, recent immobility, and presence of metastases. In the internal validation cohort (n = 508), the 22.2% of patients (113 of 508) classified as low risk by the prognostic model had a 30-day mortality of 4.4% (95% CI, 0.6%-8.2%) compared with 29.9% (95% CI, 25.4%-34.4%) in the high-risk group. In the external validation cohort, the 18% of patients (47 of 261) classified as low risk by the prognostic model had a 30-day mortality of 0%, compared with 19.6% (95% CI, 14.3%-25.0%) in the high-risk group. CONCLUSIONS: The developed clinical prediction rule accurately identifies low-risk patients with cancer and acute PE.
Resumo:
In medical imaging, merging automated segmentations obtained from multiple atlases has become a standard practice for improving the accuracy. In this letter, we propose two new fusion methods: "Global Weighted Shape-Based Averaging" (GWSBA) and "Local Weighted Shape-Based Averaging" (LWSBA). These methods extend the well known Shape-Based Averaging (SBA) by additionally incorporating the similarity information between the reference (i.e., atlas) images and the target image to be segmented. We also propose a new spatially-varying similarity-weighted neighborhood prior model, and an edge-preserving smoothness term that can be used with many of the existing fusion methods. We first present our new Markov Random Field (MRF) based fusion framework that models the above mentioned information. The proposed methods are evaluated in the context of segmentation of lymph nodes in the head and neck 3D CT images, and they resulted in more accurate segmentations compared to the existing SBA.
Resumo:
In the framework of the classical compound Poisson process in collective risk theory, we study a modification of the horizontal dividend barrier strategy by introducing random observation times at which dividends can be paid and ruin can be observed. This model contains both the continuous-time and the discrete-time risk model as a limit and represents a certain type of bridge between them which still enables the explicit calculation of moments of total discounted dividend payments until ruin. Numerical illustrations for several sets of parameters are given and the effect of random observation times on the performance of the dividend strategy is studied.
Resumo:
We asked whether locally applied recombinant-Bone Morphogenic Protein-2 (rh-BMP-2) with an absorbable Type I collagen sponge (ACS) carrier could enhance the consolidation phase in a callotasis model. We performed unilateral transverse osteotomy of the tibia in 21 immature male rabbits. After a latency period of 7 days, a 3-weeks distraction was begun at a rate of 0.5mm/12h. At the end of the distraction period (Day 28) animals were randomly divided into three groups and underwent a second surgical procedure: 6 rabbits in Group I (Control group; the callus was exposed and nothing was added), 6 rabbits in Group II (ACS group; receiving the absorbable collagen sponge soaked with saline) and 9 rabbits in Group III (rh-BMP-2/ACS group; receiving the ACS soaked with 100μg/kg of rh-BMP-2, Inductos(®), Medtronic). Starting at Day 28 we assessed quantitative and qualitative radiographic parameters as well as densitometric parameters every two weeks (Days 28, 42, 56, 70 and 84). Animals were sacrificed after 8 weeks of consolidation (Day 84). Qualitative radiographic evaluation revealed hypertrophic calluses in the Group III animals. The rh-BMP-2/ACS also influenced the development of the cortex of the calluses as shown by the modified radiographic patterns in Group III when compared to Groups I and II. Densitometric analysis revealed the bone mineral content (BMC) was significantly higher in the rh-BMP-2/ACS treated animals (Group III).
Resumo:
PURPOSE: To compare the effect of a rat anti-VEGF antibody, administered either by topical or subconjunctival (SC) routes, on a rat model of corneal transplant rejection.METHODS: Twenty-four rats underwent corneal transplantation and were randomized into four treatment groups (n=6 in each group). G1 and G2 received six SC injections (0.02 ml 10 µg/ml) of denatured (G1) or active (G2) anti-VEGF from Day 0 to Day 21 every third day. G3 and G4 were instilled three times a day with denatured (G3) or active (G4) anti-VEGF drops (10 µg/ml) from Day 0 to Day 21. Corneal mean clinical scores (MCSs) of edema (E), transparency (T), and neovessels (nv) were recorded at Days 3, 9, 15, and 21. Quantification of neovessels was performed after lectin staining of vessels on flat mounted corneas.RESULTS: Twenty-one days after surgery, MCSs differed significantly between G1 and G2, but not between G3 and G4, and the rejection rate was significantly reduced in rats receiving active antibodies regardless of the route of administration (G2=50%, G4=66.65% versus G1 and G3=100%; p<0.05). The mean surfaces of neovessels were significantly reduced in groups treated with active anti-VEGF (G2, G4). However, anti-VEGF therapy did not completely suppress corneal neovessels.CONCLUSIONS: Specific rat anti-VEGF antibodies significantly reduced neovascularization and subsequent corneal graft rejection. The SC administration of the anti-VEGF antibody was more effective than topical instillation.
Resumo:
An incentives based theory of policing is developed which can explain the phenomenon of random “crackdowns,” i.e., intermittent periods of high interdiction/surveillance. For a variety of police objective functions, random crackdowns can be part of the optimal monitoring strategy. We demonstrate support for implications of the crackdown theory using traffic data gathered by the Belgian Police Department and use the model to estimate the deterrence effectof additional resources spent on speeding interdiction.
Resumo:
Random coefficient regression models have been applied in differentfields and they constitute a unifying setup for many statisticalproblems. The nonparametric study of this model started with Beranand Hall (1992) and it has become a fruitful framework. In thispaper we propose and study statistics for testing a basic hypothesisconcerning this model: the constancy of coefficients. The asymptoticbehavior of the statistics is investigated and bootstrapapproximations are used in order to determine the critical values ofthe test statistics. A simulation study illustrates the performanceof the proposals.
Resumo:
This paper proposes a common and tractable framework for analyzingdifferent definitions of fixed and random effects in a contant-slopevariable-intercept model. It is shown that, regardless of whethereffects (i) are treated as parameters or as an error term, (ii) areestimated in different stages of a hierarchical model, or whether (iii)correlation between effects and regressors is allowed, when the sameinformation on effects is introduced into all estimation methods, theresulting slope estimator is also the same across methods. If differentmethods produce different results, it is ultimately because differentinformation is being used for each methods.
Resumo:
In many research areas (such as public health, environmental contamination, and others) one deals with the necessity of using data to infer whether some proportion (%) of a population of interest is (or one wants it to be) below and/or over some threshold, through the computation of tolerance interval. The idea is, once a threshold is given, one computes the tolerance interval or limit (which might be one or two - sided bounded) and then to check if it satisfies the given threshold. Since in this work we deal with the computation of one - sided tolerance interval, for the two-sided case we recomend, for instance, Krishnamoorthy and Mathew [5]. Krishnamoorthy and Mathew [4] performed the computation of upper tolerance limit in balanced and unbalanced one-way random effects models, whereas Fonseca et al [3] performed it based in a similar ideas but in a tow-way nested mixed or random effects model. In case of random effects model, Fonseca et al [3] performed the computation of such interval only for the balanced data, whereas in the mixed effects case they dit it only for the unbalanced data. For the computation of twosided tolerance interval in models with mixed and/or random effects we recomend, for instance, Sharma and Mathew [7]. The purpose of this paper is the computation of upper and lower tolerance interval in a two-way nested mixed effects models in balanced data. For the case of unbalanced data, as mentioned above, Fonseca et al [3] have already computed upper tolerance interval. Hence, using the notions persented in Fonseca et al [3] and Krishnamoorthy and Mathew [4], we present some results on the construction of one-sided tolerance interval for the balanced case. Thus, in order to do so at first instance we perform the construction for the upper case, and then the construction for the lower case.
Resumo:
There has been a recent revolution in the ability to manipulate micrometer-sized objects on surfaces patterned by traps or obstacles of controllable configurations and shapes. One application of this technology is to separate particles driven across such a surface by an external force according to some particle characteristic such as size or index of refraction. The surface features cause the trajectories of particles driven across the surface to deviate from the direction of the force by an amount that depends on the particular characteristic, thus leading to sorting. While models of this behavior have provided a good understanding of these observations, the solutions have so far been primarily numerical. In this paper we provide analytic predictions for the dependence of the angle between the direction of motion and the external force on a number of model parameters for periodic as well as random surfaces. We test these predictions against exact numerical simulations.
Resumo:
The effects of flow induced by a random acceleration field (g-jitter) are considered in two related situations that are of interest for microgravity fluid experiments: the random motion of isolated buoyant particles, and diffusion driven coarsening of a solid-liquid mixture. We start by analyzing in detail actual accelerometer data gathered during a recent microgravity mission, and obtain the values of the parameters defining a previously introduced stochastic model of this acceleration field. The diffusive motion of a single solid particle suspended in an incompressible fluid that is subjected to such random accelerations is considered, and mean squared velocities and effective diffusion coefficients are explicitly given. We next study the flow induced by an ensemble of such particles, and show the existence of a hydrodynamically induced attraction between pairs of particles at distances large compared with their radii, and repulsion at short distances. Finally, a mean field analysis is used to estimate the effect of g-jitter on diffusion controlled coarsening of a solid-liquid mixture. Corrections to classical coarsening rates due to the induced fluid motion are calculated, and estimates are given for coarsening of Sn-rich particles in a Sn-Pb eutectic fluid, an experiment to be conducted in microgravity in the near future.
Resumo:
A systematic assessment of global neural network connectivity through direct electrophysiological assays has remained technically infeasible, even in simpler systems like dissociated neuronal cultures. We introduce an improved algorithmic approach based on Transfer Entropy to reconstruct structural connectivity from network activity monitored through calcium imaging. We focus in this study on the inference of excitatory synaptic links. Based on information theory, our method requires no prior assumptions on the statistics of neuronal firing and neuronal connections. The performance of our algorithm is benchmarked on surrogate time series of calcium fluorescence generated by the simulated dynamics of a network with known ground-truth topology. We find that the functional network topology revealed by Transfer Entropy depends qualitatively on the time-dependent dynamic state of the network (bursting or non-bursting). Thus by conditioning with respect to the global mean activity, we improve the performance of our method. This allows us to focus the analysis to specific dynamical regimes of the network in which the inferred functional connectivity is shaped by monosynaptic excitatory connections, rather than by collective synchrony. Our method can discriminate between actual causal influences between neurons and spurious non-causal correlations due to light scattering artifacts, which inherently affect the quality of fluorescence imaging. Compared to other reconstruction strategies such as cross-correlation or Granger Causality methods, our method based on improved Transfer Entropy is remarkably more accurate. In particular, it provides a good estimation of the excitatory network clustering coefficient, allowing for discrimination between weakly and strongly clustered topologies. Finally, we demonstrate the applicability of our method to analyses of real recordings of in vitro disinhibited cortical cultures where we suggest that excitatory connections are characterized by an elevated level of clustering compared to a random graph (although not extreme) and can be markedly non-local.
Resumo:
A new solvable model of synchronization dynamics is introduced. It consists of a system of long range interacting tops or magnetic moments with random precession frequencies. The model allows for an explicit study of orientational effects in synchronization phenomena as well as nonlinear processes in resonance phenomena in strongly coupled magnetic systems. A stability analysis of the incoherent solution is performed for different types of orientational disorder. A system with orientational disorder always synchronizes in the absence of noise.
Resumo:
We present a generator of random networks where both the degree-dependent clustering coefficient and the degree distribution are tunable. Following the same philosophy as in the configuration model, the degree distribution and the clustering coefficient for each class of nodes of degree k are fixed ad hoc and a priori. The algorithm generates corresponding topologies by applying first a closure of triangles and second the classical closure of remaining free stubs. The procedure unveils an universal relation among clustering and degree-degree correlations for all networks, where the level of assortativity establishes an upper limit to the level of clustering. Maximum assortativity ensures no restriction on the decay of the clustering coefficient whereas disassortativity sets a stronger constraint on its behavior. Correlation measures in real networks are seen to observe this structural bound.
Resumo:
A dynamical model based on a continuous addition of colored shot noises is presented. The resulting process is colored and non-Gaussian. A general expression for the characteristic function of the process is obtained, which, after a scaling assumption, takes on a form that is the basis of the results derived in the rest of the paper. One of these is an expansion for the cumulants, which are all finite, subject to mild conditions on the functions defining the process. This is in contrast with the Lévy distribution¿which can be obtained from our model in certain limits¿which has no finite moments. The evaluation of the spectral density and the form of the probability density function in the tails of the distribution shows that the model exhibits a power-law spectrum and long tails in a natural way. A careful analysis of the characteristic function shows that it may be separated into a part representing a Lévy process together with another part representing the deviation of our model from the Lévy process. This