68 resultados para PROBABILITY-DISTRIBUTIONS
em Indian Institute of Science - Bangalore - Índia
Resumo:
The stochastic version of Pontryagin's maximum principle is applied to determine an optimal maintenance policy of equipment subject to random deterioration. The deterioration of the equipment with age is modelled as a random process. Next the model is generalized to include random catastrophic failure of the equipment. The optimal maintenance policy is derived for two special probability distributions of time to failure of the equipment, namely, exponential and Weibull distributions Both the salvage value and deterioration rate of the equipment are treated as state variables and the maintenance as a control variable. The result is illustrated by an example
Resumo:
This paper describes a detailed study of the structure of turbulence in boundary layers along mildly curved convex and concave surfaces. The surface curvature studied corresponds to δ/Rw = ± 0·01, δ being the boundary-layer thickness and Rw the radius of curvature of the wall, taken as positive for convex and negative for concave curvature. Measurements of turbulent energy balance, autocorrelations, auto- and cross-power spectra, amplitude probability distributions and conditional correlations are reported. It is observed that even mild curvature has very strong effects on the various aspects of the turbulent structure. For example, convex curvature suppresses the diffusion of turbulent energy away from the wall, reduces drastically the integral time scales and shifts the spectral distributions of turbulent energy and Reynolds shear stress towards high wavenumbers. Exactly opposite effects, though generally of a smaller magnitude, are produced by concave wall curvature. It is also found that curvature of either sign affects the v fluctuations more strongly than the u fluctuations and that curvature effects are more significant in the outer region of the boundary layer than in the region close to the wall. The data on the conditional correlations are used to study, in detail, the mechanism of turbulent transport in curved boundary layers. (Published Online April 12 2006)
Resumo:
We study the problem of matching applicants to jobs under one-sided preferences: that is, each applicant ranks a non-empty subset of jobs under an order of preference, possibly involving ties. A matching M is said to be rnore popular than T if the applicants that prefer M to T outnumber those that prefer T to M. A matching is said to be popular if there is no matching more popular than it. Equivalently, a matching M is popular if phi(M,T) >= phi(T, M) for all matchings T, where phi(X, Y) is the number of applicants that prefer X to Y. Previously studied solution concepts based oil the popularity criterion are either not guaranteed to exist for every instance (e.g., popular matchings) or are NP-hard to compute (e.g., least unpopular matchings). This paper addresses this issue by considering mixed matchings. A mixed matching is simply a probability distributions over matchings in the input graph. The function phi that compares two matchings generalizes in a natural manner to mixed matchings by taking expectation. A mixed matching P is popular if phi(P,Q) >= phi(Q,P) for all mixed matchings Q. We show that popular mixed matchings always exist. and we design polynomial time algorithms for finding them. Then we study their efficiency and give tight bounds on the price of anarchy and price of stability of the popular matching problem.
Resumo:
Guo and Nixon proposed a feature selection method based on maximizing I(x; Y),the multidimensional mutual information between feature vector x and class variable Y. Because computing I(x; Y) can be difficult in practice, Guo and Nixon proposed an approximation of I(x; Y) as the criterion for feature selection. We show that Guo and Nixon's criterion originates from approximating the joint probability distributions in I(x; Y) by second-order product distributions. We remark on the limitations of the approximation and discuss computationally attractive alternatives to compute I(x; Y).
Resumo:
We review some advances in the theory of homogeneous, isotropic turbulence. Our emphasis is on the new insights that have been gained from recent numerical studies of the three-dimensional Navier Stokes equation and simpler shell models for turbulence. In particular, we examine the status of multiscaling corrections to Kolmogorov scaling, extended self similarity, generalized extended self similarity, and non-Gaussian probability distributions for velocity differences and related quantities. We recount our recent proposal of a wave-vector-space version of generalized extended self similarity and show how it allows us to explore an intriguing and apparently universal crossover from inertial- to dissipation-range asymptotics.
Resumo:
In a detailed model for reservoir irrigation taking into account the soil moisture dynamics in the root zone of the crops, the data set for reservoir inflow and rainfall in the command will usually be of sufficient length to enable their variations to be described by probability distributions. However, the potential evapotranspiration of the crop itself depends on the characteristics of the crop and the reference evaporation, the quantification of both being associated with a high degree of uncertainty. The main purpose of this paper is to propose a mathematical programming model to determine the annual relative yield of crops and to determine its reliability, for a single reservoir meant for irrigation of multiple crops, incorporating variations in inflow, rainfall in the command area, and crop consumptive use. The inflow to the reservoir and rainfall in the reservoir command area are treated as random variables, whereas potential evapotranspiration is modeled as a fuzzy set. The model's application is illustrated with reference to an existing single-reservoir system in Southern India.
Resumo:
Denial-of-service (DoS) attacks form a very important category of security threats that are prevalent in MIPv6 (mobile internet protocol version 6) today. Many schemes have been proposed to alleviate such threats, including one of our own [9]. However, reasoning about the correctness of such protocols is not trivial. In addition, new solutions to mitigate attacks may need to be deployed in the network on a frequent basis as and when attacks are detected, as it is practically impossible to anticipate all attacks and provide solutions in advance. This makes it necessary to validate the solutions in a timely manner before deployment in the real network. However, threshold schemes needed in group protocols make analysis complex. Model checking threshold-based group protocols that employ cryptography have not been successful so far. Here, we propose a new simulation based approach for validation using a tool called FRAMOGR that supports executable specification of group protocols that use cryptography. FRAMOGR allows one to specify attackers and track probability distributions of values or paths. We believe that infrastructure such as FRAMOGR would be required in future for validating new group based threshold protocols that may be needed for making MIPv6 more robust.
Resumo:
Using all atom molecular dynamics simulations, we report spontaneous unzipping and strong binding of small interfering RNA (siRNA) on graphene. Our dispersion corrected density functional theory based calculations suggest that nucleosides of RNA have stronger attractive interactions with graphene as compared to DNA residues. These stronger interactions force the double stranded siRNA to spontaneously unzip and bind to the graphene surface. Unzipping always nucleates at one end of the siRNA and propagates to the other end after few base-pairs get unzipped. While both the ends get unzipped, the middle part remains in double stranded form because of torsional constraint. Unzipping probability distributions fitted to single exponential function give unzipping time (tau) of the order of few nanoseconds which decrease exponentially with temperature. From the temperature variation of unzipping time we estimate the energy barrier to unzipping. (C) 2012 American Institute of Physics. http://dx.doi.org/10.1063/1.4742189]
Resumo:
Our work is motivated by geographical forwarding of sporadic alarm packets to a base station in a wireless sensor network (WSN), where the nodes are sleep-wake cycling periodically and asynchronously. We seek to develop local forwarding algorithms that can be tuned so as to tradeoff the end-to-end delay against a total cost, such as the hop count or total energy. Our approach is to solve, at each forwarding node enroute to the sink, the local forwarding problem of minimizing one-hop waiting delay subject to a lower bound constraint on a suitable reward offered by the next-hop relay; the constraint serves to tune the tradeoff. The reward metric used for the local problem is based on the end-to-end total cost objective (for instance, when the total cost is hop count, we choose to use the progress toward sink made by a relay as the reward). The forwarding node, to begin with, is uncertain about the number of relays, their wake-up times, and the reward values, but knows the probability distributions of these quantities. At each relay wake-up instant, when a relay reveals its reward value, the forwarding node's problem is to forward the packet or to wait for further relays to wake-up. In terms of the operations research literature, our work can be considered as a variant of the asset selling problem. We formulate our local forwarding problem as a partially observable Markov decision process (POMDP) and obtain inner and outer bounds for the optimal policy. Motivated by the computational complexity involved in the policies derived out of these bounds, we formulate an alternate simplified model, the optimal policy for which is a simple threshold rule. We provide simulation results to compare the performance of the inner and outer bound policies against the simple policy, and also against the optimal policy when the source knows the exact number of relays. Observing the good performance and the ease of implementation of the simple policy, we apply it to our motivating problem, i.e., local geographical routing of sporadic alarm packets in a large WSN. We compare the end-to-end performance (i.e., average total delay and average total cost) obtained by the simple policy, when used for local geographical forwarding, against that obtained by the globally optimal forwarding algorithm proposed by Kim et al. 1].
Resumo:
The distributed, low-feedback, timer scheme is used in several wireless systems to select the best node from the available nodes. In it, each node sets a timer as a function of a local preference number called a metric, and transmits a packet when its timer expires. The scheme ensures that the timer of the best node, which has the highest metric, expires first. However, it fails to select the best node if another node transmits a packet within Delta s of the transmission by the best node. We derive the optimal metric-to-timer mappings for the practical scenario where the number of nodes is unknown. We consider two cases in which the probability distribution of the number of nodes is either known a priori or is unknown. In the first case, the optimal mapping maximizes the success probability averaged over the probability distribution. In the second case, a robust mapping maximizes the worst case average success probability over all possible probability distributions on the number of nodes. Results reveal that the proposed mappings deliver significant gains compared to the mappings considered in the literature.
Resumo:
We consider the basic bidirectional relaying problem, in which two users in a wireless network wish to exchange messages through an intermediate relay node. In the compute-and-forward strategy, the relay computes a function of the two messages using the naturally occurring sum of symbols simultaneously transmitted by user nodes in a Gaussian multiple-access channel (MAC), and the computed function value is forwarded to the user nodes in an ensuing broadcast phase. In this paper, we study the problem under an additional security constraint, which requires that each user's message be kept secure from the relay. We consider two types of security constraints: 1) perfect secrecy, in which the MAC channel output seen by the relay is independent of each user's message and 2) strong secrecy, which is a form of asymptotic independence. We propose a coding scheme based on nested lattices, the main feature of which is that given a pair of nested lattices that satisfy certain goodness properties, we can explicitly specify probability distributions for randomization at the encoders to achieve the desired security criteria. In particular, our coding scheme guarantees perfect or strong secrecy even in the absence of channel noise. The noise in the channel only affects reliability of computation at the relay, and for Gaussian noise, we derive achievable rates for reliable and secure computation. We also present an application of our methods to the multihop line network in which a source needs to transmit messages to a destination through a series of intermediate relays.
Resumo:
There is a need to use probability distributions with power-law decaying tails to describe the large variations exhibited by some of the physical phenomena. The Weierstrass Random Walk (WRW) shows promise for modeling such phenomena. The theory of anomalous diffusion is now well established. It has found number of applications in Physics, Chemistry and Biology. However, its applications are limited in structural mechanics in general, and structural engineering in particular. The aim of this paper is to present some mathematical preliminaries related to WRW that would help in possible applications. In the limiting case, it represents a diffusion process whose evolution is governed by a fractional partial differential equation. Three applications of superdiffusion processes in mechanics, illustrating their effectiveness in handling large variations, are presented.
Resumo:
The effect of fluid velocity fluctuations on the dynamics of the particles in a turbulent gas–solid suspension is analysed in the low-Reynolds-number and high Stokes number limits, where the particle relaxation time is long compared with the correlation time for the fluid velocity fluctuations, and the drag force on the particles due to the fluid can be expressed by the modified Stokes law. The direct numerical simulation procedure is used for solving the Navier–Stokes equations for the fluid, the particles are modelled as hard spheres which undergo elastic collisions and a one-way coupling algorithm is used where the force exerted by the fluid on the particles is incorporated, but not the reverse force exerted by the particles on the fluid. The particle mean and root-mean-square (RMS) fluctuating velocities, as well as the probability distribution function for the particle velocity fluctuations and the distribution of acceleration of the particles in the central region of the Couette (where the velocity profile is linear and the RMS velocities are nearly constant), are examined. It is found that the distribution of particle velocities is very different from a Gaussian, especially in the spanwise and wall-normal directions. However, the distribution of the acceleration fluctuation on the particles is found to be close to a Gaussian, though the distribution is highly anisotropic and there is a correlation between the fluctuations in the flow and gradient directions. The non-Gaussian nature of the particle velocity fluctuations is found to be due to inter-particle collisions induced by the large particle velocity fluctuations in the flow direction. It is also found that the acceleration distribution on the particles is in very good agreement with the distribution that is calculated from the velocity fluctuations in the fluid, using the Stokes drag law, indicating that there is very little correlation between the fluid velocity fluctuations and the particle velocity fluctuations in the presence of one-way coupling. All of these results indicate that the effect of the turbulent fluid velocity fluctuations can be accurately represented by an anisotropic Gaussian white noise.
Resumo:
We address the problem of designing an optimal pointwise shrinkage estimator in the transform domain, based on the minimum probability of error (MPE) criterion. We assume an additive model for the noise corrupting the clean signal. The proposed formulation is general in the sense that it can handle various noise distributions. We consider various noise distributions (Gaussian, Student's-t, and Laplacian) and compare the denoising performance of the estimator obtained with the mean-squared error (MSE)-based estimators. The MSE optimization is carried out using an unbiased estimator of the MSE, namely Stein's Unbiased Risk Estimate (SURE). Experimental results show that the MPE estimator outperforms the SURE estimator in terms of SNR of the denoised output, for low (0 -10 dB) and medium values (10 - 20 dB) of the input SNR.
Resumo:
We investigate the relaxation of long-tailed distributions under stochastic dynamics that do not support such tails. Linear relaxation is found to be a borderline case in which long tails are exponentially suppressed in time but not eliminated. Relaxation stronger than linear suppresses long tails immediately, but may lead to strong transient peaks in the probability distribution. We also find that a delta-function initial distribution under stronger than linear decay displays not one but two different regimes of diffusive spreading.