128 resultados para RANDOM REGULAR GRAPHS
Resumo:
Fabrication of devices based on thin film structures deposited using the pulsed laser deposition technique relies on reproducibility and control of deposition rates over substrate areas as large as possible. Here we present an application of the random phase plate technique to smooth and homogenize the intensity distribution of a KrF laser footprint on the surface of a target which is to be ablated. It is demonstrated that intensity distributions over millimeter-sized spots on the target can be made insensitive to the typical changes that occur in the near-field intensity distribution of the ultraviolet output from a KrF laser. (C) 1999 American Institute of Physics. [S0034-6748(99)02504-6].
Resumo:
Many studies have shown that with increasing LET of ionizing radiation the RBE (relative biological effectiveness) for dsb (double strand breaks) induction remains around 1.0 despite the increase in the RBE for cell killing. This has been attributed to an increase in the complexity of lesions, classified as dsb with current techniques, at multiply damaged sites. This study determines the molecular weight distributions of DNA from Chinese hamster V79 cells irradiated with X-rays or 110 keV/mu m alpha-particles. Two running conditions for pulsed-field gel-electrophoresis were chosen to give optimal separation of fragments either in the 225 kbp-5.7 Mbp range or the 0.3 kbp to 225 kbp range. Taking the total fraction of DNA migrating into the gel as a measure of fragmentation, the RBE for dsb induction was less than 1.0 for both molecular weight regions studied. The total yields of dsb were 8.2 x 10(-9) dsb/Gy/bp for X-rays and 7.8 x 10(-9) dsb/Gy/bp for a-particles, measured using a random breakage model. Analysis of the RBE of alpha-particles versus molecular weight gave a different response. In the 0.4 Mbp-57 Mbp region the RBE was less than 1.0; however, below 0.4 Mbp the RBE increased above 1.0. The frequency distributions of fragment sizes were found to differ from those predicted by a model assuming random breakage along the length of the DNA and the differences were greater for alpha-particles than for X-rays. An excess of fragments induced by a single-hit mechanism was found in the 8-300 kbp region and for X-rays and alpha-particles these corresponded to an extra 0.8 x 10(-9) and 3.4 x 10(-9) dsb/bp/Gy, respectively. Thus for every alpha-particle track that induces a dsb there is a 44% probability of inducing a second break within 300 kbp and for electron tracks the probability is 10%. This study shows that the distribution of damage from a high LET alpha-particle track is significantly different from that observed with low LET X-rays. In particular, it suggests that the fragmentation patterns of irradiated DNA may be related to the higher-order chromatin repealing structures found in intact cells.
Resumo:
Many studies have shown that the effectiveness of radiations of varying LET is similar when yields of dsb have been measured, despite large differences in biological response. Recent evidence has suggested however, that current techniques underestimate the yields of dsb. By monitoring the fragmentation of DNA over a wide range of fragment sizes ( 6 Mbp) by pulsed field electrophoresis, RBE values greater than 1.0 for radiations of around 100 keV/mm have been determined. The data provide evidence for the production of correlated breaks produced within cells as particle tracks traverse the nucleus. The highly ordered structure of DNA within mammalian cells may lead to clustering of breaks over distances related to the repeating unit structures of the chromatin. As well as these regionally damaged sites, a major contributor to radiation effectiveness will be the localised clustering of damage in the 1 - 20 bp region. A major effort is required to elucidate the relative importance of these levels of clustering and their importance in biological response.
Resumo:
True random number generation is crucial in hardware security applications. Proposed is a voltage-controlled true random number generator that is inherently field-programmable. This facilitates increased entropy as a randomness source because there is more than one configuration state which lends itself to more compact and low-power architectures. It is evaluated through electrical characterisation and statistically through industry-standard randomness tests. To the best of the author's knowledge, it is one of the most efficient designs to date with respect to hardware design metrics.
Resumo:
Domain patterns consisting of triangular nanodomains of less than 50 nm size, arranged into long regular vertex arrays separated by stripe domains, were observed by (scanning and high-resolution) transmission electron microscopy and piezoresponse force microscopy in BiFeO3 single crystals grown from solution flux. Piezoresponse force microscopy analysis together with crystallographic analysis by selected area and nanobeam electron diffraction indicate that these patterns consist of ferroelectric 109 degrees domains. A possibility for conserving Kittel's law is discussed in terms of the patterns being confined to the skin layer observed recently on BiFeO3 single crystals.
Resumo:
This paper introduces the discrete choice model-paradigm of Random Regret Minimization (RRM) to the field of environmental and resource economics. The RRM-approach has been very recently developed in the context of travel demand modelling and presents a tractable, regret-based alternative to the dominant choice-modelling paradigm based on Random Utility Maximization-theory (RUM-theory). We highlight how RRM-based models provide closed form, logit-type formulations for choice probabilities that allow for capturing semi-compensatory behaviour and choice set-composition effects while being equally parsimonious as their utilitarian counterparts. Using data from a Stated Choice-experiment aimed at identifying valuations of characteristics of nature parks, we compare RRM-based models and RUM-based models in terms of parameter estimates, goodness of fit, elasticities and consequential policy implications.
Resumo:
A new model to explain animal spacing, based on a trade-off between foraging efficiency and predation risk, is derived from biological principles. The model is able to explain not only the general tendency for animal groups to form, but some of the attributes of real groups. These include the independence of mean animal spacing from group population, the observed variation of animal spacing with resource availability and also with the probability of predation, and the decline in group stability with group size. The appearance of "neutral zones" within which animals are not motivated to adjust their relative positions is also explained. The model assumes that animals try to minimize a cost potential combining the loss of intake rate due to foraging interference and the risk from exposure to predators. The cost potential describes a hypothetical field giving rise to apparent attractive and repulsive forces between animals. Biologically based functions are given for the decline in interference cost and increase in the cost of predation risk with increasing animal separation. Predation risk is calculated from the probabilities of predator attack and predator detection as they vary with distance. Using example functions for these probabilities and foraging interference, we calculate the minimum cost potential for regular lattice arrangements of animals before generalizing to finite-sized groups and random arrangements of animals, showing optimal geometries in each case and describing how potentials vary with animal spacing. (C) 1999 Academic Press.</p>
Resumo:
A conceptual model is described for generating distributions of grazing animals, according to their searching behavior, to investigate the mechanisms animals may use to achieve their distributions. The model simulates behaviors ranging from random diffusion, through taxis and cognitively aided navigation (i.e., using memory), to the optimization extreme of the Ideal Free Distribution. These behaviors are generated from simulation of biased diffusion that operates at multiple scales simultaneously, formalizing ideas of multiple-scale foraging behavior. It uses probabilistic bias to represent decisions, allowing multiple search goals to be combined (e.g., foraging and social goals) and the representation of suboptimal behavior. By allowing bias to arise at multiple scales within the environment, each weighted relative to the others, the model can represent different scales of simultaneous decision-making and scale-dependent behavior. The model also allows different constraints to be applied to the animal's ability (e.g., applying food-patch accessibility and information limits). Simulations show that foraging-decision randomness and spatial scale of decision bias have potentially profound effects on both animal intake rate and the distribution of resources in the environment. Spatial variograms show that foraging strategies can differentially change the spatial pattern of resource abundance in the environment to one characteristic of the foraging strategy.</
Resumo:
A multivariate Fokker-Planck-type kinetic equation modeling a test - panicle weakly interacting with an electrostatic plasma. in the presence of a magnetic field B . is analytically solved in an Ornstein - Uhlenbeck - type approximation. A new set of analytic expressions are obtained for variable moments and panicle density as a function of time. The process is diffusive.
Resumo:
In order to relate macroscopic random motion (described e.g. by Langevin-type theories) to microscopic dynamics, we have undertaken the derivation of a Fokker-Planck-type equation from first microscopic principles. Both subsystems are subject to an external force field. Explicit expressions for the diffusion and drift coefficients are obtained, in terms of the field.
Resumo:
A new nonlinear theory for the perpendicular transport of charged particles is presented. This approach is based on an improved nonlinear treatment of field line random walk in combination with a generalized compound diffusion model. The generalized compound diffusion model is much more systematic and reliable, in comparison to previous theories. Furthermore, the new theory shows remarkably good agreement with test-particle simulations and heliospheric observations.
Resumo:
We compare two approaches for estimating the distribution of consumers' willingness to pay (WTP) in discrete choice models. The usual procedure is to estimate the distribution of the utility coefficients and then derive the distribution of WTP, which is the ratio of coefficients. The alternative is to estimate the distribution of WTP directly. We apply both approaches to data on site choice in the Alps. We find that the alternative approach fits the data better, reduces the incidence of exceedingly large estimated WTP values, and provides the analyst with greater control in specifying and testing the distribution of WTP. © 2008 Agricultural and Applied Economics Association.