945 resultados para Entropic graphs


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Most haptic environments are based on single point interactions whereas in practice, object manipulation requires multiple contact points between the object, fingers, thumb and palm. The Friction Cone Algorithm was developed specifically to work well in a multi-finger haptic environment where object manipulation would occur. However, the Friction Cone Algorithm has two shortcomings when applied to polygon meshes: there is no means of transitioning polygon boundaries or feeling non-convex edges. In order to overcome these deficiencies, Face Directed Connection Graphs have been developed as well as a robust method for applying friction to non-convex edges. Both these extensions are described herein, as well as the implementation issues associated with them.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The arrival of a student who is Blind in the School of Systems Engineering at the University of Reading has made it an interesting and challenging year for all. Visually impaired students have already graduated from other Schools of the University and the School of Systems Engineering has seen three students with visual impairment graduate recently with good degrees. These students could access materials - and do assessments - essentially by means of enlargement and judicious choice of options. The new student had previously been supported by a specialist college. She is a proficient typist and also a user of both Braille and JAWS screen reader, and she is doing a joint course in Cybernetics and Computer Science. The course requires mathematics which itself includes graphs, and also many diagrams including numerous circuit diagrams. The University bought proven equipment such as a scanner to process books into speech or Braille, and screen reading software as well as a specialist machine for producing tactile diagrams for educational use. Clearly it is also important that the student can access assessments and examinations and present answers for marking or feedback (by sighted staff). So the School also used innovative in-house tactile methods to represent diagrams. This paper discusses the success or otherwise of various modifications of course delivery and the way forward for the next three years.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this paper, we introduce two kinds of graphs: the generalized matching networks (GMNs) and the recursive generalized matching networks (RGMNs). The former generalize the hypercube-like networks (HLNs), while the latter include the generalized cubes and the star graphs. We prove that a GMN on a family of k-connected building graphs is -connected. We then prove that a GMN on a family of Hamiltonian-connected building graphs having at least three vertices each is Hamiltonian-connected. Our conclusions generalize some previously known results.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Generalized cubes are a subclass of hypercube-like networks, which include some hypercube variants as special cases. Let theta(G)(k) denote the minimum number of nodes adjacent to a set of k vertices of a graph G. In this paper, we prove theta(G)(k) >= -1/2k(2) + (2n - 3/2)k - (n(2) - 2) for each n-dimensional generalized cube and each integer k satisfying n + 2 <= k <= 2n. Our result is an extension of a result presented by Fan and Lin [J. Fan, X. Lin, The t/k-diagnosability of the BC graphs, IEEE Trans. Comput. 54 (2) (2005) 176-184]. (c) 2005 Elsevier B.V. All rights reserved.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In order to make a full evaluation of an interconnection network, it is essential to estimate the minimum size of a largest connected component of this network provided the faulty vertices in the network may break its connectedness. Star graphs are recognized as promising candidates for interconnection networks. This article addresses the size of a largest connected component of a faulty star graph. We prove that, in an n-star graph (n >= 3) with up to 2n-4 faulty vertices, all fault-free vertices but at most two form a connected component. Moreover, all fault-free vertices but exactly two form a connected component if and only if the set of all faulty vertices is equal to the neighbourhood of a pair of fault-free adjacent vertices. These results show that star graphs exhibit excellent fault-tolerant abilities in the sense that there exists a large functional network in a faulty star graph.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We have investigated the dynamic mechanical behavior of two cross-linked polymer networks with very different topologies: one made of backbones randomly linked along their length; the other with fixed-length strands uniformly cross-linked at their ends. The samples were analyzed using oscillatory shear, at very small strains corresponding to the linear regime. This was carried out at a range of frequencies, and at temperatures ranging from the glass plateau, through the glass transition, and well into the rubbery region. Through the glass transition, the data obeyed the time-temperature superposition principle, and could be analyzed using WLF treatment. At higher temperatures, in the rubbery region, the storage modulus was found to deviate from this, taking a value that is independent of frequency. This value increased linearly with temperature, as expected for the entropic rubber elasticity, but with a substantial negative offset inconsistent with straightforward enthalpic effects. Conversely, the loss modulus continued to follow time-temperature superposition, decreasing with increasing temperature, and showing a power-law dependence on frequency.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Stochastic Diffusion Search is an efficient probabilistic bestfit search technique, capable of transformation invariant pattern matching. Although inherently parallel in operation it is difficult to implement efficiently in hardware as it requires full inter-agent connectivity. This paper describes a lattice implementation, which, while qualitatively retaining the properties of the original algorithm, restricts connectivity, enabling simpler implementation on parallel hardware. Diffusion times are examined for different network topologies, ranging from ordered lattices, over small-world networks to random graphs.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In the ordered state, symmetric diblock copolymers self-assemble into an anisotropic lamellar morphology. The equilibrium thickness of the lamellae is the result of a delicate balance between enthalpic and entropic energies, which can be tuned by controlling the temperature. Here we devise a simple yet powerful method of detecting tiny changes in the lamellar thickness using optical microscopy. From such measurements we characterize the enthalpic interaction as well as the kinetics of molecules as they hop from one layer to the next in order to adjust the lamellar thickness in response to a temperature jump. The resolution of the measurements facilitate a direct comparison to predictions from self-consistent field theory.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Classical measures of network connectivity are the number of disjoint paths between a pair of nodes and the size of a minimum cut. For standard graphs, these measures can be computed efficiently using network flow techniques. However, in the Internet on the level of autonomous systems (ASs), referred to as AS-level Internet, routing policies impose restrictions on the paths that traffic can take in the network. These restrictions can be captured by the valley-free path model, which assumes a special directed graph model in which edge types represent relationships between ASs. We consider the adaptation of the classical connectivity measures to the valley-free path model, where it is -hard to compute them. Our first main contribution consists of presenting algorithms for the computation of disjoint paths, and minimum cuts, in the valley-free path model. These algorithms are useful for ASs that want to evaluate different options for selecting upstream providers to improve the robustness of their connection to the Internet. Our second main contribution is an experimental evaluation of our algorithms on four types of directed graph models of the AS-level Internet produced by different inference algorithms. Most importantly, the evaluation shows that our algorithms are able to compute optimal solutions to instances of realistic size of the connectivity problems in the valley-free path model in reasonable time. Furthermore, our experimental results provide information about the characteristics of the directed graph models of the AS-level Internet produced by different inference algorithms. It turns out that (i) we can quantify the difference between the undirected AS-level topology and the directed graph models with respect to fundamental connectivity measures, and (ii) the different inference algorithms yield topologies that are similar with respect to connectivity and are different with respect to the types of paths that exist between pairs of ASs.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Statistical graphics are a fundamental, yet often overlooked, set of components in the repertoire of data analytic tools. Graphs are quick and efficient, yet simple instruments of preliminary exploration of a dataset to understand its structure and to provide insight into influential aspects of inference such as departures from assumptions and latent patterns. In this paper, we present and assess a graphical device for choosing a method for estimating population size in capture-recapture studies of closed populations. The basic concept is derived from a homogeneous Poisson distribution where the ratios of neighboring Poisson probabilities multiplied by the value of the larger neighbor count are constant. This property extends to the zero-truncated Poisson distribution which is of fundamental importance in capture–recapture studies. In practice however, this distributional property is often violated. The graphical device developed here, the ratio plot, can be used for assessing specific departures from a Poisson distribution. For example, simple contaminations of an otherwise homogeneous Poisson model can be easily detected and a robust estimator for the population size can be suggested. Several robust estimators are developed and a simulation study is provided to give some guidance on which should be used in practice. More systematic departures can also easily be detected using the ratio plot. In this paper, the focus is on Gamma mixtures of the Poisson distribution which leads to a linear pattern (called structured heterogeneity) in the ratio plot. More generally, the paper shows that the ratio plot is monotone for arbitrary mixtures of power series densities.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We propose and analyse a class of evolving network models suitable for describing a dynamic topological structure. Applications include telecommunication, on-line social behaviour and information processing in neuroscience. We model the evolving network as a discrete time Markov chain, and study a very general framework where, conditioned on the current state, edges appear or disappear independently at the next timestep. We show how to exploit symmetries in the microscopic, localized rules in order to obtain conjugate classes of random graphs that simplify analysis and calibration of a model. Further, we develop a mean field theory for describing network evolution. For a simple but realistic scenario incorporating the triadic closure effect that has been empirically observed by social scientists (friends of friends tend to become friends), the mean field theory predicts bistable dynamics, and computational results confirm this prediction. We also discuss the calibration issue for a set of real cell phone data, and find support for a stratified model, where individuals are assigned to one of two distinct groups having different within-group and across-group dynamics.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This project is concerned with the way that illustrations, photographs, diagrams and graphs, and typographic elements interact to convey ideas on the book page. A framework for graphic description is proposed to elucidate this graphic language of ‘complex texts’. The model is built up from three main areas of study, with reference to a corpus of contemporary children’s science books. First, a historical survey puts the subjects for study in context. Then a multidisciplinary discussion of graphic communication provides a theoretical underpinning for the model; this leads to various proposals, such as the central importance of ratios and relationships among parts in creating meaning in graphic communication. Lastly a series of trials in description contribute to the structure of the model itself. At the heart of the framework is an organising principle that integrates descriptive models from fields of design, literary criticism, art history, and linguistics, among others, as well as novel categories designed specifically for book design. Broadly, design features are described in terms of elemental component parts (micro-level), larger groupings of these (macro-level), and finally in terms of overarching, ‘whole book’ qualities (meta-level). Various features of book design emerge at different levels; for instance, the presence of nested discursive structures, a form of graphic recursion in editorial design, is proposed at the macro-level. Across these three levels are the intersecting categories of ‘rule’ and ‘context’, offering different perspectives with which to describe graphic characteristics. Contextbased features are contingent on social and cultural environment, the reader’s previous knowledge, and the actual conditions of reading; rule-based features relate to the systematic or codified aspects of graphic language. The model aims to be a frame of reference for graphic description, of use in different forms of qualitative or quantitative research and as a heuristic tool in practice and teaching.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Undirected graphical models are widely used in statistics, physics and machine vision. However Bayesian parameter estimation for undirected models is extremely challenging, since evaluation of the posterior typically involves the calculation of an intractable normalising constant. This problem has received much attention, but very little of this has focussed on the important practical case where the data consists of noisy or incomplete observations of the underlying hidden structure. This paper specifically addresses this problem, comparing two alternative methodologies. In the first of these approaches particle Markov chain Monte Carlo (Andrieu et al., 2010) is used to efficiently explore the parameter space, combined with the exchange algorithm (Murray et al., 2006) for avoiding the calculation of the intractable normalising constant (a proof showing that this combination targets the correct distribution in found in a supplementary appendix online). This approach is compared with approximate Bayesian computation (Pritchard et al., 1999). Applications to estimating the parameters of Ising models and exponential random graphs from noisy data are presented. Each algorithm used in the paper targets an approximation to the true posterior due to the use of MCMC to simulate from the latent graphical model, in lieu of being able to do this exactly in general. The supplementary appendix also describes the nature of the resulting approximation.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We use a combination of ab initio calculations and statistical mechanics to investigate the substitution of Li+ for Mg2+ in magnesium hydride (MgH2) accompanied by the formation of hydrogen vacancies with positive charge (with respect to the original ion at the site). We show that the binding energy between dopants and vacancy defects leads to a significant fraction of trapped vacancies and therefore a dramatic reduction in the number of free vacancies available for diffusion. The concentration of free vacancies initially increases with dopant concentration but reaches a maximum at around 1 mol % Li doping and slowly decreases with further doping. At the optimal level of doping, the corresponding concentration of free vacancies is much higher than the equilibrium concentrations of charged and neutral vacancies in pure MgH2 at typical hydrogen storage conditions. We also show that Li-doped MgH2 is thermodynamically metastable with respect to phase separation into pure magnesium and lithium hydrides at any significant Li concentration, even after considering the stabilization provided by dopant-vacancy interactions and configurational entropic effects. Our results suggest that lithium doping may enhance hydrogen diffusion hydride but only to a limited extent determined by an optimal dopant concentration and conditioned to the stability of the doped phase.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This article considers cinematic time in James Benning’s film, casting a glance (2007), in relation to its subject, Robert Smithson’s 1970 earthwork Spiral Jetty, and his film of the same name. The radicalism of Smithson’s thinking on time has been widely acknowledged, and his influence continues to pervade contemporary artistic practice. The relationship of Benning’s films with this legacy may appear somewhat oblique, given their apparent phenomenological rendition of ‘real time’. However, closer examination of Benning’s formal strategies reveals a more complex temporal construction, characterized by uncertain intervals that interrupt the folding of cinematic time into the flow of consciousness. Smithson’s film uses cinematic analogy to gesture towards vast reaches of geological time; Benning’s film creates a simulated timescale to evoke the short history of the earthwork itself. Smithson’s embrace of the entropic was a counter-cultural stance at the end of the1960s, but under the shadow of ecological disaster, this orientation has come to appear melancholy and romantic rather than radical. Benning’s film returns the jetty to anthropic time, but raises questions about the ways we inhabit time. His practice of working with ‘borrowed time’ is particularly suited to the cultural and historical moment of his later work.