976 resultados para pseudo-groupe de Lie


Relevância:

20.00% 20.00%

Publicador:

Resumo:

The crystal structure of the modified unsymmetrically N, N'-substituted viologen chromophore, N-ethyl- N'-(2-phosphonoethyl)-4, 4'-bipyridinium dichloride 0.75 hydrate. (1) has been determined. Crystals are triclinic, space group P-1 with Z = 2 in a cell with a = 7.2550(1), b = 13.2038(5), c = 18.5752(7) Å, α = 86.495(3), β = 83.527(2), γ = 88.921(2)o. The two independent but pseudo-symmetrically related cations in the asymmetric unit form one-dimensional hydrogen-bonded chains through short homomeric phosphonic acid O-H...O links [2.455(4), 2.464(4)A] while two of the chloride anions are similarly strongly linked to phosphonic acid groups [O-H…Cl, 2.889(4), 2.896(4)Å]. The other two chloride anions together with the two water molecules of solvation (one with partial occupancy) form unusual cyclic hydrogen-bonded bis(Cl...water) dianion units which lie between the layers of bipyridylium rings of the cation chain structures with which they are weakly associated.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This article introduces a “pseudo classical” notion of modelling non-separability. This form of non-separability can be viewed as lying between separability and quantum-like non-separability. Non-separability is formalized in terms of the non-factorizabilty of the underlying joint probability distribution. A decision criterium for determining the non-factorizability of the joint distribution is related to determining the rank of a matrix as well as another approach based on the chi-square-goodness-of-fit test. This pseudo-classical notion of non-separability is discussed in terms of quantum games and concept combinations in human cognition.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The panel "Duplicity/Complicity: Performing and Misperforming Lies" at PSi #15 in Croatia in July 2009 examined the half-truths, hidden assumptions and power relations embedded in every act of performance through an analysis of the way bodies, buildings, personae and communities perform and misperform lies. It was a collection of new academic voices from Australia and Croatia, intersecting and colliding and, at times, outright lying, with each other and with commentary from Alan Read. Inspired by this successful adventure in collaborative academic mis-performance, "The ‘Dirty Work’ of the Lie" takes the challenge set by the Prelude Panel at PSI #15 and subjects the ideas emerging from this panel to "friendly fire" in order to build a multi authored response to 'performance that lies', with reference to the work of A Chorus of Women, disabled artists Bill Shannon, Aaron Williamson and Kathryn Araneillo, US dance performer Ann Liv Young and US theatre and festival director Peter Sellars. In doing so, "The 'Dirty Work' of the Lie" provides a reflexive response to the duplicity inherent in the performances, and also in our own academic analyses. With Alan Read acting as interlocutor, each contributor will creatively respond to a paper presented by another, developing the key intersecting issues that emerged through the formation of the panel. These issues include impression management, self-belief and performers who are 'taken in by their own act', the dirty work of taking others in with an act, the guerrilla dimension of lying, the productivity of the lie, and questions of audience engagement and ethics. As a result, this new paper tests how the 'misperformance' of lies across different cultural sites, be it deliberate or accidental, can become a productive – and, indeed, politicised – aspect of cultural performance, betraying accepted attitudes, ideas and structures of authority and offering alternative visions. Through it’s distinctively multi vocal texture, "The 'Dirty Work' of the Lie" also interrogates the modes of analysis available to us, questioning the 'duplicity' in our reflecting, responding and listening to each other as well as the work.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The Airy stress function, although frequently employed in classical linear elasticity, does not receive similar usage for granular media problems. For plane strain quasi-static deformations of a cohesionless Coulomb–Mohr granular solid, a single nonlinear partial differential equation is formulated for the Airy stress function by combining the equilibrium equations with the yield condition. This has certain advantages from the usual approach, in which two stress invariants and a stress angle are introduced, and a system of two partial differential equations is needed to describe the flow. In the present study, the symmetry analysis of differential equations is utilised for our single partial differential equation, and by computing an optimal system of one-dimensional Lie algebras, a complete set of group-invariant solutions is derived. By this it is meant that any group-invariant solution of the governing partial differential equation (provided it can be derived via the classical symmetries method) may be obtained as a member of this set by a suitable group transformation. For general values of the parameters (angle of internal friction and gravity g) it is found there are three distinct classes of solutions which correspond to granular flows considered previously in the literature. For the two limiting cases of high angle of internal friction and zero gravity, the governing partial differential equation admit larger families of Lie point symmetries, and from these symmetries, further solutions are derived, many of which are new. Furthermore, the majority of these solutions are exact, which is rare for granular flow, especially in the case of gravity driven flows.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The structure of the pseudo-merohedrally twinned crystal of the 1:1 proton-transfer compound of 5-sulfosalicylic acid (3-carboxy-4-hydroxybenzenesulfonic acid) with 4-aminopyridine: 4-aminopyridinium 3-carboxy-4-hydroxybenzenesulfonate sesquihydrate has been determined at 180 K and the hydrogen-bonding pattern is described. Crystals of the compound are monoclinic with space group P21/c, with unit cell dimensions a = 35.2589(8), b = 7.1948(1), c = 24.5851(5) Å, β = 110.373(2)o, and Z = 16. The monoclinic asymmetric unit comprises four cation-anion pairs and six water molecules of solvation with only the pyridinium cations having pseudo-symmetry as a result of inter-cation aromatic ring π-π stacking effects. Extensive hydrogen bonding gives a three-dimensional framework structure.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Freeways are divided roadways designed to facilitate the uninterrupted movement of motor vehicles. However, many freeways now experience demand flows in excess of capacity, leading to recurrent congestion. The Highway Capacity Manual (TRB, 1994) uses empirical macroscopic relationships between speed, flow and density to quantify freeway operations and performance. Capacity may be predicted as the maximum uncongested flow achievable. Although they are effective tools for design and analysis, macroscopic models lack an understanding of the nature of processes taking place in the system. Szwed and Smith (1972, 1974) and Makigami and Matsuo (1990) have shown that microscopic modelling is also applicable to freeway operations. Such models facilitate an understanding of the processes whilst providing for the assessment of performance, through measures of capacity and delay. However, these models are limited to only a few circumstances. The aim of this study was to produce more comprehensive and practical microscopic models. These models were required to accurately portray the mechanisms of freeway operations at the specific locations under consideration. The models needed to be able to be calibrated using data acquired at these locations. The output of the models needed to be able to be validated with data acquired at these sites. Therefore, the outputs should be truly descriptive of the performance of the facility. A theoretical basis needed to underlie the form of these models, rather than empiricism, which is the case for the macroscopic models currently used. And the models needed to be adaptable to variable operating conditions, so that they may be applied, where possible, to other similar systems and facilities. It was not possible to produce a stand-alone model which is applicable to all facilities and locations, in this single study, however the scene has been set for the application of the models to a much broader range of operating conditions. Opportunities for further development of the models were identified, and procedures provided for the calibration and validation of the models to a wide range of conditions. The models developed, do however, have limitations in their applicability. Only uncongested operations were studied and represented. Driver behaviour in Brisbane was applied to the models. Different mechanisms are likely in other locations due to variability in road rules and driving cultures. Not all manoeuvres evident were modelled. Some unusual manoeuvres were considered unwarranted to model. However the models developed contain the principal processes of freeway operations, merging and lane changing. Gap acceptance theory was applied to these critical operations to assess freeway performance. Gap acceptance theory was found to be applicable to merging, however the major stream, the kerb lane traffic, exercises only a limited priority over the minor stream, the on-ramp traffic. Theory was established to account for this activity. Kerb lane drivers were also found to change to the median lane where possible, to assist coincident mergers. The net limited priority model accounts for this by predicting a reduced major stream flow rate, which excludes lane changers. Cowan's M3 model as calibrated for both streams. On-ramp and total upstream flow are required as input. Relationships between proportion of headways greater than 1 s and flow differed for on-ramps where traffic leaves signalised intersections and unsignalised intersections. Constant departure onramp metering was also modelled. Minimum follow-on times of 1 to 1.2 s were calibrated. Critical gaps were shown to lie between the minimum follow-on time, and the sum of the minimum follow-on time and the 1 s minimum headway. Limited priority capacity and other boundary relationships were established by Troutbeck (1995). The minimum average minor stream delay and corresponding proportion of drivers delayed were quantified theoretically in this study. A simulation model was constructed to predict intermediate minor and major stream delays across all minor and major stream flows. Pseudo-empirical relationships were established to predict average delays. Major stream average delays are limited to 0.5 s, insignificant compared with minor stream delay, which reach infinity at capacity. Minor stream delays were shown to be less when unsignalised intersections are located upstream of on-ramps than signalised intersections, and less still when ramp metering is installed. Smaller delays correspond to improved merge area performance. A more tangible performance measure, the distribution of distances required to merge, was established by including design speeds. This distribution can be measured to validate the model. Merging probabilities can be predicted for given taper lengths, a most useful performance measure. This model was also shown to be applicable to lane changing. Tolerable limits to merging probabilities require calibration. From these, practical capacities can be estimated. Further calibration is required of traffic inputs, critical gap and minimum follow-on time, for both merging and lane changing. A general relationship to predict proportion of drivers delayed requires development. These models can then be used to complement existing macroscopic models to assess performance, and provide further insight into the nature of operations.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The emergence of pseudo-marginal algorithms has led to improved computational efficiency for dealing with complex Bayesian models with latent variables. Here an unbiased estimator of the likelihood replaces the true likelihood in order to produce a Bayesian algorithm that remains on the marginal space of the model parameter (with latent variables integrated out), with a target distribution that is still the correct posterior distribution. Very efficient proposal distributions can be developed on the marginal space relative to the joint space of model parameter and latent variables. Thus psuedo-marginal algorithms tend to have substantially better mixing properties. However, for pseudo-marginal approaches to perform well, the likelihood has to be estimated rather precisely. This can be difficult to achieve in complex applications. In this paper we propose to take advantage of multiple central processing units (CPUs), that are readily available on most standard desktop computers. Here the likelihood is estimated independently on the multiple CPUs, with the ultimate estimate of the likelihood being the average of the estimates obtained from the multiple CPUs. The estimate remains unbiased, but the variability is reduced. We compare and contrast two different technologies that allow the implementation of this idea, both of which require a negligible amount of extra programming effort. The superior performance of this idea over the standard approach is demonstrated on simulated data from a stochastic volatility model.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Application of `advanced analysis' methods suitable for non-linear analysis and design of steel frame structures permits direct and accurate determination of ultimate system strengths, without resort to simplified elastic methods of analysis and semi-empirical specification equations. However, the application of advanced analysis methods has previously been restricted to steel frames comprising only compact sections that are not influenced by the effects of local buckling. A concentrated plasticity method suitable for practical advanced analysis of steel frame structures comprising non-compact sections is presented in this paper. The pseudo plastic zone method implicitly accounts for the effects of gradual cross-sectional yielding, longitudinal spread of plasticity, initial geometric imperfections, residual stresses, and local buckling. The accuracy and precision of the method for the analysis of steel frames comprising non-compact sections is established by comparison with a comprehensive range of analytical benchmark frame solutions. The pseudo plastic zone method is shown to be more accurate and precise than the conventional individual member design methods based on elastic analysis and specification equations.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Monitoring stream networks through time provides important ecological information. The sampling design problem is to choose locations where measurements are taken so as to maximise information gathered about physicochemical and biological variables on the stream network. This paper uses a pseudo-Bayesian approach, averaging a utility function over a prior distribution, in finding a design which maximizes the average utility. We use models for correlations of observations on the stream network that are based on stream network distances and described by moving average error models. Utility functions used reflect the needs of the experimenter, such as prediction of location values or estimation of parameters. We propose an algorithmic approach to design with the mean utility of a design estimated using Monte Carlo techniques and an exchange algorithm to search for optimal sampling designs. In particular we focus on the problem of finding an optimal design from a set of fixed designs and finding an optimal subset of a given set of sampling locations. As there are many different variables to measure, such as chemical, physical and biological measurements at each location, designs are derived from models based on different types of response variables: continuous, counts and proportions. We apply the methodology to a synthetic example and the Lake Eacham stream network on the Atherton Tablelands in Queensland, Australia. We show that the optimal designs depend very much on the choice of utility function, varying from space filling to clustered designs and mixtures of these, but given the utility function, designs are relatively robust to the type of response variable.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Ever since Cox et. al published their paper, “A Secure, Robust Watermark for Multimedia” in 1996 [6], there has been tremendous progress in multimedia watermarking. The same pattern re-emerged with Agrawal and Kiernan publishing their work “Watermarking Relational Databases” in 2001 [1]. However, little attention has been given to primitive data collections with only a handful works of research known to the authors [11, 10]. This is primarily due to the absence of an attribute that differentiates marked items from unmarked item during insertion and detection process. This paper presents a distribution-independent, watermarking model that is secure against secondary-watermarking in addition to conventional attacks such as data addition, deletion and distortion. The low false positives and high capacity provide additional strength to the scheme. These claims are backed by experimental results provided in the paper.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In Crypto’95, Micali and Sidney proposed a method for shared generation of a pseudo-random function f(·) among n players in such a way that for all the inputs x, any u players can compute f(x) while t or fewer players fail to do so, where 0⩽tdea behind the Micali–Sidney scheme is to generate and distribute secret seeds S={s1,…,sd} of a poly-random collection of functions, among the n players, each player gets a subset of S, in such a way that any u players together hold all the secret seeds in S while any t or fewer players will lack at least one element from S. The pseudo-random function is then computed as where fsi(·)'s are poly-random functions. One question raised by Micali and Sidney is how to distribute the secret seeds satisfying the above condition such that the number of seeds, d, is as small as possible. In this paper, we continue the work of Micali and Sidney. We first provide a general framework for shared generation of pseudo-random function using cumulative maps. We demonstrate that the Micali–Sidney scheme is a special case of this general construction. We then derive an upper and a lower bound for d. Finally we give a simple, yet efficient, approximation greedy algorithm for generating the secret seeds S in which d is close to the optimum by a factor of at most u ln 2.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In Crypto’95, Micali and Sidney proposed a method for shared generation of a pseudo-random function f(·) among n players in such a way that for all the inputs x, any u players can compute f(x) while t or fewer players fail to do so, where 0 ≤ t < u ≤ n. The idea behind the Micali-Sidney scheme is to generate and distribute secret seeds S = s1, . . . , sd of a poly-random collection of functions, among the n players, each player gets a subset of S, in such a way that any u players together hold all the secret seeds in S while any t or fewer players will lack at least one element from S. The pseudo-random function is then computed as where f s i (·)’s are poly-random functions. One question raised by Micali and Sidney is how to distribute the secret seeds satisfying the above condition such that the number of seeds, d, is as small as possible. In this paper, we continue the work of Micali and Sidney. We first provide a general framework for shared generation of pseudo-random function using cumulative maps. We demonstrate that the Micali-Sidney scheme is a special case of this general construction.We then derive an upper and a lower bound for d. Finally we give a simple, yet efficient, approximation greedy algorithm for generating the secret seeds S in which d is close to the optimum by a factor of at most u ln 2.