36 resultados para Ferro de valência zero

em CentAUR: Central Archive University of Reading - UK


Relevância:

20.00% 20.00%

Publicador:

Resumo:

The Kelvin Helmholtz (KH) problem, with zero stratification, is examined as a limiting case of the Rayleigh model of a single shear layer whose width tends to zero. The transition of the Rayleigh modal dispersion relation to the KH one, as well as the disappearance of the supermodal transient growth in the KH limit, are both rationalized from the counterpropagating Rossby wave perspective.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Using a flexible chemical box model with full heterogeneous chemistry, intercepts of chemically modified Langley plots have been computed for the 5 years of zenith-sky NO2 data from Faraday in Antarctica (65°S). By using these intercepts as the effective amount in the reference spectrum, drifts in zero of total vertical NO2 were much reduced. The error in zero of total NO2 is ±0.03×1015 moleccm−2 from one year to another. This error is small enough to determine trends in midsummer and any variability in denoxification between midwinters. The technique also suggests a more sensitive method for determining N2O5 from zenith-sky NO2 data.

Relevância:

20.00% 20.00%

Publicador:

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper we consider the estimation of population size from onesource capture–recapture data, that is, a list in which individuals can potentially be found repeatedly and where the question is how many individuals are missed by the list. As a typical example, we provide data from a drug user study in Bangkok from 2001 where the list consists of drug users who repeatedly contact treatment institutions. Drug users with 1, 2, 3, . . . contacts occur, but drug users with zero contacts are not present, requiring the size of this group to be estimated. Statistically, these data can be considered as stemming from a zero-truncated count distribution.We revisit an estimator for the population size suggested by Zelterman that is known to be robust under potential unobserved heterogeneity. We demonstrate that the Zelterman estimator can be viewed as a maximum likelihood estimator for a locally truncated Poisson likelihood which is equivalent to a binomial likelihood. This result allows the extension of the Zelterman estimator by means of logistic regression to include observed heterogeneity in the form of covariates. We also review an estimator proposed by Chao and explain why we are not able to obtain similar results for this estimator. The Zelterman estimator is applied in two case studies, the first a drug user study from Bangkok, the second an illegal immigrant study in the Netherlands. Our results suggest the new estimator should be used, in particular, if substantial unobserved heterogeneity is present.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Fixed transactions costs that prohibit exchange engender bias in supply analysis due to censoring of the sample observations. The associated bias in conventional regression procedures applied to censored data and the construction of robust methods for mitigating bias have been preoccupations of applied economists since Tobin [Econometrica 26 (1958) 24]. This literature assumes that the true point of censoring in the data is zero and, when this is not the case, imparts a bias to parameter estimates of the censored regression model. We conjecture that this bias can be significant; affirm this from experiments; and suggest techniques for mitigating this bias using Bayesian procedures. The bias-mitigating procedures are based on modifications of the key step that facilitates Bayesian estimation of the censored regression model; are easy to implement; work well in both small and large samples; and lead to significantly improved inference in the censored regression model. These findings are important in light of the widespread use of the zero-censored Tobit regression and we investigate their consequences using data on milk-market participation in the Ethiopian highlands. (C) 2004 Elsevier B.V. All rights reserved.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Estimation of population size with missing zero-class is an important problem that is encountered in epidemiological assessment studies. Fitting a Poisson model to the observed data by the method of maximum likelihood and estimation of the population size based on this fit is an approach that has been widely used for this purpose. In practice, however, the Poisson assumption is seldom satisfied. Zelterman (1988) has proposed a robust estimator for unclustered data that works well in a wide class of distributions applicable for count data. In the work presented here, we extend this estimator to clustered data. The estimator requires fitting a zero-truncated homogeneous Poisson model by maximum likelihood and thereby using a Horvitz-Thompson estimator of population size. This was found to work well, when the data follow the hypothesized homogeneous Poisson model. However, when the true distribution deviates from the hypothesized model, the population size was found to be underestimated. In the search of a more robust estimator, we focused on three models that use all clusters with exactly one case, those clusters with exactly two cases and those with exactly three cases to estimate the probability of the zero-class and thereby use data collected on all the clusters in the Horvitz-Thompson estimator of population size. Loss in efficiency associated with gain in robustness was examined based on a simulation study. As a trade-off between gain in robustness and loss in efficiency, the model that uses data collected on clusters with at most three cases to estimate the probability of the zero-class was found to be preferred in general. In applications, we recommend obtaining estimates from all three models and making a choice considering the estimates from the three models, robustness and the loss in efficiency. (© 2008 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim)

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Three new metal-organic polymeric complexes, [Fe(N-3)(2)(bPP)(2)] (1), [Fe(N-3)(2)(bpe)] (2), and [Fe(N-3)(2)(phen)] (3) [bpp = (1,3-bis(4-pyridyl)-propane), bpe = (1,2-bis(4-pyridyl)-ethane), phen = 1,10-phenanthroline], have been synthesized and characterized by single-crystal X-ray diffraction studies and low-temperature magnetic measurements in the range 300-2 K. Complexes 1 and 2 crystallize in the monoclinic system, space group C2/c, with the following cell parameters: a = 19.355(4) Angstrom, b = 7.076(2) Angstrom, c = 22.549(4) Angstrom, beta = 119.50(3)degrees, Z = 4, and a = 10.007(14) Angstrom, b = 13.789(18) Angstrom, c = 10.377(14) Angstrom, beta = 103.50(1)degrees, Z = 4, respectively. Complex 3 crystallizes in the triclinic system, space group P (1) over bar, with a = 7.155(12) Angstrom, b = 10.066(14) Angstrom, c = 10.508(14) Angstrom, alpha = 109.57(1)degrees, beta = 104.57(1)degrees, gamma = 105.10(1)degrees, and Z = 2. All coordination polymers exhibit octahedral Fe(II) nodes. The structural determination of 1 reveals a parallel interpenetrated structure of 2D layers of (4,4) topology, formed by Fe(II) nodes linked through bpp ligands, while mono-coordinated azide anions are pendant from the corrugated sheet. Complex 2 has a 2D arrangement constructed through 1D double end-to-end azide bridged iron(11) chains interconnected through bpe ligands. Complex 3 shows a polymeric arrangement where the metal ions are interlinked through pairs of end-on and end-to-end azide ligands exhibiting a zigzag arrangement of metals (Fe-Fe-Fe angle of 111.18degrees) and an intermetallic separation of 3.347 Angstrom (through the EO azide) and of 5.229 Angstrom (EE azide). Variable-temperature magnetic susceptibility data suggest that there is no magnetic interaction between the metal centers in 1, whereas in 2 there is an antiferromagnetic interaction through the end-to-end azide bridge. Complex 3 shows ferro- as well as anti-ferromagnetic interactions between the metal centers generated through the alternating end-on and end-to-end azide bridges. Complex I has been modeled using the D parameter (considering distorted octahedral Fe(II) geometry and with any possible J value equal to zero) and complex 2 has been modeled as a one-dimensional system with classical and/or quantum spin where we have used two possible full diagonalization processes: without and with the D parameter, considering the important distortions of the Fe(II) ions. For complex 3, the alternating coupling model impedes a mathematical solution for the modeling as classical spins. With quantum spin, the modeling has been made as in 2.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

[Cu4L2(bpy)(4)(H2O)(3)](ClO4)(4).2.5H(2)O, 1, a new tetranuclear Cu-II cluster showing square planar geometry, formed with aspartate bridging ligand (L) has been synthesized. The global magnetic coupling is ferromagnetic but theoretical DFT/B3LYP calculations are necessary to assign which Cu-L-Cu side is ferro or antiferromagnetically coupled.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In the forecasting of binary events, verification measures that are “equitable” were defined by Gandin and Murphy to satisfy two requirements: 1) they award all random forecasting systems, including those that always issue the same forecast, the same expected score (typically zero), and 2) they are expressible as the linear weighted sum of the elements of the contingency table, where the weights are independent of the entries in the table, apart from the base rate. The authors demonstrate that the widely used “equitable threat score” (ETS), as well as numerous others, satisfies neither of these requirements and only satisfies the first requirement in the limit of an infinite sample size. Such measures are referred to as “asymptotically equitable.” In the case of ETS, the expected score of a random forecasting system is always positive and only falls below 0.01 when the number of samples is greater than around 30. Two other asymptotically equitable measures are the odds ratio skill score and the symmetric extreme dependency score, which are more strongly inequitable than ETS, particularly for rare events; for example, when the base rate is 2% and the sample size is 1000, random but unbiased forecasting systems yield an expected score of around −0.5, reducing in magnitude to −0.01 or smaller only for sample sizes exceeding 25 000. This presents a problem since these nonlinear measures have other desirable properties, in particular being reliable indicators of skill for rare events (provided that the sample size is large enough). A potential way to reconcile these properties with equitability is to recognize that Gandin and Murphy’s two requirements are independent, and the second can be safely discarded without losing the key advantages of equitability that are embodied in the first. This enables inequitable and asymptotically equitable measures to be scaled to make them equitable, while retaining their nonlinearity and other properties such as being reliable indicators of skill for rare events. It also opens up the possibility of designing new equitable verification measures.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A sparse kernel density estimator is derived based on the zero-norm constraint, in which the zero-norm of the kernel weights is incorporated to enhance model sparsity. The classical Parzen window estimate is adopted as the desired response for density estimation, and an approximate function of the zero-norm is used for achieving mathemtical tractability and algorithmic efficiency. Under the mild condition of the positive definite design matrix, the kernel weights of the proposed density estimator based on the zero-norm approximation can be obtained using the multiplicative nonnegative quadratic programming algorithm. Using the -optimality based selection algorithm as the preprocessing to select a small significant subset design matrix, the proposed zero-norm based approach offers an effective means for constructing very sparse kernel density estimates with excellent generalisation performance.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The optimal and the zero-forcing beamformers are two commonly used algorithms in the subspace-based blind beamforming technology. The optimal beamformer is regarded as the algorithm with the best output SINR. The zero-forcing algorithm emphasizes the co-channel interference cancellation. This paper compares the performance of these two algorithms under some practical conditions: the effect of the finite data length and the existence of the angle estimation error. The investigation reveals that the zero-forcing algorithm can be more robust in the practical environment than the optimal algorithm.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A new sparse kernel probability density function (pdf) estimator based on zero-norm constraint is constructed using the classical Parzen window (PW) estimate as the target function. The so-called zero-norm of the parameters is used in order to achieve enhanced model sparsity, and it is suggested to minimize an approximate function of the zero-norm. It is shown that under certain condition, the kernel weights of the proposed pdf estimator based on the zero-norm approximation can be updated using the multiplicative nonnegative quadratic programming algorithm. Numerical examples are employed to demonstrate the efficacy of the proposed approach.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Following the US model, the UK has seen considerable innovation in the funding, finance and procurement of real estate in the last decade. In the growing CMBS market asset backed securitisations have included $2.25billion secured on the Broadgate office development and issues secured on Canary Wharf and the Trafford Centre regional mall. Major occupiers (retailer Sainsbury’s, retail bank Abbey National) have engaged in innovative sale & leaseback and outsourcing schemes. Strong claims are made concerning the benefits of such schemes – e.g. British Land were reported to have reduced their weighted cost of debt by 150bp as a result of the Broadgate issue. The paper reports preliminary findings from a project funded by the Corporation of London and the RICS Research Foundation examining a number of innovative schemes to identify, within a formal finance framework, sources of added value and hidden costs. The analysis indicates that many of the gains claimed conceal costs – in terms of market value of debt or flexibility of management – while others result from unusual firm or market conditions (for example utilising the UK long lease and the unusual shape of the yield curve). Nonetheless, there are real gains resulting from the innovations, reflecting arbitrage and institutional constraints in the direct (private) real estate market