29 resultados para Fourth-order methods
Resumo:
Two common methods of accounting for electric-field-induced perturbations to molecular vibration are analyzed and compared. The first method is based on a perturbation-theoretic treatment and the second on a finite-field treatment. The relationship between the two, which is not immediately apparent, is made by developing an algebraic formalism for the latter. Some of the higher-order terms in this development are documented here for the first time. As well as considering vibrational dipole polarizabilities and hyperpolarizabilities, we also make mention of the vibrational Stark effec
Resumo:
A procedure based on quantum molecular similarity measures (QMSM) has been used to compare electron densities obtained from conventional ab initio and density functional methodologies at their respective optimized geometries. This method has been applied to a series of small molecules which have experimentally known properties and molecular bonds of diverse degrees of ionicity and covalency. Results show that in most cases the electron densities obtained from density functional methodologies are of a similar quality than post-Hartree-Fock generalized densities. For molecules where Hartree-Fock methodology yields erroneous results, the density functional methodology is shown to yield usually more accurate densities than those provided by the second order Møller-Plesset perturbation theory
Resumo:
In the present paper we discuss and compare two different energy decomposition schemes: Mayer's Hartree-Fock energy decomposition into diatomic and monoatomic contributions [Chem. Phys. Lett. 382, 265 (2003)], and the Ziegler-Rauk dissociation energy decomposition [Inorg. Chem. 18, 1558 (1979)]. The Ziegler-Rauk scheme is based on a separation of a molecule into fragments, while Mayer's scheme can be used in the cases where a fragmentation of the system in clearly separable parts is not possible. In the Mayer scheme, the density of a free atom is deformed to give the one-atom Mulliken density that subsequently interacts to give rise to the diatomic interaction energy. We give a detailed analysis of the diatomic energy contributions in the Mayer scheme and a close look onto the one-atom Mulliken densities. The Mulliken density ρA has a single large maximum around the nuclear position of the atom A, but exhibits slightly negative values in the vicinity of neighboring atoms. The main connecting point between both analysis schemes is the electrostatic energy. Both decomposition schemes utilize the same electrostatic energy expression, but differ in how fragment densities are defined. In the Mayer scheme, the electrostatic component originates from the interaction of the Mulliken densities, while in the Ziegler-Rauk scheme, the undisturbed fragment densities interact. The values of the electrostatic energy resulting from the two schemes differ significantly but typically have the same order of magnitude. Both methods are useful and complementary since Mayer's decomposition focuses on the energy of the finally formed molecule, whereas the Ziegler-Rauk scheme describes the bond formation starting from undeformed fragment densities
Resumo:
This paper investigates the effects of women‘s labour force participation on fertility, as well as the effects of the combined labour force participation of both members of a couple. It specifically focuses on such dimensions as unemployment, earnings, temporary contracts and part-time jobs, and it shows that their effects differ in accordance with national institutions and labour market regulations. Event-history methods and a longitudinal sample of the European Community Household Panel are used in the analyses, concerning the years 1993-2000. The results show that labour market insecurity of one or both members of a couple has a particularly strong impact in reducing birth rates in the Southern European countries studied. The more conventional model of men’s employment combined with housewifery has a positive impact on second or higher order births in United Kingdom, Spain and Italy, while in Denmark the effect is the opposite. These differences are consistent with different national models of combining parental responsibilities and participation by gender across the life course.
Resumo:
Using a suitable Hull and White type formula we develop a methodology to obtain asecond order approximation to the implied volatility for very short maturities. Using thisapproximation we accurately calibrate the full set of parameters of the Heston model. Oneof the reasons that makes our calibration for short maturities so accurate is that we alsotake into account the term-structure for large maturities. We may say that calibration isnot "memoryless", in the sense that the option's behavior far away from maturity doesinfluence calibration when the option gets close to expiration. Our results provide a wayto perform a quick calibration of a closed-form approximation to vanilla options that canthen be used to price exotic derivatives. The methodology is simple, accurate, fast, andit requires a minimal computational cost.
Resumo:
A `next' operator, s, is built on the set R1=(0,1]-{ 1-1/e} defining a partial order that, with the help of the axiom of choice, can be extended to a total order in R1. Besides, the orbits {sn(a)}nare all dense in R1 and are constituted by elements of the samearithmetical character: if a is an algebraic irrational of degreek all the elements in a's orbit are algebraic of degree k; if a istranscendental, all are transcendental. Moreover, the asymptoticdistribution function of the sequence formed by the elements in anyof the half-orbits is a continuous, strictly increasing, singularfunction very similar to the well-known Minkowski's ?(×) function.
Resumo:
Consider the problem of testing k hypotheses simultaneously. In this paper,we discuss finite and large sample theory of stepdown methods that providecontrol of the familywise error rate (FWE). In order to improve upon theBonferroni method or Holm's (1979) stepdown method, Westfall and Young(1993) make eective use of resampling to construct stepdown methods thatimplicitly estimate the dependence structure of the test statistics. However,their methods depend on an assumption called subset pivotality. The goalof this paper is to construct general stepdown methods that do not requiresuch an assumption. In order to accomplish this, we take a close look atwhat makes stepdown procedures work, and a key component is a monotonicityrequirement of critical values. By imposing such monotonicity on estimatedcritical values (which is not an assumption on the model but an assumptionon the method), it is demonstrated that the problem of constructing a validmultiple test procedure which controls the FWE can be reduced to the problemof contructing a single test which controls the usual probability of a Type 1error. This reduction allows us to draw upon an enormous resamplingliterature as a general means of test contruction.
Resumo:
Amorphous thin films of Fe/Sm, prepared by evaporation methods, have been magnetically characterized and the results were interpreted in terms of the random magnets theory. The samples behave as 2D and 3D random magnets depending on the total thickness of the film. From our data the existence of orientational order, which greatly influences the magnetic behavior of the films, is also clear.
Resumo:
Amorphous thin films of Fe/Sm, prepared by evaporation methods, have been magnetically characterized and the results were interpreted in terms of the random magnets theory. The samples behave as 2D and 3D random magnets depending on the total thickness of the film. From our data the existence of orientational order, which greatly influences the magnetic behavior of the films, is also clear.
Resumo:
The work presented here is part of a larger study to identify novel technologies and biomarkers for early Alzheimer disease (AD) detection and it focuses on evaluating the suitability of a new approach for early AD diagnosis by non-invasive methods. The purpose is to examine in a pilot study the potential of applying intelligent algorithms to speech features obtained from suspected patients in order to contribute to the improvement of diagnosis of AD and its degree of severity. In this sense, Artificial Neural Networks (ANN) have been used for the automatic classification of the two classes (AD and control subjects). Two human issues have been analyzed for feature selection: Spontaneous Speech and Emotional Response. Not only linear features but also non-linear ones, such as Fractal Dimension, have been explored. The approach is non invasive, low cost and without any side effects. Obtained experimental results were very satisfactory and promising for early diagnosis and classification of AD patients.
Resumo:
This paper deals with the goodness of the Gaussian assumption when designing second-order blind estimationmethods in the context of digital communications. The low- andhigh-signal-to-noise ratio (SNR) asymptotic performance of the maximum likelihood estimator—derived assuming Gaussiantransmitted symbols—is compared with the performance of the optimal second-order estimator, which exploits the actualdistribution of the discrete constellation. The asymptotic study concludes that the Gaussian assumption leads to the optimalsecond-order solution if the SNR is very low or if the symbols belong to a multilevel constellation such as quadrature-amplitudemodulation (QAM) or amplitude-phase-shift keying (APSK). On the other hand, the Gaussian assumption can yield importantlosses at high SNR if the transmitted symbols are drawn from a constant modulus constellation such as phase-shift keying (PSK)or continuous-phase modulations (CPM). These conclusions are illustrated for the problem of direction-of-arrival (DOA) estimation of multiple digitally-modulated signals.
Resumo:
Forensic Anthropology and Bioarchaeology studies depend critically on the accuracy and reliability of age-estimation techniques. In this study we have evaluated two age-estimation methods for adults based on the pubic symphysis (Suchey-Brooks) and the auricular surface (Buckberry-Chamberlain) in a current sample of 139 individuals (67 women and 72 men) from Madrid in order to verify the accuracy of both methods applied to a sample of innominate bones from the central Iberian Peninsula. Based on the overall results of this study, the Buckberry-Chamberlain method seems to be the method that provides better estimates in terms of accuracy (percentage of hits) and absolute difference to the chronological age taking into account the total sample. The percentage of hits and mean absolute difference of the Buckberry-Chamberlain and Suchey-Brooks methods are 97.3% and 11.24 years, and 85.7% and 14.38 years, respectively. However, this apparently greater applicability of the Buckberry-Chamberlain method is mainly due to the broad age ranges provided. Results indicated that Suchey-Brooks method is more appropriate for populations with a majority of young individuals, whereas Buckberry-Chamberlain method is recommended for populations with a higher percentage of individuals in the range 60-70 years. These different age estimation methodologies significantly influence the resulting demographic profile, consequently affecting the biological characteristics reconstruction of the samples in which they are applied.
Resumo:
Mixed methods research involves the combined use of quantitative and qualitative methods in the same research study, and it is becoming increasingly important in several scientific areas. The aim of this paper is to review and compare through a mixed methods multiple-case study the application of this methodology in three reputable behavioural science journals: the Journal of Organizational Behavior, Addictive Behaviors and Psicothema. A quantitative analysis was carried out to review all the papers published in these journals during the period 2003-2008 and classify them into two blocks: theoretical and empirical, with the latter being further subdivided into three subtypes (quantitative, qualitative and mixed). A qualitative analysis determined the main characteristics of the mixed methods studies identified, in order to describe in more detail the ways in which the two methods are combined based on their purpose, priority, implementation and research design. From the journals selected, a total of 1.958 articles were analysed, the majority of which corresponded to empirical studies, with only a small number referring to research that used mixed methods. Nonetheless, mixed methods research does appear in all the behavioural science journals studied within the period selected, showing a range of designs, where the sequential equal weight mixed methods research design seems to stand out.