856 resultados para time-depedency in tunnelling
Resumo:
While investigations using covert food manipulations tend to suggest that individuals are poor at adjusting for previous energy intake, in the real world adults rarely consume foods of which they are ill-informed. This study investigated the impact in fully complicit consumers of consuming commercially available dark chocolate, milk chocolate, sweet biscuits and fruit bars on subsequent appetite. Using a repeated measures design, participants received four small portions (4 × 10-11 g) of either dark chocolate, milk chocolate, sweet biscuits, fruit bars or no food throughout five separate study days (counterbalanced in order), and test meal intake, hunger, liking and acceptability were measured. Participants consumed significantly less at lunch following dark chocolate, milk chocolate and sweet biscuits compared to no food (smallest t(19) = 2.47, p = 0.02), demonstrating very good energy compensation (269-334%). No effects were found for fruit bars (t(19) = 1.76, p = 0.09), in evening meal intakes (F(4,72) = 0.62, p = 0.65) or in total intake (lunch + evening meal + food portions) (F(4,72) = 0.40, p = 0.69). No differences between conditions were found in measures of hunger (largest F(4,76) = 1.26, p = 0.29), but fruit bars were significantly less familiar than all other foods (smallest t(19) = 3.14, p = 0.01). These findings demonstrate good compensation over the short term for small portions of familiar foods in complicit consumers. Findings are most plausibly explained as a result of participant awareness and cognitions, although the nature of these cognitions cannot be discerned from this study. These findings however, also suggest that covert manipulations may have limited transfer to real world scenarios.
Resumo:
Tese de doutoramento, Informática (Engenharia Informática), Universidade de Lisboa, Faculdade de Ciências, 2014
Resumo:
Tese de Doutoramento em Biologia apresentada à Faculdade de Ciências da Universidade do Porto, 2015.
Resumo:
Financial time series have a complex dynamic nature. Many techniques were adopted having in mind standard paradigms of time flow. This paper explores an alternative route involving relativistic effects. It is observed that the measuring perspective influences the results and that we can have different time textures.
Resumo:
The use of demand response programs enables the adequate use of resources of small and medium players, bringing high benefits to the smart grid, and increasing its efficiency. One of the difficulties to proceed with this paradigm is the lack of intelligence in the management of small and medium size players. In order to make demand response programs a feasible solution, it is essential that small and medium players have an efficient energy management and a fair optimization mechanism to decrease the consumption without heavy loss of comfort, making it acceptable for the users. This paper addresses the application of real-time pricing in a house that uses an intelligent optimization module involving artificial neural networks.
Resumo:
Real-time monitoring applications may be used in a wireless sensor network (WSN) and may generate packet flows with strict quality of service requirements in terms of delay, jitter, or packet loss. When strict delays are imposed from source to destination, the packets must be delivered at the destination within an end-to-end delay (EED) hard limit in order to be considered useful. Since the WSN nodes are scarce both in processing and energy resources, it is desirable that they only transport useful data, as this contributes to enhance the overall network performance and to improve energy efficiency. In this paper, we propose a novel cross-layer admission control (CLAC) mechanism to enhance the network performance and increase energy efficiency of a WSN, by avoiding the transmission of potentially useless packets. The CLAC mechanism uses an estimation technique to preview packets EED, and decides to forward a packet only if it is expected to meet the EED deadline defined by the application, dropping it otherwise. The results obtained show that CLAC enhances the network performance by increasing the useful packet delivery ratio in high network loads and improves the energy efficiency in every network load.
Resumo:
One hundred and thirty four subjects participated in this survey. Quantitative data were obtained and correlational analyses were used to test a model to study the relationships among the achievement of work values and organizational commitment and job satisfaction and to identify the moderating effects of the meaningfulness of work and responsibility for work on these relationships. Part-time faculty in the Faculty of Continuing Education of a community college were mailed a questionnaire on all the variables of the model. Several reliable, valid instruments were used to test the variables. Data analysis through Pearson correlation and stepwise multiple regression analyses revealed that the achievement of the work values of recognition and satisfaction with promotions did predict organizational commitment and job satisfaction, although the moderating effects of the meaningfulness of work and responsibility for work was not supported in this study. This study suggests that the revised model may be used for determining the relationships between the achievement of work values and organizational commitment and job satisfaction in a community college setting.
Resumo:
Groupe de recherche sur le système nerveux central, Département d'informatique et de recherche opérationnelle, Département de physiologie.
Inference for nonparametric high-frequency estimators with an application to time variation in betas
Resumo:
We consider the problem of conducting inference on nonparametric high-frequency estimators without knowing their asymptotic variances. We prove that a multivariate subsampling method achieves this goal under general conditions that were not previously available in the literature. We suggest a procedure for a data-driven choice of the bandwidth parameters. Our simulation study indicates that the subsampling method is much more robust than the plug-in method based on the asymptotic expression for the variance. Importantly, the subsampling method reliably estimates the variability of the Two Scale estimator even when its parameters are chosen to minimize the finite sample Mean Squared Error; in contrast, the plugin estimator substantially underestimates the sampling uncertainty. By construction, the subsampling method delivers estimates of the variance-covariance matrices that are always positive semi-definite. We use the subsampling method to study the dynamics of financial betas of six stocks on the NYSE. We document significant variation in betas within year 2006, and find that tick data captures more variation in betas than the data sampled at moderate frequencies such as every five or twenty minutes. To capture this variation we estimate a simple dynamic model for betas. The variance estimation is also important for the correction of the errors-in-variables bias in such models. We find that the bias corrections are substantial, and that betas are more persistent than the naive estimators would lead one to believe.
Resumo:
We propose to show in this paper, that the time series obtained from biological systems such as human brain are invariably nonstationary because of different time scales involved in the dynamical process. This makes the invariant parameters time dependent. We made a global analysis of the EEG data obtained from the eight locations on the skull space and studied simultaneously the dynamical characteristics from various parts of the brain. We have proved that the dynamical parameters are sensitive to the time scales and hence in the study of brain one must identify all relevant time scales involved in the process to get an insight in the working of brain.
Resumo:
This thesis Entitled Studies on Quasinormal modes and Late-time tails black hole spacetimes. In this thesis, the signature of these new theories are probed on the evolution of field perturbations on the black hole spacetimes in the theory. Chapter 1 gives a general introduction to black holes and its perturbation formalism. Various concepts in the area covered by the thesis are also elucidated in this chapter. Chapter 2 describes the evolution of massive, charged scalar field perturbations around a Reissner-Nordstrom black hole surrounded by a static and spherically symmetric quintessence. Chapter 3 comprises the evolution of massless scalar, electromagnetic and gravitational fields around spherically symmetric black hole whose asymptotes are defined by the quintessence, with special interest on the late-time behavior. Chapter 4 examines the evolution of Dirac field around a Schwarzschild black hole surrounded by quintessence. Detailed numerical simulations are done to analyze the nature of field on different surfaces of constant radius . Chapter 5is dedicated to the study of the evolution of massless fields around the black hole geometry in the HL gravity.
Resumo:
The term reliability of an equipment or device is often meant to indicate the probability that it carries out the functions expected of it adequately or without failure and within specified performance limits at a given age for a desired mission time when put to use under the designated application and operating environmental stress. A broad classification of the approaches employed in relation to reliability studies can be made as probabilistic and deterministic, where the main interest in the former is to device tools and methods to identify the random mechanism governing the failure process through a proper statistical frame work, while the latter addresses the question of finding the causes of failure and steps to reduce individual failures thereby enhancing reliability. In the probabilistic attitude to which the present study subscribes to, the concept of life distribution, a mathematical idealisation that describes the failure times, is fundamental and a basic question a reliability analyst has to settle is the form of the life distribution. It is for no other reason that a major share of the literature on the mathematical theory of reliability is focussed on methods of arriving at reasonable models of failure times and in showing the failure patterns that induce such models. The application of the methodology of life time distributions is not confined to the assesment of endurance of equipments and systems only, but ranges over a wide variety of scientific investigations where the word life time may not refer to the length of life in the literal sense, but can be concieved in its most general form as a non-negative random variable. Thus the tools developed in connection with modelling life time data have found applications in other areas of research such as actuarial science, engineering, biomedical sciences, economics, extreme value theory etc.
Resumo:
El desalineamiento temporal es la incorrespondencia de dos señales debido a una distorsión en el eje temporal. La Detección y Diagnóstico de Fallas (Fault Detection and Diagnosis-FDD) permite la detección, el diagnóstico y la corrección de fallos en un proceso. La metodología usada en FDD está dividida en dos categorías: técnicas basadas en modelos y no basadas en modelos. Esta tesis doctoral trata sobre el estudio del efecto del desalineamiento temporal en FDD. Nuestra atención se enfoca en el análisis y el diseño de sistemas FDD en caso de problemas de comunicación de datos, como retardos y pérdidas. Se proponen dos técnicas para reducir estos problemas: una basada en programación dinámica y la otra en optimización. Los métodos propuestos han sido validados sobre diferentes sistemas dinámicos: control de posición de un motor de corriente continua, una planta de laboratorio y un problema de sistemas eléctricos conocido como hueco de tensión.
Resumo:
A total of 86 profiles from meat and egg strains of chickens (male and female) were used in this study. Different flexible growth functions were evaluated with regard to their ability to describe the relationship between live weight and age and were compared with the Gompertz and logistic equations, which have a fixed point of inflection. Six growth functions were used: Gompertz, logistic, Lopez, Richards, France, and von Bertalanffy. A comparative analysis was carried out based on model behavior and statistical performance. The results of this study confirmed the initial concern about the limitation of a fixed point of inflection, such as in the Gompertz equation. Therefore, consideration of flexible growth functions as an alternatives to the simpler equations (with a fixed point of inflection) for describing the relationship between live weight and age are recommended for the following reasons: they are easy to fit, they very often give a closer fit to data points because of their flexibility and therefore a smaller RSS value, than the simpler models, and they encompasses simpler models for the addition of an extra parameter, which is especially important when the behavior of a particular data set is not defined previously.
Resumo:
Capturing the pattern of structural change is a relevant task in applied demand analysis, as consumer preferences may vary significantly over time. Filtering and smoothing techniques have recently played an increasingly relevant role. A dynamic Almost Ideal Demand System with random walk parameters is estimated in order to detect modifications in consumer habits and preferences, as well as changes in the behavioural response to prices and income. Systemwise estimation, consistent with the underlying constraints from economic theory, is achieved through the EM algorithm. The proposed model is applied to UK aggregate consumption of alcohol and tobacco, using quarterly data from 1963 to 2003. Increased alcohol consumption is explained by a preference shift, addictive behaviour and a lower price elasticity. The dynamic and time-varying specification is consistent with the theoretical requirements imposed at each sample point. (c) 2005 Elsevier B.V. All rights reserved.