990 resultados para Zerocrossing sampling theory,


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Closing feedback loops using an IEEE 802.11b ad hoc wireless communication network incurs many challenges sensitivity to varying channel conditions and lower physical transmission rates tend to limit the bandwidth of the communication channel. Given that the bandwidth usage and control performance are linked, a method of adapting the sampling interval based on an 'a priori', static sampling policy has been proposed and, more significantly, assuring stability in the mean square sense using discrete-time Markov jump linear system theory. Practical issues including current limitations of the 802.11 b protocol, the sampling policy and stability are highlighted. Simulation results on a cart-mounted inverted pendulum show that closed-loop stability can be improved using sample rate adaptation and that the control design criteria can be met in the presence of channel errors and severe channel contention.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Beta diversity describes how local communities within an area or region differ in species composition/abundance. There have been attempts to use changes in beta diversity as a biotic indicator of disturbance, but lack of theory and methodological caveats have hampered progress. We here propose that the neutral theory of biodiversity plus the definition of beta diversity as the total variance of a community matrix provide a suitable, novel, starting point for ecological applications. Observed levels of beta diversity (BD) can be compared to neutral predictions with three possible outcomes: Observed BD equals neutral prediction or is larger (divergence) or smaller (convergence) than the neutral prediction. Disturbance might lead to either divergence or convergence, depending on type and strength. We here apply these ideas to datasets collected on oribatid mites (a key, very diverse soil taxon) under several regimes of disturbances. When disturbance is expected to increase the heterogeneity of soil spatial properties or the sampling strategy encompassed a range of diverging environmental conditions, we observed diverging assemblages. On the contrary, we observed patterns consistent with neutrality when disturbance could determine homogenization of soil properties in space or the sampling strategy encompassed fairly homogeneous areas. With our method, spatial and temporal changes in beta diversity can be directly and easily monitored to detect significant changes in community dynamics, although the method itself cannot inform on underlying mechanisms. However, human-driven disturbances and the spatial scales at which they operate are usually known. In this case, our approach allows the formulation of testable predictions in terms of expected changes in beta diversity, thereby offering a promising monitoring tool.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A review of the literature reveals that there are a number of children in the educational system who are characterized by Attention Deficit Disorder. Further review of the literature reveals that there are information processing programs which have had some success in increasing the learning of these children. Currently, an information processing program which is based on schema theory is being implemented in Lincoln County. Since schema theory based programs build structural, conditional, factual, and procedural schemata which assist the learner in attending to salient factors, learning should be increased. Thirty-four children were selected from a random sampling of Grade Seven classes in Lincoln County. Seventeen of these children were identified by the researcher and classroom teacher as being characterized by Attention Deficit Disorder. From the remaining population, 17 children who were not characterized by Attention Deficit Disorder were randomly selected. The data collected were compared using independent t-tests, paired t-tests, and correlation analysis. Significant differences were found in all cases. The Non-Attention Deficit Disorder children scored significantly higher on all the tests but the Attention Defici t Disorder children had a significantly higher ratio of gain between the pretests and posttests.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The attached file is created with Scientific Workplace Latex

Relevância:

30.00% 30.00%

Publicador:

Resumo:

La présente thèse se base sur les principes de la théorisation ancrée (Strauss & Corbin, 1998) afin de répondre au manque de documentation concernant les stratégies adoptées par des « agents intermédiaires » pour promouvoir l’utilisation des connaissances issues de la recherche auprès des intervenants en éducation. Le terme « agent intermédiaire » réfère aux personnes qui sont positionnées à l’interface entre les producteurs et les utilisateurs des connaissances scientifiques et qui encouragent et soutiennent les intervenants scolaires dans l’application des connaissances scientifiques dans leur pratique. L’étude s’inscrit dans le cadre d’un projet du ministère de l’Éducation, du Loisir et du Sport du Québec visant à améliorer la réussite scolaire des élèves du secondaire provenant de milieux défavorisés. Des agents intermédiaires de différents niveaux du système éducatif ayant obtenu le mandat de transférer des connaissances issues de la recherche auprès des intervenants scolaires dans les écoles visées par le projet ont été sollicités pour participer à l’étude. Une stratégie d’échantillonnage de type « boule-de-neige » (Biernacki & Waldorf, 1981; Patton, 1990) a été employée afin d’identifier les personnes reconnues par leurs pairs pour la qualité du soutien offert aux intervenants scolaires quant à l’utilisation de la recherche dans leur pratique. Seize entrevues semi-structurées ont été réalisées. L’analyse des données permet de proposer un modèle d’intervention en transfert de connaissances composé de 32 stratégies d’influence, regroupées en 6 composantes d’intervention, soit : relationnelle, cognitive, politique, facilitatrice, évaluative, de même que de soutien et de suivi continu. Les résultats suggèrent que les stratégies d’ordre relationnelle, cognitive et politique sont interdépendantes et permettent d’établir un climat favorable dans lequel les agents peuvent exercer une plus grande influence sur l’appropriation du processus de l’utilisation des connaissances des intervenants scolaire. Ils montrent en outre que la composante de soutien et de suivi continu est importante pour maintenir les changements quant à l’utilisation de la recherche dans la pratique chez les intervenants scolaires. Les implications théoriques qui découlent du modèle, ainsi que les explications des mécanismes impliqués dans les différentes composantes, sont mises en perspective tant avec la documentation scientifique en transfert de connaissances dans les secteurs de la santé et de l’éducation, qu’avec les travaux provenant de disciplines connexes (notamment la psychologie). Enfin, des pistes d’action pour la pratique sont proposées.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

En synthèse d’images, reproduire les effets complexes de la lumière sur des matériaux transluminescents, tels que la cire, le marbre ou la peau, contribue grandement au réalisme d’une image. Malheureusement, ce réalisme supplémentaire est couteux en temps de calcul. Les modèles basés sur la théorie de la diffusion visent à réduire ce coût en simulant le comportement physique du transport de la lumière sous surfacique tout en imposant des contraintes de variation sur la lumière incidente et sortante. Une composante importante de ces modèles est leur application à évaluer hiérarchiquement l’intégrale numérique de l’illumination sur la surface d’un objet. Cette thèse révise en premier lieu la littérature actuelle sur la simulation réaliste de la transluminescence, avant d’investiguer plus en profondeur leur application et les extensions des modèles de diffusion en synthèse d’images. Ainsi, nous proposons et évaluons une nouvelle technique d’intégration numérique hiérarchique utilisant une nouvelle analyse fréquentielle de la lumière sortante et incidente pour adapter efficacement le taux d’échantillonnage pendant l’intégration. Nous appliquons cette théorie à plusieurs modèles qui correspondent à l’état de l’art en diffusion, octroyant une amélioration possible à leur efficacité et précision.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

La synthèse d'images dites photoréalistes nécessite d'évaluer numériquement la manière dont la lumière et la matière interagissent physiquement, ce qui, malgré la puissance de calcul impressionnante dont nous bénéficions aujourd'hui et qui ne cesse d'augmenter, est encore bien loin de devenir une tâche triviale pour nos ordinateurs. Ceci est dû en majeure partie à la manière dont nous représentons les objets: afin de reproduire les interactions subtiles qui mènent à la perception du détail, il est nécessaire de modéliser des quantités phénoménales de géométries. Au moment du rendu, cette complexité conduit inexorablement à de lourdes requêtes d'entrées-sorties, qui, couplées à des évaluations d'opérateurs de filtrage complexes, rendent les temps de calcul nécessaires à produire des images sans défaut totalement déraisonnables. Afin de pallier ces limitations sous les contraintes actuelles, il est nécessaire de dériver une représentation multiéchelle de la matière. Dans cette thèse, nous construisons une telle représentation pour la matière dont l'interface correspond à une surface perturbée, une configuration qui se construit généralement via des cartes d'élévations en infographie. Nous dérivons notre représentation dans le contexte de la théorie des microfacettes (conçue à l'origine pour modéliser la réflectance de surfaces rugueuses), que nous présentons d'abord, puis augmentons en deux temps. Dans un premier temps, nous rendons la théorie applicable à travers plusieurs échelles d'observation en la généralisant aux statistiques de microfacettes décentrées. Dans l'autre, nous dérivons une procédure d'inversion capable de reconstruire les statistiques de microfacettes à partir de réponses de réflexion d'un matériau arbitraire dans les configurations de rétroréflexion. Nous montrons comment cette théorie augmentée peut être exploitée afin de dériver un opérateur général et efficace de rééchantillonnage approximatif de cartes d'élévations qui (a) préserve l'anisotropie du transport de la lumière pour n'importe quelle résolution, (b) peut être appliqué en amont du rendu et stocké dans des MIP maps afin de diminuer drastiquement le nombre de requêtes d'entrées-sorties, et (c) simplifie de manière considérable les opérations de filtrage par pixel, le tout conduisant à des temps de rendu plus courts. Afin de valider et démontrer l'efficacité de notre opérateur, nous synthétisons des images photoréalistes anticrenelées et les comparons à des images de référence. De plus, nous fournissons une implantation C++ complète tout au long de la dissertation afin de faciliter la reproduction des résultats obtenus. Nous concluons avec une discussion portant sur les limitations de notre approche, ainsi que sur les verrous restant à lever afin de dériver une représentation multiéchelle de la matière encore plus générale.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The variogram is essential for local estimation and mapping of any variable by kriging. The variogram itself must usually be estimated from sample data. The sampling density is a compromise between precision and cost, but it must be sufficiently dense to encompass the principal spatial sources of variance. A nested, multi-stage, sampling with separating distances increasing in geometric progression from stage to stage will do that. The data may then be analyzed by a hierarchical analysis of variance to estimate the components of variance for every stage, and hence lag. By accumulating the components starting from the shortest lag one obtains a rough variogram for modest effort. For balanced designs the analysis of variance is optimal; for unbalanced ones, however, these estimators are not necessarily the best, and the analysis by residual maximum likelihood (REML) will usually be preferable. The paper summarizes the underlying theory and illustrates its application with data from three surveys, one in which the design had four stages and was balanced and two implemented with unbalanced designs to economize when there were more stages. A Fortran program is available for the analysis of variance, and code for the REML analysis is listed in the paper. (c) 2005 Elsevier Ltd. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The variogram is essential for local estimation and mapping of any variable by kriging. The variogram itself must usually be estimated from sample data. The sampling density is a compromise between precision and cost, but it must be sufficiently dense to encompass the principal spatial sources of variance. A nested, multi-stage, sampling with separating distances increasing in geometric progression from stage to stage will do that. The data may then be analyzed by a hierarchical analysis of variance to estimate the components of variance for every stage, and hence lag. By accumulating the components starting from the shortest lag one obtains a rough variogram for modest effort. For balanced designs the analysis of variance is optimal; for unbalanced ones, however, these estimators are not necessarily the best, and the analysis by residual maximum likelihood (REML) will usually be preferable. The paper summarizes the underlying theory and illustrates its application with data from three surveys, one in which the design had four stages and was balanced and two implemented with unbalanced designs to economize when there were more stages. A Fortran program is available for the analysis of variance, and code for the REML analysis is listed in the paper. (c) 2005 Elsevier Ltd. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Asynchronous Optical Sampling (ASOPS) [1,2] and frequency comb spectrometry [3] based on dual Ti:saphire resonators operated in a master/slave mode have the potential to improve signal to noise ratio in THz transient and IR sperctrometry. The multimode Brownian oscillator time-domain response function described by state-space models is a mathematically robust framework that can be used to describe the dispersive phenomena governed by Lorentzian, Debye and Drude responses. In addition, the optical properties of an arbitrary medium can be expressed as a linear combination of simple multimode Brownian oscillator functions. The suitability of a range of signal processing schemes adopted from the Systems Identification and Control Theory community for further processing the recorded THz transients in the time or frequency domain will be outlined [4,5]. Since a femtosecond duration pulse is capable of persistent excitation of the medium within which it propagates, such approach is perfectly justifiable. Several de-noising routines based on system identification will be shown. Furthermore, specifically developed apodization structures will be discussed. These are necessary because due to dispersion issues, the time-domain background and sample interferograms are non-symmetrical [6-8]. These procedures can lead to a more precise estimation of the complex insertion loss function. The algorithms are applicable to femtosecond spectroscopies across the EM spectrum. Finally, a methodology for femtosecond pulse shaping using genetic algorithms aiming to map and control molecular relaxation processes will be mentioned.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The systematic sampling (SYS) design (Madow and Madow, 1944) is widely used by statistical offices due to its simplicity and efficiency (e.g., Iachan, 1982). But it suffers from a serious defect, namely, that it is impossible to unbiasedly estimate the sampling variance (Iachan, 1982) and usual variance estimators (Yates and Grundy, 1953) are inadequate and can overestimate the variance significantly (Särndal et al., 1992). We propose a novel variance estimator which is less biased and that can be implemented with any given population order. We will justify this estimator theoretically and with a Monte Carlo simulation study.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Evolutionary developmental genetics brings together systematists, morphologists and developmental geneticists; it will therefore impact on each of these component disciplines. The goals and methods of phylogenetic analysis are reviewed here, and the contribution of evolutionary developmental genetics to morphological systematics, in terms of character conceptualisation and primary homology assessment, is discussed. Evolutionary developmental genetics, like its component disciplines phylogenetic systematics and comparative morphology, is concerned with homology concepts. Phylogenetic concepts of homology and their limitations are considered here, and the need for independent homology statements at different levels of biological organisation is evaluated. The role of systematics in evolutionary developmental genetics is outlined. Phylogenetic systematics and comparative morphology will suggest effective sampling strategies to developmental geneticists. Phylogenetic systematics provides hypotheses of character evolution (including parallel evolution and convergence), stimulating investigations into the evolutionary gains and losses of morphologies. Comparative morphology identifies those structures that are not easily amenable to typological categorisation, and that may be of particular interest in terms of developmental genetics. The concepts of latent homology and genetic recall may also prove useful in the evolutionary interpretation of developmental genetic data.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Theories on the link between achievement goals and achievement emotions focus on their within-person functional relationship (i.e., intraindividual relations). However, empirical studies have failed to analyze these intraindividual relations and have instead examined between-person covariation of the two constructs (i.e., interindividual relations). Aiming to better connect theory and empirical research, the present study (N = 120 10th grade students) analyzed intraindividual relations by assessing students’ state goals and emotions using experience sampling (N = 1,409 assessments within persons). In order to replicate previous findings on interindividual relations, students’ trait goals and emotions were assessed using self-report questionnaires. Despite being statistically independent, both types of relations were consistent with theoretical expectations, as shown by multi-level modeling: Mastery goals were positive predictors of enjoyment and negative predictors of boredom and anger; performance-approach goals were positive predictors of pride; and performance-avoidance goals were positive predictors of anxiety and shame. Reasons for the convergence of intra- and interindividual findings, directions for future research, and implications for educational practice are discussed.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Alternative sampling procedures are compared to the pure random search method. It is shown that the efficiency of the algorithm can be improved with respect to the expected number of steps to reach an epsilon-neighborhood of the optimal point.