947 resultados para Which-way experiments
Resumo:
This study examined the effectiveness of motor-encoding activities on memory and performance of students in a Grade One reading program. There were two experiments in the study. Experiment 1 replicated a study by Eli Saltz and David Dixon (1982). The effect of motoric enactment (Le., pretend play) of sentences on memory for the sentences was investigated. Forty Grade One students performed a "memory-for-sentences" technique, devised by Saltz and Dixon. Only the experimental group used motoric enactment of the sentences. Although quantitative findings revealed no significant difference between the mean scores of the experimental group versus the control group, aspects of the experimental design could have affected the results. It was suggested that Saltz and Dixon's study could be replicated again, with more attention given to variables such as population size, nature of the test sentences, subjects' previous educational experience and conditions related to the testing environment. The second experiment was an application of Saltz and Dixon's theory that motoric imagery should facilitate memory for sentences. The intent was to apply this theory to Grade One students' ability to remember words from their reading program. An experimental gym program was developed using kinesthetic activities to reinforce the skills of the classroom reading program. The same subject group was used in Experiment 2. It was hypothesized that the subjects who experienced the experimental gym program would show greater signs of progress in reading ability, as evidenced by their scores on Form G of the Woodcock Reading Mastery Test--Revised. The data from the WRM--R were analyzed with a 3-way split-plot analysis of variance in which group (experimental vs. control) and sex were the between subjects variables and test-time (pre-test vs. post-test) was the within-subjects variable. Findings revealed the following: (a) both groups made substantial gains over time on the visual-auditory learning sub-test and the triple action of group x sex x time also was significant; (b) children in the experimental and control groups performed similarly on both the pre- and post-test of the letter identification test; (c) time was the only significant effect on subjects' performance on the word identification task; (d) work attack scores showed marked improvement in performance over time for both the experimenta+ and control groups; (e) passage comprehension scores indicated an improvement in performance for both groups over time. Similar to Experiment 1, it is suggested that several modifications in the experimental design could produce significant results. These factors are addressed with suggestions for further research in the area of active learning; more specifically, the effect of motor-encoding activities on memory and academic performance of children.
Resumo:
Traditional psychometric theory and practice classify people according to broad ability dimensions but do not examine how these mental processes occur. Hunt and Lansman (1975) proposed a 'distributed memory' model of cognitive processes with emphasis on how to describe individual differences based on the assumption that each individual possesses the same components. It is in the quality of these components ~hat individual differences arise. Carroll (1974) expands Hunt's model to include a production system (after Newell and Simon, 1973) and a response system. He developed a framework of factor analytic (FA) factors for : the purpose of describing how individual differences may arise from them. This scheme is to be used in the analysis of psychometric tes ts . Recent advances in the field of information processing are examined and include. 1) Hunt's development of differences between subjects designated as high or low verbal , 2) Miller's pursuit of the magic number seven, plus or minus two, 3) Ferguson's examination of transfer and abilities and, 4) Brown's discoveries concerning strategy teaching and retardates . In order to examine possible sources of individual differences arising from cognitive tasks, traditional psychometric tests were searched for a suitable perceptual task which could be varied slightly and administered to gauge learning effects produced by controlling independent variables. It also had to be suitable for analysis using Carroll's f ramework . The Coding Task (a symbol substitution test) found i n the Performance Scale of the WISe was chosen. Two experiments were devised to test the following hypotheses. 1) High verbals should be able to complete significantly more items on the Symbol Substitution Task than low verbals (Hunt, Lansman, 1975). 2) Having previous practice on a task, where strategies involved in the task may be identified, increases the amount of output on a similar task (Carroll, 1974). J) There should be a sUbstantial decrease in the amount of output as the load on STM is increased (Miller, 1956) . 4) Repeated measures should produce an increase in output over trials and where individual differences in previously acquired abilities are involved, these should differentiate individuals over trials (Ferguson, 1956). S) Teaching slow learners a rehearsal strategy would improve their learning such that their learning would resemble that of normals on the ,:same task. (Brown, 1974). In the first experiment 60 subjects were d.ivided·into high and low verbal, further divided randomly into a practice group and nonpractice group. Five subjects in each group were assigned randomly to work on a five, seven and nine digit code throughout the experiment. The practice group was given three trials of two minutes each on the practice code (designed to eliminate transfer effects due to symbol similarity) and then three trials of two minutes each on the actual SST task . The nonpractice group was given three trials of two minutes each on the same actual SST task . Results were analyzed using a four-way analysis of variance . In the second experiment 18 slow learners were divided randomly into two groups. one group receiving a planned strategy practioe, the other receiving random practice. Both groups worked on the actual code to be used later in the actual task. Within each group subjects were randomly assigned to work on a five, seven or nine digit code throughout. Both practice and actual tests consisted on three trials of two minutes each. Results were analyzed using a three-way analysis of variance . It was found in t he first experiment that 1) high or low verbal ability by itself did not produce significantly different results. However, when in interaction with the other independent variables, a difference in performance was noted . 2) The previous practice variable was significant over all segments of the experiment. Those who received previo.us practice were able to score significantly higher than those without it. J) Increasing the size of the load on STM severely restricts performance. 4) The effect of repeated trials proved to be beneficial. Generally, gains were made on each successive trial within each group. S) In the second experiment, slow learners who were allowed to practice randomly performed better on the actual task than subjeots who were taught the code by means of a planned strategy. Upon analysis using the Carroll scheme, individual differences were noted in the ability to develop strategies of storing, searching and retrieving items from STM, and in adopting necessary rehearsals for retention in STM. While these strategies may benef it some it was found that for others they may be harmful . Temporal aspects and perceptual speed were also found to be sources of variance within individuals . Generally it was found that the largest single factor i nfluencing learning on this task was the repeated measures . What e~ables gains to be made, varies with individuals . There are environmental factors, specific abilities, strategy development, previous learning, amount of load on STM , perceptual and temporal parameters which influence learning and these have serious implications for educational programs .
Resumo:
This paper explores the cognitive functions of the Reality Status Evaluation (RSE) system in our experiences of narrative mediated messages (NMM) (fictional, narrative, audio-visual one-way input and moving picture messages), such as fictional TV programs and films. We regard reality in mediated experiences as a special mental and emotional construction and a multi-dimensional concept. We argue that viewers' reality sense in NMM is influenced by many factors with "real - on" as the default value. Some of these factors function as primary mental processes, including the content realism factors of those messages such as Factuality (F), Social Realism (SR), Life Relevance (LR), and Perceptual Realism - involvement (PR), which would have direct impacts on reality evaluations. Other factors, such as Narrative Meaning (NM), Emotional Responses, and personality trait Absorption (AB), will influence the reality evaluations directly or through the mediations of these main dimensions. I designed a questionnaire to study this theoretical construction. I developed items to form scales and sub-scales measuring viewers' subjective experiences of reality evaluations and these factors. Pertinent statistical techniques, such as internal consistency and factorial analysis, were employed to make revisions and improve the quality of the questionnaire. In the formal experiment, after viewing two short films, which were selected as high or low narrative structure messages from previous experiments, participants were required to answer the questionnaire, Absorption questionnaire, and SAM (Self-Assessment Manikin, measuring immediate emotional responses). Results were analyzed using the EQS, structural equation modeling (SEM), and discussed in terms oflatent relations among these subjective factors in mediated experience. The present results supported most of my theoretical hypotheses. In NMM, three main jactors, or dimensions, could be extracted in viewers' subjective reality evaluations: Social Realism (combining with Factuality), Life Relevance and Perceptual Realism. I designed two ways to assess viewers' understanding of na"ative meanings in mediated messages, questionnaire (NM-Q) and rating (NM-R) measurement, and its significant influences on reality evaluations was supported in the final EQS models. Particularly in high story stnlcture messages, the effect of Narrative Meaning (NM) can rarely be explained by only these dimensions of reality evaluations. Also, Empathy seems to playa more important role in RSE of low story structure messages. Also, I focused on two other factors that were pertinent to RSE in NMM, the personality trait Absorption, and Emotional Responses (including two dimensions: Valence and Intensity). Final model results partly supported my theoretical hypotheses about the relationships among Absorption (AB), Social Realism (SR) and Life Relevance (LR); and the immediate impact of Emotional Responses on Perceptual Realism cPR).
Resumo:
Adam Seybet, Chairman.
Resumo:
July 31, 1813. Resolution agreed to and forwarded to the President.
Resumo:
July 31, 1813. Resolution agreed to and forwarded to the President. Printed by A and G Way
Resumo:
A wide range of tests for heteroskedasticity have been proposed in the econometric and statistics literature. Although a few exact homoskedasticity tests are available, the commonly employed procedures are quite generally based on asymptotic approximations which may not provide good size control in finite samples. There has been a number of recent studies that seek to improve the reliability of common heteroskedasticity tests using Edgeworth, Bartlett, jackknife and bootstrap methods. Yet the latter remain approximate. In this paper, we describe a solution to the problem of controlling the size of homoskedasticity tests in linear regression contexts. We study procedures based on the standard test statistics [e.g., the Goldfeld-Quandt, Glejser, Bartlett, Cochran, Hartley, Breusch-Pagan-Godfrey, White and Szroeter criteria] as well as tests for autoregressive conditional heteroskedasticity (ARCH-type models). We also suggest several extensions of the existing procedures (sup-type of combined test statistics) to allow for unknown breakpoints in the error variance. We exploit the technique of Monte Carlo tests to obtain provably exact p-values, for both the standard and the new tests suggested. We show that the MC test procedure conveniently solves the intractable null distribution problem, in particular those raised by the sup-type and combined test statistics as well as (when relevant) unidentified nuisance parameter problems under the null hypothesis. The method proposed works in exactly the same way with both Gaussian and non-Gaussian disturbance distributions [such as heavy-tailed or stable distributions]. The performance of the procedures is examined by simulation. The Monte Carlo experiments conducted focus on : (1) ARCH, GARCH, and ARCH-in-mean alternatives; (2) the case where the variance increases monotonically with : (i) one exogenous variable, and (ii) the mean of the dependent variable; (3) grouped heteroskedasticity; (4) breaks in variance at unknown points. We find that the proposed tests achieve perfect size control and have good power.
Resumo:
Un fichier intitulé Charbonneau_Nathalie_2008_AnimationAnnexeT accompagne la thèse. Il contient une séquence animée démontrant le type de parcours pouvant être effectué au sein des environnements numériques développés. Il s'agit d'un fichier .wmv qui a été compressé.
Resumo:
Par cette recherche, nous voulons évaluer de manière exhaustive les bénéfices qu’apporte l’ExAO (Expérimentation Assistée par Ordinateur) dans les laboratoires scolaires de sciences et technologie au Liban. Nous aimerions aussi qu’elle contribue d’une manière tangible aux recherches du laboratoire de Robotique Pédagogique de l’Université de Montréal, notamment dans le développement du µlaboratoire ExAO. Nous avons voulu tester les capacités de l’ExAO, son utilisation en situation de classe comme : 1. Substitut d’un laboratoire traditionnel dans l’utilisation de la méthode expérimentale; 2. Outil d’investigation scientifique; 3. Outil d’intégration des sciences expérimentales et des mathématiques; 4. Outil d’intégration des sciences expérimentales, des mathématiques et de la technologie dans un apprentissage technoscientifique; Pour ce faire, nous avons mobilisé 13 groupe-classes de niveaux complémentaire et secondaire, provenant de 10 écoles libanaises. Nous avons désigné leurs enseignants pour expérimenter eux-mêmes avec leurs étudiants afin d’évaluer, de manière plus réaliste les avantages d’implanter ce micro laboratoire informatisé à l’école. Les différentes mise à l’essai, évaluées à l’aide des résultats des activités d’apprentissage réalisées par les étudiants, de leurs réponses à un questionnaire et des commentaires des enseignants, nous montrent que : 1. La substitution d’un laboratoire traditionnel par un µlaboratoire ExAO ne semble pas poser de problème; dix minutes ont suffi aux étudiants pour se familiariser avec cet environnement, mentionnant que la rapidité avec laquelle les données étaient représentées sous forme graphique était plus productive. 2. Pour l’investigation d’un phénomène physique, la convivialité du didacticiel associée à la capacité d’amplifier le phénomène avant de le représenter graphiquement a permis aux étudiants de concevoir et de mettre en œuvre rapidement et de manière autonome, une expérimentation permettant de vérifier leur prédiction. 3. L’intégration des mathématiques dans une démarche expérimentale permet d’appréhender plus rapidement le phénomène. De plus, elle donne un sens aux représentations graphiques et algébriques, à l’avis des enseignants, permettant d’utiliser celle-ci comme outil cognitif pour interpréter le phénomène. 4. La démarche réalisée par les étudiants pour concevoir et construire un objet technologique, nous a montré que cette activité a été réalisée facilement par l’utilisation des capteurs universels et des amplificateurs à décalage de l’outil de modélisation graphique ainsi que la capacité du didacticiel à transformer toute variable mesurée par une autre variable (par exemple la variation de résistance en variation de température, …). Cette activité didactique nous montre que les étudiants n’ont eu aucune difficulté à intégrer dans une même activité d’apprentissage les mathématiques, les sciences expérimentales et la technologie, afin de concevoir et réaliser un objet technologique fonctionnel. µlaboratoire ExAO, en offrant de nouvelles possibilités didactiques, comme la capacité de concevoir, réaliser et valider un objet technologique, de disposer pour ce faire, des capacités nouvelles pour amplifier les mesures, modéliser les phénomènes physiques, créer de nouveaux capteurs, est un ajout important aux expériences actuellement réalisées en ExAO.
Resumo:
Le développement du logiciel actuel doit faire face de plus en plus à la complexité de programmes gigantesques, élaborés et maintenus par de grandes équipes réparties dans divers lieux. Dans ses tâches régulières, chaque intervenant peut avoir à répondre à des questions variées en tirant des informations de sources diverses. Pour améliorer le rendement global du développement, nous proposons d'intégrer dans un IDE populaire (Eclipse) notre nouvel outil de visualisation (VERSO) qui calcule, organise, affiche et permet de naviguer dans les informations de façon cohérente, efficace et intuitive, afin de bénéficier du système visuel humain dans l'exploration de données variées. Nous proposons une structuration des informations selon trois axes : (1) le contexte (qualité, contrôle de version, bogues, etc.) détermine le type des informations ; (2) le niveau de granularité (ligne de code, méthode, classe, paquetage) dérive les informations au niveau de détails adéquat ; et (3) l'évolution extrait les informations de la version du logiciel désirée. Chaque vue du logiciel correspond à une coordonnée discrète selon ces trois axes, et nous portons une attention toute particulière à la cohérence en naviguant entre des vues adjacentes seulement, et ce, afin de diminuer la charge cognitive de recherches pour répondre aux questions des utilisateurs. Deux expériences valident l'intérêt de notre approche intégrée dans des tâches représentatives. Elles permettent de croire qu'un accès à diverses informations présentées de façon graphique et cohérente devrait grandement aider le développement du logiciel contemporain.
Resumo:
Management sciences and corporate practices of companies bring extensive views on the ethics and the corporate responsibility. The practical experiments try to build, and to make appropriate themselves to itself by the actors of companies, the deontological ethics governed by principles. Dialogue and deliberation are the political frame given to this construction. A question seems then central, that of the ultimate rational foundation of the morality and its normative contents. K.O. Apel’s propositions concerning his ethics of the discussion founded in a pragmatico-transcendental way seems to me to allow an essential advance on the subject. According to Apel, this foundation should supply: 1) The bases of an ultimate rational foundation of the morality and its normative contents, 2) The bases of the universal validity of a principle of justice, solidarity and co responsibility, 3) The bases of an ethics of the responsibility, but not an deontological ethics, governed by principles which could come to complete and to legitimize socio organizational processes and systems of relations which, placed in a new ethical context, could continue. In this article, I try to show in what K.O. Apel’s propositions allow to enrich a reflection opened by the research works in business ethics and in organizational behavior.
Resumo:
La fibrillation auriculaire, l'arythmie la plus fréquente en clinique, affecte 2.3 millions de patients en Amérique du Nord. Pour en étudier les mécanismes et les thérapies potentielles, des modèles animaux de fibrillation auriculaire ont été développés. La cartographie électrique épicardique à haute densité est une technique expérimentale bien établie pour suivre in vivo l'activité des oreillettes en réponse à une stimulation électrique, à du remodelage, à des arythmies ou à une modulation du système nerveux autonome. Dans les régions qui ne sont pas accessibles par cartographie épicardique, la cartographie endocardique sans contact réalisée à l'aide d'un cathéter en forme de ballon pourrait apporter une description plus complète de l'activité auriculaire. Dans cette étude, une expérience chez le chien a été conçue et analysée. Une reconstruction électro-anatomique, une cartographie épicardique (103 électrodes), une cartographie endocardique sans contact (2048 électrodes virtuelles calculées à partir un cathéter en forme de ballon avec 64 canaux) et des enregistrements endocardiques avec contact direct ont été réalisés simultanément. Les systèmes d'enregistrement ont été également simulés dans un modèle mathématique d'une oreillette droite de chien. Dans les simulations et les expériences (après la suppression du nœud atrio-ventriculaire), des cartes d'activation ont été calculées pendant le rythme sinusal. La repolarisation a été évaluée en mesurant l'aire sous l'onde T auriculaire (ATa) qui est un marqueur de gradient de repolarisation. Les résultats montrent un coefficient de corrélation épicardique-endocardique de 0.8 (expérience) and 0.96 (simulation) entre les cartes d'activation, et un coefficient de corrélation de 0.57 (expérience) and 0.92 (simulation) entre les valeurs de ATa. La cartographie endocardique sans contact apparait comme un instrument expérimental utile pour extraire de l'information en dehors des régions couvertes par les plaques d'enregistrement épicardique.
Resumo:
This thesis is concerned with the interaction between literature and abstract thought. More specifically, it studies the epistemological charge of the literary, the type of knowledge that is carried by elements proper to fictional narratives into different disciplines. By concentrating on two different theoretical methods, the creation of thought experiments and the framing of possible worlds, methods which were elaborated and are still used today in spheres as varied as modal logics, analytic philosophy and physics, and by following their reinsertion within literary theory, the research develops the theory that both thought experiments and possible worlds are in fact short narrative stories that inform knowledge through literary means. By using two novels, Abbott’s Flatland and Vonnegut’s The Sirens of Titan, that describe extra-dimensional existence in radically different ways, respectively as a phenomenologically unknowable space and as an outward perspective on time, it becomes clear that literature is constitutive of the way in which worlds, fictive, real or otherwise, are constructed and understood. Thus dimensions, established through extensional analogies as either experimental knowledge or modal possibility for a given world, generate new directions for thought, which can then take part in the inductive/deductive process of scientia. By contrasting the dimensions of narrative with the way that dimensions were historically constituted, the research also establishes that the literary opens up an infinite potential of abstract space-time domains, defined by their specific rules and limits, and that these different experimental folds are themselves partaking in a dimensional process responsible for new forms of understanding. Over against science fiction literary theories of speculation that posit an equation between the fictive and the real, this thesis examines the complex structure of many overlapping possibilities that can organise themselves around larger compossible wholes, thus offering a theory of reading that is both non-mimetic and non-causal. It consequently examines the a dynamic process whereby literature is always reconceived through possibilities actualised by reading while never defining how the reader will ultimately understand the overarching structure. In this context, the thesis argues that a causal story can be construed out of any one interaction with a given narrative—underscoring, for example, the divinatory strength of a particular vision of the future—even as this narrative represents only a fraction of the potential knowledge of any particular literary text. Ultimately, the study concludes by tracing out how novel comprehensions of the literary, framed by the material conditions of their own space and time, endlessly renew themselves through multiple interactions, generating analogies and speculations that facilitate the creation of new knowledge.
Resumo:
Cette thèse porte sur le calcul de structures électroniques dans les solides. À l'aide de la théorie de la fonctionnelle de densité, puis de la théorie des perturbations à N-corps, on cherche à calculer la structure de bandes des matériaux de façon aussi précise et efficace que possible. Dans un premier temps, les développements théoriques ayant mené à la théorie de la fonctionnelle de densité (DFT), puis aux équations de Hedin sont présentés. On montre que l'approximation GW constitue une méthode pratique pour calculer la self-énergie, dont les résultats améliorent l'accord de la structure de bandes avec l'expérience par rapport aux calculs DFT. On analyse ensuite la performance des calculs GW dans différents oxydes transparents, soit le ZnO, le SnO2 et le SiO2. Une attention particulière est portée aux modèles de pôle de plasmon, qui permettent d'accélérer grandement les calculs GW en modélisant la matrice diélectrique inverse. Parmi les différents modèles de pôle de plasmon existants, celui de Godby et Needs s'avère être celui qui reproduit le plus fidèlement le calcul complet de la matrice diélectrique inverse dans les matériaux étudiés. La seconde partie de la thèse se concentre sur l'interaction entre les vibrations des atomes du réseau cristallin et les états électroniques. Il est d'abord montré comment le couplage électron-phonon affecte la structure de bandes à température finie et à température nulle, ce qu'on nomme la renormalisation du point zéro (ZPR). On applique ensuite la méthode GW au calcul du couplage électron-phonon dans le diamant. Le ZPR s'avère être fortement amplifié par rapport aux calculs DFT lorsque les corrections GW sont appliquées, améliorant l'accord avec les observations expérimentales.
Resumo:
Rampant increases in oil prices and detrimental effects of fossil fuels on the environment have been the main impetus for the development of environmentally friendly and sustainable energy sources. Amongst the many possibilities, microalgae have been proposed as a new alternative energy source to fossil fuels, as their growth is both sustainable and ecologically safe. By definition, microalgae are unicellular photosynthetic microorganisms containing chlorophyll a. These organisms are capable of producing large quantities of oils, surpassing that of traditional oil-seed crops, which can be transformed, through chemical processes, into biofuels such as biodiesel or bio-gasoline. Thus, recent research has gone into discovering high lipid producing algal strains, optimising growth media for increased lipid production and developing metabolic engineering to make microalgae a source of biofuel that is competitive to more traditional sources of biofuel and even to fossil fuel. In this context, the research reported here focused on using a mixotrophic growth mode as a way to increase lipid production for certain strains of microalgae. In addition, nitrogen starvation combined with mixotrophy was studied to analyse its effects on lipid production. Mixotrophy is the parallel usage of two trophic modes, in our case photoautotrophy and heterotrophy. Consequently, 12 algal strains were screened for mixotrophic growth, using glycerol as a carbon source. Glycerol is a waste product of the current biodiesel industry; it is a cheap and abundant carbon source present in many metabolic pathways. From this initial screening, several strains were chosen for subsequent experiments involving nitrogen starvation. Nitrogen starvation has been shown to induce lipid accumulation. The results obtained show that a mixotrophic growth mode, using glycerol as a carbon source, enhances lipid production for certain strains. Moreover, lipid enhancement was shown for nitrogen starvation combined with mixotrophic growth mode. This was dependant on time spent under nitrogen starvation and on initial concentrations of the nitrogen source.