987 resultados para Burt tables
Resumo:
Cette thèse porte sur l’évaluation de la cohérence du réseau conceptuel démontré par des étudiants de niveau collégial inscrits en sciences de la nature. L’évaluation de cette cohérence s’est basée sur l’analyse des tableaux de Burt issus des réponses à des questionnaires à choix multiples, sur l’étude détaillée des indices de discrimination spécifique qui seront décrits plus en détail dans le corps de l’ouvrage et sur l’analyse de séquences vidéos d’étudiants effectuant une expérimentation en contexte réel. Au terme de ce projet, quatre grands axes de recherche ont été exploré. 1) Quelle est la cohérence conceptuelle démontrée en physique newtonienne ? 2) Est-ce que la maîtrise du calcul d’incertitude est corrélée au développement de la pensée logique ou à la maîtrise des mathématiques ? 3) Quelle est la cohérence conceptuelle démontrée dans la quantification de l’incertitude expérimentale ? 4) Quelles sont les procédures concrètement mise en place par des étudiants pour quantifier l’incertitude expérimentale dans un contexte de laboratoire semi-dirigé ? Les principales conclusions qui ressortent pour chacun des axes peuvent se formuler ainsi. 1) Les conceptions erronées les plus répandues ne sont pas solidement ancrées dans un réseau conceptuel rigide. Par exemple, un étudiant réussissant une question sur la troisième loi de Newton (sujet le moins bien réussi du Force Concept Inventory) montre une probabilité à peine supérieure de réussir une autre question sur ce même sujet que les autres participants. De nombreux couples de questions révèlent un indice de discrimination spécifique négatif indiquant une faible cohérence conceptuelle en prétest et une cohérence conceptuelle légèrement améliorée en post-test. 2) Si une petite proportion des étudiants ont montré des carences marquées pour les questions reliées au contrôle des variables et à celles traitant de la relation entre la forme graphique de données expérimentales et un modèle mathématique, la majorité des étudiants peuvent être considérés comme maîtrisant adéquatement ces deux sujets. Toutefois, presque tous les étudiants démontrent une absence de maîtrise des principes sous-jacent à la quantification de l’incertitude expérimentale et de la propagation des incertitudes (ci-après appelé métrologie). Aucune corrélation statistiquement significative n’a été observée entre ces trois domaines, laissant entendre qu’il s’agit d’habiletés cognitives largement indépendantes. Le tableau de Burt a pu mettre en lumière une plus grande cohérence conceptuelle entre les questions de contrôle des variables que n’aurait pu le laisser supposer la matrice des coefficients de corrélation de Pearson. En métrologie, des questions équivalentes n’ont pas fait ressortir une cohérence conceptuelle clairement démontrée. 3) L’analyse d’un questionnaire entièrement dédié à la métrologie laisse entrevoir des conceptions erronées issues des apprentissages effectués dans les cours antérieurs (obstacles didactiques), des conceptions erronées basées sur des modèles intuitifs et une absence de compréhension globale des concepts métrologiques bien que certains concepts paraissent en voie d’acquisition. 4) Lorsque les étudiants sont laissés à eux-mêmes, les mêmes difficultés identifiées par l’analyse du questionnaire du point 3) reviennent ce qui corrobore les résultats obtenus. Cependant, nous avons pu observer d’autres comportements reliés à la mesure en laboratoire qui n’auraient pas pu être évalués par le questionnaire à choix multiples. Des entretiens d’explicitations tenus immédiatement après chaque séance ont permis aux participants de détailler certains aspects de leur méthodologie métrologique, notamment, l’emploi de procédures de répétitions de mesures expérimentales, leurs stratégies pour quantifier l’incertitude et les raisons sous-tendant l’estimation numérique des incertitudes de lecture. L’emploi des algorithmes de propagation des incertitudes a été adéquat dans l’ensemble. De nombreuses conceptions erronées en métrologie semblent résister fortement à l’apprentissage. Notons, entre autres, l’assignation de la résolution d’un appareil de mesure à affichage numérique comme valeur de l’incertitude et l’absence de procédures d’empilement pour diminuer l’incertitude. La conception que la précision d’une valeur numérique ne peut être inférieure à la tolérance d’un appareil semble fermement ancrée.
Resumo:
BACKGROUND Collaborative and active learning have been clearly identified as ways students can engage in learning with each other and the academic staff. Traditional tier based lecture theatres and the didactic style they engender are not popular with students today as evidenced by the low attendance rates for lectures. Many universities are installing spaces designed with tables for group interaction with evolutions on spaces such as the TEAL (Technology Enabled Active Learning) (Massachusetts Institute of Technology, n.d.) and SCALE-UP (Student-Centred Activities for Large-Enrolment Undergraduate Programs) (North Carolina State University, n.d.) models. Technology advances in large screen computers and applications have also aided the move to these collaborative spaces. How well have universities structured learning using these spaces and how have students engaged with the content, technology, space and each other? This paper investigates the application of collaborative learning in such spaces for a cohort of 800+ first year engineers in the context of learning about and developing professional skills representative of engineering practice. PURPOSE To determine whether moving from tiers to tables enhances the student experience. Does utilising technology rich, activity based, collaborative learning spaces lead to positive experiences and active engagement of first year undergraduate engineering students? In developing learning methodology and approach in new learning spaces, what needs to change from a more traditional lecture and tutorial configuration? DESIGN/METHOD A post delivery review and analysis of outcomes was undertaken to determine how well students and tutors engaged with learning in new collaborative learning spaces. Data was gathered via focus group and survey of tutors, students survey and attendance observations. The authors considered the unit delivery approach along with observed and surveyed outcomes then conducted further review to produce the reported results. RESULTS Results indicate high participation in the collaborative sessions while the accompanying lectures were poorly attended. Students reported a high degree of satisfaction with the learning experience; however more investigation is required to determine the degree of improvement in retained learning outcomes. Survey feedback from tutors found that students engaged well in the activities during tutorials and there was an observed improvement in the quality of professional practice modelled by students during sessions. Student feedback confirmed the positive experiences in these collaborative learning spaces with 30% improvement in satisfaction ratings from previous years. CONCLUSIONS It is concluded that the right mix of space, technology and appropriate activities does engage students, improve participation and create a rich experience to facilitate potential for improved learning outcomes. The new Collaborative Teaching Spaces, together with integrated technology and tailored activities, has transformed the delivery of this unit and improved student satisfaction in tutorials significantly.
Resumo:
Recurrence relations in mathematics form a very powerful and compact way of looking at a wide range of relationships. Traditionally, the concept of recurrence has often been a difficult one for the secondary teacher to convey to students. Closely related to the powerful proof technique of mathematical induction, recurrences are able to capture many relationships in formulas much simpler than so-called direct or closed formulas. In computer science, recursive coding often has a similar compactness property, and, perhaps not surprisingly, suffers from similar problems in the classroom as recurrences: the students often find both the basic concepts and practicalities elusive. Using models designed to illuminate the relevant principles for the students, we offer a range of examples which use the modern spreadsheet environment to powerfully illustrate the great expressive and computational power of recurrences.
Resumo:
The literacy demands of tables and graphs are different from those of prose texts such as narrative. This paper draws from part of a qualitative case study which sought to investigate strategies that scaffold and enhance the teaching and learning of varied representations in text. As indicated in the paper, the method focused on the teaching and learning of tables and graphs with use of Freebody and Luke's (1990) four resources model from literacy education.
Resumo:
Abstract-To detect errors in decision tables one needs to decide whether a given set of constraints is feasible or not. This paper describes an algorithm to do so when the constraints are linear in variables that take only integer values. Decision tables with such constraints occur frequently in business data processing and in nonnumeric applications. The aim of the algorithm is to exploit. the abundance of very simple constraints that occur in typical decision table contexts. Essentially, the algorithm is a backtrack procedure where the the solution space is pruned by using the set of simple constrains. After some simplications, the simple constraints are captured in an acyclic directed graph with weighted edges. Further, only those partial vectors are considered from extension which can be extended to assignments that will at least satisfy the simple constraints. This is how pruning of the solution space is achieved. For every partial assignment considered, the graph representation of the simple constraints provides a lower bound for each variable which is not yet assigned a value. These lower bounds play a vital role in the algorithm and they are obtained in an efficient manner by updating older lower bounds. Our present algorithm also incorporates an idea by which it can be checked whether or not an (m - 2)-ary vector can be extended to a solution vector of m components, thereby backtracking is reduced by one component.
Resumo:
We present two discriminative language modelling techniques for Lempel-Ziv-Welch (LZW) based LID system. The previous approach to LID using LZW algorithm was to directly use the LZW pattern tables forlanguage modelling. But, since the patterns in a language pattern table are shared by other language pattern tables, confusability prevailed in the LID task. For overcoming this, we present two pruning techniques (i) Language Specific (LS-LZW)-in which patterns common to more than one pattern table are removed. (ii) Length-Frequency product based (LF-LZW)-in which patterns having their length-frequency product below a threshold are removed. These approaches reduce the classification score (Compression Ratio [LZW-CR] or the weighted discriminant score [LZW-WDS]) for non native languages and increases the LID performance considerably. Also the memory and computational requirements of these techniques are much less compared to basic LZW techniques.
Resumo:
Process control rules may be specified using decision tables. Such a specification is superior when logical decisions to be taken in control dominate. In this paper we give a method of detecting redundancies, incompleteness, and contradictions in such specifications. Using such a technique thus ensures the validity of the specifications.
Resumo:
Not available.
Resumo:
Comparison between Galton equation and preston normal logarithms models allowed an empirical reconstitution of probits tables.
Resumo:
EXTRACT (SEE PDF FOR FULL ABSTRACT): The data of this paper differ from the Jones and Bradley papers [of 1982-1986] in that it represents an attempt to select thermal pollution free records rather than to include all available records. The specific long-term trends that this paper is trying to avoid are those illustrated by the heat islands of fast growing urban locations. One other major difference in this paper is that all of the records reported of this study are complete for the entire study period.
Resumo:
We present a systematic investigation of calculating quantum dots (QDs) energy levels using finite element method in the frame of eight-band k . p method. Numerical results including piezoelectricity, electron and hole levels, as yell as wave functions are achieved. In the calculation of energy levels, we do observe spurious solutions (SSs) no matter Burt-Foreman or symmetrized Hamiltonians are used. Different theories are used to analyse the SSs, we find that the ellipticity theory can give a better explanation for the origin of SSs and symmetrized Hamiltonian is easier to lead to SSs. The energy levels simulated with the two Hamiltonians are compared to each other after eliminating SSs, different Hamiltonians cause a larger difference on electron energy levels than that on hole energy levels and this difference decreases with the increase of QD size.