971 resultados para three-shell model
Resumo:
Three grade three mathematics textbooks were selected arbitrarily (every other) from a total of six currently used in the schools of Ontario. These textbooks were examined through content analysis in order to determine the extent (i. e., the frequency of occurrence) to which problem solving strategies appear in the problems and exercises of grade three mathematics textbooks, and how well they carry through the Ministry's educational goals set out in The Formative Years. Based on Polya's heuristic model, a checklist was developed by the researcher. The checklist had two main categories, textbook problems and process problems and a finer classification according to the difficulty level of a textbook problem; also six commonly used problem solving strategies for the analysis of a process problem. Topics to be analyzed were selected from the subject guideline The Formative Years, and the same topics were selected from each textbook. Frequencies of analyzed problems and exercises were compiled and tabulated textbook by textbook and topic by topic. In making comparisons, simple frequency count and percentage were used in the absence of any known criteria available for judging highor low frequency. Each textbook was coded by three coders trained to use the checklist. The results of analysis showed that while there were large numbers of exercises in each textbook, not very many were framed as problems according to Polya' s model and that process problems form a small fraction of the number of analyzed problems and exercises. There was no pattern observed as to the systematic placement of problems in the textbooks.
Resumo:
The purpose of this study was to replicate and extend a motivational model of problem drinking (Cooper, Frone, Russel, & Mudar, 1995; Read, Wood, Kahler, Maddock & Tibor, 2003), testing the notion that attachment is a common antecedent for both the affective and social paths to problem drinking. The model was tested with data from three samples, first-year university students (N=679), students about to graduate from university (N=206), and first-time clients at an addiction treatment facility (N=21 1). Participants completed a battery of questionnaires assessing alcohol use, alcohol-related consequences, drinking motives, peer models of alcohol use, positive and negative affect, attachment anxiety and attachment avoidance. Results underscored the importance of the affective path to problem drinking, while putting the social path to problem drinking into question. While drinking to cope was most prominent among the clinical sample, coping motives served as a risk factor for problem drinking for both individuals identified as problem drinkers and university students. Moreover, drinking for enhancement purposes appeared to be the strongest overall predictor of alcohol use. Results of the present study also supported the notion that attachment anxiety and avoidance are antecedents for the affective path to problem drinking, such that those with higher levels of attachment anxiety and avoidance were more vulnerable to experiencing adverse consequences related to their drinking, explained in terms of diminished affect regulation. Evidence that nonsecure attachment is a potent predictor of problem drinking was also demonstrated by the finding that attachment anxiety was directly related to alcohol-related consequences over and above its indirect relationship through affect regulation. However, results failed to show that attachment anxiety or attachment avoidance increased the risk of problem drinking via social influence.
Resumo:
The effects of a complexly worded counterattitudinal appeal on laypeople's attitudes toward a legal issue were examined, using the Elaboration Likelihood Model (ELM) of persuasion as a theoretical framework. This model states that persuasion can result from the elaboration and scrutiny of the message arguments (i.e., central route processing), or can result from less cognitively effortful strategies, such as relying on source characteristics as a cue to message validity (i.e., peripheral route processing). One hundred and sixty-seven undergraduates (85 men and 81 women) listened to eitller a low status or high status source deliver a counterattitudinal speech on a legal issue. The speech was designed to contain strong or weak arguments. These arguments were 'worded in a simple and, therefore, easy to comprehend manner, or in a complex and, therefore, difficult to comprehend manner. Thus, there were three experimental manipulations: argument comprehensibility (easy to comprehend vs. difficult to comprehend), argumel11 strength (weak vs. strong), and source status (low vs. high). After listening to tIle speec.J] participants completed a measure 'of their attitude toward the legal issue, a thought listil1g task, an argument recall task,manipulation checks, measures of motivation to process the message, and measures of mood. As a result of the failure of the argument strength manipulation, only the effects of the comprehel1sibility and source status manipulations were tested. There was, however, some evidence of more central route processing in the easy comprehension condition than in the difficult comprehension condition, as predicted. Significant correlations were found between attitude and favourable and unfavourable thoughts about the legal issue with easy to comprehend arguments; whereas, there was a correlation only between attitude and favourable thoughts 11 toward the issue with difficult to comprehend arguments, suggesting, perhaps, that central route processing, \vhich involves argument scrutiny and elaboration, occurred under conditions of easy comprehension to a greater extent than under conditions of difficult comprehension. The results also revealed, among other findings, several significant effects of gender. Men had more favourable attitudes toward the legal issue than did women, men recalled more arguments from the speech than did women, men were less frustrated while listening to the speech than were ,vomen, and men put more effort into thinking about the message arguments than did women. When the arguments were difficult to comprehend, men had more favourable thoughts and fewer unfavourable thoughts about the legal issue than did women. Men and women may have had different affective responses to the issue of plea bargaining (with women responding more negatively than men), especially in light of a local and controversial plea bargain that occurred around the time of this study. Such pre-existing gender differences may have led to tIle lower frustration, the greater effort, the greater recall, and more positive attitudes for men than for WOlnen. Results· from this study suggest that current cognitive models of persuasion may not be very applicable to controversial issues which elicit strong emotional responses. Finally, these data indicate that affective responses, the controversial and emotional nature ofthe issue, gender and other individual differences are important considerations when experts are attempting to persuade laypeople toward a counterattitudinal position.
Resumo:
EI Salvador presents an unfortunate history that includes a military regime and a civil war that together created a legacy of violence in which the country still struggle nowadays. Salud Escolar Integral (SEI) was created in 2005 as a program to combat youth violence throughout the re-formulation of physical education (PE) classes in public schools, promoting life skills learning that supports the resolution of conflicts with nonviolent ways. In 2007, SEI supported the creation of a physical e~ucation teacher education (PETE) degree at the Universidad Pedag6gica de EI Salvador (UPES), having the goal to assist pre-service teachers with a better understanding of humanistic principles. The present research analyzed if after attending all three years ofUPES PETE program, students presented high self-perception levels of competence and confidence related to attitude, skills and knowledge to teach PE within humanistic principles. Taking Personal and Social Responsibility (TPSR) was the theoretical framework used to analyze the development of humanistic principles. The study had a mixed-method longitudinal design that included questionnaires, reflection templates and interviews. In conclusion, although it is suggested that UPES should provide better support for the development of the teaching principles of empowering students and transfer learning, most of the humanistic principles were highly promoted by the program. At last, it is suggested that future research should track teachers' progress while teaching in schools, in order to analyze if the theory of promoting humanistic principles have also become a daily practice.
Resumo:
Architectural model of the Pralle-Sodaro Residence Hall, Chapman University, Orange, California, dedicated October 21, 1991. Pralle-Sodaro Hall (3 floors, 75,382 sq.ft.) is a three-story building containing one-hundred-fifty-five units. This state-of-the-art residence hall was made possible by the tremendous generosity of Bob and Helga Pralle and Don and DeeDee Sodaro. Bob Pralle served as a trustee for eighteen years beginning in 1984 and Don Sodaro as Chairman of the board of trustees and has served on the board for fourteen years beginning in 1988. Pralle-Sodaro Hall (3 floors, 75,382 sq.ft.) is a three-story building containing one-hundred-fifty-five units.
Resumo:
The purpose of this study is to examine the impact of the choice of cut-off points, sampling procedures, and the business cycle on the accuracy of bankruptcy prediction models. Misclassification can result in erroneous predictions leading to prohibitive costs to firms, investors and the economy. To test the impact of the choice of cut-off points and sampling procedures, three bankruptcy prediction models are assessed- Bayesian, Hazard and Mixed Logit. A salient feature of the study is that the analysis includes both parametric and nonparametric bankruptcy prediction models. A sample of firms from Lynn M. LoPucki Bankruptcy Research Database in the U. S. was used to evaluate the relative performance of the three models. The choice of a cut-off point and sampling procedures were found to affect the rankings of the various models. In general, the results indicate that the empirical cut-off point estimated from the training sample resulted in the lowest misclassification costs for all three models. Although the Hazard and Mixed Logit models resulted in lower costs of misclassification in the randomly selected samples, the Mixed Logit model did not perform as well across varying business-cycles. In general, the Hazard model has the highest predictive power. However, the higher predictive power of the Bayesian model, when the ratio of the cost of Type I errors to the cost of Type II errors is high, is relatively consistent across all sampling methods. Such an advantage of the Bayesian model may make it more attractive in the current economic environment. This study extends recent research comparing the performance of bankruptcy prediction models by identifying under what conditions a model performs better. It also allays a range of user groups, including auditors, shareholders, employees, suppliers, rating agencies, and creditors' concerns with respect to assessing failure risk.
Resumo:
Responding to a series of articles in sport management literature calling for more diversity in terms of areas of interest or methods, this study warns against the danger of excessively fragmenting this field of research. The works of Kuhn (1962) and Pfeffer (1993) are taken as the basis of an argument that connects convergence with scientific strength. However, being aware of the large number of counterarguments directed at this line of reasoning, a new model of convergence, which focuses on clusters of research contributions with similar areas of interest, methods, and concepts, is proposed. The existence of these clusters is determined with the help of a bibliometric analysis of publications in three sport management journals. This examination determines that there are justified reasons to be concerned about the level of convergence in the field, pointing out to a reduced ability to create large clusters of contributions in similar areas of interest.
Resumo:
Dehumanizing ideologies that explicitly liken other humans to “inferior” animals can have negative consequences for intergroup attitudes and relations. Surprisingly, very little is known about the causes of dehumanization, and essentially no research has examined strategies for reducing dehumanizing tendencies. The Interspecies Model of Prejudice specifies that animalistic dehumanization may be rooted in basic hierarchical beliefs regarding human superiority over animals. This theoretical reasoning suggests that narrowing the human-animal divide should also reduce dehumanization. The purpose of the present dissertation, therefore, was to gain a more complete understanding of the predictors of and solutions to dehumanization by examining the Interspecies Model of Prejudice, first from a layperson’s perspective and then among young children. In Study 1, laypeople strongly rejected the human-animal divide as a probable cause of, or solution to, dehumanization, despite evidence that their own personal beliefs in the human-animal divide positively predicted their dehumanization (and prejudice) scores. From Study 1, it was concluded that the human-animal divide, despite being a robust empirical predictor of dehumanization, is largely unrecognized as a probable cause of, or solution to, dehumanization by non-experts in the psychology of prejudice. Studies 2 and 3 explored the expression of dehumanization, as well as the Interspecies Model of Prejudice, among children ages six to ten years (Studies 2 and 3) and parents (Study 3). Across both studies, White children showed evidence of racial dehumanization by attributing a Black child target fewer “uniquely human” characteristics than the White child target, representing the first systematic evidence of racial dehumanization among children. In Study 3, path analyses supported the Interspecies Model of Prejudice among children. Specifically, children’s beliefs in the human-animal divide predicted greater racial prejudice, an effect explained by heightened racial dehumanization. Moreover, parents’ Social Dominance Orientation (preference for social hierarchy and inequality) positively predicted children’s human-animal divide beliefs. Critically, these effects remained significant even after controlling for established predictors of child-prejudice (i.e., parent prejudice, authoritarian parenting, and social-cognitive skills) and relevant child demographics (i.e., age and sex). Similar patterns emerged among parent participants, further supporting the Interspecies Model of Prejudice. Encouragingly, children reported narrower human-animal divide perceptions after being exposed to an experimental prime (versus control) that highlighted the similarities among humans and animals. Together the three studies reported in this dissertation offer important and novel contributions to the dehumanization and prejudice literature. Not only did we find the first systematic evidence of racial dehumanization among children, we established the human-animal divide as a meaningful dehumanization precursor. Moreover, empirical support was obtained for the Interspecies Model of Prejudice among diverse samples including university students (Study 1), children (Studies 2 and 3), and adult-aged samples (Study 3). Importantly, each study also highlights the promising social implication of targeting the human-animal divide in interventions to reduce dehumanization and other prejudicial processes.
Resumo:
In the literature, persistent neural activity over frontal and parietal areas during the delay period of oculomotor delayed response (ODR) tasks has been interpreted as an active representation of task relevant information and response preparation. Following a recent ERP study (Tekok-Kilic, Tays, & Tkach, 2011 ) that reported task related slow wave differences over frontal and parietal sites during the delay periods of three ODR tasks, the present investigation explored developmental differences in young adults and adolescents during the same ODR tasks using 128-channel dense electrode array methodology and source localization. This exploratory study showed that neural functioning underlying visual-spatial WM differed between age groups in the Match condition. More specifically, this difference is localized anteriorly during the late delay period. Given the protracted maturation of the frontal lobes, the observed variation at the frontal site may indicate that adolescents and young adults may recruit frontal-parietal resources differently.
Resumo:
Affiliation: Institut de recherche en immunologie et en cancérologie, Université de Montréal
Resumo:
Introduction: Le but de l’étude était d’examiner l’effet des matériaux à empreintes sur la précision et la fiabilité des modèles d’études numériques. Méthodes: Vingt-cinq paires de modèles en plâtre ont été choisies au hasard parmi les dossiers de la clinique d’orthodontie de l’Université de Montréal. Une empreinte en alginate (Kromopan 100), une empreinte en substitut d’alginate (Alginot), et une empreinte en PVS (Aquasil) ont été prises de chaque arcade pour tous les patients. Les empreintes ont été envoyées chez Orthobyte pour la coulée des modèles en plâtre et la numérisation des modèles numériques. Les analyses de Bolton 6 et 12, leurs mesures constituantes, le surplomb vertical (overbite), le surplomb horizontal (overjet) et la longueur d’arcade ont été utilisés pour comparaisons. Résultats : La corrélation entre mesures répétées était de bonne à excellente pour les modèles en plâtre et pour les modèles numériques. La tendance voulait que les mesures répétées sur les modèles en plâtre furent plus fiables. Il existait des différences statistiquement significatives pour l’analyse de Bolton 12, pour la longueur d’arcade mandibulaire, et pour le chevauchement mandibulaire, ce pour tous les matériaux à empreintes. La tendance observée fut que les mesures sur les modèles en plâtre étaient plus petites pour l’analyse de Bolton 12 mais plus grandes pour la longueur d’arcade et pour le chevauchement mandibulaire. Malgré les différences statistiquement significatives trouvées, ces différences n’avaient aucune signification clinique. Conclusions : La précision et la fiabilité du logiciel pour l’analyse complète des modèles numériques sont cliniquement acceptables quand on les compare avec les résultats de l’analyse traditionnelle sur modèles en plâtre.
Resumo:
While high levels of Pkd1 expression are detected in tissues of patients with autosomal dominant polycystic kidney disease (ADPKD), it is unclear whether enhanced expression could be a pathogenetic mechanism for this systemic disorder. Three transgenic mouse lines were generated from a Pkd1-BAC modified by introducing a silent tag via homologous recombination to target a sustained wild type genomic Pkd1 expression within the native tissue and temporal regulation. These mice specifically overexpressed the Pkd1 transgene in extrarenal and renal tissues from approximately 2- to 15-fold over Pkd1 endogenous levels in a copy-dependent manner. All transgenic mice reproducibly developed tubular and glomerular cysts leading to renal insufficiency. Interestingly, Pkd1(TAG) mice also exhibited renal fibrosis and calcium deposits in papilla reminiscent of nephrolithiasis as frequently observed in ADPKD. Similar to human ADPKD, these mice consistently displayed hepatic fibrosis and approximately 15% intrahepatic cysts of the bile ducts affecting females preferentially. Moreover, a significant proportion of mice developed cardiac anomalies with severe left ventricular hypertrophy, marked aortic arch distention and/or valvular stenosis and calcification that had profound functional impact. Of significance, Pkd1(TAG) mice displayed occasional cerebral lesions with evidence of ruptured and unruptured cerebral aneurysms. This Pkd1(TAG) mouse model demonstrates that overexpression of wildtype Pkd1 can trigger the typical adult renal and extrarenal phenotypes resembling human ADPKD.
Resumo:
Cette thèse contribue à une théorie générale de la conception du projet. S’inscrivant dans une demande marquée par les enjeux du développement durable, l’objectif principal de cette recherche est la contribution d’un modèle théorique de la conception permettant de mieux situer l’utilisation des outils et des normes d’évaluation de la durabilité d’un projet. Les principes fondamentaux de ces instruments normatifs sont analysés selon quatre dimensions : ontologique, méthodologique, épistémologique et téléologique. Les indicateurs de certains effets contre-productifs reliés, en particulier, à la mise en compte de ces normes confirment la nécessité d’une théorie du jugement qualitatif. Notre hypothèse principale prend appui sur le cadre conceptuel offert par la notion de « principe de précaution » dont les premières formulations remontent du début des années 1970, et qui avaient précisément pour objectif de remédier aux défaillances des outils et méthodes d’évaluation scientifique traditionnelles. La thèse est divisée en cinq parties. Commençant par une revue historique des modèles classiques des théories de la conception (design thinking) elle se concentre sur l’évolution des modalités de prise en compte de la durabilité. Dans cette perspective, on constate que les théories de la « conception verte » (green design) datant du début des années 1960 ou encore, les théories de la « conception écologique » (ecological design) datant des années 1970 et 1980, ont finalement convergé avec les récentes théories de la «conception durable» (sustainable design) à partir du début des années 1990. Les différentes approches du « principe de précaution » sont ensuite examinées sous l’angle de la question de la durabilité du projet. Les standards d’évaluation des risques sont comparés aux approches utilisant le principe de précaution, révélant certaines limites lors de la conception d’un projet. Un premier modèle théorique de la conception intégrant les principales dimensions du principe de précaution est ainsi esquissé. Ce modèle propose une vision globale permettant de juger un projet intégrant des principes de développement durable et se présente comme une alternative aux approches traditionnelles d’évaluation des risques, à la fois déterministes et instrumentales. L’hypothèse du principe de précaution est dès lors proposée et examinée dans le contexte spécifique du projet architectural. Cette exploration débute par une présentation de la notion classique de «prudence» telle qu’elle fut historiquement utilisée pour guider le jugement architectural. Qu’en est-il par conséquent des défis présentés par le jugement des projets d’architecture dans la montée en puissance des méthodes d’évaluation standardisées (ex. Leadership Energy and Environmental Design; LEED) ? La thèse propose une réinterprétation de la théorie de la conception telle que proposée par Donald A. Schön comme une façon de prendre en compte les outils d’évaluation tels que LEED. Cet exercice révèle cependant un obstacle épistémologique qui devra être pris en compte dans une reformulation du modèle. En accord avec l’épistémologie constructiviste, un nouveau modèle théorique est alors confronté à l’étude et l’illustration de trois concours d'architecture canadienne contemporains ayant adopté la méthode d'évaluation de la durabilité normalisée par LEED. Une série préliminaire de «tensions» est identifiée dans le processus de la conception et du jugement des projets. Ces tensions sont ensuite catégorisées dans leurs homologues conceptuels, construits à l’intersection du principe de précaution et des théories de la conception. Ces tensions se divisent en quatre catégories : (1) conceptualisation - analogique/logique; (2) incertitude - épistémologique/méthodologique; (3) comparabilité - interprétation/analytique, et (4) proposition - universalité/ pertinence contextuelle. Ces tensions conceptuelles sont considérées comme autant de vecteurs entrant en corrélation avec le modèle théorique qu’elles contribuent à enrichir sans pour autant constituer des validations au sens positiviste du terme. Ces confrontations au réel permettent de mieux définir l’obstacle épistémologique identifié précédemment. Cette thèse met donc en évidence les impacts généralement sous-estimés, des normalisations environnementales sur le processus de conception et de jugement des projets. Elle prend pour exemple, de façon non restrictive, l’examen de concours d'architecture canadiens pour bâtiments publics. La conclusion souligne la nécessité d'une nouvelle forme de « prudence réflexive » ainsi qu’une utilisation plus critique des outils actuels d’évaluation de la durabilité. Elle appelle une instrumentalisation fondée sur l'intégration globale, plutôt que sur l'opposition des approches environnementales.
Resumo:
Les titres financiers sont souvent modélisés par des équations différentielles stochastiques (ÉDS). Ces équations peuvent décrire le comportement de l'actif, et aussi parfois certains paramètres du modèle. Par exemple, le modèle de Heston (1993), qui s'inscrit dans la catégorie des modèles à volatilité stochastique, décrit le comportement de l'actif et de la variance de ce dernier. Le modèle de Heston est très intéressant puisqu'il admet des formules semi-analytiques pour certains produits dérivés, ainsi qu'un certain réalisme. Cependant, la plupart des algorithmes de simulation pour ce modèle font face à quelques problèmes lorsque la condition de Feller (1951) n'est pas respectée. Dans ce mémoire, nous introduisons trois nouveaux algorithmes de simulation pour le modèle de Heston. Ces nouveaux algorithmes visent à accélérer le célèbre algorithme de Broadie et Kaya (2006); pour ce faire, nous utiliserons, entre autres, des méthodes de Monte Carlo par chaînes de Markov (MCMC) et des approximations. Dans le premier algorithme, nous modifions la seconde étape de la méthode de Broadie et Kaya afin de l'accélérer. Alors, au lieu d'utiliser la méthode de Newton du second ordre et l'approche d'inversion, nous utilisons l'algorithme de Metropolis-Hastings (voir Hastings (1970)). Le second algorithme est une amélioration du premier. Au lieu d'utiliser la vraie densité de la variance intégrée, nous utilisons l'approximation de Smith (2007). Cette amélioration diminue la dimension de l'équation caractéristique et accélère l'algorithme. Notre dernier algorithme n'est pas basé sur une méthode MCMC. Cependant, nous essayons toujours d'accélérer la seconde étape de la méthode de Broadie et Kaya (2006). Afin de réussir ceci, nous utilisons une variable aléatoire gamma dont les moments sont appariés à la vraie variable aléatoire de la variance intégrée par rapport au temps. Selon Stewart et al. (2007), il est possible d'approximer une convolution de variables aléatoires gamma (qui ressemble beaucoup à la représentation donnée par Glasserman et Kim (2008) si le pas de temps est petit) par une simple variable aléatoire gamma.
Resumo:
Ce mémoire de maîtrise présente une nouvelle approche non supervisée pour détecter et segmenter les régions urbaines dans les images hyperspectrales. La méthode proposée n ́ecessite trois étapes. Tout d’abord, afin de réduire le coût calculatoire de notre algorithme, une image couleur du contenu spectral est estimée. A cette fin, une étape de réduction de dimensionalité non-linéaire, basée sur deux critères complémentaires mais contradictoires de bonne visualisation; à savoir la précision et le contraste, est réalisée pour l’affichage couleur de chaque image hyperspectrale. Ensuite, pour discriminer les régions urbaines des régions non urbaines, la seconde étape consiste à extraire quelques caractéristiques discriminantes (et complémentaires) sur cette image hyperspectrale couleur. A cette fin, nous avons extrait une série de paramètres discriminants pour décrire les caractéristiques d’une zone urbaine, principalement composée d’objets manufacturés de formes simples g ́eométriques et régulières. Nous avons utilisé des caractéristiques texturales basées sur les niveaux de gris, la magnitude du gradient ou des paramètres issus de la matrice de co-occurrence combinés avec des caractéristiques structurelles basées sur l’orientation locale du gradient de l’image et la détection locale de segments de droites. Afin de réduire encore la complexité de calcul de notre approche et éviter le problème de la ”malédiction de la dimensionnalité” quand on décide de regrouper des données de dimensions élevées, nous avons décidé de classifier individuellement, dans la dernière étape, chaque caractéristique texturale ou structurelle avec une simple procédure de K-moyennes et ensuite de combiner ces segmentations grossières, obtenues à faible coût, avec un modèle efficace de fusion de cartes de segmentations. Les expérimentations données dans ce rapport montrent que cette stratégie est efficace visuellement et se compare favorablement aux autres méthodes de détection et segmentation de zones urbaines à partir d’images hyperspectrales.