55 resultados para implicit RK
Resumo:
Protein-ligand docking has made important progress during the last decade and has become a powerful tool for drug development, opening the way to virtual high throughput screening and in silico structure-based ligand design. Despite the flattering picture that has been drawn, recent publications have shown that the docking problem is far from being solved, and that more developments are still needed to achieve high successful prediction rates and accuracy. Introducing an accurate description of the solvation effect upon binding is thought to be essential to achieve this goal. In particular, EADock uses the Generalized Born Molecular Volume 2 (GBMV2) solvent model, which has been shown to reproduce accurately the desolvation energies calculated by solving the Poisson equation. Here, the implementation of the Fast Analytical Continuum Treatment of Solvation (FACTS) as an implicit solvation model in small molecules docking calculations has been assessed using the EADock docking program. Our results strongly support the use of FACTS for docking. The success rates of EADock/FACTS and EADock/GBMV2 are similar, i.e. around 75% for local docking and 65% for blind docking. However, these results come at a much lower computational cost: FACTS is 10 times faster than GBMV2 in calculating the total electrostatic energy, and allows a speed up of EADock by a factor of 4. This study also supports the EADock development strategy relying on the CHARMM package for energy calculations, which enables straightforward implementation and testing of the latest developments in the field of Molecular Modeling.
Resumo:
The population density of an organism is one of the main aspects of its environment, and shoud therefore strongly influence its adaptive strategy. The r/K theory, based on the logistic model, was developed to formalize this influence. K-selectioon is classically thought to favour large body sizes. This prediction, however, cannot be directly derived from the logistic model: some auxiliary hypotheses are therefor implicit. These are to be made explicit if the theory is to be tested. An alternative approach, based on the Euler-Lotka equation, shows that density itself is irrelevant, but that the relative effect of density on adult and juvenile features is crucial. For instance, increasing population will select for a smaller body size if the density affects mainly juvenile growth and/or survival. In this case, density shoud indeed favour large body sizes. The theory appears nevertheless inconsistent, since a probable consequence of increasing body size will be a decrease in the carrying capacity
Resumo:
This article summarizes current concepts of the working memory with regard to its role within emotional coping strategies. In particular, it focuses on the fact that the limited capacity of the working memory to process now-relevant information can be turned into an advantage, when the individual is occupied by dealing with unpleasant emotion. Based on a phenomenon known as dual-task interference (DTI), this emotion can be chased by intense arousal due to clearly identifiable external stressors. Thus, risk perception might be used as a 'DTI inductor' that allows avoidance of unpleasant emotion. Successful mastery of risk adds a highly relevant dopaminergic component to the overall experience. The resulting mechanism of implicit learning may contribute to the development of a behavioural addiction. Besides its putative effects in the development of a behavioural addiction, the use of DTI might be of a more general interest for the clinical practice, especially in the field of psychotherapy. © 2013 S. Karger AG, Basel.
Resumo:
Social scientists often estimate models from correlational data, where the independent variable has not been exogenously manipulated; they also make implicit or explicit causal claims based on these models. When can these claims be made? We answer this question by first discussing design and estimation conditions under which model estimates can be interpreted, using the randomized experiment as the gold standard. We show how endogeneity--which includes omitted variables, omitted selection, simultaneity, common methods bias, and measurement error--renders estimates causally uninterpretable. Second, we present methods that allow researchers to test causal claims in situations where randomization is not possible or when causal interpretation is confounded, including fixed-effects panel, sample selection, instrumental variable, regression discontinuity, and difference-in-differences models. Third, we take stock of the methodological rigor with which causal claims are being made in a social sciences discipline by reviewing a representative sample of 110 articles on leadership published in the previous 10 years in top-tier journals. Our key finding is that researchers fail to address at least 66 % and up to 90 % of design and estimation conditions that make causal claims invalid. We conclude by offering 10 suggestions on how to improve non-experimental research.
Resumo:
This article describes the long-lasting psychological after-effects of a traumatic experience. There is growing lmowledge of the biomedical underpinnings of these phenomena: the underlying mechanisms belong to an implicit learning process whereby the victim remains under the influence of painful past experiences. One of these mechanisms concerns the development of a traumatic bonding which Iioticeably impedes the establishment of interpersona! relationships. The other mechanism, called "contextualisation deficit", is the difficulty of adjusting a person's emotional and behavioural reactivity to the context of present day !ife. This capacity of a traumatic experience to become incrusted long-termina human being's mind, and to haunt the victim with various forms of psychological and physical suffering, can be compared with the presence of a tumour or an abscess in somatic medicine. Th us, severe drug addiction can be conceptualised as a disorder in which the patient tries - in most cases ineffectively - to soothe the pain of today's world in cmmection with the trauma of the past. In conclusion, this article urges the development of psychiatrie care programmes which operate at the centre of the suffering encountered by the se patients, as a complement to the already well-established offers such as harm reduction, substitution therapy and social support.
Resumo:
Discriminating complex sounds relies on multiple stages of differential brain activity. The specific roles of these stages and their links to perception were the focus of the present study. We presented 250ms duration sounds of living and man-made objects while recording 160-channel electroencephalography (EEG). Subjects categorized each sound as that of a living, man-made or unknown item. We tested whether/when the brain discriminates between sound categories even when not transpiring behaviorally. We applied a single-trial classifier that identified voltage topographies and latencies at which brain responses are most discriminative. For sounds that the subjects could not categorize, we could successfully decode the semantic category based on differences in voltage topographies during the 116-174ms post-stimulus period. Sounds that were correctly categorized as that of a living or man-made item by the same subjects exhibited two periods of differences in voltage topographies at the single-trial level. Subjects exhibited differential activity before the sound ended (starting at 112ms) and on a separate period at ~270ms post-stimulus onset. Because each of these periods could be used to reliably decode semantic categories, we interpreted the first as being related to an implicit tuning for sound representations and the second as being linked to perceptual decision-making processes. Collectively, our results show that the brain discriminates environmental sounds during early stages and independently of behavioral proficiency and that explicit sound categorization requires a subsequent processing stage.
Resumo:
The multiscale finite-volume (MSFV) method is designed to reduce the computational cost of elliptic and parabolic problems with highly heterogeneous anisotropic coefficients. The reduction is achieved by splitting the original global problem into a set of local problems (with approximate local boundary conditions) coupled by a coarse global problem. It has been shown recently that the numerical errors in MSFV results can be reduced systematically with an iterative procedure that provides a conservative velocity field after any iteration step. The iterative MSFV (i-MSFV) method can be obtained with an improved (smoothed) multiscale solution to enhance the localization conditions, with a Krylov subspace method [e.g., the generalized-minimal-residual (GMRES) algorithm] preconditioned by the MSFV system, or with a combination of both. In a multiphase-flow system, a balance between accuracy and computational efficiency should be achieved by finding a minimum number of i-MSFV iterations (on pressure), which is necessary to achieve the desired accuracy in the saturation solution. In this work, we extend the i-MSFV method to sequential implicit simulation of time-dependent problems. To control the error of the coupled saturation/pressure system, we analyze the transport error caused by an approximate velocity field. We then propose an error-control strategy on the basis of the residual of the pressure equation. At the beginning of simulation, the pressure solution is iterated until a specified accuracy is achieved. To minimize the number of iterations in a multiphase-flow problem, the solution at the previous timestep is used to improve the localization assumption at the current timestep. Additional iterations are used only when the residual becomes larger than a specified threshold value. Numerical results show that only a few iterations on average are necessary to improve the MSFV results significantly, even for very challenging problems. Therefore, the proposed adaptive strategy yields efficient and accurate simulation of multiphase flow in heterogeneous porous media.
Resumo:
"IT'S THE ECONOMY STUPID", BUT CHARISMA MATTERS TOO: A DUAL PROCESS MODEL OF PRESIDENTIAL ELECTION OUTCOMES. ABSTRACT Because charisma is assumed to be an important determinant of effective leadership, the extent to which a presidential nominee is more charismatic than his opponent should be an important determinant of voter choices. We computed a composite measure of the rhetorical richness of acceptances speeches given by U.S. presidential candidates at their national party convention. We added this marker of charisma to Ray C. Fair's presidential vote-share equation (1978; 2009). We theorized that voters decide using psychological attribution (i.e., due to macroeconomics and incumbency) as well as inferential processes (i.e., due to leader charismatic behavior) when voting. Controlling for the macro-level variables and incumbency in the Fair model, our results indicated that difference between nominees' charisma is a significant determinant of electoral success, particularly in close elections. This extended model significantly improves the precision of the Fair model and correctly predicts 23 out of the last 24 U.S. presidential elections. Paper 2: IT CEO LEADERSHIP, CORPORATE SOCIAL AND FINANCIAL PERFORMANCE. ABSTRACT We investigated whether CEO leadership predicted corporate financial performance (CFP) and corporate social performance (CSP). Using longitudinal data on 258 CEOs from 117 firms across 19 countries and 10 industry sectors, we found that determinants of CEO leadership (i.e., implicit motives) significantly predicted both CFP and CSP. As expected, the most consistent positive predictor was Responsibility Disposition when interacting with n (need for) Power. n Achievement and n Affiliation were generally negatively related or unrelated to outcomes. CSP was positively related to accounting measures of CFP. Our findings suggest that executive leader characteristics have important consequences for corporate level outcomes. Paper 3. PUNISHING THE POWERFUL: ATTRIBUTIONS OF BLAME AND LEADERSHIP ABSTRACT We propose that individuals are more lenient in attributing blame to leaders than to nonleaders. We advance a motivational explanation building on the perspective of punishment and on system justification theory. We conducted two scenario experiments which supported our proposition. In study 1, wrongdoer leader status was negatively related to blame and the perceived seriousness of the wrongdoing. In study 2, controlling for the Big-Five personality factor and individual differences in moral evaluation (i.e., moral foundations), wrongdoer leader status was negatively related with desired severity of punishment, and fair punishments were perceived as more just for non-leaders than for leaders.
Resumo:
Dans le domaine de la perception, l'apprentissage est contraint par la présence d'une architecture fonctionnelle constituée d'aires corticales distribuées et très spécialisées. Dans le domaine des troubles visuels d'origine cérébrale, l'apprentissage d'un patient hémi-anopsique ou agnosique sera limité par ses capacités perceptives résiduelles, mais un déficit de reconnaissance visuelle de nature apparemment perceptive, peut également être associé à une altération des représentations en mémoire à long terme. Des réseaux neuronaux distincts pour la reconnaissance - cortex temporal - et pour la localisation des sons - cortex pariétal - ont été décrits chez l'homme. L'étude de patients cérébro-lésés confirme le rôle des indices spatiaux dans un traitement auditif explicite du « where » et dans la discrimination implicite du « what ». Cette organisation, similaire à ce qui a été décrit dans la modalité visuelle, faciliterait les apprentissages perceptifs. Plus généralement, l'apprentissage implicite fonde une grande partie de nos connaissances sur le monde en nous rendant sensible, à notre insu, aux règles et régularités de notre environnement. Il serait impliqué dans le développement cognitif, la formation des réactions émotionnelles ou encore l'apprentissage par le jeune enfant de sa langue maternelle. Le caractère inconscient de cet apprentissage est confirmé par l'étude des temps de réaction sériels de patients amnésiques dans l'acquisition d'une grammaire artificielle. Son évaluation pourrait être déterminante dans la prise en charge ré-adaptative. [In the field of perception, learning is formed by a distributed functional architecture of very specialized cortical areas. For example, capacities of learning in patients with visual deficits - hemianopia or visual agnosia - from cerebral lesions are limited by perceptual abilities. Moreover a visual deficit in link with abnormal perception may be associated with an alteration of representations in long term (semantic) memory. Furthermore, perception and memory traces rely on parallel processing. This has been recently demonstrated for human audition. Activation studies in normal subjects and psychophysical investigations in patients with focal hemispheric lesions have shown that auditory information relevant to sound recognition and that relevant to sound localisation are processed in parallel, anatomically distinct cortical networks, often referred to as the "What" and "Where" processing streams. Parallel processing may appear counterintuitive from the point of view of a unified perception of the auditory world, but there are advantages, such as rapidity of processing within a single stream, its adaptability in perceptual learning or facility of multisensory interactions. More generally, implicit learning mechanisms are responsible for the non-conscious acquisition of a great part of our knowledge about the world, using our sensitivity to the rules and regularities structuring our environment. Implicit learning is involved in cognitive development, in the generation of emotional processing and in the acquisition of natural language. Preserved implicit learning abilities have been shown in amnesic patients with paradigms like serial reaction time and artificial grammar learning tasks, confirming that implicit learning mechanisms are not sustained by the cognitive processes and the brain structures that are damaged in amnesia. In a clinical perspective, the assessment of implicit learning abilities in amnesic patients could be critical for building adapted neuropsychological rehabilitation programs.]
Resumo:
In line with educational issues involved in emergent literacy practices in preschool, in particular those concerning comprehension processes, this paper focuses on picture-based narrative comprehension during an interactive reading session of a wordless picture book, involving a group of children aged three and their teacher. Children are asked to make inferences about the meaning and outcome of the story, a procedure which gradually elicits their responses on how events link together, thus enhancing their capacity to use prior and implicit knowledge to build the story meaning. Moreover, this study highlights the importance of interaction for developing comprehension. Data collected was analysed following didactic microgenesis, an analytical approach showing that knowledge built during interaction depends on the joint construction of a zone of common meaning by which teacher and children try to adjust to each other. In order to help the process of merging different meanings of the story built online, a text written by researchers, following the narrative structure of the story, was read by the teacher after the picture-based reading. This led us to examine through interactional analysis which semiotic cues were used during recall on the following day, as an additional measure of knowledge construction.
Resumo:
BACKGROUND: An auditory perceptual learning paradigm was used to investigate whether implicit memories are formed during general anesthesia. METHODS: Eighty-seven patients who had an American Society of Anesthesiologists physical status of I-III and were scheduled to undergo an elective surgery with general anesthesia were randomly assigned to one of two groups. One group received auditory stimulation during surgery, whereas the other did not. The auditory stimulation consisted of pure tones presented via headphones. The Bispectral Index level was maintained between 40 and 50 during surgery. To assess learning, patients performed an auditory frequency discrimination task after surgery, and comparisons were made between the groups. General anesthesia was induced with thiopental and maintained with a mixture of fentanyl and sevoflurane. RESULTS: There was no difference in the amount of learning between the two groups (mean +/- SD improvement: stimulated patients 9.2 +/- 11.3 Hz, controls 9.4 +/- 14.1 Hz). There was also no difference in initial thresholds (mean +/- SD initial thresholds: stimulated patients 31.1 +/- 33.4 Hz, controls 28.4 +/- 34.2 Hz). These results suggest that perceptual learning was not induced during anesthesia. No correlation between the bispectral index and the initial level of performance was found (Pearson r = -0.09, P = 0.59). CONCLUSION: Perceptual learning was not induced by repetitive auditory stimulation during anesthesia. This result may indicate that perceptual learning requires top-down processing, which is suppressed by the anesthetic.
Resumo:
Objectives To consider the various specific substances-taking activities in sport an examination of three psychological models of doping behaviour utilised by researchers is presented in order to evaluate their real and potential impact, and to improve the relevance and efficiency of anti-doping campaigns. Design Adopting the notion of a "research program" (Lakatos, 1978) from the philosophy of science, a range of studies into the psychology of doping behaviour are classified and critically analysed. Method Theoretical and practical parameters of three research programs are critically evaluated (i) cognitive; (ii) drive; and (iii) situated-dynamic. Results The analysis reveals the diversity of theoretical commitments of the research programs and their practical consequences. The «cognitive program» assumes that athletes are accountable for their acts that reflect the endeavour to attain sporting and non-sporting goals. Attitudes, knowledge and rational decisions are understood to be the basis of doping behaviour. The «drive program» characterises the variety of traces and consequences on psychological and somatic states coming from athlete's experience with sport. Doping behaviour here is conceived of as a solution to reduce unconscious psychological and somatic distress. The «situated-dynamic program» considers a broader context of athletes' doping activity and its evolution during a sport career. Doping is considered as emergent and self-organized behaviour, grounded on temporally critical couplings between athletes' actions and situations and the specific dynamics of their development during the sporting life course. Conclusions These hypothetical, theoretical and methodological considerations offer a more nuanced understanding of doping behaviours, making an effective contribution to anti-doping education and research by enabling researchers and policy personnel to become more critically reflective about their explicit and implicit assumptions regarding models of explanations for doping behaviour.
Resumo:
Aim Structure of the Thesis In the first article, I focus on the context in which the Homo Economicus was constructed - i.e., the conception of economic actors as fully rational, informed, egocentric, and profit-maximizing. I argue that the Homo Economicus theory was developed in a specific societal context with specific (partly tacit) values and norms. These norms have implicitly influenced the behavior of economic actors and have framed the interpretation of the Homo Economicus. Different factors however have weakened this implicit influence of the broader societal values and norms on economic actors. The result is an unbridled interpretation and application of the values and norms of the Homo Economicus in the business environment, and perhaps also in the broader society. In the second article, I show that the morality of many economic actors relies on isomorphism, i.e., the attempt to fit into the group by adopting the moral norms surrounding them. In consequence, if the norms prevailing in a specific group or context (such as a specific region or a specific industry) change, it can be expected that actors with an 'isomorphism morality' will also adapt their ethical thinking and their behavior -for the 'better' or for the 'worse'. The article further describes the process through which corporations could emancipate from the ethical norms prevailing in the broader society, and therefore develop an institution with specific norms and values. These norms mainly rely on mainstream business theories praising the economic actor's self-interest and neglecting moral reasoning. Moreover, because of isomorphism morality, many economic actors have changed their perception of ethics, and have abandoned the values prevailing in the broader society in order to adopt those of the economic theory. Finally, isomorphism morality also implies that these economic actors will change their morality again if the institutional context changes. The third article highlights the role and responsibility of business scholars in promoting a systematic reflection and self-critique of the business system and develops alternative models to fill the moral void of the business institution and its inherent legitimacy crisis. Indeed, the current business institution relies on assumptions such as scientific neutrality and specialization, which seem at least partly challenged by two factors. First, self-fulfilling prophecy provides scholars with an important (even if sometimes undesired) normative influence over practical life. Second, the increasing complexity of today's (socio-political) world and interactions between the different elements constituting our society question the strong specialization of science. For instance, economic theories are not unrelated to psychology or sociology, and economic actors influence socio-political structures and processes, e.g., through lobbying (Dobbs, 2006; Rondinelli, 2002), or through marketing which changes not only the way we consume, but more generally tries to instill a specific lifestyle (Cova, 2004; M. K. Hogg & Michell, 1996; McCracken, 1988; Muniz & O'Guinn, 2001). In consequence, business scholars are key actors in shaping both tomorrow's economic world and its broader context. A greater awareness of this influence might be a first step toward an increased feeling of civic responsibility and accountability for the models and theories developed or taught in business schools.
Resumo:
In the healthcare debate, it is often stated that better quality leads to savings. Quality systems lead to additional costs for setting up, running and external evaluations. In addition, suppression of implicit rationing leads to additional costs. On the other hand, they lead to savings by procedures simplification, improvement of patients' health state and quicker integration of new collaborators. It is then logical to imagine that financial incentives could improve quality. First evidences of pay for performances initiatives show a positive impact but also some limitations. Quality and savings are linked together and require all our attention.
Resumo:
A workshop recently held at the Ecole Polytechnique Federale de Lausanne (EPFL, Switzerland) was dedicated to understanding the genetic basis of adaptive change, taking stock of the different approaches developed in theoretical population genetics and landscape genomics and bringing together knowledge accumulated in both research fields. Indeed, an important challenge in theoretical population genetics is to incorporate effects of demographic history and population structure. But important design problems (e.g. focus on populations as units, focus on hard selective sweeps, no hypothesis-based framework in the design of the statistical tests) reduce their capability of detecting adaptive genetic variation. In parallel, landscape genomics offers a solution to several of these problems and provides a number of advantages (e.g. fast computation, landscape heterogeneity integration). But the approach makes several implicit assumptions that should be carefully considered (e.g. selection has had enough time to create a functional relationship between the allele distribution and the environmental variable, or this functional relationship is assumed to be constant). To address the respective strengths and weaknesses mentioned above, the workshop brought together a panel of experts from both disciplines to present their work and discuss the relevance of combining these approaches, possibly resulting in a joint software solution in the future.