956 resultados para Meaning-Text Theory
Resumo:
We measured the movements of soccer players heading a football in a fully immersive virtual reality environment. In mid-flight the ball’s trajectory was altered from its normal quasi-parabolic path to a linear one, producing a jump in the rate of change of the angle of elevation of gaze (α) from player to ball. One reaction time later the players adjusted their speed so that the rate of change of α increased when it had been reduced and reduced it when it had been increased. Since the result of the player’s movement was to regain a value of the rate of change close to that before the disturbance, the data suggest that the players have an expectation of, and memory for, the pattern that the rate of change of α will follow during the flight. The results support the general claim that players intercepting balls use servo control strategies and are consistent with the particular claim of Optic Acceleration Cancellation theory that the servo strategy is to allow α to increase at a steadily decreasing rate.
Resumo:
Firms form consortia in order to win contracts. Once a project has been awarded to a consortium each member then concentrates on his or her own contract with the client. Therefore, consortia are marketing devices, which present the impression of teamworking, but the production process is just as fragmented as under conventional procurement methods. In this way, the consortium forms a barrier between the client and the actual construction production process. Firms form consortia, not as a simple development of normal ways of working, but because the circumstances for specific projects make it a necessary vehicle. These circumstances include projects that are too large or too complex to undertake alone or projects that require on-going services which cannot be provided by the individual firms inhouse. It is not a preferred way of working, because participants carry extra risk in the form of liability for the actions of their partners in the consortium. The behaviour of members of consortia is determined by their relative power, based on several factors, including financial commitment and ease of replacement. The level of supply chain visibility to the public sector client and to the industry is reduced by the existence of a consortium because the consortium forms an additional obstacle between the client and the firms undertaking the actual construction work. Supply chain visibility matters to the client who otherwise loses control over the process of construction or service provision, while remaining accountable for cost overruns. To overcome this separation there is a convincing argument in favour of adopting the approach put forward in the Project Partnering Contract 2000 (PPC2000) Agreement. Members of consortia do not necessarily go on to work in the same consortia again because members need to respond flexibly to opportunities as and when they arise. Decision-making processes within consortia tend to be on an ad hoc basis. Construction risk is taken by the contractor and the construction supply chain but the reputational risk is carried by all the firms associated with a consortium. There is a wide variation in the manner that consortia are formed, determined by the individual circumstances of each project; its requirements, size and complexity, and the attitude of individual project leaders. However, there are a number of close working relationships based on generic models of consortia-like arrangements for the purpose of building production, such as the Housing Corporation Guidance Notes and the PPC2000.
Resumo:
This article outlines some of the key issues involved in developing a programme of strategy training for learners of French, in listening and in writing. It highlights the theoretical perspectives and research findings on listening and writing that informed the selection of strategies to teach learners and thence the development of appropriate materials. Examples of these materials are given as well as advice regarding their use. The article concludes with suggestions for how strategy training might be incorporated into teachers' own work with learners.
Resumo:
Theory of mind ability has been associated with performance in interpersonal interactions and has been found to influence aspects such as emotion recognition, social competence, and social anxiety. Being able to attribute mental states to others requires attention to subtle communication cues such as facial emotional expressions. Decoding and interpreting emotions expressed by the face, especially those with negative valence, are essential skills to successful social interaction. The current study explored the association between theory of mind skills and attentional bias to facial emotional expressions. According to the study hypothesis, individuals with poor theory of mind skills showed preferential attention to negative faces over both non-negative faces and neutral objects. Tentative explanations for the findings are offered emphasizing the potential adaptive role of vigilance for threat as a way of allocating a limited capacity to interpret others’ mental states to obtain as much information as possible about potential danger in the social environment.
Resumo:
Experimentally and theoretically determined infrared spectra are reported for a series of straight-chain perfluorocarbons: C2F6, C3F8, C4F10, C5F12, C6F14, and C8F18. Theoretical spectra were determined using both density functional (DFT) and ab initio methods. Radiative efficiencies (REs) were determined using the method of Pinnock et al. (1995) and combined with atmospheric lifetimes from the literature to determine global warming potentials (GWPs). Theoretically determined absorption cross sections were within 10% of experimentally determined values. Despite being much less computationally expensive, DFT calculations were generally found to perform better than ab initio methods. There is a strong wavenumber dependence of radiative forcing in the region of the fundamental C-F vibration, and small differences in wavelength between band positions determined by theory and experiment have a significant impact on the REs. We apply an empirical correction to the theoretical spectra and then test this correction on a number of branched chain and cyclic perfluoroalkanes. We then compute absorption cross sections, REs, and GWPs for an additional set of perfluoroalkenes.
Resumo:
The role of the academic in the built environment seems generally to be not well understood or articulated. While this problem is not unique to our field, there are plenty of examples in a wide range of academic disciplines where the academic role has been fully articulated. But built environment academics have tended not to look beyond their own literature and their own vocational context in trying to give meaning to their academic work. The purpose of this keynote presentation is to explore the context of academic work generally and the connections between education, research and practice in the built environment, specifically. By drawing on ideas from the sociology of the professions, the role of universities, and the fundamentals of social science research, a case is made that helps to explain the kind of problems that routinely obstruct academic progress in our field. This discussion reveals that while there are likely to be great weaknesses in much of what is published and taught in the built environment, it is not too great a stretch to provide a more robust understanding and a good basis for developing our field in a way that would enable us collectively to make a major contribution to theory-building, theory-testing and to make a good stab at tackling some of the problems facing society at large. There is no reason to disregard the fundamental academic disciplines that underpin our knowledge of the built environment. If we contextualise our work in these more fundamental disciplines, there is every reason to think that we can have a much greater impact that we have experienced to date.
Resumo:
The academic discipline of television studies has been constituted by the claim that television is worth studying because it is popular. Yet this claim has also entailed a need to defend the subject against the triviality that is associated with the television medium because of its very popularity. This article analyses the many attempts in the later twentieth and twenty-first centuries to constitute critical discourses about television as a popular medium. It focuses on how the theoretical currents of Television Studies emerged and changed in the UK, where a disciplinary identity for the subject was founded by borrowing from related disciplines, yet argued for the specificity of the medium as an object of criticism. Eschewing technological determinism, moral pathologization and sterile debates about television's supposed effects, UK writers such as Raymond Williams addressed television as an aspect of culture. Television theory in Britain has been part of, and also separate from, the disciplinary fields of media theory, literary theory and film theory. It has focused its attention on institutions, audio-visual texts, genres, authors and viewers according to the ways that research problems and theoretical inadequacies have emerged over time. But a consistent feature has been the problem of moving from a descriptive discourse to an analytical and evaluative one, and from studies of specific texts, moments and locations of television to larger theories. By discussing some historically significant critical work about television, the article considers how academic work has constructed relationships between the different kinds of objects of study. The article argues that a fundamental tension between descriptive and politically activist discourses has confused academic writing about ›the popular‹. Television study in Britain arose not to supply graduate professionals to the television industry, nor to perfect the instrumental techniques of allied sectors such as advertising and marketing, but to analyse and critique the medium's aesthetic forms and to evaluate its role in culture. Since television cannot be made by ›the people‹, the empowerment that discourses of television theory and analysis aimed for was focused on disseminating the tools for critique. Recent developments in factual entertainment television (in Britain and elsewhere) have greatly increased the visibility of ›the people‹ in programmes, notably in docusoaps, game shows and other participative formats. This has led to renewed debates about whether such ›popular‹ programmes appropriately represent ›the people‹ and how factual entertainment that is often despised relates to genres hitherto considered to be of high quality, such as scripted drama and socially-engaged documentary television. A further aspect of this problem of evaluation is how television globalisation has been addressed, and the example that the issue has crystallised around most is the reality TV contest Big Brother. Television theory has been largely based on studying the texts, institutions and audiences of television in the Anglophone world, and thus in specific geographical contexts. The transnational contexts of popular television have been addressed as spaces of contestation, for example between Americanisation and national or regional identities. Commentators have been ambivalent about whether the discipline's role is to celebrate or critique television, and whether to do so within a national, regional or global context. In the discourses of the television industry, ›popular television‹ is a quantitative and comparative measure, and because of the overlap between the programming with the largest audiences and the scheduling of established programme types at the times of day when the largest audiences are available, it has a strong relationship with genre. The measurement of audiences and the design of schedules are carried out in predominantly national contexts, but the article refers to programmes like Big Brother that have been broadcast transnationally, and programmes that have been extensively exported, to consider in what ways they too might be called popular. Strands of work in television studies have at different times attempted to diagnose what is at stake in the most popular programme types, such as reality TV, situation comedy and drama series. This has centred on questions of how aesthetic quality might be discriminated in television programmes, and how quality relates to popularity. The interaction of the designations ›popular‹ and ›quality‹ is exemplified in the ways that critical discourse has addressed US drama series that have been widely exported around the world, and the article shows how the two critical terms are both distinct and interrelated. In this context and in the article as a whole, the aim is not to arrive at a definitive meaning for ›the popular‹ inasmuch as it designates programmes or indeed the medium of television itself. Instead the aim is to show how, in historically and geographically contingent ways, these terms and ideas have been dynamically adopted and contested in order to address a multiple and changing object of analysis.
Resumo:
Another Proof of the Preceding Theory was produced as part of a residency run by Artists in Archeology in conjunction with the Stonehenge Riverside project. The film explores the relationship between science, work and ritual, imagining archaeology as a future cult. As two robed disciples stray off from the dig, they are drawn to the drone of the stones and proceed to play the henge like a gigantic Theremin. Just as a Theremin is played with the hand interfering in an electric circuit and producing sound without contact, so the stones respond to the choreographed bodily proximity. Finally, one of the two continues alone to the avenue at Avebury, where the magnetic pull of the stones reaches its climax. Shot on VHS, the film features a score by Zuzushi Monkey, with percussion and theremin sounds mirroring the action. The performers are mostly artists and archeologists from the art and archaeology teams. The archeologists were encouraged to perform their normal work in the robes, in an attempt to explore the meeting points of science and ritual and interrogate our relationship to an ultimately unknowable prehistoric past where activities we do not understand are relegated to the realm of religion. Stonehenge has unique acoustic properties, it’s large sarsen stones are finely worked on the inside, left rough on the outside, intensifying sound waves within the inner horseshoe, but since their real use, having been built over centuries, remains ambiguous, the film proposes that our attempts to decode them may themselves become encoded in their cumulative meaning for future researchers.
Resumo:
Twenty first century challenges facing agriculture include climate change, threats to food security for a growing population and downward economic pressures on rural livelihoods. Addressing these challenges will require innovation in extension theory, policy and education, at a time when the dominance of the state in the provision of knowledge and information services to farmers and rural entrepreneurs continues to decline. This paper suggests that extension theory is catching up with and helping us to understand innovative extension practice, and therefore provides a platform for improving rural development policies and strategies. Innovation is now less likely to be spoken of as something to be passed on to farmers, than as a continuing process of creativity and adaptation that can be nurtured and sustained. Innovation systems and innovation platforms are concepts that recognise the multiple factors that lead to farmers’ developing, adapting and applying new ideas and the importance of linking all actors in the value chain to ensure producers can access appropriate information and advice for decision making at all stages in the production process. Concepts of social learning, group development and solidarity, social capital, collective action and empowerment all help to explain and therefore to apply more effectively group extension approaches in building confidence and sustaining innovation. A challenge facing educators is to ensure the curricula for aspiring extension professionals in our higher education institutions are regularly reviewed and keep up with current and future developments in theory, policy and practice.
Resumo:
This paper compares a number of different extreme value models for determining the value at risk (VaR) of three LIFFE futures contracts. A semi-nonparametric approach is also proposed, where the tail events are modeled using the generalised Pareto distribution, and normal market conditions are captured by the empirical distribution function. The value at risk estimates from this approach are compared with those of standard nonparametric extreme value tail estimation approaches, with a small sample bias-corrected extreme value approach, and with those calculated from bootstrapping the unconditional density and bootstrapping from a GARCH(1,1) model. The results indicate that, for a holdout sample, the proposed semi-nonparametric extreme value approach yields superior results to other methods, but the small sample tail index technique is also accurate.
Resumo:
Real estate development appraisal is a quantification of future expectations. The appraisal model relies upon the valuer/developer having an understanding of the future in terms of the future marketability of the completed development and the future cost of development. In some cases the developer has some degree of control over the possible variation in the variables, as with the cost of construction through the choice of specification. However, other variables, such as the sale price of the final product, are totally dependent upon the vagaries of the market at the completion date. To try to address the risk of a different outcome to the one expected (modelled) the developer will often carry out a sensitivity analysis on the development. However, traditional sensitivity analysis has generally only looked at the best and worst scenarios and has focused on the anticipated or expected outcomes. This does not take into account uncertainty and the range of outcomes that can happen. A fuller analysis should include examination of the uncertainties in each of the components of the appraisal and account for the appropriate distributions of the variables. Similarly, as many of the variables in the model are not independent, the variables need to be correlated. This requires a standardised approach and we suggest that the use of a generic forecasting software package, in this case Crystal Ball, allows the analyst to work with an existing development appraisal model set up in Excel (or other spreadsheet) and to work with a predetermined set of probability distributions. Without a full knowledge of risk, developers are unable to determine the anticipated level of return that should be sought to compensate for the risk. This model allows the user a better understanding of the possible outcomes for the development. Ultimately the final decision will be made relative to current expectations and current business constraints, but by assessing the upside and downside risks more appropriately, the decision maker should be better placed to make a more informed and “better”.
Resumo:
This paper investigates the frequency of extreme events for three LIFFE futures contracts for the calculation of minimum capital risk requirements (MCRRs). We propose a semiparametric approach where the tails are modelled by the Generalized Pareto Distribution and smaller risks are captured by the empirical distribution function. We compare the capital requirements form this approach with those calculated from the unconditional density and from a conditional density - a GARCH(1,1) model. Our primary finding is that both in-sample and for a hold-out sample, our extreme value approach yields superior results than either of the other two models which do not explicitly model the tails of the return distribution. Since the use of these internal models will be permitted under the EC-CAD II, they could be widely adopted in the near future for determining capital adequacies. Hence, close scrutiny of competing models is required to avoid a potentially costly misallocation capital resources while at the same time ensuring the safety of the financial system.