57 resultados para Functions of complex variables.
Resumo:
Scale functions play a central role in the fluctuation theory of spectrally negative Lévy processes and often appear in the context of martingale relations. These relations are often require excursion theory rather than Itô calculus. The reason for the latter is that standard Itô calculus is only applicable to functions with a sufficient degree of smoothness and knowledge of the precise degree of smoothness of scale functions is seemingly incomplete. The aim of this article is to offer new results concerning properties of scale functions in relation to the smoothness of the underlying Lévy measure. We place particular emphasis on spectrally negative Lévy processes with a Gaussian component and processes of bounded variation. An additional motivation is the very intimate relation of scale functions to renewal functions of subordinators. The results obtained for scale functions have direct implications offering new results concerning the smoothness of such renewal functions for which there seems to be very little existing literature on this topic.
Resumo:
Radiometric data in the visible domain acquired by satellite remote sensing have proven to be powerful for monitoring the states of the ocean, both physical and biological. With the help of these data it is possible to understand certain variations in biological responses of marine phytoplankton on ecological time scales. Here, we implement a sequential data-assimilation technique to estimate from a conventional nutrient–phytoplankton–zooplankton (NPZ) model the time variations of observed and unobserved variables. In addition, we estimate the time evolution of two biological parameters, namely, the specific growth rate and specific mortality of phytoplankton. Our study demonstrates that: (i) the series of time-varying estimates of specific growth rate obtained by sequential data assimilation improves the fitting of the NPZ model to the satellite-derived time series: the model trajectories are closer to the observations than those obtained by implementing static values of the parameter; (ii) the estimates of unobserved variables, i.e., nutrient and zooplankton, obtained from an NPZ model by implementation of a pre-defined parameter evolution can be different from those obtained on applying the sequences of parameters estimated by assimilation; and (iii) the maximum estimated specific growth rate of phytoplankton in the study area is more sensitive to the sea-surface temperature than would be predicted by temperature-dependent functions reported previously. The overall results of the study are potentially useful for enhancing our understanding of the biological response of phytoplankton in a changing environment.
Resumo:
Forgetting immediate physical reality and having awareness of one�s location in the simulated world is critical to enjoyment and performance in virtual environments be it an interactive 3D game such as Quake or an online virtual 3d community space such as Second Life. Answer to the question "where am I?" at two levels, whether the locus is in the immediate real world as opposed to the virtual world and whether one is aware of the spatial co-ordinates of that locus, hold the key to any virtual 3D experience. While 3D environments, especially virtual environments and their impact on spatial comprehension has been studied in disciplines such as architecture, it is difficult to determine the relative contributions of specific attributes such as screen size or stereoscopy towards spatial comprehension since most of them treat the technology as monolith (box-centered). Using a variable-centered approach put forth by Nass and Mason (1990) which breaks down the technology into its component variables and their corresponding values as its theoretical basis, this paper looks at the contributions of five variables (Stereoscopy, screen size, field of view, level of realism and level of detail) common to most virtual environments on spatial comprehension and presence. The variable centered approach can be daunting as the increase in the number of variables can exponentially increase the number of conditions and resources required. We overcome this drawback posed by adoption of such a theoretical approach by the use of a fractional factorial design for the experiment. This study has completed the first wave of data collection and starting the next phase in January 2007 and expected to complete by February 2007. Theoretical and practical implications of the study are discussed.
Resumo:
The relative contributions of five variables (Stereoscopy, screen size, field of view, level of realism and level of detail) of virtual reality systems on spatial comprehension and presence are evaluated here. Using a variable-centered approach instead of an object-centric view as its theoretical basis, the contributions of these five variables and their two-way interactions are estimated through a 25-1 fractional factorial experiment (screening design) of resolution V with 84 subjects. The experiment design, procedure, measures used, creation of scales and indices, results of statistical analysis, their meaning and agenda for future research are elaborated.
Resumo:
A geometrical construction of the transcomplex numbers was given elsewhere. Here we simplify the transcomplex plane and construct the set of transcomplex numbers from the set of complex numbers. Thus transcomplex numbers and their arithmetic arise as consequences of their construction, not by an axiomatic development. This simplifes transcom- plex arithmetic, compared to the previous treatment, but retains totality so that every arithmetical operation can be applied to any transcomplex number(s) such that the result is a transcomplex number. Our proof establishes the consistency of transcomplex and transreal arithmetic and establishes the expected containment relationships amongst transcomplex, complex, transreal and real numbers. We discuss some of the advantages the transarithmetics have over their partial counterparts.
Resumo:
Understanding how and why the capability of one set of business resources, its structural arrangements and mechanisms compared to another works can provide competitive advantage in terms of new business processes and product and service development. However, most business models of capability are descriptive and lack formal modelling language to qualitatively and quantifiably compare capabilities, Gibson’s theory of affordance, the potential for action, provides a formal basis for a more robust and quantitative model, but most formal affordance models are complex and abstract and lack support for real-world applications. We aim to understand the ‘how’ and ‘why’ of business capability, by developing a quantitative and qualitative model that underpins earlier work on Capability-Affordance Modelling – CAM. This paper integrates an affordance based capability model and the formalism of Coloured Petri Nets to develop a simulation model. Using the model, we show how capability depends on the space time path of interacting resources, the mechanism of transition and specific critical affordance factors relating to the values of the variables for resources, people and physical objects. We show how the model can identify the capabilities of resources to enable the capability to inject a drug and anaesthetise a patient.
Resumo:
In the UK, architectural design is regulated through a system of design control for the public interest, which aims to secure and promote ‘quality’ in the built environment. Design control is primarily implemented by locally employed planning professionals with political oversight, and independent design review panels, staffed predominantly by design professionals. Design control has a lengthy and complex history, with the concept of ‘design’ offering a range of challenges for a regulatory system of governance. A simultaneously creative and emotive discipline, architectural design is a difficult issue to regulate objectively or consistently, often leading to policy that is regarded highly discretionary and flexible. This makes regulatory outcomes difficult to predict, as approaches undertaken by the ‘agents of control’ can vary according to the individual. The role of the design controller is therefore central, tasked with the responsibility of interpreting design policy and guidance, appraising design quality and passing professional judgment. However, little is really known about what influences the way design controllers approach their task, providing a ‘veil’ over design control, shrouding the basis of their decisions. This research engaged directly with the attitudes and perceptions of design controllers in the UK, lifting this ‘veil’. Using in-depth interviews and Q-Methodology, the thesis explores this hidden element of control, revealing a number of key differences in how controllers approach and implement policy and guidance, conceptualise design quality, and rationalise their evaluations and judgments. The research develops a conceptual framework for agency in design control – this consists of six variables (Regulation; Discretion; Skills; Design Quality; Aesthetics; and Evaluation) and it is suggested that this could act as a ‘heuristic’ instrument for UK controllers, prompting more reflexivity in relation to evaluating their own position, approaches, and attitudes, leading to better practice and increased transparency of control decisions.
Resumo:
This research explores whether patterns of typographic differentiation influence readers’ impressions of documents. It develops a systematic approach to typographic investigation that considers relationships between different kinds of typographic attributes, rather than testing the influence of isolated variables. An exploratory study using multiple sort tasks and semantic differential scales identifies that readers form a variety of impressions in relation to how typographic elements are differentiated in document design. Building on the findings of the exploratory study and analysis of a sample of magazines, the research describes three patterns of typographic differentiation: high, moderate, and low. Each pattern comprises clusters of typographic attributes and organisational principles that are articulated in relation to a specified level of typographic differentiation (amplified, medium, or subtle). The patterns are applied to two sets of controlled test material. Using this purposely-designed material, the influence of patterns of typographic differentiation on readers’ impressions of documents is explored in a repertory grid analysis and a paired comparison procedure. The results of these studies indicate that patterns of typographic differentiation consistently shape readers’ impressions of documents, influencing judgments of credibility, document address, and intended readership; and suggesting particular kinds of engagement and genre associations. For example, high differentiation documents are likely to be considered casual, sensationalist, and young; moderate differentiation documents are most likely to be seen as formal and serious; and low differentiation examples are considered calm. Typographic meaning is shown to be created through complex, yet systematic, interrelationships rather than reduced to a linear model of increasing or decreasing variation. The research provides a way of describing typographic articulation that has application across a variety of disciplines and design practice. In particular, it illuminates the ways in which typographic presentation is meaningful to readers, providing knowledge that document producers can use to communicate more effectively.
Resumo:
A new coupled cloud physics–radiation parameterization of the bulk optical properties of ice clouds is presented. The parameterization is consistent with assumptions in the cloud physics scheme regarding particle size distributions (PSDs) and mass–dimensional relationships. The parameterization is based on a weighted ice crystal habit mixture model, and its bulk optical properties are parameterized as simple functions of wavelength and ice water content (IWC). This approach directly couples IWC to the bulk optical properties, negating the need for diagnosed variables, such as the ice crystal effective dimension. The parameterization is implemented into the Met Office Unified Model Global Atmosphere 5.0 (GA5) configuration. The GA5 configuration is used to simulate the annual 20-yr shortwave (SW) and longwave (LW) fluxes at the top of the atmosphere (TOA), as well as the temperature structure of the atmosphere, under various microphysical assumptions. The coupled parameterization is directly compared against the current operational radiation parameterization, while maintaining the same cloud physics assumptions. In this experiment, the impacts of the two parameterizations on the SW and LW radiative effects at TOA are also investigated and compared against observations. The 20-yr simulations are compared against the latest observations of the atmospheric temperature and radiative fluxes at TOA. The comparisons demonstrate that the choice of PSD and the assumed ice crystal shape distribution are as important as each other. Moreover, the consistent radiation parameterization removes a long-standing tropical troposphere cold temperature bias but slightly warms the southern midlatitudes by about 0.5 K.
Resumo:
Objectives Extending the roles of nurses, pharmacists and allied health professionals to include prescribing has been identified as one way of improving service provision. In the UK, over 50 000 non-medical healthcare professionals are now qualified to prescribe. Implementation of non-medical prescribing ( NMP) is crucial to realise the potential return on investment. The UK Department of Health recommends a NMP lead to be responsible for the implementation of NMP within organisations. The aim of this study was to explore the role of NMP leads in organisations across one Strategic Health Authority (SHA) and to inform future planning with regards to the criteria for those adopting this role, the scope of the role and factors enabling the successful execution of the role. Methods Thirty-nine NMP leads across one SHA were approached. Semi-structured telephone interviews were conducted. Issues explored included the perceived role of the NMP lead, safety and clinical governance procedures and facilitators to the role. Transcribed audiotapes were coded and analysed using thematic analytical techniques. Key findings In total, 27/39 (69.2%) NMP leads were interviewed. The findings highlight the key role that the NMP lead plays with regards to the support and development of NMP within National Health Service trusts. Processes used to appoint NMP leads lacked clarity and varied between trusts. Only two NMP leads had designated or protected time for their role. Strategic influence, operational management and clinical governance were identified as key functions. Factors that supported the role included organisational support, level of influence and dedicated time. Conclusion The NMP lead plays a significant role in the development and implementation of NMP. Clear national guidance is needed with regards to the functions of this role, the necessary attributes for individuals recruited into this post and the time that should be designated to it. This is important as prescribing is extended to include other groups of non-medical healthcare professionals.
Resumo:
Objective. Therapeutic alliance, modality, and ability to engage with the process of therapy have been the main focus of research into what makes psychotherapy successful. Individuals with complex trauma histories or schizophrenia are suggested to be more difficult to engage and may be less likely to benefit from therapy. This study aimed to track the in-session ‘process’ of working alliance and emotional processing of trauma memories for individuals with schizophrenia. Design. The study utilized session recordings from the treatment arm of an open randomized clinical trial investigating trauma-focused cognitive behavioural therapy (TF-CBT) for individuals with schizophrenia (N = 26). Method. Observer measures of working alliance, emotional processing, and affect arousal were rated at early and late phases of therapy. Correlation analysis was undertaken for process measures. Temporal analysis of expressed emotions was also reported. Results. Working alliance was established and maintained throughout the therapy; however, agreement on goals reduced at the late phase. The participants appeared to be able to engage in emotional processing, but not to the required level for successful cognitive restructuring. Conclusion. This study undertook novel exploration of process variables not usually explored in CBT. It is also the first study of process for TF-CBT with individuals with schizophrenia. This complex clinical sample showed no difficulty in engagement; however, they may not be able to fully undertake the cognitive–emotional demands of this type of therapy. Clinical and research implications and potential limitations of these methods are considered.
Resumo:
The present longitudinal study examines the interaction of learner variables (gender, motivation, self-efficacy and first language literacy) and their influence on second language learning outcomes. The study follows English learners of French from Year 5 in primary school (aged 9-10) to the first year in secondary school (Year 7 aged 11-12). Language outcomes were measured by two oral production tasks; a sentence repetition task and a photo description task both of which were administered at three time points. Longitudinal data on learner attitudes and motivation were collected via questionnaires. Teacher assessment data for general first language literacy attainment were also provided. The results show a great deal of variation in learner attitudes and outcomes and that there is a complex relationship between first language literacy, self-efficacy, gender and attainment. For example, in general, girls held more positive attitudes to boys and were more successful. However, the inclusion of first language ability, which explained 30-40% of variation, shows that gender differences in attitudes and outcomes are likely mediated by first language literacy and prior learning experience.