20 resultados para Driving issues
em Greenwich Academic Literature Archive - UK
Resumo:
Traditionally, when designing a ship the driving issues are seen to be powering, stability, strength and seakeeping. Issues related to ship operations and evolutions are investigated later in the design process, within the constraint of a fixed layout. This can result in operational inefficiencies and limitations, excessive crew numbers and potentially hazardous situations. University College London and the University of Greenwich are in the final year of a three year EPSRC funded research project to integrate the simulation of personnel movement into early stage ship design. This allows the assessment of onboard operations while the design is still amenable to change. The project brings together the University of Greenwich developed maritimeEXODUS personnel movement simulation software and the SURFCON implementation of the Design Building Block approach to early stage ship design, which originated with the UCL Ship Design Research team. Central to the success of this project is the definition of a suitable series of Naval Combatant Human Performance Metrics which can be used to assess the performance of the design in different operational scenarios. The paper outlines the progress made on deriving the human performance metric from human factors criteria measured in simulations and their incorporation into a Behavioural Matrix for analysis. It describes the production of a series of SURFCON ship designs based on the RN Type 22 Batch 3 frigate, and their analysis using the PARAMARINE and maritimeEXODUS software. Conclusions to date will be presented on the integration of personnel movement simulation into the preliminary ship design process.
Resumo:
Traditionally, when designing a ship the driving issues are seen to be powering, stability, strength and seakeeping. Issues related to ship operations and evolutions are investigated later in the design process, within the constraint of a fixed layout. This can result in operational inefficiencies and limitations, excessive crew numbers and potentially hazardous situations. This paper summarises work by University College London and the University of Greenwich prior to the completion of a three year EPSRC funded research project to integrate the simulation of personnel movement into early stage ship design. This integration is intended to facilitate the assessment of onboard operations while the design is still highly amenable to change. The project brings together the University of Greenwich developed maritimeEXODUS personnel movement simulation software and the SURFCON implementation of the Design Building Block approach to early stage ship design, which originated with the UCL Ship Design Research team and has been implemented within the PARAMARINE ship design system produced by Graphics Research Corporation. Central to the success of this project is the definition of a suitable series of Performance Measures (PM) which can be used to assess the human performance of the design in different operational scenarios. The paper outlines the progress made on deriving the PM from human dynamics criteria measured in simulations and their incorporation into a Human Performance Metric (HPM) for analysis. It describes the production of a series of SURFCON ship designs, based on the Royal Navy’s Type 22 Batch 3 frigate, and their analysis using the PARAMARINE and maritimeEXODUS software. Conclusions on the work to date and for the remainder of the project are presented addressing the integration of personnel movement simulation into the preliminary ship design process.
Resumo:
The manufacture of materials products involves the control of a range of interacting physical phenomena. The material to be used is synthesised and then manipulated into some component form. The structure and properties of the final component are influenced by both interactions of continuum-scale phenomena and those at an atomistic-scale level. Moreover, during the processing phase there are some properties that cannot be measured (typically the liquid-solid phase change). However, it seems there is a potential to derive properties and other features from atomistic-scale simulations that are of key importance at the continuum scale. Some of the issues that need to be resolved in this context focus upon computational techniques and software tools facilitating: (i) the multiphysics modeling at continuum scale; (ii) the interaction and appropriate degrees of coupling between the atomistic through microstructure to continuum scale; and (iii) the exploitation of high-performance parallel computing power delivering simulation results in a practical time period. This paper discusses some of the attempts to address each of the above issues, particularly in the context of materials processing for manufacture.
Resumo:
Social network analysts have tried to capture the idea of social role explicitly by proposing a framework that precisely gives conditions under which group actors are playing equivalent roles. They term these methods positional analysis techniques. The most general definition is regular equivalence which captures the idea that equivalent actors are related in a similar way to equivalent alters. Regular equivalence gives rise to a whole class of partitions on a network. Given a network we have two different computational problems. The first is how to find a particular regular equivalence. An algorithm exists to find the largest regular partition but there are not efficient algorithms to test whether there is a regular k-partition. That is a partition in k groups that is regular. In addition, when dealing with real data, it is unlikely that any regular partitions exist. To overcome this problem relaxations of regular equivalence have been proposed along with optimisation techniques to find nearly regular partitions. In this paper we review the algorithms that have developed to find particular regular equivalences and look at some of the recent theoretical results which give an insight into the complexity of finding regular partitions.
Resumo:
The factors that are driving the development and use of grids and grid computing, such as size, dynamic features, distribution and heterogeneity, are also pushing to the forefront service quality issues. These include performance, reliability and security. Although grid middleware can address some of these issues on a wider scale, it has also become imperative to ensure adequate service provision at local level. Load sharing in clusters can contribute to the provision of a high quality service, by exploiting both static and dynamic information. This paper is concerned with the presentation of a load sharing scheme, that can satisfy grid computing requirements. It follows a proactive, non preemptive and distributed approach. Load information is gathered continuously before it is needed, and a task is allocated to the most appropriate node for execution. Performance and reliability are enhanced by the decentralised nature of the scheme and the symmetric roles of the nodes. In addition, the scheme exhibits transparency characteristics that facilitate integration with the grid.
Resumo:
The scalability of a computer system is its response to growth. It is also depended on its hardware, its operating system and the applications it is running. Most distributed systems technology today still depends on bus-based shared memory which do not scale well, and systems based on the grid or hypercube scheme requires significantly less connections than a full inter-connection that would exhibit a quadratic growth rate. The rapid convergence of mobile communication, digital broadcasting and network infrastructures calls for rich multimedia content that is adaptive and responsive to the needs of individuals, businesses and the public organisations. This paper will discuss the emergence of mobile Multimedia systems and provides an overview of the issues regarding design and delivery of multimedia content to mobile devices.
Resumo:
Today most of the IC and board designs are undertaken using two-dimensional graphics tools and rule checks. System-in-package is driving three-dimensional design concepts and this is posing a number of challenges for electronic design automation (EDA) software vendors. System-in-package requires three-dimensional EDA tools and design collaboration systems with appropriate manufacturing and assembly rules for these expanding technologies. Simulation and Analysis tools today focus on one aspect of the design requirement, for example, thermal, electrical or mechanical. System-in-Package requires analysis and simulation tools that can easily capture the complex three dimensional structures and provided integrated fast solutions to issues such as thermal management, reliability, electromagnetic interference, etc. This paper discusses some of the challenges faced by the design and analysis community in providing appropriate tools to engineers for System-in-Package design
Resumo:
Light has the greatest information carrying potential of all the perceivable interconnect mediums; consequently, optical fiber interconnects rapidly replaced copper in telecommunications networks, providing bandwidth capacity far in excess of its predecessors. As a result the modern telecommunications infrastructure has evolved into a global mesh of optical networks with VCSEL’s (Vertical Cavity Surface Emitting Lasers) dominating the short-link markets, predominately due to their low-cost. This cost benefit of VCSELs has allowed optical interconnects to again replace bandwidth limited copper as bottlenecks appear on VSR (Very Short Reach) interconnects between co-located equipment inside the CO (Central-Office). Spurred by the successful deployment in the VSR domain and in response to both intra-board backplane applications and inter-board requirements to extend the bandwidth between IC’s (Integrated Circuits), current research is migrating optical links toward board level USR (Ultra Short Reach) interconnects. Whilst reconfigurable Free Space Optical Interconnect (FSOI) are an option, they are complicated by precise line-of-sight alignment conditions hence benefits exist in developing guided wave technologies, which have been classified into three generations. First and second generation technologies are based upon optical fibers and are both capable of providing a suitable platform for intra-board applications. However, to allow component assembly, an integral requirement for inter-board applications, 3rd generation Opto-Electrical Circuit Boards (OECB’s) containing embedded waveguides are desirable. Currently, the greatest challenge preventing the deployment of OECB’s is achieving the out-of-plane coupling to SMT devices. With the most suitable low-cost platform being to integrate the optics into the OECB manufacturing process, several research avenues are being explored although none to date have demonstrated sufficient coupling performance. Once in place, the OECB assemblies will generate new reliability issues such as assembly configurations, manufacturing tolerances, and hermetic requirements that will also require development before total off-chip photonic interconnection can truly be achieved
Resumo:
This paper reports on the findings of a study on improving interaction design for visually impaired students, focusing upon the cognitive criteria for information visualisation.
Resumo:
Evaluating ship layout for human factors (HF) issues using simulation software such as maritimeEXODUS can be a long and complex process. The analysis requires the identification of relevant evaluation scenarios; encompassing evacuation and normal operations; the development of appropriate measures which can be used to gauge the performance of crew and vessel and finally; the interpretation of considerable simulation data. In this paper we present a systematic and transparent methodology for assessing the HF performance of ship design which is both discriminating and diagnostic. The methodology is demonstrated using two variants of a hypothetical naval ship.
Resumo:
Benati provides clarity about the characteristics and notion of language proficiency in the field of second language acquisition. He looks at four areas of research paradigmatically related to the role of proficiency: theorizing and measuring second language proficiency; the dimensions of L2 proficiency; factors contributing to the attainment of L2 proficiency and attaining L2 proficiency in the classroom. It also contains a variety of research accounts about the specific factors which have an effect on proficiency together with a theorised measurement of proficiency in second language research. It will be required reading for researchers in applied linguistics and second language acquisition.
Resumo:
This paper looks at the application of some of the assessment methods in practice with the view to enhance students’ learning in mathematics and statistics. It explores the effective application of assessment methods and highlights the issues or problems, and ways of avoiding them, related to some of the common methods of assessing mathematical and statistical learning. Some observations made by the author on good assessment practice and useful approaches employed at his institution in designing and applying assessment methods are discussed. Successful strategies in implementing assessment methods at different levels are described.
Resumo:
A detailed review of the status of the world nuclear industry with particular emphasis on economic aspects.