23 resultados para Free space communication

em Aston University Research Archive


Relevância:

100.00% 100.00%

Publicador:

Resumo:

A free space quantum key distribution system has been demonstrated. Consideration has been given to factors such as field of view and spectral width, to cut down the deleterious effect from background light levels. Suitable optical sources such as lasers and RCLEDs have been investigated as well as optimal wavelength choices, always with a view to building a compact and robust system. The implementation of background reduction measures resulted in a system capable of operating in daylight conditions. An autonomous system was left running and generating shared key material continuously for over 7 days. © 2009 Published by Elsevier B.V..

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We describe a free space quantum cryptography system which is designed to allow continuous unattended key exchanges for periods of several days, and over ranges of a few kilometres. The system uses a four-laser faint-pulse transmission system running at a pulse rate of 10MHz to generate the required four alternative polarization states. The receiver module similarly automatically selects a measurement basis and performs polarization measurements with four avalanche photodiodes. The controlling software can implement the full key exchange including sifting, error correction, and privacy amplification required to generate a secure key.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

A sizeable amount of the testing in eye care, requires either the identification of targets such as letters to assess functional vision, or the subjective evaluation of imagery by an examiner. Computers can render a variety of different targets on their monitors and can be used to store and analyse ophthalmic images. However, existing computing hardware tends to be large, screen resolutions are often too low, and objective assessments of ophthalmic images unreliable. Recent advances in mobile computing hardware and computer-vision systems can be used to enhance clinical testing in optometry. High resolution touch screens embedded in mobile devices, can render targets at a wide variety of distances and can be used to record and respond to patient responses, automating testing methods. This has opened up new opportunities in computerised near vision testing. Equally, new image processing techniques can be used to increase the validity and reliability of objective computer vision systems. Three novel apps for assessing reading speed, contrast sensitivity and amplitude of accommodation were created by the author to demonstrate the potential of mobile computing to enhance clinical measurement. The reading speed app could present sentences effectively, control illumination and automate the testing procedure for reading speed assessment. Meanwhile the contrast sensitivity app made use of a bit stealing technique and swept frequency target, to rapidly assess a patient’s full contrast sensitivity function at both near and far distances. Finally, customised electronic hardware was created and interfaced to an app on a smartphone device to allow free space amplitude of accommodation measurement. A new geometrical model of the tear film and a ray tracing simulation of a Placido disc topographer were produced to provide insights on the effect of tear film breakdown on ophthalmic images. Furthermore, a new computer vision system, that used a novel eye-lash segmentation technique, was created to demonstrate the potential of computer vision systems for the clinical assessment of tear stability. Studies undertaken by the author to assess the validity and repeatability of the novel apps, found that their repeatability was comparable to, or better, than existing clinical methods for reading speed and contrast sensitivity assessment. Furthermore, the apps offered reduced examination times in comparison to their paper based equivalents. The reading speed and amplitude of accommodation apps correlated highly with existing methods of assessment supporting their validity. Their still remains questions over the validity of using a swept frequency sine-wave target to assess patient’s contrast sensitivity functions as no clinical test provides the range of spatial frequencies and contrasts, nor equivalent assessment at distance and near. A validation study of the new computer vision system found that the authors tear metric correlated better with existing subjective measures of tear film stability than those of a competing computer-vision system. However, repeatability was poor in comparison to the subjective measures due to eye lash interference. The new mobile apps, computer vision system, and studies outlined in this thesis provide further insight into the potential of applying mobile and image processing technology to enhance clinical testing by eye care professionals.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Cascade transitions of rare earth ions involved in infrared host fiber provide the potential to generate dual or multiple wavelength lasing at mid-infrared region. In addition, the fast development of saturable absorber (SA) towards the long wavelengths motivates the realization of passively switched mid-infrared pulsed lasers. In this work, by combing the above two techniques, a new phenomenon of passively Q-switched ~3 μm and gain-switched ~2 μm pulses in a shared cavity was demonstrated with a Ho3+-doped fluoride fiber and a specifically designed semiconductor saturable absorber (SESAM) as the SA. The repetition rate of ~2 μm pulses can be tuned between half and same as that of ~3 μm pulses by changing the pump power. The proposed method here will add new capabilities and more flexibility for generating mid-infrared multiple wavelength pulses simultaneously that has important potential applications for laser surgery, material processing, laser radar, and free-space communications, and other areas.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Agents inhabiting large scale environments are faced with the problem of generating maps by which they can navigate. One solution to this problem is to use probabilistic roadmaps which rely on selecting and connecting a set of points that describe the interconnectivity of free space. However, the time required to generate these maps can be prohibitive, and agents do not typically know the environment in advance. In this paper we show that the optimal combination of different point selection methods used to create the map is dependent on the environment, no point selection method dominates. This motivates a novel self-adaptive approach for an agent to combine several point selection methods. The success rate of our approach is comparable to the state of the art and the generation cost is substantially reduced. Self-adaptation therefore enables a more efficient use of the agent's resources. Results are presented for both a set of archetypal scenarios and large scale virtual environments based in Second Life, representing real locations in London.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

A novel highly efficient, fiber-compatible spectrally encoded imaging (SEI) system using a 45° tilted fiber grating (TFG) is proposed and experimentally demonstrated for the first time, to the best of our knowledge. The TFG serves as an in-fiber lateral diffraction element, eliminating the need for bulky and lossy free-space diffraction gratings in conventional SEI systems. Under proper polarization control, due to the strong tilted reflection, the 45° TFG offers a diffraction efficiency as high as 93.5%. Our new design significantly reduces the volume of the SEI system and improves energy efficiency and system stability. As a proof-ofprinciple experiment, spectrally encoded imaging of a customer-designed sample (9.6 mm x 3.0 mm) using the TFG-based system is demonstrated. The lateral resolution of the SEI system is measured to be 42 μm in our experiment.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

We propose and demonstrate, for the first time to our best knowledge, the use of a 45° tilted fiber grating (TFG) as an infiber lateral diffraction element in an efficient and fiber-compatible spectrally encoded imaging (SEI) system. Under proper polarization control, the TFG has significantly enhanced diffraction efficiency (93.5%) due to strong tilted reflection. Our conceptually new fiber-topics-based design eliminates the need for bulky and lossy free-space diffraction gratings, significantly reduces the volume and cost of the imaging system, improves energy efficiency, and increases system stability. As a proof-of-principle experiment, we use the proposed system to perform an one dimensional (1D) line scan imaging of a customer-designed three-slot sample and the results show that the constructed image matches well with the actual sample. The angular dispersion of the 45° TFG is measured to be 0.054°/nm and the lateral resolution of the SEI system is measured to be 28 μm in our experiment.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The efficacy of a specially constructed Gallager-type error-correcting code to communication in a Gaussian channel is examined. The construction is based on the introduction of complex matrices, used in both encoding and decoding, which comprise sub-matrices of cascading connection values. The finite-size effects are estimated for comparing the results with the bounds set by Shannon. The critical noise level achieved for certain code rates and infinitely large systems nearly saturates the bounds set by Shannon even when the connectivity used is low.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Using current software engineering technology, the robustness required for safety critical software is not assurable. However, different approaches are possible which can help to assure software robustness to some extent. For achieving high reliability software, methods should be adopted which avoid introducing faults (fault avoidance); then testing should be carried out to identify any faults which persist (error removal). Finally, techniques should be used which allow any undetected faults to be tolerated (fault tolerance). The verification of correctness in system design specification and performance analysis of the model, are the basic issues in concurrent systems. In this context, modeling distributed concurrent software is one of the most important activities in the software life cycle, and communication analysis is a primary consideration to achieve reliability and safety. By and large fault avoidance requires human analysis which is error prone; by reducing human involvement in the tedious aspect of modelling and analysis of the software it is hoped that fewer faults will persist into its implementation in the real-time environment. The Occam language supports concurrent programming and is a language where interprocess interaction takes place by communications. This may lead to deadlock due to communication failure. Proper systematic methods must be adopted in the design of concurrent software for distributed computing systems if the communication structure is to be free of pathologies, such as deadlock. The objective of this thesis is to provide a design environment which ensures that processes are free from deadlock. A software tool was designed and used to facilitate the production of fault-tolerant software for distributed concurrent systems. Where Occam is used as a design language then state space methods, such as Petri-nets, can be used in analysis and simulation to determine the dynamic behaviour of the software, and to identify structures which may be prone to deadlock so that they may be eliminated from the design before the program is ever run. This design software tool consists of two parts. One takes an input program and translates it into a mathematical model (Petri-net), which is used for modeling and analysis of the concurrent software. The second part is the Petri-net simulator that takes the translated program as its input and starts simulation to generate the reachability tree. The tree identifies `deadlock potential' which the user can explore further. Finally, the software tool has been applied to a number of Occam programs. Two examples were taken to show how the tool works in the early design phase for fault prevention before the program is ever run.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Whilst some authors have portrayed the Internet as a powerful tool for business and political institutions, others have highlighted the potential of this technology for those vying to constrain or counter-balance the power of organizations, through e-collectivism and on-line action. What appears to be emerging is a contested space that has the potential to simultaneously enhance the power of organizations, whilst also acting as an enabling technology for the empowerment of grass-root networks. In this struggle, organizations are fighting for the retention of “old economy” positions, as well as the development of “new economy” power-bases. In realizing these positions, organizations and institutions are strategizing and manoeuvering in order to shape on-line networks and communications. For example, the on-line activities of individuals can be contained through various technological means, such as surveillance, and the structuring of the virtual world through the use of portals and “walled gardens”. However, loose groupings of individuals are also strategizing to ensure there is a liberation of their communication paths and practices, and to maintain the potential for mobilization within and across traditional boundaries. In this article, the unique nature and potential of the Internet are evaluated, and the struggle over this contested virtual space is explored.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Third Generation cellular communication systems are expected to support mixed cell architecture in which picocells, microcells and macrocells are used to achieve full coverage and increase the spectral capacity. Supporting higher numbers of mobile terminals and the use of smaller cells will result in an increase in the number of handovers, and consequently an increase in the time delays required to perform these handovers. Higher time delays will generate call interruptions and forced terminations, particularly for time sensitive applications like real-time multimedia and data services. Currently in the Global System for Mobile communications (GSM), the handover procedure is initiated and performed by the fixed part of the Public Land Mobile Network (PLMN). The mobile terminal is only capable of detecting candidate base stations suitable for the handover; it is the role of the network to interrogate a candidate base station for a free channel. Handover signalling is exchanged via the fixed network and the time delay required to perform the handover is greatly affected by the levels of teletraffic handled by the network. In this thesis, a new handover strategy is developed to reduce the total time delay for handovers in a microcellular system. The handover signalling is diverted from the fixed network to the air interface to prevent extra delays due to teletraffic congestion, and to allow the mobile terminal to exchange signalling directly with the candidate base station. The new strategy utilises Packet Reservation Multiple Access (PRMA) technique as a mechanism to transfer the control of the handover procedure from the fixed network to the mobile terminal. Simulation results are presented to show a dramatic reduction in the handover delay as compared to those obtained using fixed channel allocation and dynamic channel allocation schemes.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A method is proposed to offer privacy in computer communications, using symmetric product block ciphers. The security protocol involved a cipher negotiation stage, in which two communicating parties select privately a cipher from a public cipher space. The cipher negotiation process includes an on-line cipher evaluation stage, in which the cryptographic strength of the proposed cipher is estimated. The cryptographic strength of the ciphers is measured by confusion and diffusion. A method is proposed to describe quantitatively these two properties. For the calculation of confusion and diffusion a number of parameters are defined, such as the confusion and diffusion matrices and the marginal diffusion. These parameters involve computationally intensive calculations that are performed off-line, before any communication takes place. Once they are calculated, they are used to obtain estimation equations, which are used for on-line, fast evaluation of the confusion and diffusion of the negotiated cipher. A technique proposed in this thesis describes how to calculate the parameters and how to use the results for fast estimation of confusion and diffusion for any cipher instance within the defined cipher space.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We address the question of how to communicate among distributed processes valuessuch as real numbers, continuous functions and geometrical solids with arbitrary precision, yet efficiently. We extend the established concept of lazy communication using streams of approximants by introducing explicit queries. We formalise this approach using protocols of a query-answer nature. Such protocols enable processes to provide valid approximations with certain accuracy and focusing on certain locality as demanded by the receiving processes through queries. A lattice-theoretic denotational semantics of channel and process behaviour is developed. Thequery space is modelled as a continuous lattice in which the top element denotes the query demanding all the information, whereas other elements denote queries demanding partial and/or local information. Answers are interpreted as elements of lattices constructed over suitable domains of approximations to the exact objects. An unanswered query is treated as an error anddenoted using the top element. The major novel characteristic of our semantic model is that it reflects the dependency of answerson queries. This enables the definition and analysis of an appropriate concept of convergence rate, by assigning an effort indicator to each query and a measure of information content to eachanswer. Thus we capture not only what function a process computes, but also how a process transforms the convergence rates from its inputs to its outputs. In future work these indicatorscan be used to capture further computational complexity measures. A robust prototype implementation of our model is available.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Der vorliegende Beitrag untersucht die Frage, in welchem Maße sich Institutionen, die niederdeutsche Kulturszene und individuelle Sprecher des Niederdeutschen moderne Kommunikationstechnologien wie das Internet zunutze machen und ob computervermittelte Kommunikation helfen kann, dem Rückgang des Niederdeutschen Einhalt zu gebieten. Die grundsätzliche Herangehensweise ist eine soziolinguistische, die das Internet als sozialen Handlungsraum versteht, in dem Individuen und Institutionen kommunizieren. Für eine derartige Perspektive stehen weniger das Medium oder das Genre im Mittelpunkt des Interesses als vielmehr das kommunizierende Individuum und die Sprachgemeinschaft, in diesem Fall die virtuelle Sprachgemeinschaft. Based on studies that analyse the potential of computer-mediated communication (cmc) to help fight language shift in lesser-used languages, this paper discusses the situation of Low German in Northern Germany. Over the last three decades, Low German has lost more than half of its active speakers. The article raises the question of whether and, if so, how Low German speakers make use of cmc to stem this tide. Following a sociolinguistic approach focussed on the individual speakers who use the Internet as a space for social interaction, it gives an overview of the discursive field of Low German on the internet and analyses in detail the most popular Low German discussion board. It shows that one of the main obstacles to a more successful use of cmc can be found in speakers' complex attitude toward written Low German. © Franz Steiner Verlag Stuttgart.