912 resultados para nodes and links
Resumo:
One relatively unexplored question about the Internet's physical structure concerns the geographical location of its components: routers, links and autonomous systems (ASes). We study this question using two large inventories of Internet routers and links, collected by different methods and about two years apart. We first map each router to its geographical location using two different state-of-the-art tools. We then study the relationship between router location and population density; between geographic distance and link density; and between the size and geographic extent of ASes. Our findings are consistent across the two datasets and both mapping methods. First, as expected, router density per person varies widely over different economic regions; however, in economically homogeneous regions, router density shows a strong superlinear relationship to population density. Second, the probability that two routers are directly connected is strongly dependent on distance; our data is consistent with a model in which a majority (up to 75-95%) of link formation is based on geographical distance (as in the Waxman topology generation method). Finally, we find that ASes show high variability in geographic size, which is correlated with other measures of AS size (degree and number of interfaces). Among small to medium ASes, ASes show wide variability in their geographic dispersal; however, all ASes exceeding a certain threshold in size are maximally dispersed geographically. These findings have many implications for the next generation of topology generators, which we envisage as producing router-level graphs annotated with attributes such as link latencies, AS identifiers and geographical locations.
Resumo:
Recent advances in processor speeds, mobile communications and battery life have enabled computers to evolve from completely wired to completely mobile. In the most extreme case, all nodes are mobile and communication takes place at available opportunities – using both traditional communication infrastructure as well as the mobility of intermediate nodes. These are mobile opportunistic networks. Data communication in such networks is a difficult problem, because of the dynamic underlying topology, the scarcity of network resources and the lack of global information. Establishing end-to-end routes in such networks is usually not feasible. Instead a store-and-carry forwarding paradigm is better suited for such networks. This dissertation describes and analyzes algorithms for forwarding of messages in such networks. In order to design effective forwarding algorithms for mobile opportunistic networks, we start by first building an understanding of the set of all paths between nodes, which represent the available opportunities for any forwarding algorithm. Relying on real measurements, we enumerate paths between nodes and uncover what we refer to as the path explosion effect. The term path explosion refers to the fact that the number of paths between a randomly selected pair of nodes increases exponentially with time. We draw from the theory of epidemics to model and explain the path explosion effect. This is the first contribution of the thesis, and is a key observation that underlies subsequent results. Our second contribution is the study of forwarding algorithms. For this, we rely on trace driven simulations of different algorithms that span a range of design dimensions. We compare the performance (success rate and average delay) of these algorithms. We make the surprising observation that most algorithms we consider have roughly similar performance. We explain this result in light of the path explosion phenomenon. While the performance of most algorithms we studied was roughly the same, these algorithms differed in terms of cost. This prompted us to focus on designing algorithms with the explicit intent of reducing costs. For this, we cast the problem of forwarding as an optimal stopping problem. Our third main contribution is the design of strategies based on optimal stopping principles which we refer to as Delegation schemes. Our analysis shows that using a delegation scheme reduces cost over naive forwarding by a factor of O(√N), where N is the number of nodes in the network. We further validate this result on real traces, where the cost reduction observed is even greater. Our results so far include a key assumption, which is unbounded buffers on nodes. Next, we relax this assumption, so that the problem shifts to one of prioritization of messages for transmission and dropping. Our fourth contribution is the study of message prioritization schemes, combined with forwarding. Our main result is that one achieves higher performance by assigning higher priorities to young messages in the network. We again interpret this result in light of the path explosion effect.
Resumo:
Controlling the mobility pattern of mobile nodes (e.g., robots) to monitor a given field is a well-studied problem in sensor networks. In this setup, absolute control over the nodes’ mobility is assumed. Apart from the physical ones, no other constraints are imposed on planning mobility of these nodes. In this paper, we address a more general version of the problem. Specifically, we consider a setting in which mobility of each node is externally constrained by a schedule consisting of a list of locations that the node must visit at particular times. Typically, such schedules exhibit some level of slack, which could be leveraged to achieve a specific coverage distribution of a field. Such a distribution defines the relative importance of different field locations. We define the Constrained Mobility Coordination problem for Preferential Coverage (CMC-PC) as follows: given a field with a desired monitoring distribution, and a number of nodes n, each with its own schedule, we need to coordinate the mobility of the nodes in order to achieve the following two goals: 1) satisfy the schedules of all nodes, and 2) attain the required coverage of the given field. We show that the CMC-PC problem is NP-complete (by reduction to the Hamiltonian Cycle problem). Then we propose TFM, a distributed heuristic to achieve field coverage that is as close as possible to the required coverage distribution. We verify the premise of TFM using extensive simulations, as well as taxi logs from a major metropolitan area. We compare TFM to the random mobility strategy—the latter provides a lower bound on performance. Our results show that TFM is very successful in matching the required field coverage distribution, and that it provides, at least, two-fold query success ratio for queries that follow the target coverage distribution of the field.
Resumo:
Routing protocols for ad-hoc networks assume that the nodes forming the network are either under a single authority, or else that they would be altruistically forwarding data for other nodes with no expectation of a return. These assumptions are unrealistic since in ad-hoc networks, nodes are likely to be autonomous and rational (selfish), and thus unwilling to help unless they have an incentive to do so. Providing such incentives is an important aspect that should be considered when designing ad-hoc routing protocols. In this paper, we propose a dynamic, decentralized routing protocol for ad-hoc networks that provides incentives in the form of payments to intermediate nodes used to forward data for others. In our Constrained Selfish Routing (CSR) protocol, game-theoretic approaches are used to calculate payments (incentives) that ensure both the truthfulness of participating nodes and the fairness of the CSR protocol. We show through simulations that CSR is an energy efficient protocol and that it provides lower communication overhead in the best and average cases compared to existing approaches.
Resumo:
We propose a novel data-delivery method for delay-sensitive traffic that significantly reduces the energy consumption in wireless sensor networks without reducing the number of packets that meet end-to-end real-time deadlines. The proposed method, referred to as SensiQoS, leverages the spatial and temporal correlation between the data generated by events in a sensor network and realizes energy savings through application-specific in-network aggregation of the data. SensiQoS maximizes energy savings by adaptively waiting for packets from upstream nodes to perform in-network processing without missing the real-time deadline for the data packets. SensiQoS is a distributed packet scheduling scheme, where nodes make localized decisions on when to schedule a packet for transmission to meet its end-to-end real-time deadline and to which neighbor they should forward the packet to save energy. We also present a localized algorithm for nodes to adapt to network traffic to maximize energy savings in the network. Simulation results show that SensiQoS improves the energy savings in sensor networks where events are sensed by multiple nodes, and spatial and/or temporal correlation exists among the data packets. Energy savings due to SensiQoS increase with increase in the density of the sensor nodes and the size of the sensed events. © 2010 Harshavardhan Sabbineni and Krishnendu Chakrabarty.
Resumo:
The performance of the register insertion protocol for mixed voice-data traffic is investigated by simulation. The simulation model incorporates a common insertion buffer for station and ring packets. Bandwidth allocation is achieved by imposing a queue limit at each node. A simple priority scheme is introduced by allowing the queue limit to vary from node to node. This enables voice traffic to be given priority over data. The effect on performance of various operational and design parameters such as ratio of voice to data traffic, queue limit and voice packet size is investigated. Comparisons are made where possible with related work on other protocols proposed for voice-data integration. The main conclusions are: (a) there is a general degradation of performance as the ratio of voice traffic to data traffic increases, (b) substantial improvement in performance can be achieved by restricting the queue length at data nodes and (c) for a given ring utilisation, smaller voice packets result in lower delays for both voice and data traffic.
Resumo:
The performance of loadsharing algorithms for heterogeneous distributed systems is investigated by simulation. The systems considered are networks of workstations (nodes) which differ in processing power. Two parameters are proposed for characterising system heterogeneity, namely the variance and skew of the distribution of processing power among the network nodes. A variety of networks are investigated, with the same number of nodes and total processing power, but with the processing power distributed differently among the nodes. Two loadsharing algorithms are evaluated, at overall system loadings of 50% and 90%, using job response time as the performance metric. Comparison is made with the ideal situation of ‘perfect sharing’, where it is assumed that the communication delays are zero and that complete knowledge is available about job lengths and the loading at the different nodes, so that an arriving job can be sent to the node where it will be completed in the shortest time. The algorithms studied are based on those already in use for homogeneous networks, but were adapted to take account of system heterogeneity. Both algorithms take into account the differences in the processing powers of the nodes in their location policies, but differ in the extent to which they ‘discriminate’ against the slower nodes. It is seen that the relative performance of the two is strongly influenced by the system utilisation and the distribution of processing power among the nodes.
Resumo:
Written by leading paramedics and academic subject experts, with a foreword by Paul Bates, this is the first foundation level text to introduce the main theoretical subjects studied on the new higher education paramedic science programmes. It underpins practice and links ways of working to the crucial theory base. The numerous examples and case studies focus specifically on pre-hospital and emergency care. Key topics covered include: * NHS and ambulance service history * Ethics and law for the paramedic * Communication and reflective practice * Professional issues, including evidence-based practice, anti-discriminatory practice, clinical governance and clinical audit * Psychological perspectives on health and ill health * Safeguarding children * Sociological perspectives on health and ill health and social policy * Managing change and leadership theory * Continuing professional development The text prompts you to stop and think about what you have just read and relate this to your role. Throughout the book, the author carefully considers how the theory being discussed relates to the role of the ambulance clinician. Foundations for Paramedic Practice: A Theoretical Perspective is a core text for students of paramedic science and will also provide a valuable resource for students of allied health professions. This text provides a stepping stone to further reading and investigation. - Taken from the back cover of: Amanda Blaber - Foundations for paramedic practice: a theoretical perspective
Resumo:
Langerin is a C-type lectin expressed by a subset of dendritic leukocytes, the Langerhans cells (LC). Langerin is a cell surface receptor that induces the formation of an LC-specific organelle, the Birbeck granule (BG). We generated a langerin(-/-) mouse on a C57BL/6 background which did not display any macroscopic aberrant development. In the absence of langerin, LC were detected in normal numbers in the epidermis but the cells lacked BG. LC of langerin(-/-) mice did not present other phenotypic alterations compared to wild-type littermates. Functionally, the langerin(-/-) LC were able to capture antigen, to migrate towards skin draining lymph nodes, and to undergo phenotypic maturation. In addition, langerin(-/-) mice were not impaired in their capacity to process native OVA protein for I-A(b)-restricted presentation to CD4(+) T lymphocytes or for H-2K(b)-restricted cross-presentation to CD8(+) T lymphocytes. langerin(-/-) mice inoculated with mannosylated or skin-tropic microorganisms did not display an altered pathogen susceptibility. Finally, chemical mutagenesis resulted in a similar rate of skin tumor development in langerin(-/-) and wild-type mice. Overall, our data indicate that langerin and BG are dispensable for a number of LC functions. The langerin(-/-) C57BL/6 mouse should be a valuable model for further functional exploration of langerin and the role of BG.
Resumo:
This paper investigates the learning of a wide class of single-hidden-layer feedforward neural networks (SLFNs) with two sets of adjustable parameters, i.e., the nonlinear parameters in the hidden nodes and the linear output weights. The main objective is to both speed up the convergence of second-order learning algorithms such as Levenberg-Marquardt (LM), as well as to improve the network performance. This is achieved here by reducing the dimension of the solution space and by introducing a new Jacobian matrix. Unlike conventional supervised learning methods which optimize these two sets of parameters simultaneously, the linear output weights are first converted into dependent parameters, thereby removing the need for their explicit computation. Consequently, the neural network (NN) learning is performed over a solution space of reduced dimension. A new Jacobian matrix is then proposed for use with the popular second-order learning methods in order to achieve a more accurate approximation of the cost function. The efficacy of the proposed method is shown through an analysis of the computational complexity and by presenting simulation results from four different examples.
Resumo:
We introduce a novel graph class we call universal hierarchical graphs (UHG) whose topology can be found numerously in problems representing, e.g., temporal, spacial or general process structures of systems. For this graph class we show, that we can naturally assign two probability distributions, for nodes and for edges, which lead us directly to the definition of the entropy and joint entropy and, hence, mutual information establishing an information theory for this graph class. Furthermore, we provide some results under which conditions these constraint probability distributions maximize the corresponding entropy. Also, we demonstrate that these entropic measures can be computed efficiently which is a prerequisite for every large scale practical application and show some numerical examples. (c) 2007 Elsevier Inc. All rights reserved.
Resumo:
Between 2006 and 2007, the Prisons Memory Archive (PMA) filmed participants, including former prisoners, prison staff, teachers, chaplains, visitors, solicitors and welfare workers back inside the Maze/Long Kesh Prison and Armagh Gaol. They shared the memory of the time spent in these prisons during the period of political violence from 1970 - 2000 in Northern Ireland, commonly known as the Troubles. Underpinning the overall methodology is co-ownership of the material, which gives participants the right to veto as well as to participate in the processes of editing and exhibiting their stories, so prioritising the value of co-authorship of their stories. The PMA adopted life-story interviewing techniques with the empty sites stimulating participants’ memory while they walked and talked their way around the empty sites. A third feature is inclusivity: the archive holds stories from across the full spectrum of the prison experience. A selection of the material, with accompanying context and links is available online www.prisonsmemoryarchive.com
Further Information:
The protocols of inclusivity, co-ownership and life-story telling make this collection significant as an initiative that engages with contemporary problems of how to negotiate narratives about a conflicted past in a society emerging out of violence. Inclusivity means that prison staff, prisoners, governors, chaplains, tutors and visitors have participated, relating their individual and collective experiences, which sit side by side on the PMA website. Co-ownership addresses the issues of ethics and sensitivity, allowing key constituencies to be involved. Life-story telling, based on oral history methodologies allows participants to be the authors of their own stories, crucial when dealing with sensitive issues from a violent past. The website hosts a selection of excerpts, e.g. the Armagh Stories page shows excerpts from 15 participants, while the Maze and Long Kesh Prison page offers interactive access to 24 participants from that prison. Using an interactive documentary structure, the site offers users opportunities to navigate their own way through the material and encourages them to hear and see the ‘other’, central to attempts at encouraging dialogue in a divided society. Further, public discussions have been held after screening of excerpts with community groups in the following locations - Belfast, Newtownabbey, Derry, Armagh, Enniskillen, London, Cork, Maynooth, Clones, and Monaghan. Extracts have been screened at international academic conferences in Valencia, Australia, Tartu, Estonia, Prague, and York. A dataset of the content, with description and links, is available for REF purposes.
Resumo:
This paper describes the fractographic analysis of five CFRP post-buckled skin/stringer panels that were tested to failure in compression. The detailed damage mechanisms for skin/stiffener detachment in an undamaged panel were characterised and related to the stress conditions during post-buckling; in particular the sites of peak twist (at buckling nodes) and peak bending moments (at buckling anti-nodes). The initial event was intralaminar splitting of the +45 degrees plies adjacent to the skin/stiffener interface, induced by high twist at a nodeline. This was followed by mode II delamination, parallel to +/- 45 degrees plies and then lengthwise (0 degrees) shear along the stiffener centreline. The presence of defects or damage was found to influence this failure process, leading to a reduction in strength. This research provides an insight into the processes that control post-buckled performance of stiffened panels and suggests that 2D models and element tests do not capture the true physics of skin/stiffener detachment: a full 3D approach is required.
Resumo:
We consider the distribution of entanglement from a multimode optical driving source to a network of remote and independent optomechanical systems. By focusing on the tripartite case, we analyse the effects that the features of the optical input states have on the degree and sharing structure of the distributed, fully mechanical, entanglement. This study, which is conducted looking at the mechanical steady state, highlights the structure of the entanglement distributed among the nodes and determines the relative efficiency between bipartite and tripartite entanglement transfer. We discuss a few open points, some of which are directed towards the bypassing of such limitations.
Resumo:
We consider the problem of self-healing in networks that are reconfigurable in the sense that they can change their topology during an attack. Our goal is to maintain connectivity in these networks, even in the presence of repeated adversarial node deletion, by carefully adding edges after each attack. We present a new algorithm, DASH, that provably ensures that: 1) the network stays connected even if an adversary deletes up to all nodes in the network; and 2) no node ever increases its degree by more than 2 log n, where n is the number of nodes initially in the network. DASH is fully distributed; adds new edges only among neighbors of deleted nodes; and has average latency and bandwidth costs that are at most logarithmic in n. DASH has these properties irrespective of the topology of the initial network, and is thus orthogonal and complementary to traditional topology- based approaches to defending against attack. We also prove lower-bounds showing that DASH is asymptotically optimal in terms of minimizing maximum degree increase over multiple attacks. Finally, we present empirical results on power-law graphs that show that DASH performs well in practice, and that it significantly outperforms naive algorithms in reducing maximum degree increase.