899 resultados para Computer Networks and Communications


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Digitization, sophisticated fiber-optic networks and the resultant convergence of the media, communications and information technology industries have completely transformed the communications ecosystem in the last couple of decades. New contingent business and social models were created that have been mirrored in the amended communications regimes. Yet, despite an overhaul of the communications regulation paradigm, the status of and the rules on universal service have remained surprisingly intact, both during and after the liberalization exercise. The present paper looks into this paradox and examines the sustainability of the existing concept of universal service. It suggests that there is a need for a novel concept of universal service in the digital networked communications environment, whose objectives go beyond the conventional internalizing and redistributional rationales and concentrate on communication and information networks as a public good, where not only access to infrastructure but also access to content may be essential.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The development of electrophoretic computer models and their use for simulation of electrophoretic processes has increased significantly during the last few years. Recently, GENTRANS and SIMUL5 were extended with algorithms that describe chemical equilibria between solutes and a buffer additive in a fast 1:1 interaction process, an approach that enables simulation of the electrophoretic separation of enantiomers. For acidic cationic systems with sodium and H3 0(+) as leading and terminating components, respectively, acetic acid as counter component, charged weak bases as samples, and a neutral CD as chiral selector, the new codes were used to investigate the dynamics of isotachophoretic adjustment of enantiomers, enantiomer separation, boundaries between enantiomers and between an enantiomer and a buffer constituent of like charge, and zone stability. The impact of leader pH, selector concentration, free mobility of the weak base, mobilities of the formed complexes and complexation constants could thereby be elucidated. For selected examples with methadone enantiomers as analytes and (2-hydroxypropyl)-β-CD as selector, simulated zone patterns were found to compare well with those monitored experimentally in capillary setups with two conductivity detectors or an absorbance and a conductivity detector. Simulation represents an elegant way to provide insight into the formation of isotachophoretic boundaries and zone stability in presence of complexation equilibria in a hitherto inaccessible way.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The prenatal development of neural circuits must provide sufficient configuration to support at least a set of core postnatal behaviors. Although knowledge of various genetic and cellular aspects of development is accumulating rapidly, there is less systematic understanding of how these various processes play together in order to construct such functional networks. Here we make some steps toward such understanding by demonstrating through detailed simulations how a competitive co-operative ('winner-take-all', WTA) network architecture can arise by development from a single precursor cell. This precursor is granted a simplified gene regulatory network that directs cell mitosis, differentiation, migration, neurite outgrowth and synaptogenesis. Once initial axonal connection patterns are established, their synaptic weights undergo homeostatic unsupervised learning that is shaped by wave-like input patterns. We demonstrate how this autonomous genetically directed developmental sequence can give rise to self-calibrated WTA networks, and compare our simulation results with biological data.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The central assumption in the literature on collaborative networks and policy networks is that political outcomes are affected by a variety of state and nonstate actors. Some of these actors are more powerful than others and can therefore have a considerable effect on decision making. In this article, we seek to provide a structural and institutional explanation for these power differentials in policy networks and support the explanation with empirical evidence. We use a dyadic measure of influence reputation as a proxy for power, and posit that influence reputation over the political outcome is related to vertical integration into the political system by means of formal decision-making authority, and to horizontal integration by means of being well embedded into the policy network. Hence, we argue that actors are perceived as influential because of two complementary factors: (a) their institutional roles and (b) their structural positions in the policy network. Based on temporal and cross-sectional exponential random graph models, we compare five cases about climate, telecommunications, flood prevention, and toxic chemicals politics in Switzerland and Germany. The five networks cover national and local networks at different stages of the policy cycle. The results confirm that institutional and structural drivers seem to have a crucial impact on how an actor is perceived in decision making and implementation and, therefore, their ability to significantly shape outputs and service delivery.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Mobile networks usage rapidly increased over the years, with great consequences in terms of performance requirements. In this paper, we propose mechanisms to use Information-Centric Networking to perform load balancing in mobile networks, providing content delivery over multiple radio technologies at the same time and thus efficiently using resources and improving the overall performance of content transfer. Meaningful results were obtained by comparing content transfer over single radio links with typical strategies to content transfer over multiple radio links with Information-Centric Networking load balancing. Results demonstrate that Information-Centric Networking load balancing increases the performance and efficiency of 3GPP Long Term Evolution mobile networks while greatly improving the network perceived quality for end users.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Information-centric networking (ICN) is a new communication paradigm that aims at increasing security and efficiency of content delivery in communication networks. In recent years, many research efforts in ICN have focused on caching strategies to reduce traffic and increase overall performance by decreasing download times. Since caches need to operate at line speed, they have only a limited size and content can only be stored for a short time. However, if content needs to be available for a longer time, e.g., for delay-tolerant networking or to provide high content availability similar to content delivery networks (CDNs), persistent caching is required. We base our work on the Content-Centric Networking (CCN) architecture and investigate persistent caching by extending the current repository implementation in CCNx. We show by extensive evaluations in a YouTube and webserver traffic scenario that repositories can be efficiently used to increase content availability by significantly increasing cache hit rates.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Information-centric networking (ICN) enables communication in isolated islands, where fixed infrastructure is not available, but also supports seamless communication if the infrastructure is up and running again. In disaster scenarios, when a fixed infrastructure is broken, content discovery algorit hms are required to learn what content is locally available. For example, if preferred content is not available, users may also be satisfied with second best options. In this paper, we describe a new content discovery algorithm and compare it to existing Depth-first and Breadth-first traversal algorithms. Evaluations in mobile scenarios with up to 100 nodes show that it results in better performance, i.e., faster discovery time and smaller traffic overhead, than existing algorithms.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

BACKGROUND The diagnostic performance of biochemical scores and artificial neural network models for portal hypertension and cirrhosis is not well established. AIMS To assess diagnostic accuracy of six serum scores, artificial neural networks and liver stiffness measured by transient elastography, for diagnosing cirrhosis, clinically significant portal hypertension and oesophageal varices. METHODS 202 consecutive compensated patients requiring liver biopsy and hepatic venous pressure gradient measurement were included. Several serum tests (alone and combined into scores) and liver stiffness were measured. Artificial neural networks containing or not liver stiffness as input variable were also created. RESULTS The best non-invasive method for diagnosing cirrhosis, portal hypertension and oesophageal varices was liver stiffness (C-statistics=0.93, 0.94, and 0.90, respectively). Among serum tests/scores the best for diagnosing cirrhosis and portal hypertension and oesophageal varices were, respectively, Fibrosis-4, and Lok score. Artificial neural networks including liver stiffness had high diagnostic performance for cirrhosis, portal hypertension and oesophageal varices (accuracy>80%), but were not statistically superior to liver stiffness alone. CONCLUSIONS Liver stiffness was the best non-invasive method to assess the presence of cirrhosis, portal hypertension and oesophageal varices. The use of artificial neural networks integrating different non-invasive tests did not increase the diagnostic accuracy of liver stiffness alone.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We propose WEAVE, a geographical 2D/3D routing protocol that maintains information on a small number of waypoints and checkpoints for forwarding packets to any destination. Nodes obtain the routing information from partial traces gathered in incoming packets and use a system of checkpoints along with the segments of routes to weave end-to-end paths close to the shortest ones. WEAVE does not generate any control traffic, it is suitable for routing in both 2D and 3D networks, and does not require any strong assumption on the underlying network graph such as the Unit Disk or a Planar Graph. WEAVE compares favorably with existing protocols in both testbed experiments and simulations.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The shift from host-centric to information-centric networking (ICN) promises seamless communication in mobile networks. However, most existing works either consider well-connected networks with high node density or introduce modifications to {ICN} message processing for delay-tolerant Networking (DTN). In this work, we present agent-based content retrieval, which provides information-centric {DTN} support as an application module without modifications to {ICN} message processing. This enables flexible interoperability in changing environments. If no content source can be found via wireless multi-hop routing, requesters may exploit the mobility of neighbor nodes (called agents) by delegating content retrieval to them. Agents that receive a delegation and move closer to content sources can retrieve data and return it back to requesters. We show that agent-based content retrieval may be even more efficient in scenarios where multi-hop communication is possible. Furthermore, we show that broadcast communication may not be necessarily the best option since dynamic unicast requests have little overhead and can better exploit short contact times between nodes (no broadcast delays required for duplicate suppression).

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Abstract Information-centric networking (ICN) offers new perspectives on mobile ad-hoc communication because routing is based on names but not on endpoint identifiers. Since every content object has a unique name and is signed, authentic content can be stored and cached by any node. If connectivity to a content source breaks, it is not necessarily required to build a new path to the same source but content can also be retrieved from a closer node that provides the same content copy. For example, in case of collisions, retransmissions do not need to be performed over the entire path but due to caching only over the link where the collision occurred. Furthermore, multiple requests can be aggregated to improve scalability of wireless multi-hop communication. In this work, we base our investigations on Content-Centric Networking (CCN), which is a popular {ICN} architecture. While related works in wireless {CCN} communication are based on broadcast communication exclusively, we show that this is not needed for efficient mobile ad-hoc communication. With Dynamic Unicast requesters can build unicast paths to content sources after they have been identified via broadcast. We have implemented Dynamic Unicast in CCNx, which provides a reference implementation of the {CCN} concepts, and performed extensive evaluations in diverse mobile scenarios using NS3-DCE, the direct code execution framework for the {NS3} network simulator. Our evaluations show that Dynamic Unicast can result in more efficient communication than broadcast communication, but still supports all {CCN} advantages such as caching, scalability and implicit content discovery.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The majority of sensor network research deals with land-based networks, which are essentially two-dimensional, and thus the majority of simulation and animation tools also only handle such networks. Underwater sensor networks on the other hand, are essentially 3D networks because the depth at which a sensor node is located needs to be considered as well. Due to that additional dimension, specialized tools need to be used when conducting simulations for experimentation. The School of Engineering’s Underwater Sensor Network (UWSN) lab is conducting research on underwater sensor networks and requires simulation tools for 3D networks. The lab has extended NS-2, a widely used network simulator, so that it can simulate three-dimensional networks. However, NAM, a widely used network animator, currently only supports two-dimensional networks and no extensions have been implemented to give it three-dimensional capabilities. In this project, we develop a network visualization tool that functions similarly to NAM but is able to render network environments in full 3-D. It is able to take as input a NS-2 trace file (the same file taken as input by NAM), create the environment, position the sensor nodes, and animate the events of the simulation. Further, the visualization tool is easy to use, especially friendly to NAM users, as it is designed to follow the interfaces and functions similar to NAM. So far, the development has fulfilled the basic functionality. Future work includes fully functional capabilities for visualization and much improved user interfaces.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The emerging use of real-time 3D-based multimedia applications imposes strict quality of service (QoS) requirements on both access and core networks. These requirements and their impact to provide end-to-end 3D videoconferencing services have been studied within the Spanish-funded VISION project, where different scenarios were implemented showing an agile stereoscopic video call that might be offered to the general public in the near future. In view of the requirements, we designed an integrated access and core converged network architecture which provides the requested QoS to end-to-end IP sessions. Novel functional blocks are proposed to control core optical networks, the functionality of the standard ones is redefined, and the signaling improved to better meet the requirements of future multimedia services. An experimental test-bed to assess the feasibility of the solution was also deployed. In such test-bed, set-up and release of end-to-end sessions meeting specific QoS requirements are shown and the impact of QoS degradation in terms of the user perceived quality degradation is quantified. In addition, scalability results show that the proposed signaling architecture is able to cope with large number of requests introducing almost negligible delay.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Wireless communication is the transfer of information from one place to another without using wires. From the earliest times, humans have felt the need to develop techniques of remote communication. From this need arose the smoke signals, communication by sun reflection in mirrors and so on. But today the telecommunications electronic devices such as telephone, television, radio or computer. Radio and television are used for one-way communication. Telephone and computer are used for two-way communication. In wireless networks there is almost unlimited mobility, we can access the network almost anywhere or anytime. In wired networks we have the restriction of using the services in fixed area services. The demand of the wireless is increasing very fast; everybody wants broadband services anywhere anytime. WiMAX (Worldwide Interoperability for Microwave Access) is a broadband wireless technology based on IEEE 802.16-2004 and IEEE 802.16e-2005 that appears to solve this demand. WIMAX is a system that allows wireless data transmission in areas of up to 48 km of radius. It is designed as a wireless alternative to ADSL and a way to connect nodes in wireless metropolitan areas network. Unlike wireless systems that are limited in most cases, about 100 meter, providing greater coverage and more bandwidth. WIMAX promises to achieve high data transmission rates over large areas with a great amount of users. This alternative to the networks of broadband access common as DSL o Wi-Fi, can give broadband access to places quickly to rural areas and developing areas around the world. This paper is a study of WIMAX technology and market situation. First, the paper is responsible for explaining the technical aspects of WIMAX. For this gives an overview of WIMAX standards, physical layer, MAC layer and WiMAX, Technology and Market Beijing University of Post and Telecommunications 2 WIMAX network architecture. Second, the paper address the issue of market in which provides an overview of development and deployment of WIMAX to end the future development trend of WIMAX is addressed. RESUMEN: Por comunicaciones inalámbricas se entiende la transferencia de información desde un lugar a otro sin la necesidad de un soporte físico como es por ejemplo el cable. Por lo que remontándose a los principios de la existencia del ser humano, nos damos cuenta de que el ser humano siempre ha sentido la necesidad de desarrollar técnicas para lograr comunicarse a distancia con sus semejantes. De dicha necesidad, surgieron técnicas tan ancestrales como puede ser la comunicación mediante señales de humo o por reflexión de los rayos solares en espejos entre otras. La curiosidad del ser humano y la necesidad de comunicarse a distancia fue la que llevó a Alexander Graham Bell a inventar el teléfono en 1876. La aparición de un dispositivo que permitía comunicarse a distancia permitiendo escuchar la voz de aquella persona con la que se quería hablar, supuso una revolución no solo en el panorama tecnológico, si no también en el panorama social. Pues a parte de permitir comunicaciones a larga distancia, solventó el problema de la comunicación en “tiempo real”. A raíz de este invento, la tecnología en materia de comunicación ha ido avanzando significativamente, más concretamente en lo referido a las comunicaciones inalámbricas. En 1973 se realizó la primera llamada desde un terminal móvil aunque no fue hasta 1983 cuando se empezó a comercializar dicho terminal, lo que supuso un cambio de hábitos y costumbres para la sociedad. Desde la aparición del primer móvil el crecimiento del mercado ha sido exponencial, lo que ha repercutido en una demanda impensable de nuevas aplicaciones integradas en dichos dispositivos móviles que satisfagan las necesidades que día a día autogenera la sociedad. Tras conseguir realizar llamadas a larga distancia de forma inalámbrica, el siguiente paso fue la creación de los SMS (Short Message System) lo que supuso una nueva revolución además de abaratar costes al usuario a la hora de comunicarse. Pero el gran reto para la industria de las comunicaciones móviles surgió con la aparición de internet. Todo el mundo sentía la necesidad de poder conectarse a esa gran base de datos que es internet en cualquier parte y en cualquier momento. Las primeras conexiones a internet desde dispositivos móviles se realizaron a través de la tecnología WAP (Wireless Application Protocol) hasta la aparición de la tecnología GPRS que permitía la conexión mediante protocolo TCP/IP. A partir de estas conexiones han surgido otras tecnologías, como EDGE, HSDPA, etc., que permitían y permiten la conexión a internet desde dispositivos móviles. Hoy en día la demanda de servicios de red inalámbrica crece de forma rápida y exponencial, todo el mundo quiere servicios de banda ancha en cualquier lugar y en cualquier momento. En este documento se analiza la tecnología WiMAX ( Worldwide Interoperability for Microwave Access) que es una tecnología de banda ancha basada en el estándar IEEE 802.16 creada para brindar servicios a la demanda emergente en la banda ancha desde un punto de vista tecnológico, donde se da una visión de la parte técnica de la tecnología; y desde el punto de vista del mercado, donde se analiza el despliegue y desarrollo de la tecnología desde el punto de vista de negocio. WiMAX es una tecnología que permite la transmisión inalámbrica de datos en áreas de hasta 48Km de radio y que está diseñada como alternativa inalámbrica para ADSL y para conectar nodos de red inalámbrica en áreas metropolitanas. A diferencia de los sistemas inalámbricos existentes que están limitados en su mayoría a unos cientos de metros, WiMAX ofrece una mayor cobertura y un mayor ancho de banda que permita dar soporte a nuevas aplicaciones, además de alcanzar altas tasas de transmisión de datos en grandes áreas con una gran cantidad de usuarios. Se trata de una alternativa a las redes de acceso de banda ancha como DSL o Wi-Fi, que puede dar acceso de banda ancha a lugares tales como zonas rurales o zonas en vías de desarrollo por todo el mundo con rapidez. Existen dos tecnologías de WiMAX, WiMAX fijo (basado en el estándar IEEE 802.16d-2004) y WiMAX móvil (basado en el estándar IEEE 802.16e-2005). La tecnología fija está diseñada para comunicaciones punto a multipunto, mientras que la fija lo está para comunicaciones multipunto a multipunto. WiMAX móvil se basa en la tecnología OFDM que ofrece ventajas en términos de latencia, eficiencia en el uso del espectro y soporte avanzado para antenas. La modulación OFDM es muy robusta frente al multitrayecto, que es muy habitual en los canales de radiodifusión, frente al desvanecimiento debido a las condiciones meteorológicas y frente a las interferencias de RF. Una vez creada la tecnología WiMAX, poseedora de las características idóneas para solventar la demanda del mercado, ha de darse el siguiente paso, hay que convencer a la industria de las telecomunicaciones de que dicha tecnología realmente es la solución para que apoyen su implantación en el mercado de la banda ancha para las redes inalámbricas. Es aquí donde entra en juego el estudio del mercado que se realiza en este documento. WiMAX se enfrenta a un mercado exigente en el que a parte de tener que dar soporte a la demanda técnica, ha de ofrecer una rentabilidad económica a la industria de las comunicaciones móviles y más concretamente a las operadoras móviles que son quienes dentro del sector de las telecomunicaciones finalmente han de confiar en la tecnología para dar soporte a sus usuarios ya que estos al fin y al cabo lo único que quieren es que su dispositivo móvil satisfaga sus necesidades independientemente de la tecnología que utilicen para tener acceso a la red inalámbrica de banda ancha. Quizás el mayor problema al que se ha enfrentado WiMAX haya sido la situación económica en la que se encuentra el mundo. WiMAX a comenzado su andadura en uno de los peores momentos, pero aun así se presenta como una tecnología capaz de ayudar al mundo a salir hacia delante en estos tiempos tan duros. Finalmente se analiza uno de los debates existentes hoy en día en el sector de las comunicaciones móviles, WiMAX vs. LTE. Como se puede observar en el documento realmente una tecnología no saldrá victoriosa frente a la otra, si no que ambas tecnologías podrán coexistir y trabajar de forma conjunta.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Systems biology techniques are a topic of recent interest within the neurological field. Computational intelligence (CI) addresses this holistic perspective by means of consensus or ensemble techniques ultimately capable of uncovering new and relevant findings. In this paper, we propose the application of a CI approach based on ensemble Bayesian network classifiers and multivariate feature subset selection to induce probabilistic dependences that could match or unveil biological relationships. The research focuses on the analysis of high-throughput Alzheimer's disease (AD) transcript profiling. The analysis is conducted from two perspectives. First, we compare the expression profiles of hippocampus subregion entorhinal cortex (EC) samples of AD patients and controls. Second, we use the ensemble approach to study four types of samples: EC and dentate gyrus (DG) samples from both patients and controls. Results disclose transcript interaction networks with remarkable structures and genes not directly related to AD by previous studies. The ensemble is able to identify a variety of transcripts that play key roles in other neurological pathologies. Classical statistical assessment by means of non-parametric tests confirms the relevance of the majority of the transcripts. The ensemble approach pinpoints key metabolic mechanisms that could lead to new findings in the pathogenesis and development of AD