920 resultados para bandwidth


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Next Generation Access Networks (NGAN) are the new step forward to deliver broadband services and to facilitate the integration of different technologies. It is plausible to assume that, from a technological standpoint, the Future Internet will be composed of long-range high-speed optical networks; a number of wireless networks at the edge; and, in between, several access technologies, among which, the Passive Optical Networks (xPON) are very likely to succeed, due to their simplicity, low-cost, and increased bandwidth. Among the different PON technologies, the Ethernet-PON (EPON) is the most promising alternative to satisfy operator and user needs, due to its cost, flexibility and interoperability with other technologies. One of the most interesting challenges in such technologies relates to the scheduling and allocation of resources in the upstream (shared) channel. The aim of this research project is to study and evaluate current contributions and propose new efficient solutions to address the resource allocation issues in Next Generation EPON (NG-EPON). Key issues in this context are future end-user needs, integrated quality of service (QoS) support and optimized service provisioning for real time and elastic flows. This project will unveil research opportunities, issue recommendations and propose novel mechanisms associated with the convergence within heterogeneous access networks and will thus serve as a basis for long-term research projects in this direction. The project has served as a platform for the generation of new concepts and solutions that were published in national and international conferences, scientific journals and also in book chapter. We expect some more research publications in addition to the ones mentioned to be generated in a few months.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

La tomodensitométrie (CT) est une technique d'imagerie dont l'intérêt n'a cessé de croître depuis son apparition dans le début des années 70. Dans le domaine médical, son utilisation est incontournable à tel point que ce système d'imagerie pourrait être amené à devenir victime de son succès si son impact au niveau de l'exposition de la population ne fait pas l'objet d'une attention particulière. Bien évidemment, l'augmentation du nombre d'examens CT a permis d'améliorer la prise en charge des patients ou a rendu certaines procédures moins invasives. Toutefois, pour assurer que le compromis risque - bénéfice soit toujours en faveur du patient, il est nécessaire d'éviter de délivrer des doses non utiles au diagnostic.¦Si cette action est importante chez l'adulte elle doit être une priorité lorsque les examens se font chez l'enfant, en particulier lorsque l'on suit des pathologies qui nécessitent plusieurs examens CT au cours de la vie du patient. En effet, les enfants et jeunes adultes sont plus radiosensibles. De plus, leur espérance de vie étant supérieure à celle de l'adulte, ils présentent un risque accru de développer un cancer radio-induit dont la phase de latence peut être supérieure à vingt ans. Partant du principe que chaque examen radiologique est justifié, il devient dès lors nécessaire d'optimiser les protocoles d'acquisitions pour s'assurer que le patient ne soit pas irradié inutilement. L'avancée technologique au niveau du CT est très rapide et depuis 2009, de nouvelles techniques de reconstructions d'images, dites itératives, ont été introduites afin de réduire la dose et améliorer la qualité d'image.¦Le présent travail a pour objectif de déterminer le potentiel des reconstructions itératives statistiques pour réduire au minimum les doses délivrées lors d'examens CT chez l'enfant et le jeune adulte tout en conservant une qualité d'image permettant le diagnostic, ceci afin de proposer des protocoles optimisés.¦L'optimisation d'un protocole d'examen CT nécessite de pouvoir évaluer la dose délivrée et la qualité d'image utile au diagnostic. Alors que la dose est estimée au moyen d'indices CT (CTDIV0| et DLP), ce travail a la particularité d'utiliser deux approches radicalement différentes pour évaluer la qualité d'image. La première approche dite « physique », se base sur le calcul de métriques physiques (SD, MTF, NPS, etc.) mesurées dans des conditions bien définies, le plus souvent sur fantômes. Bien que cette démarche soit limitée car elle n'intègre pas la perception des radiologues, elle permet de caractériser de manière rapide et simple certaines propriétés d'une image. La seconde approche, dite « clinique », est basée sur l'évaluation de structures anatomiques (critères diagnostiques) présentes sur les images de patients. Des radiologues, impliqués dans l'étape d'évaluation, doivent qualifier la qualité des structures d'un point de vue diagnostique en utilisant une échelle de notation simple. Cette approche, lourde à mettre en place, a l'avantage d'être proche du travail du radiologue et peut être considérée comme méthode de référence.¦Parmi les principaux résultats de ce travail, il a été montré que les algorithmes itératifs statistiques étudiés en clinique (ASIR?, VEO?) ont un important potentiel pour réduire la dose au CT (jusqu'à-90%). Cependant, par leur fonctionnement, ils modifient l'apparence de l'image en entraînant un changement de texture qui pourrait affecter la qualité du diagnostic. En comparant les résultats fournis par les approches « clinique » et « physique », il a été montré que ce changement de texture se traduit par une modification du spectre fréquentiel du bruit dont l'analyse permet d'anticiper ou d'éviter une perte diagnostique. Ce travail montre également que l'intégration de ces nouvelles techniques de reconstruction en clinique ne peut se faire de manière simple sur la base de protocoles utilisant des reconstructions classiques. Les conclusions de ce travail ainsi que les outils développés pourront également guider de futures études dans le domaine de la qualité d'image, comme par exemple, l'analyse de textures ou la modélisation d'observateurs pour le CT.¦-¦Computed tomography (CT) is an imaging technique in which interest has been growing since it first began to be used in the early 1970s. In the clinical environment, this imaging system has emerged as the gold standard modality because of its high sensitivity in producing accurate diagnostic images. However, even if a direct benefit to patient healthcare is attributed to CT, the dramatic increase of the number of CT examinations performed has raised concerns about the potential negative effects of ionizing radiation on the population. To insure a benefit - risk that works in favor of a patient, it is important to balance image quality and dose in order to avoid unnecessary patient exposure.¦If this balance is important for adults, it should be an absolute priority for children undergoing CT examinations, especially for patients suffering from diseases requiring several follow-up examinations over the patient's lifetime. Indeed, children and young adults are more sensitive to ionizing radiation and have an extended life span in comparison to adults. For this population, the risk of developing cancer, whose latency period exceeds 20 years, is significantly higher than for adults. Assuming that each patient examination is justified, it then becomes a priority to optimize CT acquisition protocols in order to minimize the delivered dose to the patient. Over the past few years, CT advances have been developing at a rapid pace. Since 2009, new iterative image reconstruction techniques, called statistical iterative reconstructions, have been introduced in order to decrease patient exposure and improve image quality.¦The goal of the present work was to determine the potential of statistical iterative reconstructions to reduce dose as much as possible without compromising image quality and maintain diagnosis of children and young adult examinations.¦The optimization step requires the evaluation of the delivered dose and image quality useful to perform diagnosis. While the dose is estimated using CT indices (CTDIV0| and DLP), the particularity of this research was to use two radically different approaches to evaluate image quality. The first approach, called the "physical approach", computed physical metrics (SD, MTF, NPS, etc.) measured on phantoms in well-known conditions. Although this technique has some limitations because it does not take radiologist perspective into account, it enables the physical characterization of image properties in a simple and timely way. The second approach, called the "clinical approach", was based on the evaluation of anatomical structures (diagnostic criteria) present on patient images. Radiologists, involved in the assessment step, were asked to score image quality of structures for diagnostic purposes using a simple rating scale. This approach is relatively complicated to implement and also time-consuming. Nevertheless, it has the advantage of being very close to the practice of radiologists and is considered as a reference method.¦Primarily, this work revealed that the statistical iterative reconstructions studied in clinic (ASIR? and VECO have a strong potential to reduce CT dose (up to -90%). However, by their mechanisms, they lead to a modification of the image appearance with a change in image texture which may then effect the quality of the diagnosis. By comparing the results of the "clinical" and "physical" approach, it was showed that a change in texture is related to a modification of the noise spectrum bandwidth. The NPS analysis makes possible to anticipate or avoid a decrease in image quality. This project demonstrated that integrating these new statistical iterative reconstruction techniques can be complex and cannot be made on the basis of protocols using conventional reconstructions. The conclusions of this work and the image quality tools developed will be able to guide future studies in the field of image quality as texture analysis or model observers dedicated to CT.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Abstract This thesis proposes a set of adaptive broadcast solutions and an adaptive data replication solution to support the deployment of P2P applications. P2P applications are an emerging type of distributed applications that are running on top of P2P networks. Typical P2P applications are video streaming, file sharing, etc. While interesting because they are fully distributed, P2P applications suffer from several deployment problems, due to the nature of the environment on which they perform. Indeed, defining an application on top of a P2P network often means defining an application where peers contribute resources in exchange for their ability to use the P2P application. For example, in P2P file sharing application, while the user is downloading some file, the P2P application is in parallel serving that file to other users. Such peers could have limited hardware resources, e.g., CPU, bandwidth and memory or the end-user could decide to limit the resources it dedicates to the P2P application a priori. In addition, a P2P network is typically emerged into an unreliable environment, where communication links and processes are subject to message losses and crashes, respectively. To support P2P applications, this thesis proposes a set of services that address some underlying constraints related to the nature of P2P networks. The proposed services include a set of adaptive broadcast solutions and an adaptive data replication solution that can be used as the basis of several P2P applications. Our data replication solution permits to increase availability and to reduce the communication overhead. The broadcast solutions aim, at providing a communication substrate encapsulating one of the key communication paradigms used by P2P applications: broadcast. Our broadcast solutions typically aim at offering reliability and scalability to some upper layer, be it an end-to-end P2P application or another system-level layer, such as a data replication layer. Our contributions are organized in a protocol stack made of three layers. In each layer, we propose a set of adaptive protocols that address specific constraints imposed by the environment. Each protocol is evaluated through a set of simulations. The adaptiveness aspect of our solutions relies on the fact that they take into account the constraints of the underlying system in a proactive manner. To model these constraints, we define an environment approximation algorithm allowing us to obtain an approximated view about the system or part of it. This approximated view includes the topology and the components reliability expressed in probabilistic terms. To adapt to the underlying system constraints, the proposed broadcast solutions route messages through tree overlays permitting to maximize the broadcast reliability. Here, the broadcast reliability is expressed as a function of the selected paths reliability and of the use of available resources. These resources are modeled in terms of quotas of messages translating the receiving and sending capacities at each node. To allow a deployment in a large-scale system, we take into account the available memory at processes by limiting the view they have to maintain about the system. Using this partial view, we propose three scalable broadcast algorithms, which are based on a propagation overlay that tends to the global tree overlay and adapts to some constraints of the underlying system. At a higher level, this thesis also proposes a data replication solution that is adaptive both in terms of replica placement and in terms of request routing. At the routing level, this solution takes the unreliability of the environment into account, in order to maximize reliable delivery of requests. At the replica placement level, the dynamically changing origin and frequency of read/write requests are analyzed, in order to define a set of replica that minimizes communication cost.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

El creciente uso de dispositivos móviles y el gran avance en la mejora de las aplicaciones y sistemas inalámbricos ha impulsado la demanda de filtros paso banda miniaturizados, que trabajen a altas frecuencias y tengan unas prestaciones elevadas. Los filtros basados en resonadores Bulk Acoustic Wave (BAW) están siendo la mejor alternativa a los filtros Surface Acoustic Wave (SAW), ya que funcionan a frecuencias superiores, pueden trabajar a mayores niveles de potencia y son compatibles con la tecnología CMOS. El filtro en escalera, que utiliza resonadores BAW, es de momento la mejor opción, debido a su facilidad de diseño y su bajo coste de fabricación. Aunque el filtro con resonadores acoplados (CRF) presenta mejores prestaciones como mayor ancho de banda, menor tamaño y conversión de modos. El problema de este tipo de filtros reside en su complejidad de diseño y su elevado coste. Este trabajo lleva a cabo el diseño de un CRF a partir de unas especificaciones bastante estrictas, demostrando sus altas prestaciones a pesar de su mayor inconveniente: el coste de fabricación.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper applies random matrix theory to obtain analytical characterizations of the capacity of correlated multiantenna channels. The analysis is not restricted to the popular separable correlation model, but rather it embraces a more general representation that subsumesmost of the channel models that have been treated in the literature. For arbitrary signal-to-noise ratios (SNR), the characterization is conducted in the regime of large numbers of antennas. For the low- and high-SNR regions, in turn, we uncover compact capacity expansions that are valid for arbitrary numbers of antennas and that shed insight on how antenna correlation impacts the tradeoffs between power, bandwidth and rate.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

ADSL is becoming the standard form of residential and small-business broadband Internet access due to, primarily, its low deployment cost. These ADSL residential lines are often deployed with 802.11 Access Points (AP) that providewireless connectivity. Given the density of ADSL deployment, it is often possible for a residential wireless client to be in range of several other APs, belonging to neighbors, with ADSL connectivity. While the ADSL technology has showed evident limits in terms of capacity (with speeds ranging 1-10 Mbps), the short-range wireless communication can guarantee a muchhigher capacity (up to 20 Mbps). Furthermore, the ADSL links in the neighborhood are generally under-utilized, since ADSL subscribers do not connect 100% of the time. Therefore, it is possible for a wireless client to simultaneously connect to several APs in range and effectively aggregate their available ADSL bandwidth.In this paper, we introduce ClubADSL, a wireless client that can simultaneously connect to several APs in range on different frequencies and aggregate both their downlink and uplink capacity. ClubADSL is a software that runs locally on the client-side, and it requires neither modification to the existing Internet infrastructure, nor any hardware/protocol upgradesto the 802.11 local area network. We show the feasibility of ClubADSL in seamlessly transmitting TCP traffic, and validate its implementation both in controlled scenarios and with current applications over real ADSL lines. In particular we show that a ClubADSL client can greatly benefit from the aggregated download bandwidth in the case of server-client applications such as video streaming, but can also take advantage of the increased upload bandwidth greatly reducing download times with incentive-based P2P applications such as BitTorrent.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This project focuses on studying and testing the benefits of the NX Remote Desktop technology in administrative use for Finnish Meteorological Institutes existing Linux Terminal Service Project environment. This was done due to the criticality of the system caused by growing number of users as the Linux Terminal Service Project system expands. Although many of the supporting tasks can be done via Secure Shell connection, testing graphical programs or desktop behaviour in such a way is impossible. At first basic technologies behind the NX Remote Desktop were studied, and after that started the testing of two possible programs, FreeNX and NoMachine NX server. Testing the functionality and bandwidth demands were first done in a closed local area network, and results were studied. The better candidate was then installed in a virtual server simulating actual Linux Terminal Service Project server at Finnish Meteorological Institute and connection from Internet was tested to see was there any problems with firewalls and security policies. The results are reported in this study. Studying and testing the two different candidates of NX Remote Desktop showed, that NoMachine NX Server provides better customer support and documentation. Security aspects of the Finnish Meteorological Institute had also to be considered, and since updates along with the new developing tools are announced in next version of the program, this version was the choice. Studies also show that even NoMachine promises a swift connection over an average of 20Kbit/s bandwidth, at least double of that is needed. This project gives an overview of available remote desktop products along their benefits. NX Remote Desktop technology is studied, and installation instructions are included. Testing is done in both, closed and the actual environment and problems and suggestions are studied and analyzed. The installation to the actual LTSP server is not yet made, but a virtual server is put up in the same place in the view of network topology. This ensures, that if the administrators are satisfied with the system, installation and setting up the system will go as described in this report.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Objective: Although 24-hour arterial blood pressure can be monitored in a free-moving animal using pressure telemetric transmitter mostly from Data Science International (DSI), accurate monitoring of 24-hour mouse left ventricular pressure (LVP) is not available because of its insufficient frequency response to a high frequency signal such as the maximum derivative of mouse LVP (LVdP/dtmax and LVdP/dtmin). The aim of the study was to develop a tiny implantable flow-through LVP telemetric transmitter for small rodent animals, which can be potentially adapted for human 24 hour BP and LVP accurate monitoring. Design and Method: The mouse LVP telemetric transmitter (Diameter: _12 mm, _0.4 g) was assembled by a pressure sensor, a passive RF telemetry chip, and to a 1.2F Polyurethane (PU) catheter tip. The device was developed in two configurations and compared with existing DSI system: (a) prototype-I: a new flow-through pressure sensor with wire link and (b) prototype-II: prototype-I plus a telemetry chip and its receiver. All the devices were applied in C57BL/6J mice. Data are mean_SEM. Results: A high frequency response (>100 Hz) PU heparin saline-filled catheter was inserted into mouse left ventricle via right carotid artery and implanted, LV systolic pressure (LVSP), LVdP/dtmax, and LVdP/dtmin were recorded on day2, 3, 4, 5, and 7 in conscious mice. The hemodynamic values were consistent and comparable (139_4 mmHg, 16634_319, - 12283_184 mmHg/s, n¼5) to one recorded by a validated Pebax03 catheter (138_2mmHg, 16045_443 and -12112_357 mmHg/s, n¼9). Similar LV hemodynamic values were obtained with Prototype-I. The same LVP waveforms were synchronically recorded by Notocord wire and Senimed wireless software through prototype-II in anesthetized mice. Conclusion: An implantable flow-through LVP transmitter (prototype-I) is generated for LVP accurate assessment in conscious mice. The prototype-II needs a further improvement on data transmission bandwidth and signal coupling distance to its receiver for accurate monitoring of LVP in a freemoving mouse.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

For the standard kernel density estimate, it is known that one can tune the bandwidth such that the expected L1 error is within a constant factor of the optimal L1 error (obtained when one is allowed to choose the bandwidth with knowledge of the density). In this paper, we pose the same problem for variable bandwidth kernel estimates where the bandwidths are allowed to depend upon the location. We show in particular that for positive kernels on the real line, for any data-based bandwidth, there exists a densityfor which the ratio of expected L1 error over optimal L1 error tends to infinity. Thus, the problem of tuning the variable bandwidth in an optimal manner is ``too hard''. Moreover, from the class of counterexamples exhibited in the paper, it appears thatplacing conditions on the densities (monotonicity, convexity, smoothness) does not help.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We continue the development of a method for the selection of a bandwidth or a number of design parameters in density estimation. We provideexplicit non-asymptotic density-free inequalities that relate the $L_1$ error of the selected estimate with that of the best possible estimate,and study in particular the connection between the richness of the classof density estimates and the performance bound. For example, our methodallows one to pick the bandwidth and kernel order in the kernel estimatesimultaneously and still assure that for {\it all densities}, the $L_1$error of the corresponding kernel estimate is not larger than aboutthree times the error of the estimate with the optimal smoothing factor and kernel plus a constant times $\sqrt{\log n/n}$, where $n$ is the sample size, and the constant only depends on the complexity of the family of kernels used in the estimate. Further applications include multivariate kernel estimates, transformed kernel estimates, and variablekernel estimates.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

OBJECTIVE. The purpose of this study was to improve the blood-pool signal-to-noise ratio (SNR) and blood-myocardium contrast-to-noise ratio (CNR) of slow-infusion 3-T whole-heart coronary MR angiography (MRA).SUBJECTS AND METHODS. In 2D sensitivity encoding (SENSE), the number of acquired k-space lines is reduced, allowing less radiofrequency excitation per cardiac cycle and a longer TR. The former can be exploited for signal enhancement with a higher radiofrequency excitation angle, and the latter leads to noise reduction due to lower data-sampling bandwidth. Both effects contribute to SNR gain in coronary MRA when spatial and temporal resolution and acquisition time remain identical. Numeric simulation was performed to select the optimal 2D SENSE pulse sequence parameters and predict the SNR gain. Eleven patients underwent conventional unenhanced and the proposed 2D SENSE contrast-enhanced coronary MRA acquisition. Blood-pool SNR, blood-myocardium CNR, visible vessel length, vessel sharpness, and number of side branches were evaluated.RESULTS. Consistent with the numeric simulation, using 2D SENSE in contrast-enhanced coronary MRA resulted in significant improvement in aortic blood-pool SNR (unenhanced vs contrast-enhanced, 37.5 +/- 14.7 vs 121.3 +/- 44.0; p < 0.05) and CNR (14.4 +/- 6.9 vs 101.5 +/- 40.8; p < 0.05) in the patient sample. A longer length of left anterior descending coronary artery was visualized, but vessel sharpness, coronary artery coverage, and image quality score were not improved with the proposed approach.CONCLUSION. In combination with contrast administration, 2D SENSE was found effective in improving SNR and CNR in 3-T whole-heart coronary MRA. Further investigation of cardiac motion compensation is necessary to exploit the SNR and CNR advantages and to achieve submillimeter spatial resolution.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This PhD thesis addresses the issue of scalable media streaming in large-scale networking environments. Multimedia streaming is one of the largest sink of network resources and this trend is still growing as testified by the success of services like Skype, Netflix, Spotify and Popcorn Time (BitTorrent-based). In traditional client-server solutions, when the number of consumers increases, the server becomes the bottleneck. To overcome this problem, the Content-Delivery Network (CDN) model was invented. In CDN model, the server copies the media content to some CDN servers, which are located in different strategic locations on the network. However, they require heavy infrastructure investment around the world, which is too expensive. Peer-to-peer (P2P) solutions are another way to achieve the same result. These solutions are naturally scalable, since each peer can act as both a receiver and a forwarder. Most of the proposed streaming solutions in P2P networks focus on routing scenarios to achieve scalability. However, these solutions cannot work properly in video-on-demand (VoD) streaming, when resources of the media server are not sufficient. Replication is a solution that can be used in these situations. This thesis specifically provides a family of replication-based media streaming protocols, which are scalable, efficient and reliable in P2P networks. First, it provides SCALESTREAM, a replication-based streaming protocol that adaptively replicates media content in different peers to increase the number of consumers that can be served in parallel. The adaptiveness aspect of this solution relies on the fact that it takes into account different constraints like bandwidth capacity of peers to decide when to add or remove replicas. SCALESTREAM routes media blocks to consumers over a tree topology, assuming a reliable network composed of homogenous peers in terms of bandwidth. Second, this thesis proposes RESTREAM, an extended version of SCALESTREAM that addresses the issues raised by unreliable networks composed of heterogeneous peers. Third, this thesis proposes EAGLEMACAW, a multiple-tree replication streaming protocol in which two distinct trees, named EAGLETREE and MACAWTREE, are built in a decentralized manner on top of an underlying mesh network. These two trees collaborate to serve consumers in an efficient and reliable manner. The EAGLETREE is in charge of improving efficiency, while the MACAWTREE guarantees reliability. Finally, this thesis provides TURBOSTREAM, a hybrid replication-based streaming protocol in which a tree overlay is built on top of a mesh overlay network. Both these overlays cover all peers of the system and collaborate to improve efficiency and low-latency in streaming media to consumers. This protocol is implemented and tested in a real networking environment using PlanetLab Europe testbed composed of peers distributed in different places in Europe.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

RESUM Un transmissor d’AM (modulació per amplitud), utilitza una de les moltes tècniques de modulació existents avui en dia. És molta la importància que té la modulació de senyals i aquests en són alguns exemples: -Facilita la propagació del senyal per cable o per aire. -Ordena l’espectre, distribuint Canals a les diferents informacions. -Disminueix la dimensió de les antenes. -Optimitza l’ample de banda de cada canal. -Evita interferències entre Canals. -Protegeix la informació de les degradacions per soroll. -Defineix la qualitat de la informació transmesa. L’objectiu principal d’aquest treball, serà realitzar un transmissor d’AM utilitzant components electrònics disponibles al mercat. Això es realitzarà mitjançant diversos procediments de disseny. Es realitzarà un procediment de disseny teòric, tot utilitzant els “datasheets” dels diferents components. Es realitzarà un procediment de disseny mitjançant la simulació, gràcies al qual es podrà provar el disseny del dispositiu i realitzar-ne algunes parts impossibles a reproduir teòricament. I finalment es realitzarà el dispositiu a la pràctica. Entre les conclusions més rellevants obtingudes en aquest treball, voldríem destacar la importància de la simulació per poder dissenyar circuits de radiofreqüència. En aquest treball s’ha demostrat que gràcies a una bona simulació, el primer prototip de dispositiu creat ens ha funcionat a la perfecció. D’altre banda, també comentar la importància d’un disseny adequat d’antena per poder aprofitar al màxim el rendiment del nostre dispositiu. Per concloure, la realització d’un aparell transmissor aporta unes nocions equilibrades d’electrònica i telecomunicacions importants per al disseny de dispositius de comunicació.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Use of ICN’s Internet and data services have continued to increase exponentially, which reflects the capacity needed for greater access to high-speed Internet (Broadband). Users are incorporating more web-based applications, which uses larger amounts of bandwidth; such as transmitting hospital MRIs, video streaming, and web-based systems.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We report the design and validation of simple magnetic tweezers for oscillating ferromagnetic beads in the piconewton and nanometer scales. The system is based on a single pair of coaxial coils operating in two sequential modes: permanent magnetization of the beads through a large and brief pulse of magnetic field and generation of magnetic gradients to produce uniaxial oscillatory forces. By using this two step method, the magnetic moment of the beads remains constant during measurements. Therefore, the applied force can be computed and varies linearly with the driving signal. No feedback control is required to produce well defined force oscillations over a wide bandwidth. The design of the coils was optimized to obtain high magnetic fields (280 mT) and gradients (2 T/m) with high homogeneity (5% variation) within the sample. The magnetic tweezers were implemented in an inverted optical microscope with a videomicroscopy-based multiparticle tracking system. The apparatus was validated with 4.5 ¿m magnetite beads obtaining forces up to ~2 pN and subnanometer resolution. The applicability of the device includes microrheology of biopolymer and cell cytoplasm, molecular mechanics, and mechanotransduction in living cells.