882 resultados para cooling chip for handheld electronic devices
Resumo:
L’innovation pédagogique pour elle-même s’avère parfois discutable, mais elle se justifie quand les enseignants se heurtent aux difficultés d’apprentissage de leurs étudiants. En particulier, certaines notions de physique sont réputées difficiles à appréhender par les étudiants, comme c’est le cas pour l’effet photoélectrique qui n’est pas souvent compris par les étudiants au niveau collégial. Cette recherche tente de déterminer si, dans le cadre d’un cours de physique, la simulation de l’effet photoélectrique et l’utilisation des dispositifs mobiles et en situation de collaboration favorisent une évolution des conceptions des étudiants au sujet de la lumière. Nous avons ainsi procédé à l’élaboration d’un scénario d’apprentissage collaboratif intégrant une simulation de l’effet photoélectrique sur un ordinateur de poche. La conception du scénario a d’abord été influencée par notre vision socioconstructiviste de l’apprentissage. Nous avons effectué deux études préliminaires afin de compléter notre scénario d’apprentissage et valider la plateforme MobileSim et l’interface du simulateur, que nous avons utilisées dans notre expérimentation : la première avec des ordinateurs de bureau et la seconde avec des ordinateurs de poche. Nous avons fait suivre à deux groupes d’étudiants deux cours différents, l’un portant sur une approche traditionnelle d’enseignement, l’autre basé sur le scénario d’apprentissage collaboratif élaboré. Nous leur avons fait passer un test évaluant l’évolution conceptuelle sur la nature de la lumière et sur le phénomène de l’effet photoélectrique et concepts connexes, à deux reprises : la première avant que les étudiants ne s’investissent dans le cours et la seconde après la réalisation des expérimentations. Nos résultats aux prétest et post-test sont complétés par des entrevues individuelles semi-dirigées avec tous les étudiants, par des enregistrements vidéo et par des traces récupérées des fichiers logs ou sur papier. Les étudiants du groupe expérimental ont obtenu de très bons résultats au post-test par rapport à ceux du groupe contrôle. Nous avons enregistré un gain moyen d’apprentissage qualifié de niveau modéré selon Hake (1998). Les résultats des entrevues ont permis de repérer quelques difficultés conceptuelles d’apprentissage chez les étudiants. L’analyse des données recueillies des enregistrements des séquences vidéo, des questionnaires et des traces récupérées nous a permis de mieux comprendre le processus d’apprentissage collaboratif et nous a dévoilé que le nombre et la durée des interactions entre les étudiants sont fortement corrélés avec le gain d’apprentissage. Ce projet de recherche est d’abord une réussite sur le plan de la conception d’un scénario d’apprentissage relatif à un phénomène aussi complexe que l’effet photoélectrique, tout en respectant de nombreux critères (collaboration, simulation, dispositifs mobiles) qui nous paraissaient extrêmement utopiques de réunir dans une situation d’apprentissage en classe. Ce scénario pourra être adapté pour l’apprentissage d’autres notions de la physique et pourra être considéré pour la conception des environnements collaboratifs d’apprentissage mobile innovants, centrés sur les besoins des apprenants et intégrant les technologies au bon moment et pour la bonne activité.
Resumo:
Les facteurs de transcription sont des protéines spécialisées qui jouent un rôle important dans différents processus biologiques tel que la différenciation, le cycle cellulaire et la tumorigenèse. Ils régulent la transcription des gènes en se fixant sur des séquences d’ADN spécifiques (éléments cis-régulateurs). L’identification de ces éléments est une étape cruciale dans la compréhension des réseaux de régulation des gènes. Avec l’avènement des technologies de séquençage à haut débit, l’identification de tout les éléments fonctionnels dans les génomes, incluant gènes et éléments cis-régulateurs a connu une avancée considérable. Alors qu’on est arrivé à estimer le nombre de gènes chez différentes espèces, l’information sur les éléments qui contrôlent et orchestrent la régulation de ces gènes est encore mal définie. Grace aux techniques de ChIP-chip et de ChIP-séquençage il est possible d’identifier toutes les régions du génome qui sont liées par un facteur de transcription d’intérêt. Plusieurs approches computationnelles ont été développées pour prédire les sites fixés par les facteurs de transcription. Ces approches sont classées en deux catégories principales: les algorithmes énumératifs et probabilistes. Toutefois, plusieurs études ont montré que ces approches génèrent des taux élevés de faux négatifs et de faux positifs ce qui rend difficile l’interprétation des résultats et par conséquent leur validation expérimentale. Dans cette thèse, nous avons ciblé deux objectifs. Le premier objectif a été de développer une nouvelle approche pour la découverte des sites de fixation des facteurs de transcription à l’ADN (SAMD-ChIP) adaptée aux données de ChIP-chip et de ChIP-séquençage. Notre approche implémente un algorithme hybride qui combine les deux stratégies énumérative et probabiliste, afin d’exploiter les performances de chacune d’entre elles. Notre approche a montré ses performances, comparée aux outils de découvertes de motifs existants sur des jeux de données simulées et des jeux de données de ChIP-chip et de ChIP-séquençage. SAMD-ChIP présente aussi l’avantage d’exploiter les propriétés de distributions des sites liés par les facteurs de transcription autour du centre des régions liées afin de limiter la prédiction aux motifs qui sont enrichis dans une fenêtre de longueur fixe autour du centre de ces régions. Les facteurs de transcription agissent rarement seuls. Ils forment souvent des complexes pour interagir avec l’ADN pour réguler leurs gènes cibles. Ces interactions impliquent des facteurs de transcription dont les sites de fixation à l’ADN sont localisés proches les uns des autres ou bien médier par des boucles de chromatine. Notre deuxième objectif a été d’exploiter la proximité spatiale des sites liés par les facteurs de transcription dans les régions de ChIP-chip et de ChIP-séquençage pour développer une approche pour la prédiction des motifs composites (motifs composés par deux sites et séparés par un espacement de taille fixe). Nous avons testé ce module pour prédire la co-localisation entre les deux demi-sites ERE qui forment le site ERE, lié par le récepteur des œstrogènes ERα. Ce module a été incorporé à notre outil de découverte de motifs SAMD-ChIP.
Resumo:
La lithographie et la loi de Moore ont permis des avancées extraordinaires dans la fabrication des circuits intégrés. De nos jours, plusieurs systèmes très complexes peuvent être embarqués sur la même puce électronique. Les contraintes de développement de ces systèmes sont tellement grandes qu’une bonne planification dès le début de leur cycle de développement est incontournable. Ainsi, la planification de la gestion énergétique au début du cycle de développement est devenue une phase importante dans la conception de ces systèmes. Pendant plusieurs années, l’idée était de réduire la consommation énergétique en ajoutant un mécanisme physique une fois le circuit créé, comme par exemple un dissipateur de chaleur. La stratégie actuelle est d’intégrer les contraintes énergétiques dès les premières phases de la conception des circuits. Il est donc essentiel de bien connaître la dissipation d’énergie avant l’intégration des composantes dans une architecture d’un système multiprocesseurs de façon à ce que chaque composante puisse fonctionner efficacement dans les limites de ses contraintes thermiques. Lorsqu’une composante fonctionne, elle consomme de l’énergie électrique qui est transformée en dégagement de chaleur. Le but de ce mémoire est de trouver une affectation efficace des composantes dans une architecture de multiprocesseurs en trois dimensions en tenant compte des limites des facteurs thermiques de ce système.
Resumo:
Màster en Nanociència i Nanotecnologia. Curs 2007-2008. Directors: Francesca Peiró i Martínez and Jordi Arbiol i Cobos
Resumo:
From the early stages of the twentieth century, polyaniline (PANI), a well-known and extensively studied conducting polymer has captured the attention of scientific community owing to its interesting electrical and optical properties. Starting from its structural properties, to the currently pursued optical, electrical and electrochemical properties, extensive investigations on pure PANI and its composites are still much relevant to explore its potentialities to the maximum extent. The synthesis of highly crystalline PANI films with ordered structure and high electrical conductivity has not been pursued in depth yet. Recently, nanostructured PANI and the nanocomposites of PANI have attracted a great deal of research attention owing to the possibilities of applications in optical switching devices, optoelectronics and energy storage devices. The work presented in the thesis is centered around the realization of highly conducting and structurally ordered PANI and its composites for applications mainly in the areas of nonlinear optics and electrochemical energy storage. Out of the vast variety of application fields of PANI, these two areas are specifically selected for the present studies, because of the following observations. The non-linear optical properties and the energy storing properties of PANI depend quite sensitively on the extent of conjugation of the polymer structure, the type and concentration of the dopants added and the type and size of the nano particles selected for making the nanocomposites. The first phase of the work is devoted to the synthesis of highly ordered and conducting films of PANI doped with various dopants and the structural, morphological and electrical characterization followed by the synthesis of metal nanoparticles incorporated PANI samples and the detailed optical characterization in the linear and nonlinear regimes. The second phase of the work comprises the investigations on the prospects of PANI in realizing polymer based rechargeable lithium ion cells with the inherent structural flexibility of polymer systems and environmental safety and stability. Secondary battery systems have become an inevitable part of daily life. They can be found in most of the portable electronic gadgets and recently they have started powering automobiles, although the power generated is low. The efficient storage of electrical energy generated from solar cells is achieved by using suitable secondary battery systems. The development of rechargeable battery systems having excellent charge storage capacity, cyclability, environmental friendliness and flexibility has yet to be realized in practice. Rechargeable Li-ion cells employing cathode active materials like LiCoO2, LiMn2O4, LiFePO4 have got remarkable charge storage capacity with least charge leakage when not in use. However, material toxicity, chance of cell explosion and lack of effective cell recycling mechanism pose significant risk factors which are to be addressed seriously. These cells also lack flexibility in their design due to the structural characteristics of the electrode materials. Global research is directed towards identifying new class of electrode materials with less risk factors and better structural stability and flexibility. Polymer based electrode materials with inherent flexibility, stability and eco-friendliness can be a suitable choice. One of the prime drawbacks of polymer based cathode materials is the low electronic conductivity. Hence the real task with this class of materials is to get better electronic conductivity with good electrical storage capability. Electronic conductivity can be enhanced by using proper dopants. In the designing of rechargeable Li-ion cells with polymer based cathode active materials, the key issue is to identify the optimum lithiation of the polymer cathode which can ensure the highest electronic conductivity and specific charge capacity possible The development of conducting polymer based rechargeable Li-ion cells with high specific capacity and excellent cycling characteristics is a highly competitive area among research and development groups, worldwide. Polymer based rechargeable batteries are specifically attractive due to the environmentally benign nature and the possible constructional flexibility they offer. Among polymers having electrical transport properties suitable for rechargeable battery applications, polyaniline is the most favoured one due to its tunable electrical conducting properties and the availability of cost effective precursor materials for its synthesis. The performance of a battery depends significantly on the characteristics of its integral parts, the cathode, anode and the electrolyte, which in turn depend on the materials used. Many research groups are involved in developing new electrode and electrolyte materials to enhance the overall performance efficiency of the battery. Currently explored electrolytes for Li ion battery applications are in liquid or gel form, which makes well-defined sealing essential. The use of solid electrolytes eliminates the need for containment of liquid electrolytes, which will certainly simplify the cell design and improve the safety and durability. The other advantages of polymer electrolytes include dimensional stability, safety and the ability to prevent lithium dendrite formation. One of the ultimate aims of the present work is to realize all solid state, flexible and environment friendly Li-ion cells with high specific capacity and excellent cycling stability. Part of the present work is hence focused on identifying good polymer based solid electrolytes essential for realizing all solid state polymer based Li ion cells.The present work is an attempt to study the versatile roles of polyaniline in two different fields of technological applications like nonlinear optics and energy storage. Conducting form of doped PANI films with good extent of crystallinity have been realized using a level surface assisted casting method in addition to the generally employed technique of spin coating. Metal nanoparticles embedded PANI offers a rich source for nonlinear optical studies and hence gold and silver nanoparticles have been used for making the nanocomposites in bulk and thin film forms. These PANI nanocomposites are found to exhibit quite dominant third order optical non-linearity. The highlight of these studies is the observation of the interesting phenomenon of the switching between saturable absorption (SA) and reverse saturable absorption (RSA) in the films of Ag/PANI and Au/PANI nanocomposites, which offers prospects of applications in optical switching. The investigations on the energy storage prospects of PANI were carried out on Li enriched PANI which was used as the cathode active material for assembling rechargeable Li-ion cells. For Li enrichment or Li doping of PANI, n-Butyllithium (n-BuLi) in hexanes was used. The Li doping as well as the Li-ion cell assembling were carried out in an argon filled glove box. Coin cells were assembled with Li doped PANI with different doping concentrations, as the cathode, LiPF6 as the electrolyte and Li metal as the anode. These coin cells are found to show reasonably good specific capacity around 22mAh/g and excellent cycling stability and coulombic efficiency around 99%. To improve the specific capacity, composites of Li doped PANI with inorganic cathode active materials like LiFePO4 and LiMn2O4 were synthesized and coin cells were assembled as mentioned earlier to assess the electrochemical capability. The cells assembled using the composite cathodes are found to show significant enhancement in specific capacity to around 40mAh/g. One of the other interesting observations is the complete blocking of the adverse effects of Jahn-Teller distortion, when the composite cathode, PANI-LiMn2O4 is used for assembling the Li-ion cells. This distortion is generally observed, near room temperature, when LiMn2O4 is used as the cathode, which significantly reduces the cycling stability of the cells.
Resumo:
The rapid growth of the optical communication branches and the enormous demand for more bandwidth require novel networks such as dense wavelength division multiplexing (DWDM). These networks enable higher bitrate transmission using the existing optical fibers. Micromechanically tunable optical microcavity devices like VCSELs, Fabry-Pérot filters and photodetectors are core components of these novel DWDM systems. Several air-gap based tunable devices were successfully implemented in the last years. Even though these concepts are very promising, two main disadvantages are still remaining. On the one hand, the high fabrication and integration cost and on the other hand the undesired adverse buckling of the suspended membranes. This thesis addresses these two problems and consists of two main parts: • PECVD dielectric material investigation and stress control resulting in membranes shape engineering. • Implementation and characterization of novel tunable optical devices with tailored shapes of the suspended membranes. For this purposes, low-cost PECVD technology is investigated and developed in detail. The macro- and microstress of silicon nitride and silicon dioxide are controlled over a wide range. Furthermore, the effect of stress on the optical and mechanical properties of the suspended membranes and on the microcavities is evaluated. Various membrane shapes (concave, convex and planar) with several radii of curvature are fabricated. Using this resonator shape engineering, microcavity devices such as non tunable and tunable Fabry-Pérot filters, VCSELs and PIN photodetectors are succesfully implemented. The fabricated Fabry-Pérot filters cover a spectral range of over 200nm and show resonance linewidths down to 1.5nm. By varying the stress distribution across the vertical direction within a DBR, the shape and the radius of curvature of the top membrane are explicitely tailored. By adjusting the incoming light beam waist to the curvature, the fundamental resonant mode is supported and the higher order ones are suppressed. For instance, a tunable VCSEL with 26 nm tuning range, 400µW maximal output power, 47nm free spectral range and over 57dB side mode suppresion ratio (SMSR) is demonstrated. Other technologies, such as introducing light emitting organic materials in microcavities are also investigated.
Resumo:
Die in dieser Arbeit durchgeführten Untersuchungen zeigen, daß es möglich ist, komplexe thermische Systeme, unter Verwendung der thermisch-, elektrischen Analogien, mit PSpice zu simulieren. Im Mittelpunkt der Untersuchungen standen hierbei Strangkühlkörper zur Kühlung von elektronischen Bauelementen. Es konnte gezeigt werden,daß alle Wärmeübertragungsarten, (Wärmeleitung, Konvektion und Wärmestrahlung) in der Simulation berücksichtigt werden können. Für die Berechnung der Konvektion wurden verschiedene Methoden hergeleitet. Diese gelten zum einen für verschiedene Kühlkörpergeometrien, wie z.B. ebene Flächen und Kühlrippenzwischenräume, andererseits unterscheiden sie sich, je nachdem, ob freie oder erzwungene Konvektion betrachtet wird. Für die Wärmestrahlung zwischen den Kühlrippen wurden verschiedenen Berechnungsmethoden entwickelt. Für die Simulation mit PSpice wurde die Berechnung der Wärmestrahlung zwischen den Kühlrippen vereinfacht. Es konnte gezeigt werden, daß die Fehler, die durch die Vereinfachung entstehen, vernachlässigbar klein sind. Für das thermische Verhalten einer zu kühlenden Wärmequelle wurde ein allgemeines Modell entworfen. Zur Bestimmung der Modellparameter wurden verschiedene Meßverfahren entwickelt. Für eine im Fachgebiet Elektromechanik entwickelte Wärmequelle zum Test von Kühlvorrichtungen wurde mit Hilfe dieser Meßverfahren eine Parameterbestimmung durchgeführt. Die Erstellung des thermischen Modells eines Kühlkörpers für die Simulation in PSpice erfordert die Analyse der Kühlkörpergeometrie. Damit diese Analyse weitestgehend automatisiert werden kann, wurden verschiedene Algorithmen unter Matlab entwickelt. Es wurde ein Algorithmus entwickelt, der es ermöglicht, den Kühlkörper in Elementarzellen zu zerlegen, die für die Erstellung des Simulationsmodells benötigt werden. Desweiteren ist es für die Simulation notwendig zu wissen, welche der Elementarzellen am Rand des Kühlkörpers liegen, welche der Elementarzellen an einem Kühlrippenzwischenraum liegen und welche Kühlkörperkanten schräg verlaufen. Auch zur Lösung dieser Aufgaben wurden verschiedene Algorithmen entwickelt. Diese Algorithmen wurden zu einem Programm zusammengefaßt, das es gestattet, unterschiedliche Strangkühlkörper zu simulieren und die Simulationsergebnisse in Form der Temperaturverteilung auf der Montagefläche des Kühlkörpers grafisch darzustellen. Es können stationäre und transiente Simulationen durchgeführt werden. Desweiteren kann der thermische Widerstand des Kühlkörpers RthK als Funktion der Verlustleistung der Wärmequelle dargestellt werden. Zur Verifikation der Simulationsergebnisse wurden Temperaturmessungen an Kühlkörpern durchgeführt und mit den Simulationsergebnissen verglichen. Diese Vergleiche zeigen, daß die Abweichungen im Bereich der Streuung der Temperaturmessung liegen. Das hier entwickelte Verfahren zur thermischen Simulation von Strangkühlkörpern kann somit als gut bewertet werden.
Resumo:
One of the most prominent industrial applications of heat transfer science and engineering has been electronics thermal control. Driven by the relentless increase in spatial density of microelectronic devices, integrated circuit chip powers have risen by a factor of 100 over the past twenty years, with a somewhat smaller increase in heat flux. The traditional approaches using natural convection and forced-air cooling are becoming less viable as power levels increase. This paper provides a high-level overview of the thermal management problem from the perspective of a practitioner, as well as speculation on the prospects for electronics thermal engineering in years to come.
Resumo:
The use of electronic documents is constantly growing and the necessity to implement an ad-hoc eCertificate which manages access to private information is not only required but also necessary. This paper presents a protocol for the management of electronic identities (eIDs), meant as a substitute for the paper-based IDs, in a mobile environment with a user-centric approach. Mobile devices have been chosen because they provide mobility, personal use and high computational complexity. The inherent user-centricity also allows the user to personally manage the ID information and to display only what is required. The chosen path to develop the protocol is to migrate the existing eCert technologies implemented by the Learning Societies Laboratory in Southampton. By comparing this protocol with the analysis of the eID problem domain, a new solution has been derived which is compatible with both systems without loss of features.
Resumo:
Background: Sexual risk behaviors associated with poor information on sexuality have contributed to major public health problems in the area of sexual and reproductive health in teenagers and young adults in Colombia. Objective: To measure the perception of changes in sexual and reproductive risk behavior after the use of a teleconsultation service via mobile devices in a sample of young adults. Methods: A before and after observational study was designed, where a mobile application to inquire about sexual and reproductive health was developed. The perception of changes in sexual and reproductive health risk behaviors in a sample of young adults after the use of the application was measured using the validated survey “Family Health International (FHI) – Behavioral Surveillance Survey (BSS) – Survey for Adults between 15 to 40 Years”. Non-probabilistic convenience recruitment was undertaken through the study´s web page. Participants answered the survey online before and after the use of the mobile application for a six month period (intervention). For the inferential analysis, data was divided into three groups (dichotomous data, discrete quantitative data, and ordinal data), to compare the results of the questions between the first and the second survey. For all tests, a confidence interval of 95% was established. For dichotomous data, the Chi-squared test was used. For quantitative data, we used the Student’s t-test, and for ordinal data, the Mann-Whitney-Wilcoxon test. Results: A total of 257 subjects were registered in the study and met the selection criteria. The pre-intervention survey was answered by 232 subjects, and 127 completely answered the post-intervention survey, of which 54.3% did not use the application, leaving an effective population of 58 subjects for analysis. 53% (n=31) were female, and 47% (n=27) were male. The mean age was 21 years, ranging between 18 and 40 years. The differences between the answers on the first and the second survey were not statistically significant. The main risk behaviors identified in the population were homosexual relations, non-use of condoms, sexual relations with non-regular and commercial partners, the use of psychoactive substances, and ignorance about the symptoms of sexually transmitted diseases and HIV transmission. Conclusions: Although there were no differences between the pre- and post-intervention results, the study revealed different risk behaviors among the participating subjects. These findings highlight the importance of promoting educational strategies on this matter and the importance of providing patients with easily accessible tools with reliable health information.
Resumo:
El test de circuits és una fase del procés de producció que cada vegada pren més importància quan es desenvolupa un nou producte. Les tècniques de test i diagnosi per a circuits digitals han estat desenvolupades i automatitzades amb èxit, mentre que aquest no és encara el cas dels circuits analògics. D'entre tots els mètodes proposats per diagnosticar circuits analògics els més utilitzats són els diccionaris de falles. En aquesta tesi se'n descriuen alguns, tot analitzant-ne els seus avantatges i inconvenients. Durant aquests últims anys, les tècniques d'Intel·ligència Artificial han esdevingut un dels camps de recerca més importants per a la diagnosi de falles. Aquesta tesi desenvolupa dues d'aquestes tècniques per tal de cobrir algunes de les mancances que presenten els diccionaris de falles. La primera proposta es basa en construir un sistema fuzzy com a eina per identificar. Els resultats obtinguts son força bons, ja que s'aconsegueix localitzar la falla en un elevat tant percent dels casos. Per altra banda, el percentatge d'encerts no és prou bo quan a més a més s'intenta esbrinar la desviació. Com que els diccionaris de falles es poden veure com una aproximació simplificada al Raonament Basat en Casos (CBR), la segona proposta fa una extensió dels diccionaris de falles cap a un sistema CBR. El propòsit no és donar una solució general del problema sinó contribuir amb una nova metodologia. Aquesta consisteix en millorar la diagnosis dels diccionaris de falles mitjançant l'addició i l'adaptació dels nous casos per tal d'esdevenir un sistema de Raonament Basat en Casos. Es descriu l'estructura de la base de casos així com les tasques d'extracció, de reutilització, de revisió i de retenció, fent èmfasi al procés d'aprenentatge. En el transcurs del text s'utilitzen diversos circuits per mostrar exemples dels mètodes de test descrits, però en particular el filtre biquadràtic és l'utilitzat per provar les metodologies plantejades, ja que és un dels benchmarks proposats en el context dels circuits analògics. Les falles considerades son paramètriques, permanents, independents i simples, encara que la metodologia pot ser fàcilment extrapolable per a la diagnosi de falles múltiples i catastròfiques. El mètode es centra en el test dels components passius, encara que també es podria extendre per a falles en els actius.
Resumo:
One of the major differences undergraduates experience during the transition to university is the style of teaching. In schools and colleges most students study key stage 5 subjects in relatively small informal groups where teacher–pupil interaction is encouraged and two-way feedback occurs through question and answer type delivery. On starting in HE students are amazed by the sizes of the classes. For even a relatively small chemistry department with an intake of 60-70 students, biologists, pharmacists, and other first year undergraduates requiring chemistry can boost numbers in the lecture hall to around 200 or higher. In many universities class sizes of 400 are not unusual for first year groups where efficiency is crucial. Clearly the personalised classroom-style delivery is not practical and it is a brave student who shows his ignorance by venturing to ask a question in front of such an audience. In these environments learning can be a very passive process, the lecture acts as a vehicle for the conveyance of information and our students are expected to reinforce their understanding by ‘self-study’, a term, the meaning of which, many struggle to understand. The use of electronic voting systems (EVS) in such situations can vastly change the students’ learning experience from a passive to a highly interactive process. This principle has already been demonstrated in Physics, most notably in the work of Bates and colleagues at Edinburgh.1 These small hand-held devices, similar to those which have become familiar through programmes such as ‘Who Wants to be a Millionaire’ can be used to provide instant feedback to students and teachers alike. Advances in technology now allow them to be used in a range of more sophisticated settings and comprehensive guides on use have been developed for even the most techno-phobic staff.
Resumo:
This paper analyzes the convergence behavior of the least mean square (LMS) filter when used in an adaptive code division multiple access (CDMA) detector consisting of a tapped delay line with adjustable tap weights. The sampling rate may be equal to or higher than the chip rate, and these correspond to chip-spaced (CS) and fractionally spaced (FS) detection, respectively. It is shown that CS and FS detectors with the same time-span exhibit identical convergence behavior if the baseband received signal is strictly bandlimited to half the chip rate. Even in the practical case when this condition is not met, deviations from this observation are imperceptible unless the initial tap-weight vector gives an extremely large mean squared error (MSE). This phenomenon is carefully explained with reference to the eigenvalues of the correlation matrix when the input signal is not perfectly bandlimited. The inadequacy of the eigenvalue spread of the tap-input correlation matrix as an indicator of the transient behavior and the influence of the initial tap weight vector on convergence speed are highlighted. Specifically, a initialization within the signal subspace or to the origin leads to very much faster convergence compared with initialization in the a noise subspace.
Resumo:
A new electronic software distribution (ESD) life cycle analysis (LCA)methodology and model structure were constructed to calculate energy consumption and greenhouse gas (GHG) emissions. In order to counteract the use of high level, top-down modeling efforts, and to increase result accuracy, a focus upon device details and data routes was taken. In order to compare ESD to a relevant physical distribution alternative,physical model boundaries and variables were described. The methodology was compiled from the analysis and operational data of a major online store which provides ESD and physical distribution options. The ESD method included the calculation of power consumption of data center server and networking devices. An in-depth method to calculate server efficiency and utilization was also included to account for virtualization and server efficiency features. Internet transfer power consumption was analyzed taking into account the number of data hops and networking devices used. The power consumed by online browsing and downloading was also factored into the model. The embedded CO2e of server and networking devices was proportioned to each ESD process. Three U.K.-based ESD scenarios were analyzed using the model which revealed potential CO2e savings of 83% when ESD was used over physical distribution. Results also highlighted the importance of server efficiency and utilization methods.
Resumo:
Owing to continuous advances in the computational power of handheld devices like smartphones and tablet computers, it has become possible to perform Big Data operations including modern data mining processes onboard these small devices. A decade of research has proved the feasibility of what has been termed as Mobile Data Mining, with a focus on one mobile device running data mining processes. However, it is not before 2010 until the authors of this book initiated the Pocket Data Mining (PDM) project exploiting the seamless communication among handheld devices performing data analysis tasks that were infeasible until recently. PDM is the process of collaboratively extracting knowledge from distributed data streams in a mobile computing environment. This book provides the reader with an in-depth treatment on this emerging area of research. Details of techniques used and thorough experimental studies are given. More importantly and exclusive to this book, the authors provide detailed practical guide on the deployment of PDM in the mobile environment. An important extension to the basic implementation of PDM dealing with concept drift is also reported. In the era of Big Data, potential applications of paramount importance offered by PDM in a variety of domains including security, business and telemedicine are discussed.