926 resultados para Core


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Il core catalitico della DNA polimerasi III, composto dalle tre subunità α, ε e θ, è il complesso minimo responsabile della replicazione del DNA cromosomiale in Escherichia coli. Nell'oloenzima, α ed ε possiedono rispettivamente un'attività 5'-3' polimerasica ed un'attività 3'-5' esonucleasica, mentre θ non ha funzioni enzimatiche. Il presente studio si è concentrato sulle regioni del core che interagiscono direttamente con ε, ovvero θ (interagente all'estremità N-terminale di ε) e il dominio PHP di α (interagente all'estremità C-terminale di ε), delle quali non è stato sinora identificato il ruolo. Al fine di assegnare loro una funzione sono state seguite tre linee di ricerca parallele. Innanzitutto il ruolo di θ è stato studiato utilizzando approcci ex-vivo ed in vivo. I risultati presentati in questo studio mostrano che θ incrementa significativamente la stabilità della subunità ε, intrinsecamente labile. Durante gli esperimenti condotti è stata anche identificata una nuova forma dimerica di ε. Per quanto la funzione del dimero non sia definita, si è dimostrato che esso è attivamente dissociato da θ, che potrebbe quindi fungere da suo regolatore. Inoltre, è stato ritrovato e caratterizzato il primo fenotipo di θ associato alla crescita. Per quanto concerne il dominio PHP, si è dimostrato che esso possiede un'attività pirofosfatasica utilizzando un nuovo saggio, progettato per seguire le cinetiche di reazione catalizzate da enzimi rilascianti fosfato o pirofosfato. L'idrolisi del pirofosfato catalizzata dal PHP è stata dimostrata in grado di sostenere l'attività polimerasica di α in vitro, il che suggerisce il suo possibile ruolo in vivo durante la replicazione del DNA. Infine, è stata messa a punto una nuova procedura per la coespressione e purificazione del complesso α-ε-θ

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The aim of this thesis was to design, synthesize and develop a nanoparticle based system to be used as a chemosensor or as a label in bioanalytical applications. A versatile fluorescent functionalizable nanoarchitecture has been effectively produced based on the hydrolysis and condensation of TEOS in direct micelles of Pluronic® F 127, obtaining highly monodisperse silica - core / PEG - shell nanoparticles with a diameter of about 20 nm. Surface functionalized nanoparticles have been obtained in a one-pot procedure by chemical modification of the hydroxyl terminal groups of the surfactant. To make them fluorescent, a whole library of triethoxysilane fluorophores (mainly BODIPY based), encompassing the whole visible spectrum has been synthesized: this derivatization allows a high degree of doping, but the close proximity of the molecules inside the silica matrix leads to the development of self - quenching processes at high doping levels, with the concomitant fall of the fluorescence signal intensity. In order to bypass this parasite phenomenon, multichromophoric systems have been prepared, where highly efficient FRET processes occur, showing that this energy pathway is faster than self - quenching, recovering the fluorescence signal. The FRET efficiency remains very high even four dye nanoparticles, increasing the pseudo Stokes shift of the system, attractive feature for multiplexing analysis. These optimized nanoparticles have been successfully exploited in molecular imaging applications such as in vitro, in vivo and ex vivo imaging, proving themselves superior to conventional molecular fluorophores as signaling units.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This study focuses on the processes of change that firms undertake to overcome conditions of organizational rigidity and develop new dynamic capabilities, thanks to the contribution of external knowledge. When external contingencies highlight firms’ core rigidities, external actors can intervene in change projects, providing new competences to firms’ managers. Knowledge transfer and organizational learning processes can lead to the development of new dynamic capabilities. Existing literature does not completely explain how these processes develop and how external knowledge providers, as management consultants, influence them. Dynamic capabilities literature has become very rich in the last years; however, the models that explain how dynamic capabilities evolve are not particularly investigated. Adopting a qualitative approach, this research proposes four relevant case studies in which external actors introduce new knowledge within organizations, activating processes of change. Each case study consists of a management consulting project. Data are collected through in-depth interviews with consultants and managers. A large amount of documents supports evidences from interviews. A narrative approach is adopted to account for change processes and a synthetic approach is proposed to compare case studies along relevant dimensions. This study presents a model of capabilities evolution, supported by empirical evidence, to explain how external knowledge intervenes in capabilities evolution processes: first, external actors solve gaps between environmental demands and firms’ capabilities, changing organizational structures and routines; second, a knowledge transfer between consultants and managers leads to the creation of new ordinary capabilities; third, managers can develop new dynamic capabilities through a deliberate learning process that internalizes new tacit knowledge from consultants. After the end of the consulting project, two elements can influence the deliberate learning process: new external contingencies and changes in the perceptions about external actors.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Spinal cord injury (SCI) results not only in paralysis; but it is also associated with a range of autonomic dysregulation that can interfere with cardiovascular, bladder, bowel, temperature, and sexual function. The entity of the autonomic dysfunction is related to the level and severity of injury to descending autonomic (sympathetic) pathways. For many years there was limited awareness of these issues and the attention given to them by the scientific and medical community was scarce. Yet, even if a new system to document the impact of SCI on autonomic function has recently been proposed, the current standard of assessment of SCI (American Spinal Injury Association (ASIA) examination) evaluates motor and sensory pathways, but not severity of injury to autonomic pathways. Beside the severe impact on quality of life, autonomic dysfunction in persons with SCI is associated with increased risk of cardiovascular disease and mortality. Therefore, obtaining information regarding autonomic function in persons with SCI is pivotal and clinical examinations and laboratory evaluations to detect the presence of autonomic dysfunction and quantitate its severity are mandatory. Furthermore, previous studies demonstrated that there is an intimate relationship between the autonomic nervous system and sleep from anatomical, physiological, and neurochemical points of view. Although, even if previous epidemiological studies demonstrated that sleep problems are common in spinal cord injury (SCI), so far only limited polysomnographic (PSG) data are available. Finally, until now, circadian and state dependent autonomic regulation of blood pressure (BP), heart rate (HR) and body core temperature (BcT) were never assessed in SCI patients. Aim of the current study was to establish the association between the autonomic control of the cardiovascular function and thermoregulation, sleep parameters and increased cardiovascular risk in SCI patients.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The efficient emulation of a many-core architecture is a challenging task, each core could be emulated through a dedicated thread and such threads would be interleaved on an either single-core or a multi-core processor. The high number of context switches will results in an unacceptable performance. To support this kind of application, the GPU computational power is exploited in order to schedule the emulation threads on the GPU cores. This presents a non trivial divergence issue, since GPU computational power is offered through SIMD processing elements, that are forced to synchronously execute the same instruction on different memory portions. Thus, a new emulation technique is introduced in order to overcome this limitation: instead of providing a routine for each ISA opcode, the emulator mimics the behavior of the Micro Architecture level, here instructions are date that a unique routine takes as input. Our new technique has been implemented and compared with the classic emulation approach, in order to investigate the chance of a hybrid solution.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

During the last years great effort has been devoted to the fabrication of superhydrophobic surfaces because of their self-cleaning properties. A water drop on a superhydrophobic surface rolls off even at inclinations of only a few degrees while taking up contaminants encountered on its way. rnSuperhydrophobic, self-cleaning coatings are desirable for convenient and cost-effective maintenance of a variety of surfaces. Ideally, such coatings should be easy to make and apply, mechanically resistant, and long-term stable. None of the existing methods have yet mastered the challenge of meeting all of these criteria.rnSuperhydrophobicity is associated with surface roughness. The lotus leave, with its dual scale roughness, is one of the most efficient examples of superhydrophobic surface. This thesis work proposes a novel technique to prepare superhydrophobic surfaces that introduces the two length scale roughness by growing silica particles (~100 nm in diameter) onto micrometer-sized polystyrene particles using the well-established Stöber synthesis. Mechanical resistance is conferred to the resulting “raspberries” by the synthesis of a thin silica shell on their surface. Besides of being easy to make and handle, these particles offer the possibility for improving suitability or technical applications: since they disperse in water, multi-layers can be prepared on substrates by simple drop casting even on surfaces with grooves and slots. The solution of the main problem – stabilizing the multilayer – also lies in the design of the particles: the shells – although mechanically stable – are porous enough to allow for leakage of polystyrene from the core. Under tetrahydrofuran vapor polystyrene bridges form between the particles that render the multilayer-film stable. rnMulti-layers are good candidate to design surfaces whose roughness is preserved after scratch. If the top-most layer is removed, the roughness can still be ensured by the underlying layer.rnAfter hydrophobization by chemical vapor deposition (CVD) of a semi-fluorinated silane, the surfaces are superhydrophobic with a tilting angle of a few degrees. rnrnrn

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Modern embedded systems embrace many-core shared-memory designs. Due to constrained power and area budgets, most of them feature software-managed scratchpad memories instead of data caches to increase the data locality. It is therefore programmers’ responsibility to explicitly manage the memory transfers, and this make programming these platform cumbersome. Moreover, complex modern applications must be adequately parallelized before they can the parallel potential of the platform into actual performance. To support this, programming languages were proposed, which work at a high level of abstraction, and rely on a runtime whose cost hinders performance, especially in embedded systems, where resources and power budget are constrained. This dissertation explores the applicability of the shared-memory paradigm on modern many-core systems, focusing on the ease-of-programming. It focuses on OpenMP, the de-facto standard for shared memory programming. In a first part, the cost of algorithms for synchronization and data partitioning are analyzed, and they are adapted to modern embedded many-cores. Then, the original design of an OpenMP runtime library is presented, which supports complex forms of parallelism such as multi-level and irregular parallelism. In the second part of the thesis, the focus is on heterogeneous systems, where hardware accelerators are coupled to (many-)cores to implement key functional kernels with orders-of-magnitude of speedup and energy efficiency compared to the “pure software” version. However, three main issues rise, namely i) platform design complexity, ii) architectural scalability and iii) programmability. To tackle them, a template for a generic hardware processing unit (HWPU) is proposed, which share the memory banks with cores, and the template for a scalable architecture is shown, which integrates them through the shared-memory system. Then, a full software stack and toolchain are developed to support platform design and to let programmers exploiting the accelerators of the platform. The OpenMP frontend is extended to interact with it.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The aim of this work was to study the dense cloud structures and to obtain the mass distribution of the dense cores (CMF) within the NGC6357 complex, from observations of the dust continuum at 450 and 850~$\mu$m of a 30 $\times$ 30 arcmin$^2$ region containing the H\textsc{ii} regions, G353.2+0.9 and G353.1+0.6.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Despite the several issues faced in the past, the evolutionary trend of silicon has kept its constant pace. Today an ever increasing number of cores is integrated onto the same die. Unfortunately, the extraordinary performance achievable by the many-core paradigm is limited by several factors. Memory bandwidth limitation, combined with inefficient synchronization mechanisms, can severely overcome the potential computation capabilities. Moreover, the huge HW/SW design space requires accurate and flexible tools to perform architectural explorations and validation of design choices. In this thesis we focus on the aforementioned aspects: a flexible and accurate Virtual Platform has been developed, targeting a reference many-core architecture. Such tool has been used to perform architectural explorations, focusing on instruction caching architecture and hybrid HW/SW synchronization mechanism. Beside architectural implications, another issue of embedded systems is considered: energy efficiency. Near Threshold Computing is a key research area in the Ultra-Low-Power domain, as it promises a tenfold improvement in energy efficiency compared to super-threshold operation and it mitigates thermal bottlenecks. The physical implications of modern deep sub-micron technology are severely limiting performance and reliability of modern designs. Reliability becomes a major obstacle when operating in NTC, especially memory operation becomes unreliable and can compromise system correctness. In the present work a novel hybrid memory architecture is devised to overcome reliability issues and at the same time improve energy efficiency by means of aggressive voltage scaling when allowed by workload requirements. Variability is another great drawback of near-threshold operation. The greatly increased sensitivity to threshold voltage variations in today a major concern for electronic devices. We introduce a variation-tolerant extension of the baseline many-core architecture. By means of micro-architectural knobs and a lightweight runtime control unit, the baseline architecture becomes dynamically tolerant to variations.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This thesis deals with heterogeneous architectures in standard workstations. Heterogeneous architectures represent an appealing alternative to traditional supercomputers because they are based on commodity components fabricated in large quantities. Hence their price-performance ratio is unparalleled in the world of high performance computing (HPC). In particular, different aspects related to the performance and consumption of heterogeneous architectures have been explored. The thesis initially focuses on an efficient implementation of a parallel application, where the execution time is dominated by an high number of floating point instructions. Then the thesis touches the central problem of efficient management of power peaks in heterogeneous computing systems. Finally it discusses a memory-bounded problem, where the execution time is dominated by the memory latency. Specifically, the following main contributions have been carried out: A novel framework for the design and analysis of solar field for Central Receiver Systems (CRS) has been developed. The implementation based on desktop workstation equipped with multiple Graphics Processing Units (GPUs) is motivated by the need to have an accurate and fast simulation environment for studying mirror imperfection and non-planar geometries. Secondly, a power-aware scheduling algorithm on heterogeneous CPU-GPU architectures, based on an efficient distribution of the computing workload to the resources, has been realized. The scheduler manages the resources of several computing nodes with a view to reducing the peak power. The two main contributions of this work follow: the approach reduces the supply cost due to high peak power whilst having negligible impact on the parallelism of computational nodes. from another point of view the developed model allows designer to increase the number of cores without increasing the capacity of the power supply unit. Finally, an implementation for efficient graph exploration on reconfigurable architectures is presented. The purpose is to accelerate graph exploration, reducing the number of random memory accesses.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

During the last few decades an unprecedented technological growth has been at the center of the embedded systems design paramount, with Moore’s Law being the leading factor of this trend. Today in fact an ever increasing number of cores can be integrated on the same die, marking the transition from state-of-the-art multi-core chips to the new many-core design paradigm. Despite the extraordinarily high computing power, the complexity of many-core chips opens the door to several challenges. As a result of the increased silicon density of modern Systems-on-a-Chip (SoC), the design space exploration needed to find the best design has exploded and hardware designers are in fact facing the problem of a huge design space. Virtual Platforms have always been used to enable hardware-software co-design, but today they are facing with the huge complexity of both hardware and software systems. In this thesis two different research works on Virtual Platforms are presented: the first one is intended for the hardware developer, to easily allow complex cycle accurate simulations of many-core SoCs. The second work exploits the parallel computing power of off-the-shelf General Purpose Graphics Processing Units (GPGPUs), with the goal of an increased simulation speed. The term Virtualization can be used in the context of many-core systems not only to refer to the aforementioned hardware emulation tools (Virtual Platforms), but also for two other main purposes: 1) to help the programmer to achieve the maximum possible performance of an application, by hiding the complexity of the underlying hardware. 2) to efficiently exploit the high parallel hardware of many-core chips in environments with multiple active Virtual Machines. This thesis is focused on virtualization techniques with the goal to mitigate, and overtake when possible, some of the challenges introduced by the many-core design paradigm.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Nell'ambito delle nanostrutture, un ruolo primario è svolto dai punti quantici. In questo lavoro siamo interessati all'analisi teorica del processo di creazione dei punti quantici: esso può avvenire per eteroepitassia, in particolare secondo il metodo studiato da Stranski-Krastanov. Un film di Germanio viene depositato su un substrato di Silicio in modo coerente, cioè senza dislocazioni, e, a causa del misfit tra le maglie dei due materiali, c'è un accumulo di energia elastica nel film. A una certa altezza critica questa energia del film può essere ridotta se il film si organizza in isole (punti quantici), dove la tensione può essere rilassata lateralmente. L'altezza critica dipende dai moduli di Young (E, υ), dal misfit tra le maglie (m) e dalla tensione superficiali (γ). Il trasporto di materiale nel film è portato avanti per diffusione superficiale. Il punto focale nell'analisi delle instabilità indotte dal misfit tra le maglie dei materiali è la ricerca delle caratteristiche che individuano il modo di crescita più rapido dei punti quantici. In questo lavoro siamo interessati ad un caso particolare: la crescita di punti quantici non su una superficie piana ma sulla superficie di un nanofilo quantico a geometria cilindrica. L'analisi delle instabilità viene condotta risolvendo le equazioni all'equilibrio: a tal fine sono state calcolate le distribuzioni del tensore delle deformazioni e degli sforzo di un nanofilo core-shell con una superficie perturbata al primo ordine rispetto all'ampiezza della perturbazione. L'analisi è stata condotta con particolari condizioni al contorno ed ipotesi geometriche, e diverse scelte dello stato di riferimento del campo degli spostamenti. Risolto il problema elastico, è stata studiata l'equazione dinamica di evoluzione descrivente la diffusione di superficie. Il risultato dell'analisi di instabilità è il tasso di crescita in funzione del numero d'onda q, con diversi valori del raggio del core, spessore dello shell e modo normale n, al fine di trovare il più veloce modo di crescita della perturbazione.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Combinatorial Optimization is becoming ever more crucial, in these days. From natural sciences to economics, passing through urban centers administration and personnel management, methodologies and algorithms with a strong theoretical background and a consolidated real-word effectiveness is more and more requested, in order to find, quickly, good solutions to complex strategical problems. Resource optimization is, nowadays, a fundamental ground for building the basements of successful projects. From the theoretical point of view, Combinatorial Optimization rests on stable and strong foundations, that allow researchers to face ever more challenging problems. However, from the application point of view, it seems that the rate of theoretical developments cannot cope with that enjoyed by modern hardware technologies, especially with reference to the one of processors industry. In this work we propose new parallel algorithms, designed for exploiting the new parallel architectures available on the market. We found that, exposing the inherent parallelism of some resolution techniques (like Dynamic Programming), the computational benefits are remarkable, lowering the execution times by more than an order of magnitude, and allowing to address instances with dimensions not possible before. We approached four Combinatorial Optimization’s notable problems: Packing Problem, Vehicle Routing Problem, Single Source Shortest Path Problem and a Network Design problem. For each of these problems we propose a collection of effective parallel solution algorithms, either for solving the full problem (Guillotine Cuts and SSSPP) or for enhancing a fundamental part of the solution method (VRP and ND). We endorse our claim by presenting computational results for all problems, either on standard benchmarks from the literature or, when possible, on data from real-world applications, where speed-ups of one order of magnitude are usually attained, not uncommonly scaling up to 40 X factors.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Gli ammassi di galassie sono gli oggetti gravitazionalmente legati più grandi dell’Universo. Questi emettono principalmente in banda X tramite bremsstrahlung. Una frazione mostra anche emissione radio diffusa da parte di elettroni relativistici che spiraleggiano nel campo magnetico. Si possono classificare tre tipi di emissione: alon, relitti e mini-aloni radio (MH). I MH sono sorgenti radio su scale di ≥ 200 – 500 kpc, osservate al centro di ammassi caratterizzati dalla presenze di cool-core (CC). L’origine dei MH non è ancora chiara. Gli elettroni relativistici che emettono in banda radio hanno tempi di vita radiativi di molto inferiori a quelli necessari per diffondere sulle scale dell’emissione diffusa. Quindi non sono semplicemente iniettati dalle galassie presenti negli ammassi ed è necessario un meccanismo di accelerazione “in-situ” nell’ICM. I MH testimoniano la presenza di meccanismi che canalizzano parte del budget energetico disponibile nei CC nell’ICM.Quindi lo studio è importante per comprendere la fisica dell’ICM e l’interazione fra le componenti non termiche e termiche. I MH si formano attraverso la riaccelerazione delle particelle relativistiche ad opera della turbolenza del gas. L’origine di questa turbolenza tuttavia non è ancora ben compresa. Gli ammassi CC sono caratterizzati da un picco della brillanza X nelle regioni centrali e da un drop della temperatura verso il centro accompagnata da aumento della densità del gas. Si ritiene che questo sia dovuto al raffreddamento del gas che quindi fluisce nelle zone centrali. Recenti osservazioni in X risultan inconsistenti con il modello classico di CF, suggerendo la presenza di una sorgente di riscaldamento del gas su scale del core degli ammassi. Recentemente Zhuravleva (2014) hanno mostrato che il riscaldamento dovuto alla turbolenza prodotta dall'AGN centrale è in grado di bilanciare il processo di raffreddamento. Abbiamo assunto che la turbolenza responsabile del riscaldamento del gas è anche responsabile dell’accelerazione delle particelle nei MH. Nell’ambito di questo scenario ci si aspetta una correlazione tra la potenza del cooling flow, PCF, che è una misura del tasso di energia emessa dal gas che raffredda nei CC, e la luminosità radio, che è una frazione dell’energia della turbolenza che è canalizzata nell’accelerazione delle particelle. In questo lavoro di tesi abbiamo utilizzato il più grande campione disponibile di MH, allo scopo di studiare la connessione fra le proprietà dei MH e quelle del gas termico nei core degli ammassi che li ospitano. Abbiamo analizzato i dati di 21 ammassi e ricavato i parametri fisici all’interno del raggio di cooling e del MH. Abbiamo ricavato la correlazione fra luminosità radio, e PCF. Abbiamo trovato che le due quantità correlano in modo quasi-lineare confermando i risultati precedenti. Tale correlazione suggerisce uno stretto legame fra le proprietà del gas nei CC e l’origine dei MH.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Questa tesi si pone l'obiettivo di esplorare alcuni aspetti di uno dei settori più in crescita in questi anni (e nei prossimi) in ambito informatico: \textbf{Internet of Things}, con un occhio rivolto in particolar modo a quelle che sono le piattaforme di sviluppo disponibili in questo ambito. Con queste premesse, si coglie l'occasione per addentrarsi nella scoperta della piattaforma realizzata e rilasciata da pochi mesi da uno dei colossi del mercato IT: Microsoft. Nel primo capitolo verrà trattato Internet of Things in ambito generale, attraverso una panoramica iniziale seguita da un'analisi approfondita dei principali protocolli sviluppati per questa tecnologia. Nel secondo capitolo verranno elencate una serie di piattaforme open source disponibili ad oggi per lo sviluppo di sistemi IoT. Dal terzo capitolo verrà incentrata l'attenzione sulle tecnologie Microsoft, in particolare prima si tratterà Windows 10 in generale, comprendendo \emph{UWP Applications}. Di seguito, nel medesimo capitolo, sarà focalizzata l'attenzione su Windows IoT Core, esplorandolo dettagliatamente (Windows Remote Arduino, Modalità Headed/Headless, etc.). Il capitolo a seguire concernerà la parte progettuale della tesi, comprendendo lo sviluppo del progetto \textbf{Smart Parking} in tutte le sue fasi (dei Requisiti fino ad Implementazione e Testing). Nel quinto (ed ultimo) capitolo, saranno esposte le conclusioni relative a Windows IoT Core e i suoi vantaggi/svantaggi.