924 resultados para Programmable Logic Array


Relevância:

20.00% 20.00%

Publicador:

Resumo:

The advent of distributed and heterogeneous systems has laid the foundation for the birth of new architectural paradigms, in which many separated and autonomous entities collaborate and interact to the aim of achieving complex strategic goals, impossible to be accomplished on their own. A non exhaustive list of systems targeted by such paradigms includes Business Process Management, Clinical Guidelines and Careflow Protocols, Service-Oriented and Multi-Agent Systems. It is largely recognized that engineering these systems requires novel modeling techniques. In particular, many authors are claiming that an open, declarative perspective is needed to complement the closed, procedural nature of the state of the art specification languages. For example, the ConDec language has been recently proposed to target the declarative and open specification of Business Processes, overcoming the over-specification and over-constraining issues of classical procedural approaches. On the one hand, the success of such novel modeling languages strongly depends on their usability by non-IT savvy: they must provide an appealing, intuitive graphical front-end. On the other hand, they must be prone to verification, in order to guarantee the trustworthiness and reliability of the developed model, as well as to ensure that the actual executions of the system effectively comply with it. In this dissertation, we claim that Computational Logic is a suitable framework for dealing with the specification, verification, execution, monitoring and analysis of these systems. We propose to adopt an extended version of the ConDec language for specifying interaction models with a declarative, open flavor. We show how all the (extended) ConDec constructs can be automatically translated to the CLIMB Computational Logic-based language, and illustrate how its corresponding reasoning techniques can be successfully exploited to provide support and verification capabilities along the whole life cycle of the targeted systems.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Human reasoning is a fascinating and complex cognitive process that can be applied in different research areas such as philosophy, psychology, laws and financial. Unfortunately, developing supporting software (to those different areas) able to cope such as complex reasoning it’s difficult and requires a suitable logic abstract formalism. In this thesis we aim to develop a program, that has the job to evaluate a theory (a set of rules) w.r.t. a Goal, and provide some results such as “The Goal is derivable from the KB5 (of the theory)”. In order to achieve this goal we need to analyse different logics and choose the one that best meets our needs. In logic, usually, we try to determine if a given conclusion is logically implied by a set of assumptions T (theory). However, when we deal with programming logic we need an efficient algorithm in order to find such implications. In this work we use a logic rather similar to human logic. Indeed, human reasoning requires an extension of the first order logic able to reach a conclusion depending on not definitely true6 premises belonging to a incomplete set of knowledge. Thus, we implemented a defeasible logic7 framework able to manipulate defeasible rules. Defeasible logic is a non-monotonic logic designed for efficient defeasible reasoning by Nute (see Chapter 2). Those kind of applications are useful in laws area especially if they offer an implementation of an argumentation framework that provides a formal modelling of game. Roughly speaking, let the theory is the set of laws, a keyclaim is the conclusion that one of the party wants to prove (and the other one wants to defeat) and adding dynamic assertion of rules, namely, facts putted forward by the parties, then, we can play an argumentative challenge between two players and decide if the conclusion is provable or not depending on the different strategies performed by the players. Implementing a game model requires one more meta-interpreter able to evaluate the defeasible logic framework; indeed, according to Göedel theorem (see on page 127), we cannot evaluate the meaning of a language using the tools provided by the language itself, but we need a meta-language able to manipulate the object language8. Thus, rather than a simple meta-interpreter, we propose a Meta-level containing different Meta-evaluators. The former has been explained above, the second one is needed to perform the game model, and the last one will be used to change game execution and tree derivation strategies.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Es wurde ein genomischer DNA-Array der Modellpflanze Arabidopsis thaliana mit einer 13.800 EST-Klone umfassenden cDNA-Bibliothek entwickelt und in der Genexpressionsanalyse der pflanzlichen Pathogenabwehr eingesetzt. Mittels PCR-Amplifikation sind 13.000 PCR-Produkte der cDNA-Fragmente hergestellt worden, mit denen 66 genomische Arabidopsis-Arrays auf Nylon und Polypropylen als Trägermaterial hergestellt werden konnten. Die Validierung mit Fluoreszenz- und Radiaktivhybridisierung sowie der Vergleich von drei Normalisierungsmethoden führte zu reproduzierbaren Ergebnissen bei hohem Korrelationskoeffizienten. Die etablierte DNA-Array-Technologie wurde zur Genexpressionsanalyse der pathogeninduzierten Abwehrmechanismen der Pflanze Arabidopsis thaliana in den ersten 24 Stunden nach Infektion mit dem avirulenten Bakterium Pseudomonas syringae pv. tomato eingesetzt. In einer Auswahl von 75 Genen der Stoffwechselwege Glycolyse, Citrat-Cyclus, Pentosephosphat-Cyclus und Glyoxylatmetabolismus konnte für 25 % der Gene, im Shikimat-, Tryptophan- und Phenylpropanoidsyntheseweg für 60 % der Gene eine erhöhte Transkriptionsrate nachgewiesen werden. Die Ergebnisse dieser Arbeit stimmen mit experimentellen Daten verschiedener unabhängiger Studien zur pflanzlichen Pathogenantwort überein. Darüberhinaus sind erstmals Transkriptionsprofile von bisher auf Transkriptionsebene nicht untersuchten Genen erstellt worden. Diese Ergebnisse bestätigen die transkriptionelle Aktivierung ganzer Stoffwechselwege und gewähren erstmals einen Einblick in die koordinierte differentielle Transkription ganzer Stoffwechselwege während der Pathogenabwehr.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Recently in most of the industrial automation process an ever increasing degree of automation has been observed. This increasing is motivated by the higher requirement of systems with great performance in terms of quality of products/services generated, productivity, efficiency and low costs in the design, realization and maintenance. This trend in the growth of complex automation systems is rapidly spreading over automated manufacturing systems (AMS), where the integration of the mechanical and electronic technology, typical of the Mechatronics, is merging with other technologies such as Informatics and the communication networks. An AMS is a very complex system that can be thought constituted by a set of flexible working stations, one or more transportation systems. To understand how this machine are important in our society let considerate that every day most of us use bottles of water or soda, buy product in box like food or cigarets and so on. Another important consideration from its complexity derive from the fact that the the consortium of machine producers has estimated around 350 types of manufacturing machine. A large number of manufacturing machine industry are presented in Italy and notably packaging machine industry,in particular a great concentration of this kind of industry is located in Bologna area; for this reason the Bologna area is called “packaging valley”. Usually, the various parts of the AMS interact among them in a concurrent and asynchronous way, and coordinate the parts of the machine to obtain a desiderated overall behaviour is an hard task. Often, this is the case in large scale systems, organized in a modular and distributed manner. Even if the success of a modern AMS from a functional and behavioural point of view is still to attribute to the design choices operated in the definition of the mechanical structure and electrical electronic architecture, the system that governs the control of the plant is becoming crucial, because of the large number of duties associated to it. Apart from the activity inherent to the automation of themachine cycles, the supervisory system is called to perform other main functions such as: emulating the behaviour of traditional mechanical members thus allowing a drastic constructive simplification of the machine and a crucial functional flexibility; dynamically adapting the control strategies according to the different productive needs and to the different operational scenarios; obtaining a high quality of the final product through the verification of the correctness of the processing; addressing the operator devoted to themachine to promptly and carefully take the actions devoted to establish or restore the optimal operating conditions; managing in real time information on diagnostics, as a support of the maintenance operations of the machine. The kind of facilities that designers can directly find on themarket, in terms of software component libraries provides in fact an adequate support as regard the implementation of either top-level or bottom-level functionalities, typically pertaining to the domains of user-friendly HMIs, closed-loop regulation and motion control, fieldbus-based interconnection of remote smart devices. What is still lacking is a reference framework comprising a comprehensive set of highly reusable logic control components that, focussing on the cross-cutting functionalities characterizing the automation domain, may help the designers in the process of modelling and structuring their applications according to the specific needs. Historically, the design and verification process for complex automated industrial systems is performed in empirical way, without a clear distinction between functional and technological-implementation concepts and without a systematic method to organically deal with the complete system. Traditionally, in the field of analog and digital control design and verification through formal and simulation tools have been adopted since a long time ago, at least for multivariable and/or nonlinear controllers for complex time-driven dynamics as in the fields of vehicles, aircrafts, robots, electric drives and complex power electronics equipments. Moving to the field of logic control, typical for industrial manufacturing automation, the design and verification process is approached in a completely different way, usually very “unstructured”. No clear distinction between functions and implementations, between functional architectures and technological architectures and platforms is considered. Probably this difference is due to the different “dynamical framework”of logic control with respect to analog/digital control. As a matter of facts, in logic control discrete-events dynamics replace time-driven dynamics; hence most of the formal and mathematical tools of analog/digital control cannot be directly migrated to logic control to enlighten the distinction between functions and implementations. In addition, in the common view of application technicians, logic control design is strictly connected to the adopted implementation technology (relays in the past, software nowadays), leading again to a deep confusion among functional view and technological view. In Industrial automation software engineering, concepts as modularity, encapsulation, composability and reusability are strongly emphasized and profitably realized in the so-calledobject-oriented methodologies. Industrial automation is receiving lately this approach, as testified by some IEC standards IEC 611313, IEC 61499 which have been considered in commercial products only recently. On the other hand, in the scientific and technical literature many contributions have been already proposed to establish a suitable modelling framework for industrial automation. During last years it was possible to note a considerable growth in the exploitation of innovative concepts and technologies from ICT world in industrial automation systems. For what concerns the logic control design, Model Based Design (MBD) is being imported in industrial automation from software engineering field. Another key-point in industrial automated systems is the growth of requirements in terms of availability, reliability and safety for technological systems. In other words, the control system should not only deal with the nominal behaviour, but should also deal with other important duties, such as diagnosis and faults isolations, recovery and safety management. Indeed, together with high performance, in complex systems fault occurrences increase. This is a consequence of the fact that, as it typically occurs in reliable mechatronic systems, in complex systems such as AMS, together with reliable mechanical elements, an increasing number of electronic devices are also present, that are more vulnerable by their own nature. The diagnosis problem and the faults isolation in a generic dynamical system consists in the design of an elaboration unit that, appropriately processing the inputs and outputs of the dynamical system, is also capable of detecting incipient faults on the plant devices, reconfiguring the control system so as to guarantee satisfactory performance. The designer should be able to formally verify the product, certifying that, in its final implementation, it will perform itsrequired function guarantying the desired level of reliability and safety; the next step is that of preventing faults and eventually reconfiguring the control system so that faults are tolerated. On this topic an important improvement to formal verification of logic control, fault diagnosis and fault tolerant control results derive from Discrete Event Systems theory. The aimof this work is to define a design pattern and a control architecture to help the designer of control logic in industrial automated systems. The work starts with a brief discussion on main characteristics and description of industrial automated systems on Chapter 1. In Chapter 2 a survey on the state of the software engineering paradigm applied to industrial automation is discussed. Chapter 3 presentes a architecture for industrial automated systems based on the new concept of Generalized Actuator showing its benefits, while in Chapter 4 this architecture is refined using a novel entity, the Generalized Device in order to have a better reusability and modularity of the control logic. In Chapter 5 a new approach will be present based on Discrete Event Systems for the problemof software formal verification and an active fault tolerant control architecture using online diagnostic. Finally conclusive remarks and some ideas on new directions to explore are given. In Appendix A are briefly reported some concepts and results about Discrete Event Systems which should help the reader in understanding some crucial points in chapter 5; while in Appendix B an overview on the experimental testbed of the Laboratory of Automation of University of Bologna, is reported to validated the approach presented in chapter 3, chapter 4 and chapter 5. In Appendix C some components model used in chapter 5 for formal verification are reported.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Two of the main features of today complex software systems like pervasive computing systems and Internet-based applications are distribution and openness. Distribution revolves around three orthogonal dimensions: (i) distribution of control|systems are characterised by several independent computational entities and devices, each representing an autonomous and proactive locus of control; (ii) spatial distribution|entities and devices are physically distributed and connected in a global (such as the Internet) or local network; and (iii) temporal distribution|interacting system components come and go over time, and are not required to be available for interaction at the same time. Openness deals with the heterogeneity and dynamism of system components: complex computational systems are open to the integration of diverse components, heterogeneous in terms of architecture and technology, and are dynamic since they allow components to be updated, added, or removed while the system is running. The engineering of open and distributed computational systems mandates for the adoption of a software infrastructure whose underlying model and technology could provide the required level of uncoupling among system components. This is the main motivation behind current research trends in the area of coordination middleware to exploit tuple-based coordination models in the engineering of complex software systems, since they intrinsically provide coordinated components with communication uncoupling and further details in the references therein. An additional daunting challenge for tuple-based models comes from knowledge-intensive application scenarios, namely, scenarios where most of the activities are based on knowledge in some form|and where knowledge becomes the prominent means by which systems get coordinated. Handling knowledge in tuple-based systems induces problems in terms of syntax - e.g., two tuples containing the same data may not match due to differences in the tuple structure - and (mostly) of semantics|e.g., two tuples representing the same information may not match based on a dierent syntax adopted. Till now, the problem has been faced by exploiting tuple-based coordination within a middleware for knowledge intensive environments: e.g., experiments with tuple-based coordination within a Semantic Web middleware (surveys analogous approaches). However, they appear to be designed to tackle the design of coordination for specic application contexts like Semantic Web and Semantic Web Services, and they result in a rather involved extension of the tuple space model. The main goal of this thesis was to conceive a more general approach to semantic coordination. In particular, it was developed the model and technology of semantic tuple centres. It is adopted the tuple centre model as main coordination abstraction to manage system interactions. A tuple centre can be seen as a programmable tuple space, i.e. an extension of a Linda tuple space, where the behaviour of the tuple space can be programmed so as to react to interaction events. By encapsulating coordination laws within coordination media, tuple centres promote coordination uncoupling among coordinated components. Then, the tuple centre model was semantically enriched: a main design choice in this work was to try not to completely redesign the existing syntactic tuple space model, but rather provide a smooth extension that { although supporting semantic reasoning { keep the simplicity of tuple and tuple matching as easier as possible. By encapsulating the semantic representation of the domain of discourse within coordination media, semantic tuple centres promote semantic uncoupling among coordinated components. The main contributions of the thesis are: (i) the design of the semantic tuple centre model; (ii) the implementation and evaluation of the model based on an existent coordination infrastructure; (iii) a view of the application scenarios in which semantic tuple centres seem to be suitable as coordination media.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Implicazioni tettoniche ed estetiche delle logiche monoscocca integrate e stress lines analysis in architettura.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

As the elastic response of cell membranes to mechanical stimuli plays a key role in various cellular processes, novel biophysical strategies to quantify the elasticity of native membranes under physiological conditions at a nanometer scale are gaining interest. In order to investigate the elastic response of apical membranes, elasticity maps of native membrane sheets, isolated from MDCK II (Madine Darby Canine kidney strain II) epithelial cells, were recorded by local indentation with an Atomic Force Microscope (AFM). To exclude the underlying substrate effect on membrane indentation, a highly ordered gold coated porous array with a pore diameter of 1.2 μm was used to support apical membranes. Overlays of fluorescence and AFM images show that intact apical membrane sheets are attached to poly-D-lysine coated porous substrate. Force indentation measurements reveal an extremely soft elastic membrane response if it is indented at the center of the pore in comparison to a hard repulsion on the adjacent rim used to define the exact contact point. A linear dependency of force versus indentation (-dF/dh) up to 100 nm penetration depth enabled us to define an apparent membrane spring constant (kapp) as the slope of a linear fit with a stiffness value of for native apical membrane in PBS. A correlation between fluorescence intensity and kapp is also reported. Time dependent hysteresis observed with native membranes is explained by a viscoelastic solid model of a spring connected to a Kelvin-Voight solid with a time constant of 0.04 s. No hysteresis was reported with chemically fixated membranes. A combined linear and non linear elastic response is suggested to relate the experimental data of force indentation curves to the elastic modulus and the membrane thickness. Membrane bending is the dominant contributor to linear elastic indentation at low loads, whereas stretching is the dominant contributor for non linear elastic response at higher loads. The membrane elastic response was controlled either by stiffening with chemical fixatives or by softening with F-actin disrupters. Overall, the presented setup is ideally suitable to study the interactions of the apical membrane with the underlying cytoskeleton by means of force indentation elasticity maps combined with fluorescence imaging.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Nel primo capitolo viene introdotto lo studio eff�ettuato e descritto un metodo di misure successivo alla caratterizzazione della super�ficie. Nel secondo capitolo vengono descritti i campioni analizzati e, nello speci�fico, la crescita attraverso MaCE dei nanofi�li di silicio. Nel terzo capitolo viene descritto lo strumento AFM utilizzato e la teoria della caratterizzazione alla base dello studio condotto. Nella quarta sezione vengono descritti i risultati ottenuti mentre nelle conclusioni viene tratto il risultato dei valori ottenuti di RMS roughness e roughness exponent.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The thesis analyses the hydrodynamic induced by an array of Wave energy Converters (WECs), under an experimental and numerical point of view. WECs can be considered an innovative solution able to contribute to the green energy supply and –at the same time– to protect the rear coastal area under marine spatial planning considerations. This research activity essentially rises due to this combined concept. The WEC under exam is a floating device belonging to the Wave Activated Bodies (WAB) class. Experimental data were performed at Aalborg University in different scales and layouts, and the performance of the models was analysed under a variety of irregular wave attacks. The numerical simulations performed with the codes MIKE 21 BW and ANSYS-AQWA. Experimental results were also used to calibrate the numerical parameters and/or to directly been compared to numerical results, in order to extend the experimental database. Results of the research activity are summarized in terms of device performance and guidelines for a future wave farm installation. The device length should be “tuned” based on the local climate conditions. The wave transmission behind the devices is pretty high, suggesting that the tested layout should be considered as a module of a wave farm installation. Indications on the minimum inter-distance among the devices are provided. Furthermore, a CALM mooring system leads to lower wave transmission and also larger power production than a spread mooring. The two numerical codes have different potentialities. The hydrodynamics around single and multiple devices is obtained with MIKE 21 BW, while wave loads and motions for a single moored device are derived from ANSYS-AQWA. Combining the experimental and numerical it is suggested –for both coastal protection and energy production– to adopt a staggered layout, which will maximise the devices density and minimize the marine space required for the installation.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The upgrade of the Mainz Mikrotron (MAMI) electron accelerator facility in 2007 which raised the beam energy up to 1.5,GeV, gives the opportunity to study strangeness production channels through electromagnetic process. The Kaon Spectrometer (KAOS) managed by the A1 Collaboration, enables the efficient detection of the kaons associated with strangeness electroproduction. Used as a single arm spectrometer, it can be combined with the existing high-resolution spectrometers for exclusive measurements in the kinematic domain accessible to them.rnrnFor studying hypernuclear production in the ^A Z(e,e'K^+) _Lambda ^A(Z-1) reaction, the detection of electrons at very forward angles is needed. Therefore, the use of KAOS as a double-arm spectrometer for detection of kaons and the electrons at the same time is mandatory. Thus, the electron arm should be provided with a new detector package, with high counting rate capability and high granularity for a good spatial resolution. To this end, a new state-of-the-art scintillating fiber hodoscope has been developed as an electron detector.rnrnThe hodoscope is made of two planes with a total of 18432 scintillating double-clad fibers of 0.83 mm diameter. Each plane is formed by 72 modules. Each module is formed from a 60deg slanted multi-layer bundle, where 4 fibers of a tilted column are connected to a common read out. The read-out is made with 32 channels of linear array multianode photomultipliers. Signal processing makes use of newly developed double-threshold discriminators. The discriminated signal is sent in parallel to dead-time free time-to-digital modules and to logic modules for triggering purposes.rnrnTwo fiber modules were tested with a carbon beam at GSI, showing a time resolution of 220 ps (FWHM) and a position residual of 270 microm m (FWHM) with a detection efficiency epsilon>99%.rnrnThe characterization of the spectrometer arm has been achieved through simulations calculating the transfer matrix of track parameters from the fiber detector focal plane to the primary vertex. This transfer matrix has been calculated to first order using beam transport optics and has been checked by quasielastic scattering off a carbon target, where the full kinematics is determined by measuring the recoil proton momentum. The reconstruction accuracy for the emission parameters at the quasielastic vertex was found to be on the order of 0.3 % in first test realized.rnrnThe design, construction process, commissioning, testing and characterization of the fiber hodoscope are presented in this work which has been developed at the Institut für Kernphysik of the Johannes Gutenberg - Universität Mainz.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Lo scopo di questo lavoro di tesi è indagare le capacità di ALMA di osservare il continuo e le righe molecolari di emissione di HCN, HCO+ e HNC, in galassie lensate ad alto z, nelle prime fasi della loro formazione. Per farlo vengono utilizzate osservazioni ALMA di righe di emissione molecolare, relative a dati pubblici di Ciclo 0. Queste osservazioni vengono utilizzate per simulare in modo realistico l’emissione da parte di galassie ad alto z e la risposta osservativa del telescopio, assumendo diverse possibili geometrie di lensing. Inoltre le recenti osservazioni ALMA sono state utilizzate per aggiornare le relazioni già esistenti tra la luminosità FIR e la luminosità delle righe molecolari. Queste recenti osservazioni con maggiore sensibilità e risoluzione angolare, sembrano essere in buon accordo con le precedenti osservazioni. Per realizzare questo progetto di tesi, sono stati scaricati dall’archivio ALMA i dati non calibrati relativi a due sorgenti ben studiate: NGC1614, rappresentativa delle galassie di tipo starburst ed IRAS 20551-4250, un AGN oscurato. Il processo di calibrazione è stato ripetuto per esaminare le proprietà dei cubi di dati utilizzando il pacchetto Common Astronomy Software Applications (CASA). Le righe spettrali osservate e l’emissione del continuo sono state successivamente estrapolate ad alto z riscalando adeguatamente le distanze, le dimensioni delle sorgenti e le frequenze di emissione. In seguito è stato applicato un modello di lensing gravitazionale basato su quello di Lapi et al. (2012). Sono state considerate diverse configurazioni tra lente e sorgente per ottenere diverse geometrie delle immagini e diversi fattori di amplificazione. Infine stato utilizzato il software CASA per simulare le osservazioni ALMA in modo da verificare le capacità osservative del telescopio al variare della geometria del sistema. Per ogni riga molecolare e per l’emissione del continuo sono state determinate la sensibilit e la risoluzione che possono essere raggiunte con le osservazioni ALMA e sono state analizzate alcune strategie osservative per effettuare survey di righe spettrali in oggetti lensati. Inoltre stata analizzata la possibilit di caratterizzare oggetti starburst ed AGN dai rapporti tra le righe di emissione delle molecole traccianti di alta densit. Le prestazioni di ALMA consentiranno di distinguere e stimare i contributi relativi di SB ed AGN in galassie lensate a z > 2.5, quindi vicine alla loro presunta epoca di formazione (Lapi et al. 2010), in meno di 5 minuti di osservazione per qualsiasi fattore di magnificazione. Nel presente lavoro sono state inoltre discusse alcune strategie osservative per condurre survey o followup per osservare le righe di HCN(4-3), HCO+(4-3) e HNC(4-3) in galassie lensate a redshift 2.5 < z < 3, dimostrando che sono possibili per campioni statisticamente significativi in tempi relativamente brevi.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The energy harvesting research field has grown considerably in the last decade due to increasing interests in energy autonomous sensing systems, which require smart and efficient interfaces for extracting power from energy source and power management (PM) circuits. This thesis investigates the design trade-offs for minimizing the intrinsic power of PM circuits, in order to allow operation with very weak energy sources. For validation purposes, three different integrated power converter and PM circuits for energy harvesting applications are presented. They have been designed for nano-power operations and single-source converters can operate with input power lower than 1 μW. The first IC is a buck-boost converter for piezoelectric transducers (PZ) implementing Synchronous Electrical Charge Extraction (SECE), a non-linear energy extraction technique. Moreover, Residual Charge Inversion technique is exploited for extracting energy from PZ with weak and irregular excitations (i.e. lower voltage), and the implemented PM policy, named Two-Way Energy Storage, considerably reduces the start-up time of the converter, improving the overall conversion efficiency. The second proposed IC is a general-purpose buck-boost converter for low-voltage DC energy sources, up to 2.5 V. An ultra-low-power MPPT circuit has been designed in order to track variations of source power. Furthermore, a capacitive boost circuit has been included, allowing the converter start-up from a source voltage VDC0 = 223 mV. A nano-power programmable linear regulator is also included in order to provide a stable voltage to the load. The third IC implements an heterogeneous multisource buck-boost converter. It provides up to 9 independent input channels, of which 5 are specific for PZ (with SECE) and 4 for DC energy sources with MPPT. The inductor is shared among channels and an arbiter, designed with asynchronous logic to reduce the energy consumption, avoids simultaneous access to the buck-boost core, with a dynamic schedule based on source priority.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Nowadays the rise of non-recurring engineering (NRE) costs associated with complexity is becoming a major factor in SoC design, limiting both scaling opportunities and the flexibility advantages offered by the integration of complex computational units. The introduction of embedded programmable elements can represent an appealing solution, able both to guarantee the desired flexibility and upgradabilty and to widen the SoC market. In particular embedded FPGA (eFPGA) cores can provide bit-level optimization for those applications which benefits from synthesis, paying on the other side in terms of performance penalties and area overhead with respect to standard cell ASIC implementations. In this scenario this thesis proposes a design methodology for a synthesizable programmable device designed to be embedded in a SoC. A soft-core embedded FPGA (eFPGA) is hence presented and analyzed in terms of the opportunities given by a fully synthesizable approach, following an implementation flow based on Standard-Cell methodology. A key point of the proposed eFPGA template is that it adopts a Multi-Stage Switching Network (MSSN) as the foundation of the programmable interconnects, since it can be efficiently synthesized and optimized through a standard cell based implementation flow, ensuring at the same time an intrinsic congestion-free network topology. The evaluation of the flexibility potentialities of the eFPGA has been performed using different technology libraries (STMicroelectronics CMOS 65nm and BCD9s 0.11μm) through a design space exploration in terms of area-speed-leakage tradeoffs, enabled by the full synthesizability of the template. Since the most relevant disadvantage of the adopted soft approach, compared to a hardcore, is represented by a performance overhead increase, the eFPGA analysis has been made targeting small area budgets. The generation of the configuration bitstream has been obtained thanks to the implementation of a custom CAD flow environment, and has allowed functional verification and performance evaluation through an application-aware analysis.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this thesis work we will explore and discuss the properties of the gamma-ray sources included in the first Fermi-LAT catalog of sources above 10 GeV (1FHL), by considering both blazars and the non negligible fraction of still unassociated gamma-ray sources (UGS, 13%). We perform a statistical analysis of a complete sample of hard gamma-ray sources, included in the 1FHL catalog, mostly composed of HSP blazars, and we present new VLBI observations of the faintest members of the sample. The new VLBI data, complemented by an extensive search of the archives for brighter sources, are essential to gather a sample as large as possible for the assessment of the significance of the correlation between radio and very high energy (E>100 GeV) emission bands. After the characterization of the statistical properties of HSP blazars and UGS, we use a complementary approach, by focusing on an intensive multi-frequency observing VLBI and gamma-ray campaign carried out for one of the most remarkable and closest HSP blazar Markarian 421.