8 resultados para Computing and software systems
em AMS Tesi di Laurea - Alm@DL - Università di Bologna
Resumo:
The hypothalamus-pituitary-interrenal axis is involved in stress response regulation. In addition, arginine vasotocin (AVT) and isotocin (IT) are also considered as important players in this stress regulation. The present study assessed, using the teleost gilthead sea bream (Sparus aurata) as a biological model, hypothalamic mRNA expression changes of AVT and IT and their receptors at hepatic level after an acute stress situation. Specimens were submitted to air for 3 min and place back in their respective tanks after that, being sampled at different times (15 min, 30 min, 1, 2, 4 and 8 hours post-stress) in order to study the time course response. Plasma cortisol values increased after few minutes post-exposure, decreasing during the experimental time while a metabolic reorganization occurred in both plasmatic and hepatic levels. At hypothalamic level, acute stress affects mRNA expression of AVT and IT precursors, as well as hepatic expression of their receptors, suggesting the involvement of both vasotocinergic and isotocinergic systems in the acute stress response. Our results demonstrate the activation and involvement of both endocrine pathways in the regulation of metabolic and stress systems of Sparus aurata, which is stated, at least, through changes in mRNA expression levels of these genes analysed.
Resumo:
This thesis is dedicated to the Tits-Kantor-Koecher (TKK) construction which establishes a bijective correspondence between unital Jordan algebras and shortly graded Lie algebras with Z-grading induced by an sl_2-triple. It is based on the observation that if g is a Lie algebra with a short Z-grading and f lies in g_1, then the formula ab=[[a,f],b] defines a structure of a Jordan algebra on g_{-1}. The TKK construction has been extended to Jordan triple systems and, more recently, to the so-called Kantor triple systems. These generalizations are studied in the thesis.
Resumo:
One of the main practical implications of quantum mechanical theory is quantum computing, and therefore the quantum computer. Quantum computing (for example, with Shor’s algorithm) challenges the computational hardness assumptions, such as the factoring problem and the discrete logarithm problem, that anchor the safety of cryptosystems. So the scientific community is studying how to defend cryptography; there are two defense strategies: the quantum cryptography (which involves the use of quantum cryptographic algorithms on quantum computers) and the post-quantum cryptography (based on classical cryptographic algorithms, but resistant to quantum computers). For example, National Institute of Standards and Technology (NIST) is collecting and standardizing the post-quantum ciphers, as it established DES and AES as symmetric cipher standards, in the past. In this thesis an introduction on quantum mechanics was given, in order to be able to talk about quantum computing and to analyze Shor’s algorithm. The differences between quantum and post-quantum cryptography were then analyzed. Subsequently the focus was given to the mathematical problems assumed to be resistant to quantum computers. To conclude, post-quantum digital signature cryptographic algorithms selected by NIST were studied and compared in order to apply them in today’s life.
Resumo:
The study is divided into two main part: one focused on the GEO Satellite IoT and the other on the LEO Satellite IoT. Concerning the GEO Satellite IoT, the activity has been developed in the context of EUMETSAT Data Collection Service (DCS) by investigating the performance at the receiver within challenging scenarios. DCS are provided by several GEO Satellite operators, giving almost total coverage around the world. In this study firstly an overview of the DCS end-to-end architecture is given followed by a detailed description of both the tools used for the simulations: the DCP-TST (message generator and transmitter) and the DCP-RX (receiver). After generating several test messages, the performances have been evaluated with the addition of impairments (CW and sweeping interferences) and considerations in terms of BER and Good Messages are produced. Furthermore, a study on the PLL System is also conducted together with evaluations on the effectiveness of tuning the PLL Bw on the overall performance. Concerning the LEO Satellite IoT, the activity was carried out in the framework of the ASI Bidirectional IoT Satellite Service (BISS) Project. The elaborate covers a survey about the possible services that the project can accomplish and a technical analysis on the uplink MA. In particular, the LR-FHSS is proved to be a valid alternative for the uplink through an extensive analysis on its Network capacity and through the study of an analytic model for Success Probability with its Matlab implementation.
Resumo:
Communication and coordination are two key-aspects in open distributed agent system, being both responsible for the system’s behaviour integrity. An infrastructure capable to handling these issues, like TuCSoN, should to be able to exploit modern technologies and tools provided by fast software engineering contexts. Thesis aims to demonstrate TuCSoN infrastructure’s abilities to cope new possibilities, hardware and software, offered by mobile technology. The scenarios are going to configure, are related to the distributed nature of multi-agent systems where an agent should be located and runned just on a mobile device. We deal new mobile technology frontiers concerned with smartphones using Android operating system by Google. Analysis and deployment of a distributed agent-based system so described go first to impact with quality and quantity considerations about available resources. Engineering issue at the base of our research is to use TuCSoN against to reduced memory and computing capability of a smartphone, without the loss of functionality, efficiency and integrity for the infrastructure. Thesis work is organized on two fronts simultaneously: the former is the rationalization process of the available hardware and software resources, the latter, totally orthogonal, is the adaptation and optimization process about TuCSoN architecture for an ad-hoc client side release.
Resumo:
Information technology (IT) is on the verge of another revolution. Driven by the increasing capabilities and ever declining costs of computing and communications devices, IT is being embedded into a growing range of physical devices linked together through networks and will become ever more pervasive as the component technologies become smaller, faster, and cheaper. [..] These networked systems of embedded computers, referred to as EmNets throughout this report, have the potential to change radically the way people interact with their environment by linking together a range of devices and sensors that will allow information to be collected, shared, and processed in unprecedented ways.[..] The use of EmNets throughout society could well dwarf previous milestones in the information revolution.[..] IT will eventually become \textbf{an invisible component of almost everything} in everyone`s surroundings. Con il ridursi dei costi e l'aumentare della capacità di computazione dei componenti elettronici sono proliferate piattaforme che permettono al bambino come all'ingegnere di sviluppare un'idea che trasversalmente taglia il mondo reale e quello virtuale. Una collisione tra due mondi che fino a poco tempo fa era consentita esclusivamente a professionisti. Oggetti che possono acquisire o estendere funzionalità, che ci permettono di estendere la nostra percezione del mondo e di rivalutarne i suoi limiti. Oggetti connessi alla 'rete delle reti' che condividono ed elaborano dati per un nuovo utilizzo delle informazioni. Con questa tesi si vuole andare ad esplorare l'applicazione degli agenti software alle nuove piattaforme dei sistemi embedded e dell'Internet of Things, tecnologie abbastanza mature eppure non ancora esplorate a fondo. Ha senso modellare un sistema embedded con gli agenti?
Resumo:
X-ray absorption spectroscopy (XAS) is a powerful means of investigation of structural and electronic properties in condensed -matter physics. Analysis of the near edge part of the XAS spectrum, the so – called X-ray Absorption Near Edge Structure (XANES), can typically provide the following information on the photoexcited atom: - Oxidation state and coordination environment. - Speciation of transition metal compounds. - Conduction band DOS projected on the excited atomic species (PDOS). Analysis of XANES spectra is greatly aided by simulations; in the most common scheme the multiple scattering framework is used with the muffin tin approximation for the scattering potential and the spectral simulation is based on a hypothetical, reference structure. This approach has the advantage of requiring relatively little computing power but in many cases the assumed structure is quite different from the actual system measured and the muffin tin approximation is not adequate for low symmetry structures or highly directional bonds. It is therefore very interesting and justified to develop alternative methods. In one approach, the spectral simulation is based on atomic coordinates obtained from a DFT (Density Functional Theory) optimized structure. In another approach, which is the object of this thesis, the XANES spectrum is calculated directly based on an ab – initio DFT calculation of the atomic and electronic structure. This method takes full advantage of the real many-electron final wavefunction that can be computed with DFT algorithms that include a core-hole in the absorbing atom to compute the final cross section. To calculate the many-electron final wavefunction the Projector Augmented Wave method (PAW) is used. In this scheme, the absorption cross section is written in function of several contributions as the many-electrons function of the finale state; it is calculated starting from pseudo-wavefunction and performing a reconstruction of the real-wavefunction by using a transform operator which contains some parameters, called partial waves and projector waves. The aim of my thesis is to apply and test the PAW methodology to the calculation of the XANES cross section. I have focused on iron and silicon structures and on some biological molecules target (myoglobin and cytochrome c). Finally other inorganic and biological systems could be taken into account for future applications of this methodology, which could become an important improvement with respect to the multiscattering approach.
Resumo:
Our generation of computational scientists is living in an exciting time: not only do we get to pioneer important algorithms and computations, we also get to set standards on how computational research should be conducted and published. From Euclid’s reasoning and Galileo’s experiments, it took hundreds of years for the theoretical and experimental branches of science to develop standards for publication and peer review. Computational science, rightly regarded as the third branch, can walk the same road much faster. The success and credibility of science are anchored in the willingness of scientists to expose their ideas and results to independent testing and replication by other scientists. This requires the complete and open exchange of data, procedures and materials. The idea of a “replication by other scientists” in reference to computations is more commonly known as “reproducible research”. In this context the journal “EAI Endorsed Transactions on Performance & Modeling, Simulation, Experimentation and Complex Systems” had the exciting and original idea to make the scientist able to submit simultaneously the article and the computation materials (software, data, etc..) which has been used to produce the contents of the article. The goal of this procedure is to allow the scientific community to verify the content of the paper, reproducing it in the platform independently from the OS chosen, confirm or invalidate it and especially allow its reuse to reproduce new results. This procedure is therefore not helpful if there is no minimum methodological support. In fact, the raw data sets and the software are difficult to exploit without the logic that guided their use or their production. This led us to think that in addition to the data sets and the software, an additional element must be provided: the workflow that relies all of them.