9 resultados para Functional Performance
em AMS Tesi di Dottorato - Alm@DL - Università di Bologna
Resumo:
The study is aimed to calculate an innovative numerical index for bit performance evaluation called Bit Index (BI), applied on a new type of bit database named Formation Drillability Catalogue (FDC). A dedicated research programme (developed by Eni E&P and the University of Bologna) studied a drilling model for bit performance evaluation named BI, derived from data recorded while drilling (bit records, master log, wireline log, etc.) and dull bit evaluation. This index is calculated with data collected inside the FDC, a novel classification of Italian formations aimed to the geotechnical and geomechanical characterization and subdivisions of the formations, called Minimum Interval (MI). FDC was conceived and prepared at Eni E&P Div., and contains a large number of significant drilling parameters. Five wells have been identified inside the FDC and have been tested for bit performance evaluation. The values of BI are calculated for each bit run and are compared with the values of the cost per metre. The case study analyzes bits of the same type, diameters and run in the same formation. The BI methodology implemented on MI classification of FDC can improve consistently the bit performances evaluation, and it helps to identify the best performer bits. Moreover, FDC turned out to be functional to BI, since it discloses and organizes formation details that are not easily detectable or usable from bit records or master logs, allowing for targeted bit performance evaluations. At this stage of development, the BI methodology proved to be economic and reliable. The quality of bit performance analysis obtained with BI seems also more effective than the traditional “quick look” analysis, performed on bit records, or on the pure cost per metre evaluation.
Resumo:
Aging is a physiological process characterized by a progressive decline of the “cellular homeostatic reserve”, refereed as the capability to respond suitably to exogenous and endogenous stressful stimuli. Due to their high energetic requests and post-mitotic nature, neurons are peculiarly susceptible to this phenomenon. However, the aged brain maintains a certain level of adaptive capacities and if properly stimulated may warrant a considerable functional recovery. Aim of the present research was to verify the plastic potentialities of the aging brain of rats subjected to two kind of exogenous stimuli: A) the replacement of the standard diet with a ketogenic regimen (the change forces the brain to use ketone bodies (KB) in alternative to glucose to satisfy the energetic needs) and B) a behavioural task able to induce the formation of inhibitory avoidance memory. A) Fifteen male Wistar rats of 19 months of age were divided into three groups (average body weight pair-matched), and fed for 8 weeks with different dietary regimens: i) diet containing 10% medium chain triglycerides (MCT); ii) diet containing 20% MCT; iii) standard commercial chow. Five young (5 months of age) and five old (26-27 months of age) animals fed with the standard diet were used as further controls. The following morphological parameters reflecting synaptic plasticity were evaluated in the stratum moleculare of the hippocampal CA1 region (SM CA1), in the outer molecular layer of the hippocampal dentate gyrus (OML DG), and in the granule cell layer of the cerebellar cortex (GCL-CCx): average area (S), numeric density (Nvs), and surface density (Sv) of synapses, and average volume (V), numeric density (Nvm), and volume density (Vv) of synaptic mitochondria. Moreover, succinic dehydrogenase (SDH) activity was cytochemically determined in Purkinje cells (PC) and V, Nvm, Vv, and cytochemical precipitate area/mitochondrial area (R) of SDH-positive mitochondria were evaluated. In SM CA1, MCT-KDs induced the early appearance of the morphological patterns typical of old animals: higher S and V, and lower Nvs and Nvm. On the contrary, in OML DG, Sv and Vv of MCT-KDs-fed rats were higher (as a result of higher Nvs and Nvm) vs. controls; these modifications are known to improve synaptic function and metabolic supply. The opposite effects of MCT-KDs might reflect the different susceptibility of these brain regions to the aging processes: OML DG is less vulnerable than SM CA1, and the reactivation of ketone bodies uptake and catabolism might occur more efficiently in this region, allowing the exploitation of their peculiar metabolic properties. In GCL-CCx, the results described a new scenario in comparison to that found in the hippocampal formation: 10%MCT-KD induced the early appearance of senescent patterns (decreased Nvs and Nvm; increased V), whereas 20%MCT-KD caused no changes. Since GCL-CCx is more vulnerable to age than DG, and less than CA1, these data further support the hypothesis that MCT-KDs effects in the aging brain critically depend on neuronal vulnerability to age, besides MCT percentage. Regarding PC, it was decided to evaluate only the metabolic effect of the dietetic regimen (20%MCT-KD) characterized by less side effects. KD counteracted age-related decrease in numeric density of SDH-positive mitochondria, and enhanced their energetic efficiency (R was significantly higher in MCT-KD-fed rats vs. all the controls). Since it is well known that Purkinje and dentate gyrus cells are less vulnerable to aging than CA1 neurons, these results corroborate our previous hypothesis. In conclusion, the A) experimental line provides the first evidence that morphological and functional parameters reflecting synaptic plasticity and mitochondrial metabolic competence may be modulated by MCT-KDs in the pre-senescent central nervous system, and that the effects may be heterogeneous in different brain regions. MCT-KDs seem to supply high energy metabolic intermediates and to be beneficial (“anti-aging”) for those neurons that maintain the capability to exploit them. This implies risks but also promising potentialities for the therapeutic use of these diets during aging B) Morphological parameters of synapses and synaptic mitochondria in SM CA1 were investigated in old (26-27 month-old) female Wistar rats following a single trial inhibitory avoidance task. In this memory protocol animals learn to avoid a dark compartment in which they received a mild, inescapable foot-shock. Rats were tested 3 and 6 or 9 hours after the training, divided into good and bad responders according to their performance (retention times above or below 100 s, respectively) and immediately sacrificed. Nvs, S, Sv, Nvm, V, and Vv were evaluated. In the good responder group, the numeric density of synapses and mitochondria was significantly higher and the average mitochondrial volume was significantly smaller 9 hours vs. 6 hours after the training. No significant differences were observed among bad responders. Thus, better performances in passive avoidance memory task are correlated with more efficient plastic remodeling of synaptic contacts and mitochondria in hippocampal CA1. These findings indicate that maintenance of synaptic plastic reactivity during aging is a critical requirement for preserving long-term memory consolidation.
Resumo:
Recently in most of the industrial automation process an ever increasing degree of automation has been observed. This increasing is motivated by the higher requirement of systems with great performance in terms of quality of products/services generated, productivity, efficiency and low costs in the design, realization and maintenance. This trend in the growth of complex automation systems is rapidly spreading over automated manufacturing systems (AMS), where the integration of the mechanical and electronic technology, typical of the Mechatronics, is merging with other technologies such as Informatics and the communication networks. An AMS is a very complex system that can be thought constituted by a set of flexible working stations, one or more transportation systems. To understand how this machine are important in our society let considerate that every day most of us use bottles of water or soda, buy product in box like food or cigarets and so on. Another important consideration from its complexity derive from the fact that the the consortium of machine producers has estimated around 350 types of manufacturing machine. A large number of manufacturing machine industry are presented in Italy and notably packaging machine industry,in particular a great concentration of this kind of industry is located in Bologna area; for this reason the Bologna area is called “packaging valley”. Usually, the various parts of the AMS interact among them in a concurrent and asynchronous way, and coordinate the parts of the machine to obtain a desiderated overall behaviour is an hard task. Often, this is the case in large scale systems, organized in a modular and distributed manner. Even if the success of a modern AMS from a functional and behavioural point of view is still to attribute to the design choices operated in the definition of the mechanical structure and electrical electronic architecture, the system that governs the control of the plant is becoming crucial, because of the large number of duties associated to it. Apart from the activity inherent to the automation of themachine cycles, the supervisory system is called to perform other main functions such as: emulating the behaviour of traditional mechanical members thus allowing a drastic constructive simplification of the machine and a crucial functional flexibility; dynamically adapting the control strategies according to the different productive needs and to the different operational scenarios; obtaining a high quality of the final product through the verification of the correctness of the processing; addressing the operator devoted to themachine to promptly and carefully take the actions devoted to establish or restore the optimal operating conditions; managing in real time information on diagnostics, as a support of the maintenance operations of the machine. The kind of facilities that designers can directly find on themarket, in terms of software component libraries provides in fact an adequate support as regard the implementation of either top-level or bottom-level functionalities, typically pertaining to the domains of user-friendly HMIs, closed-loop regulation and motion control, fieldbus-based interconnection of remote smart devices. What is still lacking is a reference framework comprising a comprehensive set of highly reusable logic control components that, focussing on the cross-cutting functionalities characterizing the automation domain, may help the designers in the process of modelling and structuring their applications according to the specific needs. Historically, the design and verification process for complex automated industrial systems is performed in empirical way, without a clear distinction between functional and technological-implementation concepts and without a systematic method to organically deal with the complete system. Traditionally, in the field of analog and digital control design and verification through formal and simulation tools have been adopted since a long time ago, at least for multivariable and/or nonlinear controllers for complex time-driven dynamics as in the fields of vehicles, aircrafts, robots, electric drives and complex power electronics equipments. Moving to the field of logic control, typical for industrial manufacturing automation, the design and verification process is approached in a completely different way, usually very “unstructured”. No clear distinction between functions and implementations, between functional architectures and technological architectures and platforms is considered. Probably this difference is due to the different “dynamical framework”of logic control with respect to analog/digital control. As a matter of facts, in logic control discrete-events dynamics replace time-driven dynamics; hence most of the formal and mathematical tools of analog/digital control cannot be directly migrated to logic control to enlighten the distinction between functions and implementations. In addition, in the common view of application technicians, logic control design is strictly connected to the adopted implementation technology (relays in the past, software nowadays), leading again to a deep confusion among functional view and technological view. In Industrial automation software engineering, concepts as modularity, encapsulation, composability and reusability are strongly emphasized and profitably realized in the so-calledobject-oriented methodologies. Industrial automation is receiving lately this approach, as testified by some IEC standards IEC 611313, IEC 61499 which have been considered in commercial products only recently. On the other hand, in the scientific and technical literature many contributions have been already proposed to establish a suitable modelling framework for industrial automation. During last years it was possible to note a considerable growth in the exploitation of innovative concepts and technologies from ICT world in industrial automation systems. For what concerns the logic control design, Model Based Design (MBD) is being imported in industrial automation from software engineering field. Another key-point in industrial automated systems is the growth of requirements in terms of availability, reliability and safety for technological systems. In other words, the control system should not only deal with the nominal behaviour, but should also deal with other important duties, such as diagnosis and faults isolations, recovery and safety management. Indeed, together with high performance, in complex systems fault occurrences increase. This is a consequence of the fact that, as it typically occurs in reliable mechatronic systems, in complex systems such as AMS, together with reliable mechanical elements, an increasing number of electronic devices are also present, that are more vulnerable by their own nature. The diagnosis problem and the faults isolation in a generic dynamical system consists in the design of an elaboration unit that, appropriately processing the inputs and outputs of the dynamical system, is also capable of detecting incipient faults on the plant devices, reconfiguring the control system so as to guarantee satisfactory performance. The designer should be able to formally verify the product, certifying that, in its final implementation, it will perform itsrequired function guarantying the desired level of reliability and safety; the next step is that of preventing faults and eventually reconfiguring the control system so that faults are tolerated. On this topic an important improvement to formal verification of logic control, fault diagnosis and fault tolerant control results derive from Discrete Event Systems theory. The aimof this work is to define a design pattern and a control architecture to help the designer of control logic in industrial automated systems. The work starts with a brief discussion on main characteristics and description of industrial automated systems on Chapter 1. In Chapter 2 a survey on the state of the software engineering paradigm applied to industrial automation is discussed. Chapter 3 presentes a architecture for industrial automated systems based on the new concept of Generalized Actuator showing its benefits, while in Chapter 4 this architecture is refined using a novel entity, the Generalized Device in order to have a better reusability and modularity of the control logic. In Chapter 5 a new approach will be present based on Discrete Event Systems for the problemof software formal verification and an active fault tolerant control architecture using online diagnostic. Finally conclusive remarks and some ideas on new directions to explore are given. In Appendix A are briefly reported some concepts and results about Discrete Event Systems which should help the reader in understanding some crucial points in chapter 5; while in Appendix B an overview on the experimental testbed of the Laboratory of Automation of University of Bologna, is reported to validated the approach presented in chapter 3, chapter 4 and chapter 5. In Appendix C some components model used in chapter 5 for formal verification are reported.
Resumo:
This thesis is a collection of essays related to the topic of innovation in the service sector. The choice of this structure is functional to the purpose of single out some of the relevant issues and try to tackle them, revising first the state of the literature and then proposing a way forward. Three relevant issues has been therefore selected: (i) the definition of innovation in the service sector and the connected question of measurement of innovation; (ii) the issue of productivity in services; (iii) the classification of innovative firms in the service sector. Facing the first issue, chapter II shows how the initial width of the original Schumpeterian definition of innovation has been narrowed and then passed to the service sector form the manufacturing one in a reduce technological form. Chapter III tackle the issue of productivity in services, discussing the difficulties for measuring productivity in a context where the output is often immaterial. We reconstruct the dispute on the Baumol’s cost disease argument and propose two different ways to go forward in the research on productivity in services: redefining the output along the line of a characteristic approach; and redefining the inputs, particularly analysing which kind of input it’s worth saving. Chapter IV derives an integrated taxonomy of innovative service and manufacturing firms, using data coming from the 2008 CIS survey for Italy. This taxonomy is based on the enlarged definition of “innovative firm” deriving from the Schumpeterian definition of innovation and classify firms using a cluster analysis techniques. The result is the emergence of a four cluster solution, where firms are differentiated by the breadth of the innovation activities in which they are involved. Chapter 5 reports some of the main conclusions of each singular previous chapter and the points worth of further research in the future.
Resumo:
The diagnosis, grading and classification of tumours has benefited considerably from the development of DCE-MRI which is now essential to the adequate clinical management of many tumour types due to its capability in detecting active angiogenesis. Several strategies have been proposed for DCE-MRI evaluation. Visual inspection of contrast agent concentration curves vs time is a very simple yet operator dependent procedure, therefore more objective approaches have been developed in order to facilitate comparison between studies. In so called model free approaches, descriptive or heuristic information extracted from time series raw data have been used for tissue classification. The main issue concerning these schemes is that they have not a direct interpretation in terms of physiological properties of the tissues. On the other hand, model based investigations typically involve compartmental tracer kinetic modelling and pixel-by-pixel estimation of kinetic parameters via non-linear regression applied on region of interests opportunely selected by the physician. This approach has the advantage to provide parameters directly related to the pathophysiological properties of the tissue such as vessel permeability, local regional blood flow, extraction fraction, concentration gradient between plasma and extravascular-extracellular space. Anyway, nonlinear modelling is computational demanding and the accuracy of the estimates can be affected by the signal-to-noise ratio and by the initial solutions. The principal aim of this thesis is investigate the use of semi-quantitative and quantitative parameters for segmentation and classification of breast lesion. The objectives can be subdivided as follow: describe the principal techniques to evaluate time intensity curve in DCE-MRI with focus on kinetic model proposed in literature; to evaluate the influence in parametrization choice for a classic bi-compartmental kinetic models; to evaluate the performance of a method for simultaneous tracer kinetic modelling and pixel classification; to evaluate performance of machine learning techniques training for segmentation and classification of breast lesion.
Resumo:
Despite several clinical tests that have been developed to qualitatively describe complex motor tasks by functional testing, these methods often depend on clinicians' interpretation, experience and training, which make the assessment results inconsistent, without the precision required to objectively assess the effect of the rehabilitative intervention. A more detailed characterization is required to fully capture the various aspects of motor control and performance during complex movements of lower and upper limbs. The need for cost-effective and clinically applicable instrumented tests would enable quantitative assessment of performance on a subject-specific basis, overcoming the limitations due to the lack of objectiveness related to individual judgment, and possibly disclosing subtle alterations that are not clearly visible to the observer. Postural motion measurements at additional locations, such as lower and upper limbs and trunk, may be necessary in order to obtain information about the inter-segmental coordination during different functional tests involved in clinical practice. With these considerations in mind, this Thesis aims: i) to suggest a novel quantitative assessment tool for the kinematics and dynamics evaluation of a multi-link kinematic chain during several functional motor tasks (i.e. squat, sit-to-stand, postural sway), using one single-axis accelerometer per segment, ii) to present a novel quantitative technique for the upper limb joint kinematics estimation, considering a 3-link kinematic chain during the Fugl-Meyer Motor Assessment and using one inertial measurement unit per segment. The suggested methods could have several positive feedbacks from clinical practice. The use of objective biomechanical measurements, provided by inertial sensor-based technique, may help clinicians to: i) objectively track changes in motor ability, ii) provide timely feedback about the effectiveness of administered rehabilitation interventions, iii) enable intervention strategies to be modified or changed if found to be ineffective, and iv) speed up the experimental sessions when several subjects are asked to perform different functional tests.
Resumo:
The present Thesis reports on the various research projects to which I have contributed during my PhD period, working with several research groups, and whose results have been communicated in a number of scientific publications. The main focus of my research activity was to learn, test, exploit and extend the recently developed vdW-DFT (van der Waals corrected Density Functional Theory) methods for computing the structural, vibrational and electronic properties of ordered molecular crystals from first principles. A secondary, and more recent, research activity has been the analysis with microelectrostatic methods of Molecular Dynamics (MD) simulations of disordered molecular systems. While only very unreliable methods based on empirical models were practically usable until a few years ago, accurate calculations of the crystal energy are now possible, thanks to very fast modern computers and to the excellent performance of the best vdW-DFT methods. Accurate energies are particularly important for describing organic molecular solids, since they often exhibit several alternative crystal structures (polymorphs), with very different packing arrangements but very small energy differences. Standard DFT methods do not describe the long-range electron correlations which give rise to the vdW interactions. Although weak, these interactions are extremely sensitive to the packing arrangement, and neglecting them used to be a problem. The calculations of reliable crystal structures and vibrational frequencies has been made possible only recently, thanks to development of some good representations of the vdW contribution to the energy (known as “vdW corrections”).
Resumo:
Nello sport di alto livello l’uso della tecnologia ha raggiunto un ruolo di notevole importanza per l’analisi e la valutazione della prestazione. Negli ultimi anni sono emerse nuove tecnologie e sono migliorate quelle pre-esistenti (i.e. accelerometri, giroscopi e software per l’analisi video) in termini di campionamento, acquisizione dati, dimensione dei sensori che ha permesso la loro “indossabilità” e l’inserimento degli stessi all’interno degli attrezzi sportivi. La tecnologia è sempre stata al servizio degli atleti come strumento di supporto per raggiungere l’apice dei risultati sportivi. Per questo motivo la valutazione funzionale dell’atleta associata all’uso di tecnologie si pone lo scopo di valutare i miglioramenti degli atleti misurando la condizione fisica e/o la competenza tecnica di una determinata disciplina sportiva. L’obiettivo di questa tesi è studiare l’utilizzo delle applicazioni tecnologiche e individuare nuovi metodi di valutazione della performance in alcuni sport acquatici. La prima parte (capitoli 1-5), si concentra sulla tecnologia prototipale chiamata E-kayak e le varie applicazioni nel kayak di velocità. In questi lavori è stata verificata l’attendibilità dei dati forniti dal sistema E-kayak con i sistemi presenti in letteratura. Inoltre, sono stati indagati nuovi parametri utili a comprendere il modello di prestazione del paddler. La seconda parte (capitolo 6), si riferisce all’analisi cinematica della spinta verticale del pallanuotista, attraverso l’utilizzo della video analisi 2D, per l’individuazione delle relazioni Forza-velocità e Potenza-velocità direttamente in acqua. Questo studio pilota, potrà fornire indicazioni utili al monitoraggio e condizionamento di forza e potenza da svolgere direttamente in acqua. Infine la terza parte (capitoli 7-8), si focalizza sull’individuazione della sequenza di Fibonacci (sequenza divina) nel nuoto a stile libero e a farfalla. I risultati di questi studi suggeriscono che il ritmo di nuotata tenuto durante le medie/lunghe distanze gioca un ruolo chiave. Inoltre, il livello di autosomiglianza (self-similarity) aumenta con la tecnica del nuoto.
Resumo:
Interfacing materials with different intrinsic chemical-physical characteristics allows for the generation of a new system with multifunctional features. Here, this original concept is implemented for tailoring the functional properties of bi-dimensional black phosphorus (2D bP or phosphorene) and organic light-emitting transistors (OLETs). Phosphorene is highly reactive under atmospheric conditions and its small-area/lab-scale deposition techniques have hampered the introduction of this material in real-world applications so far. The protection of 2D bP against the oxygen by means of functionalization with alkane molecules and pyrene derivatives, showed long-term stability with respect to the bare 2D bP by avoiding remarkable oxidation up to 6 months, paving the way towards ultra-sensitive oxygen chemo-sensors. A new approach of deposition-precipitation heterogeneous reaction was developed to decorate 2D bP with Au nanoparticles (NP)s, obtaining a “stabilizer-free” that may broaden the possible applications of the 2D bP/Au NPs interface in catalysis and biodiagnostics. Finally, 2D bP was deposited by electrospray technique, obtaining oxidized-phosphorous flakes as wide as hundreds of µm2 and providing for the first time a phosphorous-based bidimensional system responsive to electromechanical stimuli. The second part of the thesis focuses on the study of organic heterostructures in ambipolar OLET devices, intriguing optoelectronic devices that couple the micro-scaled light-emission with electrical switching. Initially, an ambipolar single-layer OLET based on a multifunctional organic semiconductor, is presented. The bias-depending light-emission shifted within the transistor channel, as expected in well-balanced ambipolar OLETs. However, the emitted optical power of the single layer-based device was unsatisfactory. To improve optoelectronic performance of the device, a multilayer organic architecture based on hole-transporting semiconductor, emissive donor-acceptor blend and electron-transporting semiconductor was optimized. We showed that the introduction of a suitable electron-injecting layer at the interface between the electron-transporting and light-emission layers may enable a ≈ 2× improvement of efficiency at reduced applied bias.