917 resultados para reliability test system
Resumo:
Since the first underground nuclear explosion, carried out in 1958, the analysis of seismic signals generated by these sources has allowed seismologists to refine the travel times of seismic waves through the Earth and to verify the accuracy of the location algorithms (the ground truth for these sources was often known). Long international negotiates have been devoted to limit the proliferation and testing of nuclear weapons. In particular the Treaty for the comprehensive nuclear test ban (CTBT), was opened to signatures in 1996, though, even if it has been signed by 178 States, has not yet entered into force, The Treaty underlines the fundamental role of the seismological observations to verify its compliance, by detecting and locating seismic events, and identifying the nature of their sources. A precise definition of the hypocentral parameters represents the first step to discriminate whether a given seismic event is natural or not. In case that a specific event is retained suspicious by the majority of the State Parties, the Treaty contains provisions for conducting an on-site inspection (OSI) in the area surrounding the epicenter of the event, located through the International Monitoring System (IMS) of the CTBT Organization. An OSI is supposed to include the use of passive seismic techniques in the area of the suspected clandestine underground nuclear test. In fact, high quality seismological systems are thought to be capable to detect and locate very weak aftershocks triggered by underground nuclear explosions in the first days or weeks following the test. This PhD thesis deals with the development of two different seismic location techniques: the first one, known as the double difference joint hypocenter determination (DDJHD) technique, is aimed at locating closely spaced events at a global scale. The locations obtained by this method are characterized by a high relative accuracy, although the absolute location of the whole cluster remains uncertain. We eliminate this problem introducing a priori information: the known location of a selected event. The second technique concerns the reliable estimates of back azimuth and apparent velocity of seismic waves from local events of very low magnitude recorded by a trypartite array at a very local scale. For the two above-mentioned techniques, we have used the crosscorrelation technique among digital waveforms in order to minimize the errors linked with incorrect phase picking. The cross-correlation method relies on the similarity between waveforms of a pair of events at the same station, at the global scale, and on the similarity between waveforms of the same event at two different sensors of the try-partite array, at the local scale. After preliminary tests on the reliability of our location techniques based on simulations, we have applied both methodologies to real seismic events. The DDJHD technique has been applied to a seismic sequence occurred in the Turkey-Iran border region, using the data recorded by the IMS. At the beginning, the algorithm was applied to the differences among the original arrival times of the P phases, so the cross-correlation was not used. We have obtained that the relevant geometrical spreading, noticeable in the standard locations (namely the locations produced by the analysts of the International Data Center (IDC) of the CTBT Organization, assumed as our reference), has been considerably reduced by the application of our technique. This is what we expected, since the methodology has been applied to a sequence of events for which we can suppose a real closeness among the hypocenters, belonging to the same seismic structure. Our results point out the main advantage of this methodology: the systematic errors affecting the arrival times have been removed or at least reduced. The introduction of the cross-correlation has not brought evident improvements to our results: the two sets of locations (without and with the application of the cross-correlation technique) are very similar to each other. This can be commented saying that the use of the crosscorrelation has not substantially improved the precision of the manual pickings. Probably the pickings reported by the IDC are good enough to make the random picking error less important than the systematic error on travel times. As a further justification for the scarce quality of the results given by the cross-correlation, it should be remarked that the events included in our data set don’t have generally a good signal to noise ratio (SNR): the selected sequence is composed of weak events ( magnitude 4 or smaller) and the signals are strongly attenuated because of the large distance between the stations and the hypocentral area. In the local scale, in addition to the cross-correlation, we have performed a signal interpolation in order to improve the time resolution. The algorithm so developed has been applied to the data collected during an experiment carried out in Israel between 1998 and 1999. The results pointed out the following relevant conclusions: a) it is necessary to correlate waveform segments corresponding to the same seismic phases; b) it is not essential to select the exact first arrivals; and c) relevant information can be also obtained from the maximum amplitude wavelet of the waveforms (particularly in bad SNR conditions). Another remarkable point of our procedure is that its application doesn’t demand a long time to process the data, and therefore the user can immediately check the results. During a field survey, such feature will make possible a quasi real-time check allowing the immediate optimization of the array geometry, if so suggested by the results at an early stage.
Resumo:
The research activity carried out during the PhD course in Electrical Engineering belongs to the branch of electric and electronic measurements. The main subject of the present thesis is a distributed measurement system to be installed in Medium Voltage power networks, as well as the method developed to analyze data acquired by the measurement system itself and to monitor power quality. In chapter 2 the increasing interest towards power quality in electrical systems is illustrated, by reporting the international research activity inherent to the problem and the relevant standards and guidelines emitted. The aspect of the quality of voltage provided by utilities and influenced by customers in the various points of a network came out only in recent years, in particular as a consequence of the energy market liberalization. Usually, the concept of quality of the delivered energy has been associated mostly to its continuity. Hence the reliability was the main characteristic to be ensured for power systems. Nowadays, the number and duration of interruptions are the “quality indicators” commonly perceived by most customers; for this reason, a short section is dedicated also to network reliability and its regulation. In this contest it should be noted that although the measurement system developed during the research activity belongs to the field of power quality evaluation systems, the information registered in real time by its remote stations can be used to improve the system reliability too. Given the vast scenario of power quality degrading phenomena that usually can occur in distribution networks, the study has been focused on electromagnetic transients affecting line voltages. The outcome of such a study has been the design and realization of a distributed measurement system which continuously monitor the phase signals in different points of a network, detect the occurrence of transients superposed to the fundamental steady state component and register the time of occurrence of such events. The data set is finally used to locate the source of the transient disturbance propagating along the network lines. Most of the oscillatory transients affecting line voltages are due to faults occurring in any point of the distribution system and have to be seen before protection equipment intervention. An important conclusion is that the method can improve the monitored network reliability, since the knowledge of the location of a fault allows the energy manager to reduce as much as possible both the area of the network to be disconnected for protection purposes and the time spent by technical staff to recover the abnormal condition and/or the damage. The part of the thesis presenting the results of such a study and activity is structured as follows: chapter 3 deals with the propagation of electromagnetic transients in power systems by defining characteristics and causes of the phenomena and briefly reporting the theory and approaches used to study transients propagation. Then the state of the art concerning methods to detect and locate faults in distribution networks is presented. Finally the attention is paid on the particular technique adopted for the same purpose during the thesis, and the methods developed on the basis of such approach. Chapter 4 reports the configuration of the distribution networks on which the fault location method has been applied by means of simulations as well as the results obtained case by case. In this way the performance featured by the location procedure firstly in ideal then in realistic operating conditions are tested. In chapter 5 the measurement system designed to implement the transients detection and fault location method is presented. The hardware belonging to the measurement chain of every acquisition channel in remote stations is described. Then, the global measurement system is characterized by considering the non ideal aspects of each device that can concur to the final combined uncertainty on the estimated position of the fault in the network under test. Finally, such parameter is computed according to the Guide to the Expression of Uncertainty in Measurements, by means of a numeric procedure. In the last chapter a device is described that has been designed and realized during the PhD activity aiming at substituting the commercial capacitive voltage divider belonging to the conditioning block of the measurement chain. Such a study has been carried out aiming at providing an alternative to the used transducer that could feature equivalent performance and lower cost. In this way, the economical impact of the investment associated to the whole measurement system would be significantly reduced, making the method application much more feasible.
Resumo:
Questa tesi di laurea è stata redatta presso l’azienda Sacmi Imola S.C. ed in particolare all’interno della divisione Closures, che si occupa della progettazione e della realizzazione di linee per la produzione di varie tipologie di capsule. Lo scopo dell’elaborato è descrivere lo sviluppo di un sistema di tracciabilità di prodotto; sistemi di questo tipo, adottati inizialmente nel settore alimentare, stanno acquisendo sempre maggiore importanza anche in altri campi produttivi, poiché rivestono un ruolo strategico al fine della realizzazione di prodotti caratterizzati da livelli elevati di performance e di qualità, capaci di emergere nel mercato moderno caratterizzato da una concorrenza estesa a livello mondiale e molto attento alle esigenze dei clienti. Nel caso specifico di Sacmi il sistema di tracciabilità si rivolge ad una pressa, la CCM (Continuous Compression Moulder), realizzata dall’azienda per la produzione di capsule in materiale termoplastico tramite la tecnologia dello stampaggio a compressione. In particolare il sistema si concentra sugli stampi della macchina CCM, i quali ne rappresentano gli elementi critici dal punto di vista sia tecnico che economico. A livello generale, un sistema di tracciabilità è costituito da due componenti fondamentali: il primo è un sistema di identificazione che permetta di rendere distinguibili ed individuabili le unità da tracciare, mentre il secondo è un sistema di raccolta dati in grado di raccogliere le informazioni desiderate. Queste sono poi archiviate in un apposito database ed attribuite alle entità corrispondenti sfruttando le proprietà del sistema di identificazione. Il primo passo da compiere quando si intende sviluppare un sistema di tracciabilità all’interno di un contesto produttivo già consolidato è la ricostruzione del processo produttivo presente in azienda: si tratta di individuare tutti gli enti aziendali che concorrono al processo e che saranno interessati dall’introduzione del nuovo sistema. Una volta definiti gli attori, è necessario anche capire come questi siano collegati dai flussi di materiale e di informazioni. Il processo produttivo di Sacmi era caratterizzato dalla quasi totale assenza di un flusso strutturato di informazioni a supporto di quello di materiale, ed il sistema di tracciabilità ha provveduto a colmare proprio questa mancanza. Il sistema deve essere in grado di integrarsi perfettamente nel contesto produttivo aziendale: è necessario trovare il giusto compromesso per quanto riguarda la quantità di informazioni da raccogliere, che devono garantire una corretta copertura di tutto il processo senza però appesantirlo eccessivamente. E’ bene che la raccolta dati sia basata su procedure standard che assicurino la ripetibilità delle operazioni di prelievo delle informazioni. Come è logico immaginarsi, l’introduzione di numerose novità nell’ambito di un contesto già strutturato ha fatto emergere un certo numero di problematiche, come ad esempio difficoltà nello stoccaggio e ritardi di produzione; queste devono essere risolte chiedendo uno sforzo aggiuntivo agli enti interessati o, nel medio/lungo periodo, evolvendo ed affinando il sistema con soluzioni più snelle.
Resumo:
Fra i sistemi di propulsione elettrica per satelliti, il Pulsed Plasma Thruster, PPT, è quello dal design più semplice. È anche il primo sistema di propulsione elettrica utilizzato in un satellite artificiale, ossia ZOND-2 lanciato nel 1964 dall’Unione Sovietica. Tuttavia, dopo circa 50 anni di ricerca, la comprensione teorica e sperimentale di questo dispositivo rimane limitata. Questo elaborato di tesi magistrale indaga sul sottosistema di accensione del PPT, cercando di mettere in luce alcuni aspetti legati al lifetime della spark plug, SP. Tale SP, o candela, è l’attuatore del sottosistema di accensione. Questa produce una scintilla sulla sua superficie, la quale permette la realizzazione della scarica elettrica principale fra i due elettrodi del motore. Questa scarica crea una sottile parete di plasma che, per mezzo della forza elettromagnetica di Lorentz, produce la spinta del PPT. Poiché la SP si trova all’interno del catodo del motore e si affaccia nella camera di scarica, questa soffre di fenomeni di corrosione e di deposizione carbonacea proveniente dal propellente. Questi fenomeni possono limitare notevolmente il lifetime della SP. I parametri connessi alla vita operativa della SP sono numerosi. In questo elaborato si è analizzata la possibilità di utilizzare una elettronica di accensione della candela alternativa alla classica soluzione che utilizza un trasformatore. Il sottosistema di accensione classico e quello nuovo sono stati realizzati e testati, per metterne in luce le differenze ed i possibili vantaggi/svantaggi.
Resumo:
Recent studies have shown that the nociceptive withdrawal reflex threshold (NWR-T) and the electrical pain threshold (EP-T) are reliable measures in pain-free populations. However, it is necessary to investigate the reliability of these measures in patients with chronic pain in order to translate these techniques from laboratory to clinic. The aims of this study were to determine the test-retest reliability of the NWR-T and EP-T after single and repeated (temporal summation) electrical stimulation in a group of patients with chronic low back pain, and to investigate the association between the NWR-T and the EP-T. To this end, 25 patients with chronic pain participated in three identical sessions, separated by 1 week in average, in which the NWR-T and the EP-T to single and repeated stimulation were measured. Test-retest reliability was assessed using intra-class correlation coefficient (ICC), coefficient of variation (CV), and Bland-Altman analysis. The association between the thresholds was assessed using the coefficient of determination (r (2)). The results showed good-to-excellent reliability for both NWR-T and EP-T in all cases, with average ICC values ranging 0.76-0.90 and average CV values ranging 12.0-17.7%. The association between thresholds was better after repeated stimulation than after single stimulation, with average r (2) values of 0.83 and 0.56, respectively. In conclusion, the NWR-T and the EP-T are reliable assessment tools for assessing the sensitivity of spinal nociceptive pathways in patients with chronic pain.
Resumo:
End-stage ankle arthritis should have an appropriate classification to assist surgeons in the management of end-stage ankle arthritis. Outcomes research also requires a classification system to stratify patients appropriately.
Resumo:
End-stage ankle arthritis is operatively treated with numerous designs of total ankle replacement and different techniques for ankle fusion. For superior comparison of these procedures, outcome research requires a classification system to stratify patients appropriately. A postoperative 4-type classification system was designed by 6 fellowship-trained foot and ankle surgeons. Four surgeons reviewed blinded patient profiles and radiographs on 2 occasions to determine the interobserver and intraobserver reliability of the classification. Excellent interobserver reliability (κ = .89) and intraobserver reproducibility (κ = .87) were demonstrated for the postoperative classification system. In conclusion, the postoperative Canadian Orthopaedic Foot and Ankle Society (COFAS) end-stage ankle arthritis classification system appears to be a valid tool to evaluate the outcome of patients operated for end-stage ankle arthritis.
Resumo:
The pattern of contact sensitization to the supposedly most important allergens assembled in the baseline series differs between countries, presumably at least partly because of exposure differences. Objectives. To describe the prevalence of contact sensitization to allergens tested in consecutive patients in the years 2007 and 2008, and to discuss possible differences.
Resumo:
There is a steadily increasing pressure on cost-savings and productivity growth in sectors of order-picking such that the wish for rationalization by automation is rising. Special problems are faced trying to automatize handling operations of order-picking articles packed in bags. The mechanical properties of the objects and their hard-to-predict shape and position represent obstacles and are complicating handling operations. A systematic approach in system design is required. This article deals with the properties of such products under aspects of difficulties arising in automated handling and points out a useful system development methodology.
Resumo:
A patient classification system was developed integrating a patient acuity instrument with a computerized nursing distribution method based on a linear programming model. The system was designed for real-time measurement of patient acuity (workload) and allocation of nursing personnel to optimize the utilization of resources.^ The acuity instrument was a prototype tool with eight categories of patients defined by patient severity and nursing intensity parameters. From this tool, the demand for nursing care was defined in patient points with one point equal to one hour of RN time. Validity and reliability of the instrument was determined as follows: (1) Content validity by a panel of expert nurses; (2) predictive validity through a paired t-test analysis of preshift and postshift categorization of patients; (3) initial reliability by a one month pilot of the instrument in a practice setting; and (4) interrater reliability by the Kappa statistic.^ The nursing distribution system was a linear programming model using a branch and bound technique for obtaining integer solutions. The objective function was to minimize the total number of nursing personnel used by optimally assigning the staff to meet the acuity needs of the units. A penalty weight was used as a coefficient of the objective function variables to define priorities for allocation of staff.^ The demand constraints were requirements to meet the total acuity points needed for each unit and to have a minimum number of RNs on each unit. Supply constraints were: (1) total availability of each type of staff and the value of that staff member (value was determined relative to that type of staff's ability to perform the job function of an RN (i.e., value for eight hours RN = 8 points, LVN = 6 points); (2) number of personnel available for floating between units.^ The capability of the model to assign staff quantitatively and qualitatively equal to the manual method was established by a thirty day comparison. Sensitivity testing demonstrated appropriate adjustment of the optimal solution to changes in penalty coefficients in the objective function and to acuity totals in the demand constraints.^ Further investigation of the model documented: correct adjustment of assignments in response to staff value changes; and cost minimization by an addition of a dollar coefficient to the objective function. ^