984 resultados para Experimental validation


Relevância:

60.00% 60.00%

Publicador:

Resumo:

Abstract Due to recent scientific and technological advances in information sys¬tems, it is now possible to perform almost every application on a mobile device. The need to make sense of such devices more intelligent opens an opportunity to design data mining algorithm that are able to autonomous execute in local devices to provide the device with knowledge. The problem behind autonomous mining deals with the proper configuration of the algorithm to produce the most appropriate results. Contextual information together with resource information of the device have a strong impact on both the feasibility of a particu¬lar execution and on the production of the proper patterns. On the other hand, performance of the algorithm expressed in terms of efficacy and efficiency highly depends on the features of the dataset to be analyzed together with values of the parameters of a particular implementation of an algorithm. However, few existing approaches deal with autonomous configuration of data mining algorithms and in any case they do not deal with contextual or resources information. Both issues are of particular significance, in particular for social net¬works application. In fact, the widespread use of social networks and consequently the amount of information shared have made the need of modeling context in social application a priority. Also the resource consumption has a crucial role in such platforms as the users are using social networks mainly on their mobile devices. This PhD thesis addresses the aforementioned open issues, focusing on i) Analyzing the behavior of algorithms, ii) mapping contextual and resources information to find the most appropriate configuration iii) applying the model for the case of a social recommender. Four main contributions are presented: - The EE-Model: is able to predict the behavior of a data mining algorithm in terms of resource consumed and accuracy of the mining model it will obtain. - The SC-Mapper: maps a situation defined by the context and resource state to a data mining configuration. - SOMAR: is a social activity (event and informal ongoings) recommender for mobile devices. - D-SOMAR: is an evolution of SOMAR which incorporates the configurator in order to provide updated recommendations. Finally, the experimental validation of the proposed contributions using synthetic and real datasets allows us to achieve the objectives and answer the research questions proposed for this dissertation.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Laser material processing is being extensively used in photovoltaic applications for both the fabrication of thin film modules and the enhancement of the crystalline silicon solar cells. The two temperature model for thermal diffusion was numerically solved in this paper. Laser pulses of 1064, 532 or 248 nm with duration of 35, 26 or 10 ns were considered as the thermal source leading to the material ablation. Considering high irradiance levels (108–109 W cm−2), a total absorption of the energy during the ablation process was assumed in the model. The materials analysed in the simulation were aluminium (Al) and silver (Ag), which are commonly used as metallic electrodes in photovoltaic devices. Moreover, thermal diffusion was also simulated for crystalline silicon (c-Si). A similar trend of temperature as a function of depth and time was found for both metals and c-Si regardless of the employed wavelength. For each material, the ablation depth dependence on laser pulse parameters was determined by means of an ablation criterion. Thus, after the laser pulse, the maximum depth for which the total energy stored in the material is equal to the vaporisation enthalpy was considered as the ablation depth. For all cases, the ablation depth increased with the laser pulse fluence and did not exhibit a clear correlation with the radiation wavelength. Finally, the experimental validation of the simulation results was carried out and the ability of the model with the initial hypothesis of total energy absorption to closely fit experimental results was confirmed.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Ubiquitous computing software needs to be autonomous so that essential decisions such as how to configure its particular execution are self-determined. Moreover, data mining serves an important role for ubiquitous computing by providing intelligence to several types of ubiquitous computing applications. Thus, automating ubiquitous data mining is also crucial. We focus on the problem of automatically configuring the execution of a ubiquitous data mining algorithm. In our solution, we generate configuration decisions in a resource aware and context aware manner since the algorithm executes in an environment in which the context often changes and computing resources are often severely limited. We propose to analyze the execution behavior of the data mining algorithm by mining its past executions. By doing so, we discover the effects of resource and context states as well as parameter settings on the data mining quality. We argue that a classification model is appropriate for predicting the behavior of an algorithm?s execution and we concentrate on decision tree classifier. We also define taxonomy on data mining quality so that tradeoff between prediction accuracy and classification specificity of each behavior model that classifies by a different abstraction of quality, is scored for model selection. Behavior model constituents and class label transformations are formally defined and experimental validation of the proposed approach is also performed.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This work presents a single stage converter for a high bandwidth and a high efficiency envelope amplifier. The current ripple cancellation technique is applied to a synchronous buck converter to cancel the output current ripple and to decrease the switching frequency without a reduction in the large signal bandwidth. The converter is modeled and the new design with ripple cancellation circuit is detailed. The advantages of the proposed design are presented and validated experimentally. The transfer function of the output filter of the buck converter with ripple cancellation circuit has been modeled and compared to measurements, showing a good correspondence. Experimental validation is provided at 4MHz of switching frequency for DC and variable output voltage for a sinusoidal and a 64QAM signal. Additional experimental validation of the efficiency improvement is provided, compared to the equivalent design (same bandwidth and output voltage ripple) of the conventional buck converter.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Los conjuntos bacterianos son sistemas dinámicos difíciles de modelar debido a que las bacterias colaboran e intercambian información entre sí. Estos microorganismos procariotas pueden tomar decisiones por mayoría e intercambiar información genética importante que, por ejemplo, las haga resistentes a un antibiótico. El proceso de conjugación consiste en el intercambio de un plásmido de una bacteria con otra, permitiendo así que se transfieran propiedades. Estudios recientes han demostrado que estos plásmidos pueden ser reprogramados artificialmente para que la bacteria que lo contenga realice una función específica [1]. Entre la multitud de aplicaciones que supone esta idea, el proyecto europeo PLASWIRES está intentando demostrar que es posible usar organismos vivos como computadores distribuidos en paralelo y plásmidos como conexión entre ellos mediante conjugación. Por tanto, mediante una correcta programación de un plásmido, se puede conseguir, por ejemplo, hacer que una colonia de bacterias haga la función de un antibiótico o detecte otros plásmidos peligrosos en bacterias virulentas. El proceso experimental para demostrar esta idea puede llegar a ser algo lento y tedioso, por lo que es necesario el uso de simuladores que predigan su comportamiento. Debido a que el proyecto PLASWIRES se basa en la conjugación bacteriana, surge la necesidad de un simulador que reproduzca esta operación. El presente trabajo surge debido a la deficiencia del simulador GRO para reproducir la conjugación. En este documento se detallan las modificaciones necesarias para que GRO pueda representar este proceso, así como analizar los datos obtenidos e intentar ajustar el modelo a los datos obtenidos por el Instituto de Biomedicina y Biotecnología de Cantabria (IBBTEC). ---ABSTRACT---Bacterial colonies are dynamical systems difficult to model because bacteria collaborate and exchange information with each other. These prokaryotic organisms can make decisions by majority and exchange important genetic information, for example, make them resistant to an antibiotic. The conjugation process is the exchange of a plasmid from one bacterium to another, allowing both to have the same properties. Recent studies have shown that these plasmids can be artificially reprogrammed to make the bacteria that contain it to perform a specific function [1]. Among the multitude of applications involved in this idea, the European project PLASWIRES is attempting to prove that it is possible to use living organisms as parallel and distributed computers with plasmids acting as connectors between them through conjugation. Thus, by properly programming a plasmid, you can get a colony of bacteria that work as an antibiotic or detect hazardous plasmids in virulent bacteria. The experimental process to prove this idea can be slow and tedious, so the use of simulators to predict their behavior is required. Since PLASWIRES project is based on bacterial conjugation, a simulator that can reproduce this operation is required. This work arises due to the absence of the conjugation process in the simulator GRO. This document details the changes made to GRO to represent this process, analyze the data and try to adjust the model to the data obtained by the Institute of Biomedicine and Biotechnology of Cantabria ( IBBTEC ). This project has two main objectives, the first is to add the functionality of intercellular communication by conjugation to the simulator GRO, and the second is to use the experimental data obtained by the IBBTEC. To do this, the following points should be followed: • Study of conjugation biology as a mechanism of intercellular communication. • Design and implementation of the algorithm that simulates conjugation. • Experimental validation and model adjust to the experimental data on rates of conjugation and bacterial growth.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

La calidad del hormigón prefabricado se determina mediante ensayos de rotura a compresión en probetas transcurridos los 28 días de curado, según establece la EHE-08. Sin embargo, en la plantas de prefabricados es necesario además saber cuándo el hormigón está listo para ser procesado (destensado, cortado, trasladado), por lo que es necesario hacer ensayos de resistencia a la compresión entre las 48 y 72 horas, este tiempo se determina a partir de la experiencia previa adquirida y depende de las condiciones de cada planta. Si las probetas no han alcanzado el valor establecido, normalmente debido a un cambio en las condiciones climatológicas o en los materiales utilizados como el tipo de cemento o agregados, la solución adoptada suele ser dejar curar el material más horas en la pista para que alcance la resistencia necesaria para ser procesado. Si sigue sin alcanzarla, lo cual sucede muy ocasionalmente, se intenta analizar cuál ha sido el motivo, pudiéndose tirar toda la producción de ese día si se comprueba que ha sido un fallo en la fabricación de la línea, y no un fallo de la probeta. Por tanto, esta metodología de control de calidad, basada en técnicas destructivas, supone dos tipos de problemas, costes y representatividad. Los métodos no destructivos que más se han aplicado para caracterizar el proceso de curado del hormigón son los ultrasónicos y la medida de la temperatura como se recoge en la bibliografía consultada. Hay diferentes modelos que permiten establecer una relación entre la temperatura y el tiempo de curado para estimar la resistencia a compresión del material, y entre la velocidad de propagación ultrasónica y la resistencia. Aunque estas relaciones no son generales, se han obtenido muy buenos resultados, ejemplo de ello es el modelo basado en la temperatura, Maturity Method, que forma parte de la norma de la ASTM C 1074 y en el mercado hay disponibles equipos comerciales (maturity meters) para medir el curado del hormigón. Además, es posible diseñar sistemas de medida de estos dos parámetros económicos y robustos; por lo cual es viable la realización de una metodología para el control de calidad del curado que pueda ser implantado en las plantas de producción de prefabricado. En este trabajo se ha desarrollado una metodología que permite estimar la resistencia a la compresión del hormigón durante el curado, la cual consta de un procedimiento para el control de calidad del prefabricado y un sistema inalámbrico de sensores para la medida de la temperatura y la velocidad ultrasónica. El procedimiento para el control de calidad permite realizar una predicción de la resistencia a compresión a partir de un modelo basado en la temperatura de curado y otros dos basados en la velocidad, método de tiempo equivalente y método lineal. El sistema inalámbrico de sensores desarrollado, WilTempUS, integra en el mismo dispositivo sensores de temperatura, humedad relativa y ultrasonidos. La validación experimental se ha realizado mediante monitorizaciones en probetas y en las líneas de prefabricados. Los resultados obtenidos con los modelos de estimación y el sistema de medida desarrollado muestran que es posible predecir la resistencia en prefabricados de hormigón en planta con errores comparables a los aceptables por norma en los ensayos de resistencia a compresión en probetas. ABSTRACT Precast concrete quality is determined by compression tests breakage on specimens after 28 days of curing, as established EHE-08. However, in the precast plants is also necessary to know when the concrete is ready to be processed (slack, cut, moved), so it is necessary to test the compressive strength between 48 and 72 hours. This time is determined from prior experience and depends on the conditions of each plant. If the samples have not reached the set value, usually due to changes in the weather conditions or in the materials used as for example the type of cement or aggregates, the solution usually adopted is to cure the material on track during more time to reach the required strength for processing. If the material still does not reach this strength, which happens very occasionally, the reason of this behavior is analyzed , being able to throw the entire production of that day if there was a failure in the manufacturing line, not a failure of the specimen. Therefore, this method of quality control, using destructive techniques, involves two kinds of problems, costs and representativeness. The most used non-destructive methods to characterize the curing process of concrete are those based on ultrasonic and temperature measurement as stated in the literature. There are different models to establish a relationship between temperature and the curing time to estimate the compressive strength of the material, and between the ultrasonic propagation velocity and the compressive strength. Although these relationships are not general, they have been very successful, for example the Maturity Method is based on the temperature measurements. This method is part of the standards established in ASTM C 1074 and there are commercial equipments available (maturity meters) in the market to measure the concrete curing. Furthermore, it is possible to design inexpensive and robust systems to measure ultrasounds and temperature. Therefore is feasible to determine a method for quality control of curing to be implanted in the precast production plants. In this work, it has been developed a methodology which allows to estimate the compressive strength of concrete during its curing process. This methodology consists of a procedure for quality control of the precast concrete and a wireless sensor network to measure the temperature and ultrasonic velocity. The procedure for quality control allows to predict the compressive strength using a model based on the curing temperature and two other models based on ultrasonic velocity, the equivalent time method and the lineal one. The wireless sensor network, WilTempUS, integrates is the same device temperature, relative humidity and ultrasonic sensors. The experimental validation has been carried out in cubic specimens and in the production plants. The results obtained with the estimation models and the measurement system developed in this thesis show that it is possible to predict the strength in precast concrete plants with errors within the limits of the standards for testing compressive strength specimens.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

En el campo de la fusión nuclear y desarrollándose en paralelo a ITER (International Thermonuclear Experimental Reactor), el proyecto IFMIF (International Fusion Material Irradiation Facility) se enmarca dentro de las actividades complementarias encaminadas a solucionar las barreras tecnológicas que aún plantea la fusión. En concreto IFMIF es una instalación de irradiación cuya misión es caracterizar materiales resistentes a condiciones extremas como las esperadas en los futuros reactores de fusión como DEMO (DEMOnstration power plant). Consiste de dos aceleradores de deuterones que proporcionan un haz de 125 mA y 40 MeV cada uno, que al colisionar con un blanco de litio producen un flujo neutrónico intenso (1017 neutrones/s) con un espectro similar al de los neutrones de fusión [1], [2]. Dicho flujo neutrónico es empleado para irradiar los diferentes materiales candidatos a ser empleados en reactores de fusión, y las muestras son posteriormente examinadas en la llamada instalación de post-irradiación. Como primer paso en tan ambicioso proyecto, una fase de validación y diseño llamada IFMIFEVEDA (Engineering Validation and Engineering Design Activities) se encuentra actualmente en desarrollo. Una de las actividades contempladas en esta fase es la construcción y operación de una acelarador prototipo llamado LIPAc (Linear IFMIF Prototype Accelerator). Se trata de un acelerador de deuterones de alta intensidad idéntico a la parte de baja energía de los aceleradores de IFMIF. Los componentes del LIPAc, que será instalado en Japón, son suministrados por diferentes países europeos. El acelerador proporcionará un haz continuo de deuterones de 9 MeV con una potencia de 1.125 MW que tras ser caracterizado con diversos instrumentos deberá pararse de forma segura. Para ello se requiere un sistema denominado bloque de parada (Beam Dump en inglés) que absorba la energía del haz y la transfiera a un sumidero de calor. España tiene el compromiso de suministrar este componente y CIEMAT (Centro de Investigaciones Energéticas Medioambientales y Tecnológicas) es responsable de dicha tarea. La pieza central del bloque de parada, donde se para el haz de iones, es un cono de cobre con un ángulo de 3.5o, 2.5 m de longitud y 5 mm de espesor. Dicha pieza está refrigerada por agua que fluye en su superficie externa por el canal que se forma entre el cono de cobre y otra pieza concéntrica con éste. Este es el marco en que se desarrolla la presente tesis, cuyo objeto es el diseño del sistema de refrigeración del bloque de parada del LIPAc. El diseño se ha realizado utilizando un modelo simplificado unidimensional. Se han obtenido los parámetros del agua (presión, caudal, pérdida de carga) y la geometría requerida en el canal de refrigeración (anchura, rugosidad) para garantizar la correcta refrigeración del bloque de parada. Se ha comprobado que el diseño permite variaciones del haz respecto a la situación nominal siendo el flujo crítico calorífico al menos 2 veces superior al nominal. Se han realizado asimismo simulaciones fluidodinámicas 3D con ANSYS-CFX en aquellas zonas del canal de refrigeración que lo requieren. El bloque de parada se activará como consecuencia de la interacción del haz de partículas lo que impide cualquier cambio o reparación una vez comenzada la operación del acelerador. Por ello el diseño ha de ser muy robusto y todas las hipótesis utilizadas en la realización de éste deben ser cuidadosamente comprobadas. Gran parte del esfuerzo de la tesis se centra en la estimación del coeficiente de transferencia de calor que es determinante en los resultados obtenidos, y que se emplea además como condición de contorno en los cálculos mecánicos. Para ello por un lado se han buscado correlaciones cuyo rango de aplicabilidad sea adecuado para las condiciones del bloque de parada (canal anular, diferencias de temperatura agua-pared de decenas de grados). En un segundo paso se han comparado los coeficientes de película obtenidos a partir de la correlación seleccionada (Petukhov-Gnielinski) con los que se deducen de simulaciones fluidodinámicas, obteniendo resultados satisfactorios. Por último se ha realizado una validación experimental utilizando un prototipo y un circuito hidráulico que proporciona un flujo de agua con los parámetros requeridos en el bloque de parada. Tras varios intentos y mejoras en el experimento se han obtenido los coeficientes de película para distintos caudales y potencias de calentamiento. Teniendo en cuenta la incertidumbre de las medidas, los valores experimentales concuerdan razonablemente bien (en el rango de 15%) con los deducidos de las correlaciones. Por motivos radiológicos es necesario controlar la calidad del agua de refrigeración y minimizar la corrosión del cobre. Tras un estudio bibliográfico se identificaron los parámetros del agua más adecuados (conductividad, pH y concentración de oxígeno disuelto). Como parte de la tesis se ha realizado asimismo un estudio de la corrosión del circuito de refrigeración del bloque de parada con el doble fin de determinar si puede poner en riesgo la integridad del componente, y de obtener una estimación de la velocidad de corrosión para dimensionar el sistema de purificación del agua. Se ha utilizado el código TRACT (TRansport and ACTivation code) adaptándalo al caso del bloque de parada, para lo cual se trabajó con el responsable (Panos Karditsas) del código en Culham (UKAEA). Los resultados confirman que la corrosión del cobre en las condiciones seleccionadas no supone un problema. La Tesis se encuentra estructurada de la siguiente manera: En el primer capítulo se realiza una introducción de los proyectos IFMIF y LIPAc dentro de los cuales se enmarca esta Tesis. Además se describe el bloque de parada, siendo el diseño del sistema de rerigeración de éste el principal objetivo de la Tesis. En el segundo y tercer capítulo se realiza un resumen de la base teórica así como de las diferentes herramientas empleadas en el diseño del sistema de refrigeración. El capítulo cuarto presenta los resultados del relativos al sistema de refrigeración. Tanto los obtenidos del estudio unidimensional, como los obtenidos de las simulaciones fluidodinámicas 3D mediante el empleo del código ANSYS-CFX. En el quinto capítulo se presentan los resultados referentes al análisis de corrosión del circuito de refrigeración del bloque de parada. El capítulo seis se centra en la descripción del montaje experimental para la obtención de los valores de pérdida de carga y coeficiente de transferencia del calor. Asimismo se presentan los resultados obtenidos en dichos experimentos. Finalmente encontramos un capítulo de apéndices en el que se describen una serie de experimentos llevados a cabo como pasos intermedios en la obtención del resultado experimental del coeficiente de película. También se presenta el código informático empleado para el análisis unidimensional del sistema de refrigeración del bloque de parada llamado CHICA (Cooling and Heating Interaction and Corrosion Analysis). ABSTRACT In the nuclear fusion field running in parallel to ITER (International Thermonuclear Experimental Reactor) as one of the complementary activities headed towards solving the technological barriers, IFMIF (International Fusion Material Irradiation Facility) project aims to provide an irradiation facility to qualify advanced materials resistant to extreme conditions like the ones expected in future fusion reactors like DEMO (DEMOnstration Power Plant). IFMIF consists of two constant wave deuteron accelerators delivering a 125 mA and 40 MeV beam each that will collide on a lithium target producing an intense neutron fluence (1017 neutrons/s) with a similar spectra to that of fusion neutrons [1], [2]. This neutron flux is employed to irradiate the different material candidates to be employed in the future fusion reactors, and the samples examined after irradiation at the so called post-irradiative facilities. As a first step in such an ambitious project, an engineering validation and engineering design activity phase called IFMIF-EVEDA (Engineering Validation and Engineering Design Activities) is presently going on. One of the activities consists on the construction and operation of an accelerator prototype named LIPAc (Linear IFMIF Prototype Accelerator). It is a high intensity deuteron accelerator identical to the low energy part of the IFMIF accelerators. The LIPAc components, which will be installed in Japan, are delivered by different european countries. The accelerator supplies a 9 MeV constant wave beam of deuterons with a power of 1.125 MW, which after being characterized by different instruments has to be stopped safely. For such task a beam dump to absorb the beam energy and take it to a heat sink is needed. Spain has the compromise of delivering such device and CIEMAT (Centro de Investigaciones Energéticas Medioambientales y Tecnológicas) is responsible for such task. The central piece of the beam dump, where the ion beam is stopped, is a copper cone with an angle of 3.5o, 2.5 m long and 5 mm width. This part is cooled by water flowing on its external surface through the channel formed between the copper cone and a concentric piece with the latter. The thesis is developed in this realm, and its objective is designing the LIPAc beam dump cooling system. The design has been performed employing a simplified one dimensional model. The water parameters (pressure, flow, pressure loss) and the required annular channel geometry (width, rugoisty) have been obtained guaranteeing the correct cooling of the beam dump. It has been checked that the cooling design allows variations of the the beam with respect to the nominal position, being the CHF (Critical Heat Flux) at least twice times higher than the nominal deposited heat flux. 3D fluid dynamic simulations employing ANSYS-CFX code in the beam dump cooling channel sections which require a more thorough study have also been performed. The beam dump will activateasaconsequenceofthe deuteron beam interaction, making impossible any change or maintenance task once the accelerator operation has started. Hence the design has to be very robust and all the hypotheses employed in the design mustbecarefully checked. Most of the work in the thesis is concentrated in estimating the heat transfer coefficient which is decisive in the obtained results, and is also employed as boundary condition in the mechanical analysis. For such task, correlations which applicability range is the adequate for the beam dump conditions (annular channel, water-surface temperature differences of tens of degrees) have been compiled. In a second step the heat transfer coefficients obtained from the selected correlation (Petukhov- Gnielinski) have been compared with the ones deduced from the 3D fluid dynamic simulations, obtaining satisfactory results. Finally an experimental validation has been performed employing a prototype and a hydraulic circuit that supplies a flow with the requested parameters in the beam dump. After several tries and improvements in the experiment, the heat transfer coefficients for different flows and heating powers have been obtained. Considering the uncertainty in the measurements the experimental values agree reasonably well (in the order of 15%) with the ones obtained from the correlations. Due to radiological reasons the quality of the cooling water must be controlled, hence minimizing the copper corrosion. After performing a bibligraphic study the most adequate water parameters were identified (conductivity, pH and dissolved oxygen concentration). As part of this thesis a corrosion study of the beam dump cooling circuit has been performed with the double aim of determining if corrosion can pose a risk for the copper beam dump , and obtaining an estimation of the corrosion velocitytodimension the water purification system. TRACT code(TRansport and ACTivation) has been employed for such study adapting the code for the beam dump case. For such study a collaboration with the code responsible (Panos Karditsas) at Culham (UKAEA) was established. The work developed in this thesis has supposed the publication of three articles in JCR journals (”Journal of Nuclear Materials” y ”Fusion Engineering and Design”), as well as presentations in more than four conferences and relevant meetings.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Recent disasters have shown that having clearly defined preventive procedures and decisions is a critical component that minimizes evacuation hazards and ensures a rapid and successful evolution of evacuation plans. In this context, we present our Situation-Aware System for enhancing Evacuation Plans (SASEP) system, which allows creating end-user business rules that technically support the specific events, conditions and actions related to evacuation plans. An experimental validation was carried out where 32 people faced a simulated emergency situation, 16 of them using SASEP and the other 16 using a legacy system based on static signs. From the results obtained, we compare both techniques and discuss in which situations SASEP offers a better evacuation route option, confirming that it is highly valuable when there is a threat in the evacuation route. In addition, a study about user satisfaction using both systems is presented showing in which cases the systems are assessed as satisfactory, relevant and not frustrating.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

En esta tesis se desarrolla una metodología alternativa para la determinación de la dureza Brinell a partir de imágenes obtenidas mediante microscopía confocal, que se ha mostrado robusta para mejorar los resultados de medición del diámetro en condiciones de reproducibilidad. Las validaciones realizadas evidencian su posibilidad real de implementación, especialmente para la certificación de patrones de dureza. Los estudios experimentales realizados ponen de manifiesto que la medición del diámetro de una huella de dureza Brinell, siguiendo la metodología tradicional, depende de la posición del patrón, de las características del equipo empleado y del propio operador. Dicha medida resulta crítica y las dificultades para identificar el borde de la huella incorporan a menudo una fuente adicional de incertidumbre difícil de soslayar. En esta investigación se han desarrollado dos modelos matemáticos que permiten identificar de forma unívoca el diámetro de la huella en el punto donde se produce el límite de contacto entre el indentador y el material de la probeta durante la realización del ensayo. Ambos modelos han sido implementados en Matlab® y se ha verificado su validez mediante datos sintéticos. Asimismo, se ha realizado una validación experimental sobre patrones de dureza certificados, empleando un microscopio confocal marca Leica, modelo DCM 3D disponible en el Laboratorio de Investigación de Materiales de Interés Tecnológico (LIMIT) de la Escuela Técnica Superior de Ingeniería y Diseño Industrial de la Universidad Politécnica de Madrid (ETSIDI – UPM). Dicha validación ha puesto de manifiesto la utilidad de esta nueva metodología por cuanto permite caracterizar las huellas, estimar las incertidumbres de medida y garantizar la trazabilidad metrológica de los resultados. ABSTRACT This PhD thesis presents an alternative methodology to determine the Brinell hardness from the images obtained by confocal microscopy that has proved to be robust to improve the results of indentation diameter measurements in reproducibility conditions. The validations carried out show the real possibility of its implementation, especially for calibration of hardness reference blocks. Experimental studies performed worldwide show that the measurement of the indentation diameter in a Brinell hardness test depends, when the traditional methodology is applied, on the position of the test block, the equipment characteristics and the operator. This measurement is critical and the difficulties to identify the edge of the indentation often bring an additional source of uncertainty with them that is hard to avoid. In this research two specific mathematical models have been developed to identify unambiguously the indentation diameter at the point where the edge of the boundary between the indenter and the test block is found during the test. Both models have been implemented on Matlab® and their validity has been verified by synthetic data An additional experimental validation with calibrated hardness reference blocks has been carried out using a Leica-brand confocal microscope, model DCM 3D, available in the Laboratory for Research on Materials of Technological Interest (LIMIT in its Spanish acronym) of the Escuela Técnica Superior de Ingeniería y Diseño Industrial de la Universidad Politécnica de Madrid (ETSIDI-UPM). This validation has shown the utility of this new methodology since it allows to characterize the indentation, to estimate the measurement uncertainties and to ensure the metrological traceability of the results.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Stream-mining approach is defined as a set of cutting-edge techniques designed to process streams of data in real time, in order to extract knowledge. In the particular case of classification, stream-mining has to adapt its behaviour to the volatile underlying data distributions, what has been called concept drift. Moreover, it is important to note that concept drift may lead to situations where predictive models become invalid and have therefore to be updated to represent the actual concepts that data poses. In this context, there is a specific type of concept drift, known as recurrent concept drift, where the concepts represented by data have already appeared in the past. In those cases the learning process could be saved or at least minimized by applying a previously trained model. This could be extremely useful in ubiquitous environments that are characterized by the existence of resource constrained devices. To deal with the aforementioned scenario, meta-models can be used in the process of enhancing the drift detection mechanisms used by data stream algorithms, by representing and predicting when the change will occur. There are some real-world situations where a concept reappears, as in the case of intrusion detection systems (IDS), where the same incidents or an adaptation of them usually reappear over time. In these environments the early prediction of drift by means of a better knowledge of past models can help to anticipate to the change, thus improving efficiency of the model regarding the training instances needed. By means of using meta-models as a recurrent drift detection mechanism, the ability to share concepts representations among different data mining processes is open. That kind of exchanges could improve the accuracy of the resultant local model as such model may benefit from patterns similar to the local concept that were observed in other scenarios, but not yet locally. This would also improve the efficiency of training instances used during the classification process, as long as the exchange of models would aid in the application of already trained recurrent models, that have been previously seen by any of the collaborative devices. Which it is to say that the scope of recurrence detection and representation is broaden. In fact the detection, representation and exchange of concept drift patterns would be extremely useful for the law enforcement activities fighting against cyber crime. Being the information exchange one of the main pillars of cooperation, national units would benefit from the experience and knowledge gained by third parties. Moreover, in the specific scope of critical infrastructures protection it is crucial to count with information exchange mechanisms, both from a strategical and technical scope. The exchange of concept drift detection schemes in cyber security environments would aid in the process of preventing, detecting and effectively responding to threads in cyber space. Furthermore, as a complement of meta-models, a mechanism to assess the similarity between classification models is also needed when dealing with recurrent concepts. In this context, when reusing a previously trained model a rough comparison between concepts is usually made, applying boolean logic. The introduction of fuzzy logic comparisons between models could lead to a better efficient reuse of previously seen concepts, by applying not just equal models, but also similar ones. This work faces the aforementioned open issues by means of: the MMPRec system, that integrates a meta-model mechanism and a fuzzy similarity function; a collaborative environment to share meta-models between different devices; a recurrent drift generator that allows to test the usefulness of recurrent drift systems, as it is the case of MMPRec. Moreover, this thesis presents an experimental validation of the proposed contributions using synthetic and real datasets.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Thesis (Ph.D.)--University of Washington, 2016-06

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Detection of a fatigue crack in a welded frame structure is studied in this paper using coupled response measurements. Similarity to real engineering structures is maintained in the fabrication of the test frame with hollow section chords and branch members. The fatigue crack was created by a special reciprocating mechanism that generates cyclic stress on a beam member of the structure. The methodology of coupled response measurements is first demonstrated on a single hollow section beam by analytical simulation and experimental validation. The issues of using this approach for fatigue crack detection in real structures are then examined. Finally, the experimental results of the frame under different scenarioes are presented. The existence of the crack is clearly observable from the FRF plots. It is suggested that this approach offers the potential to detect cracks in welded frame structures and is a useful tool for routine maintenance work and health assessment.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

In this work, a working model is proposed of molecular sieve silica (MSS) multistage membrane systems for CO cleanup at high temperatures (up to 500 degrees C) in a simulated fuel cell fuel processing system. Gases are described as having little interactions with each other relative to the pore walls due to low isosteric heat of adsorption on silica surfaces and high temperatures. The Arrhenius function for activated transport of pure gases was used to predict mixture concentration in the permeate and retentate streams. Simulation predicted CO could be reduced to levels below the required 50 ppmv for polymer electrolyte membrane fuel cell anodes at a stage H-2/CO selectivity of higher than 40 in 4 series membrane units. Experimental validation showed predicting mixture concentrations required only pure gas permeation data. This model has significant application for setting industrial stretch targets and as a robust basis for complex membrane model configurations. (c) 2006 American Institute of Chemical Engineers.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Non- protein- coding RNAs ( ncRNAs) are increasingly being recognized as having important regulatory roles. Although much recent attention has focused on tiny 22- to 25- nucleotide microRNAs, several functional ncRNAs are orders of magnitude larger in size. Examples of such macro ncRNAs include Xist and Air, which in mouse are 18 and 108 kilobases ( Kb), respectively. We surveyed the 102,801 FANTOM3 mouse cDNA clones and found that Air and Xist were present not as single, full- length transcripts but as a cluster of multiple, shorter cDNAs, which were unspliced, had little coding potential, and were most likely primed from internal adenine- rich regions within longer parental transcripts. We therefore conducted a genome- wide search for regional clusters of such cDNAs to find novel macro ncRNA candidates. Sixty- six regions were identified, each of which mapped outside known protein- coding loci and which had a mean length of 92 Kb. We detected several known long ncRNAs within these regions, supporting the basic rationale of our approach. In silico analysis showed that many regions had evidence of imprinting and/ or antisense transcription. These regions were significantly associated with microRNAs and transcripts from the central nervous system. We selected eight novel regions for experimental validation by northern blot and RT- PCR and found that the majority represent previously unrecognized noncoding transcripts that are at least 10 Kb in size and predominantly localized in the nucleus. Taken together, the data not only identify multiple new ncRNAs but also suggest the existence of many more macro ncRNAs like Xist and Air.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Three-tiered kinase modules, such as the Raf-MEK (mitogen-activated or extracellular signal-regulated protein kinase kinase)-ERK (extracellular signal-regulated kinase) mitogen-activated protein kinase pathway, are widespread in biology, suggesting that this structure conveys evolutionarily advantageous properties. We show that the three-tiered kinase amplifier module combined with negative feedback recapitulates the design principles of a negative feedback amplifier (NFA), which is used in electronic circuits to confer robustness, output stabilization, and linearization of nonlinear signal amplification. We used mathematical modeling and experimental validation to demonstrate that the ERK pathway has properties of an NFA that (i) converts intrinsic switch-like activation kinetics into graded linear responses, (ii) conveys robustness to changes in rates of reactions within the NFA module, and (iii) stabilizes outputs in response to drug-induced perturbations of the amplifier. These properties determine biological behavior, including activation kinetics and the response to drugs.