26 resultados para Multilevel thresholding
em Universidad Politécnica de Madrid
Resumo:
Soil erosion is a complex phenomenon involving the detachment and transport of soil particles, storage and runoff of rainwater, and infiltration. The relative magnitude and importance of these processes depends on several factors being one of them surface micro-topography, usually quanti[U+FB01]ed trough soil surface roughness (SSR). SSR greatly affects surface sealing and runoff generation, yet little information is available about the effect of roughness on the spatial distribution of runoff and on flow concentration. The methods commonly used to measure SSR involve measuring point elevation using a pin roughness meter or laser, both of which are labor intensive and expensive. Lately a simple and inexpensive technique based on percentage of shadow in soil surface image has been developed to determine SSR in the field in order to obtain measurement for wide spread application. One of the first steps in this technique is image de-noising and thresholding to estimate the percentage of black pixels in the studied area. In this work, a series of soil surface images have been analyzed applying several de-noising wavelet analysis and thresholding algorithms to study the variation in percentage of shadows and the shadows size distribution
Resumo:
Landcover is subject to continuous changes on a wide variety of temporal and spatial scales. Those changes produce significant effects in human and natural activities. Maintaining an updated spatial database with the occurred changes allows a better monitoring of the Earth?s resources and management of the environment. Change detection (CD) techniques using images from different sensors, such as satellite imagery, aerial photographs, etc., have proven to be suitable and secure data sources from which updated information can be extracted efficiently, so that changes can also be inventoried and monitored. In this paper, a multisource CD methodology for multiresolution datasets is applied. First, different change indices are processed, then different thresholding algorithms for change/no_change are applied to these indices in order to better estimate the statistical parameters of these categories, finally the indices are integrated into a change detection multisource fusion process, which allows generating a single CD result from several combination of indices. This methodology has been applied to datasets with different spectral and spatial resolution properties. Then, the obtained results are evaluated by means of a quality control analysis, as well as with complementary graphical representations. The suggested methodology has also been proved efficiently for identifying the change detection index with the higher contribution.
Resumo:
We have determined the cross-section σ for color center generation under single Br ion impacts on amorphous SiO2. The evolution of the cross-sections, σ(E) and σ(Se), show an initial flat stage that we associate to atomic collision mechanisms. Above a certain threshold value (Se > 2 keV/nm), roughly coinciding with that reported for the onset of macroscopic disorder (compaction), σ shows a marked increase due to electronic processes. In this regime, a energetic cost of around 7.5 keV is necessary to create a non bridging oxygen hole center-E′ (NBOHC/E′) pair, whatever the input energy. The data appear consistent with a non-radiative decay of self-trapped excitons.
Resumo:
This paper presents an envelope amplifier solution for envelope elimination and restoration (EER), that consists of a series combination of a switch-mode power supply (SMPS), based on three-level voltage cells and a linear regulator. This cell topology offers several advantages over a previously presented envelope amplifier based on a different multilevel topology (two-level voltage cells). The topology of the multilevel converter affects to the whole design of the envelope amplifier and a comparison between both design alternatives regarding the size, complexity and the efficiency of the solution is done. Both envelope amplifier solutions have a bandwidth of 2 MHz with an instantaneous maximum power of 50 W. It is also analyzed the linearity of the three-level cell solution, with critical importance in the EER technique implementation. Additionally, considerations to optimize the design of the envelope amplifier and experimental comparison between both cell topologies are included.
Resumo:
Se evalúa con indicadores de gobernanza urbana la sostenibilidad de las formas de hacer ciudad hibrida compleja del gobierno de la gestión visible (GGV). Argumenta que el GGV hace ciudad para legitimarse por desempeño y fortalecer la gobernanza local, en un contexto de mutaciones múltiples y radicales que tienden a diluir y centralizar el poder local y fractalizar la ciudad, profundizando la segregación sociopolítica-territorial y la ingobernabilidad genética de la ciudad hibrida, poniendo en riesgo el Estado federal descentralizado, el derecho a la ciudad, al gobierno local y la gobernanza urbana y multinivel (hipótesis). La estrategia de evaluación de gobernanza innovadora (EEG+i) diseñada para evaluar la relación entre las formas de hacer ciudad hibrida (variables espaciales) y gobernanza (variable a-espacial) es transversal, multidimensional y se construye desde la complejidad, el análisis de escenarios, formulación de constructos, modelos e indicadores de gobernanza, entretejiendo tres campos de conocimiento, gobierno, ciudad y sostenibilidad, en cuatro fases. La Fase 1, contextualiza la gobernanza en la dramática del siglo XXI. La Fase 2, desarrolla la fundamentación teórico-práctica, nuevos conceptos y un abordaje analítico propio ‘genética territorial’, para analizar y comprehender la complejidad de la ciudad hibrida de países en desarrollo, tejiendo ontogenética territorial y el carácter autopoiético del gen informal. En la Fase 3, se caracterizan las formas de hacer ciudad desde la genética del territorio, se formulan modelos e indicadores de gobernanza con los que se evalúan, aplicando un delphi y cuestionarios, los genes tipológicos-formas de hacer ciudad y validan las conclusiones. En la Fase 4, se correlacionan los resultados de los instrumentos aplicados con la praxis urbana del GGV, durante cuatro periodos de gobierno (1996-2010). Concluyendo que, la estrategia de evaluación comprobó las hipótesis y demostró la correlación transversal y multinivel existente entre, las mutaciones en curso que contradicen el modelo de gobernanza constitucional, el paisaje de gobernanza latinoamericano y venezolano, la praxis de los regímenes híbridos ricos en recursos naturales, las perspectivas de desarrollo globales y se expresa sociopolíticamente en déficit de gobernanza, Estado de derecho y cohesión-capital social y, espaciolocalmente, en la ciudad hibrida dispersa y diluida (compleja) y en el gobierno del poder diluido centralizado. La confrontación de flujos de poder centrípetos y centrífugos en la ciudad profundiza la fragmentación socioespacial y política y el deterioro de la calidad de vida, incrementando las protestas ciudadanas e ingobernabilidad que obstaculiza la superación de la pobreza y gobernanza urbana y multinivel. La evaluación de la praxis urbana del GGV evidenció que la correlación entre gobernanza, la producción de genes formales y la ciudad por iniciativa privada tiende a ser positiva y entre gobernanza, genes y producción de ciudad informal negativa, por el carácter autopoiético-autogobernable del gen informal y de los nuevos gobiernos sublocales que dificulta gobernar en gobernanza. La praxis del GGV es contraria al modelo de gobernanza formulado y la disolución centralizada del gobierno local y de la ciudad hibrida-dispersa es socio-espacial y políticamente insostenible. Se proponen estrategias y tácticas de gobernanza multinivel para recuperar la cohesión social y de planificación de la gestión innovadora (EG [PG] +i) para orquestar, desde el Consejo Local de Gobernanza (CLG) y con la participación de los espacios y gobiernos sublocales, un proyecto de ciudad compartido y sostenible. ABSTRACT The sustainability of the forms of making the hybrid-complex city by the visible management government (VMG) is evaluated using urban governance indicators. Argues that the VMG builds city to legitimate itself by performance and to strengthen local governance in a context of multiple and radical mutations that tend to dilute and centralize local power and fractalize the city, deepening the socio-spatial and political segregation, the genetic ingovernability of the hybrid city and placing the decentralized federal State, the right to city, local government and urban governance at risk (hypothesis). The innovative governance evaluation strategy (GES+i) designed to assess the relationship between the forms of making the hybrid city (spatial variables) and governance (a-spatial variable) is transversal, multidimensional; is constructed from complexity, scenario analysis, the formulation of concepts, models and governance indicators, weaving three fields of knowledge, government, city and sustainability in four phases. Phase 1, contextualizes governance in the dramatic of the twenty-first century. Phase 2, develops the theoretical and practical foundations, new concepts and a proper analytical approach to comprehend the complexity of the hybrid city from developing countries, weaving territorial ontogenetic with the autopiethic character of the informal city gen. In Phase 3, the ways of making city are characterized from the genetics of territory; governance indicators and models are formulated to evaluate, using delphi and questionnaires, the ways of making city and validate the conclusions. In Phase 4, the results of the instruments applied are correlated with the urban praxis of the VMG during the four periods of government analyzed (1996-2010). Concluding that, the evaluation strategy proved the hypothesis and showed the transversal and multilevel correlation between, mutations that contradict the constitutional governance model, the governance landscape of Latinamerica and the country, the praxis of the hybrid regimes rich in natural resources, the perspectives of the glocal economy and expresses socio-politically the governance and rule of law and social capital-cohesion deficit and spatial-temporarily the hybrid disperse and diluted city (complex) and the diluted-centralized local government. The confrontation of flows of power centripetal and centrifugal in the city deepens the socio-spatial and political fragmentation and deterioration of the quality of life, increasing citizens' protests and ingovernability which hinders poverty eradication and, multilevel and urban governance. The evaluation of the VMG urban praxis showed the correlation between governance, the production of formal genes and city by private initiative tended to be positive and, between informal genes-city production and governance negative, due to its autopiethic-self governable character that hinders governance. The urban praxis of the VMG contradicts the formulated governance model and thecentralized dissolution of the local government and hybrid city are socio-spatial and politically unsustainable. Multiscale governance strategies are proposed to recreate social cohesion and a management planning innovative method (EG [PG] + i) to orchestrate, from the Local Governance Council (LGC) and with the participation of sublocal governments and spaces, a shared and sustainable city project.
Resumo:
The challenges regarding seamless integration of distributed, heterogeneous and multilevel data arising in the context of contemporary, post-genomic clinical trials cannot be effectively addressed with current methodologies. An urgent need exists to access data in a uniform manner, to share information among different clinical and research centers, and to store data in secure repositories assuring the privacy of patients. Advancing Clinico-Genomic Trials (ACGT) was a European Commission funded Integrated Project that aimed at providing tools and methods to enhance the efficiency of clinical trials in the -omics era. The project, now completed after four years of work, involved the development of both a set of methodological approaches as well as tools and services and its testing in the context of real-world clinico-genomic scenarios. This paper describes the main experiences using the ACGT platform and its tools within one such scenario and highlights the very promising results obtained.
Resumo:
Soil voids manifest the cumulative effect of local pedogenic processes and ultimately influence soil behavior - especially as it pertains to aeration and hydrophysical properties. Because of the relatively weak attenuation of X-rays by air, compared with liquids or solids, non-disruptive CT scanning has become a very attractive tool for generating three-dimensional imagery of soil voids. One of the main steps involved in this analysis is the thresholding required to transform the original (greyscale) images into the type of binary representation (e.g., pores in white, solids in black) needed for fractal analysis or simulation with Lattice?Boltzmann models (Baveye et al., 2010). The objective of the current work is to apply an innovative approach to quantifying soil voids and pore networks in original X-ray CT imagery using Relative Entropy (Bird et al., 2006; Tarquis et al., 2008). These will be illustrated using typical imagery representing contrasting soil structures. Particular attention will be given to the need to consider the full 3D context of the CT imagery, as well as scaling issues, in the application and interpretation of this index.
Resumo:
One important issue emerging strongly in agriculture is related with the automatization of tasks, where the optical sensors play an important role. They provide images that must be conveniently processed. The most relevantimage processing procedures require the identification of green plants, in our experiments they come from barley and corn crops including weeds, so that some types of action can be carried out, including site-specific treatments with chemical products or mechanical manipulations. Also the identification of textures belonging to the soil could be useful to know some variables, such as humidity, smoothness or any others. Finally, from the point of view of the autonomous robot navigation, where the robot is equipped with the imaging system, some times it is convenient to know not only the soil information and the plants growing in the soil but also additional information supplied by global references based on specific areas. This implies that the images to be processed contain textures of three main types to be identified: green plants, soil and sky if any. This paper proposes a new automatic approach for segmenting these main textures and also to refine the identification of sub-textures inside the main ones. Concerning the green identification, we propose a new approach that exploits the performance of existing strategies by combining them. The combination takes into account the relevance of the information provided by each strategy based on the intensity variability. This makes an important contribution. The combination of thresholding approaches, for segmenting the soil and the sky, makes the second contribution; finally the adjusting of the supervised fuzzy clustering approach for identifying sub-textures automatically, makes the third finding. The performance of the method allows to verify its viability for automatic tasks in agriculture based on image processing
Resumo:
El presente trabajo describe una nueva metodología para la detección automática del espacio glotal de imágenes laríngeas tomadas a partir de 15 vídeos grabados por el servicio ORL del hospital Gregorio Marañón de Madrid con luz estroboscópica. El sistema desarrollado está basado en el modelo de contornos activos (snake). El algoritmo combina en el pre-procesado, algunas técnicas tradicionales (umbralización y filtro de mediana) con técnicas más sofisticadas tales como filtrado anisotrópico. De esta forma, se obtiene una imagen apropiada para el uso de las snakes. El valor escogido para el umbral es del 85% del pico máximo del histograma de la imagen; sobre este valor la información de los píxeles no es relevante. El filtro anisotrópico permite distinguir dos niveles de intensidad, uno es el fondo y el otro es la glotis. La inicialización se basa en obtener el módulo del campo GVF; de esta manera se asegura un proceso automático para la selección del contorno inicial. El rendimiento del algoritmo se valida usando los coeficientes de Pratt y se compara contra una segmentación realizada manualmente y otro método automático basado en la transformada de watershed. SUMMARY: The present work describes a new methodology for the automatic detection of the glottal space from laryngeal images taken from 15 videos recorded by the ENT service of the Gregorio Marañon Hospital in Madrid with videostroboscopic equipment. The system is based on active contour models (snakes). The algorithm combines for the pre-processing, some traditional techniques (thresholding and median filter) with more sophisticated techniques such as anisotropic filtering. In this way, we obtain an appropriate image for the use of snake. The value selected for the threshold is 85% of the maximum peak of the image histogram; over this point the information of the pixels is not relevant. The anisotropic filter permits to distinguish two intensity levels, one is the background and the other one is the glottis. The initialization is based on the obtained magnitude by GVF field; in this manner an automatic process for the initial contour selection will be assured. The performance of the algorithm is tested using the Pratt coefficient and compared against a manual segmentation and another automatic method based on the watershed transformation.
Resumo:
OntoTag - A Linguistic and Ontological Annotation Model Suitable for the Semantic Web
1. INTRODUCTION. LINGUISTIC TOOLS AND ANNOTATIONS: THEIR LIGHTS AND SHADOWS
Computational Linguistics is already a consolidated research area. It builds upon the results of other two major ones, namely Linguistics and Computer Science and Engineering, and it aims at developing computational models of human language (or natural language, as it is termed in this area). Possibly, its most well-known applications are the different tools developed so far for processing human language, such as machine translation systems and speech recognizers or dictation programs.
These tools for processing human language are commonly referred to as linguistic tools. Apart from the examples mentioned above, there are also other types of linguistic tools that perhaps are not so well-known, but on which most of the other applications of Computational Linguistics are built. These other types of linguistic tools comprise POS taggers, natural language parsers and semantic taggers, amongst others. All of them can be termed linguistic annotation tools.
Linguistic annotation tools are important assets. In fact, POS and semantic taggers (and, to a lesser extent, also natural language parsers) have become critical resources for the computer applications that process natural language. Hence, any computer application that has to analyse a text automatically and ‘intelligently’ will include at least a module for POS tagging. The more an application needs to ‘understand’ the meaning of the text it processes, the more linguistic tools and/or modules it will incorporate and integrate.
However, linguistic annotation tools have still some limitations, which can be summarised as follows:
1. Normally, they perform annotations only at a certain linguistic level (that is, Morphology, Syntax, Semantics, etc.).
2. They usually introduce a certain rate of errors and ambiguities when tagging. This error rate ranges from 10 percent up to 50 percent of the units annotated for unrestricted, general texts.
3. Their annotations are most frequently formulated in terms of an annotation schema designed and implemented ad hoc.
A priori, it seems that the interoperation and the integration of several linguistic tools into an appropriate software architecture could most likely solve the limitations stated in (1). Besides, integrating several linguistic annotation tools and making them interoperate could also minimise the limitation stated in (2). Nevertheless, in the latter case, all these tools should produce annotations for a common level, which would have to be combined in order to correct their corresponding errors and inaccuracies. Yet, the limitation stated in (3) prevents both types of integration and interoperation from being easily achieved.
In addition, most high-level annotation tools rely on other lower-level annotation tools and their outputs to generate their own ones. For example, sense-tagging tools (operating at the semantic level) often use POS taggers (operating at a lower level, i.e., the morphosyntactic) to identify the grammatical category of the word or lexical unit they are annotating. Accordingly, if a faulty or inaccurate low-level annotation tool is to be used by other higher-level one in its process, the errors and inaccuracies of the former should be minimised in advance. Otherwise, these errors and inaccuracies would be transferred to (and even magnified in) the annotations of the high-level annotation tool.
Therefore, it would be quite useful to find a way to
(i) correct or, at least, reduce the errors and the inaccuracies of lower-level linguistic tools;
(ii) unify the annotation schemas of different linguistic annotation tools or, more generally speaking, make these tools (as well as their annotations) interoperate.
Clearly, solving (i) and (ii) should ease the automatic annotation of web pages by means of linguistic tools, and their transformation into Semantic Web pages (Berners-Lee, Hendler and Lassila, 2001). Yet, as stated above, (ii) is a type of interoperability problem. There again, ontologies (Gruber, 1993; Borst, 1997) have been successfully applied thus far to solve several interoperability problems. Hence, ontologies should help solve also the problems and limitations of linguistic annotation tools aforementioned.
Thus, to summarise, the main aim of the present work was to combine somehow these separated approaches, mechanisms and tools for annotation from Linguistics and Ontological Engineering (and the Semantic Web) in a sort of hybrid (linguistic and ontological) annotation model, suitable for both areas. This hybrid (semantic) annotation model should (a) benefit from the advances, models, techniques, mechanisms and tools of these two areas; (b) minimise (and even solve, when possible) some of the problems found in each of them; and (c) be suitable for the Semantic Web. The concrete goals that helped attain this aim are presented in the following section.
2. GOALS OF THE PRESENT WORK
As mentioned above, the main goal of this work was to specify a hybrid (that is, linguistically-motivated and ontology-based) model of annotation suitable for the Semantic Web (i.e. it had to produce a semantic annotation of web page contents). This entailed that the tags included in the annotations of the model had to (1) represent linguistic concepts (or linguistic categories, as they are termed in ISO/DCR (2008)), in order for this model to be linguistically-motivated; (2) be ontological terms (i.e., use an ontological vocabulary), in order for the model to be ontology-based; and (3) be structured (linked) as a collection of ontology-based
Resumo:
Este Proyecto Fin de Carrera trata sobre el reconocimiento e identificación de caracteres de matrículas de automóviles. Este tipo de sistemas de reconocimiento también se los conoce mundialmente como sistemas ANPR ("Automatic Number Plate Recognition") o LPR ("License Plate Recognition"). La gran cantidad de vehículos y logística que se mueve cada segundo por todo el planeta, hace necesaria su registro para su tratamiento y control. Por ello, es necesario implementar un sistema que pueda identificar correctamente estos recursos, para su posterior procesado, construyendo así una herramienta útil, ágil y dinámica. El presente trabajo ha sido estructurado en varias partes. La primera de ellas nos muestra los objetivos y las motivaciones que se persiguen con la realización de este proyecto. En la segunda, se abordan y desarrollan todos los diferentes procesos teóricos y técnicos, así como matemáticos, que forman un sistema ANPR común, con el fin de implementar una aplicación práctica que pueda demostrar la utilidad de estos en cualquier situación. En la tercera, se desarrolla esa parte práctica en la que se apoya la base teórica del trabajo. En ésta se describen y desarrollan los diversos algoritmos, creados con el fin de estudiar y comprobar todo lo planteado hasta ahora, así como observar su comportamiento. Se implementan varios procesos característicos del reconocimiento de caracteres y patrones, como la detección de áreas o patrones, rotado y transformación de imágenes, procesos de detección de bordes, segmentación de caracteres y patrones, umbralización y normalización, extracción de características y patrones, redes neuronales, y finalmente el reconocimiento óptico de caracteres o comúnmente conocido como OCR. La última parte refleja los resultados obtenidos a partir del sistema de reconocimiento de caracteres implementado para el trabajo y se exponen las conclusiones extraídas a partir de éste. Finalmente se plantean las líneas futuras de mejora, desarrollo e investigación, para poder realizar un sistema más eficiente y global. This Thesis deals about license plate characters recognition and identification. These kinds of systems are also known worldwide as ANPR systems ("Automatic Number Plate Recognition") or LPR ("License Plate Recognition"). The great number of vehicles and logistics moving every second all over the world, requires a registration for treatment and control. Thereby, it’s therefore necessary to implement a system that can identify correctly these resources, for further processing, thus building a useful, flexible and dynamic tool. This work has been structured into several parts. The first one shows the objectives and motivations attained by the completion of this project. In the second part, it’s developed all the different theoretical and technical processes, forming a common ANPR system in order to implement a practical application that can demonstrate the usefulness of these ones on any situation. In the third, the practical part is developed, which is based on the theoretical work. In this one are described and developed various algorithms, created to study and verify all the questions until now suggested, and complain the behavior of these systems. Several recognition of characters and patterns characteristic processes are implemented, such as areas or patterns detection, image rotation and transformation, edge detection processes, patterns and character segmentation, thresholding and normalization, features and patterns extraction, neural networks, and finally the optical character recognition or commonly known like OCR. The last part shows the results obtained from the character recognition system implemented for this thesis and the outlines conclusions drawn from it. Finally, future lines of improvement, research and development are proposed, in order to make a more efficient and comprehensive system.
Resumo:
La mejora en la eficiencia energética y la reducción de la tasa de fallos en los contactos lubricados son aspectos que resultan de gran interés en numerosos sectores de la industria, y plantean en estos momentos nuevas dificultades operativas y retos para un futuro próximo. Los avances tecnológicos han incrementado las exigencias técnicas que se requieren a los aceites para cumplir su función al extender sus variables operativas a un mayor espectro de aplicaciones, tanto de condiciones de funcionamiento como a la gran variedad de nuevos materiales constitutivos de los engranajes en los que se tiene que utilizar. Por ello, actualmente se está incentivado el desarrollo de nuevos procedimientos que permitan comprender el comportamiento de este tipo de contactos lubricados, con el fin de lograr mejoras técnicas en su diseño y la correcta selección del aceite. En esta Tesis Doctoral se presenta una metodología de cálculo numérico que permite simular el comportamiento de contactos elastohidrodinámicos (EHD) puntuales, como puede ser el caso de un rodamiento. La resolución de este problema presenta diversas complejidades matemáticas y exige el desarrollo de un elaborado procedimiento de cálculo basado en técnicas multinivel. Para hacer del procedimiento una herramienta válida en un gran número de condiciones de funcionamiento y tipos de lubricantes, se ha tenido en cuenta en el cálculo la posible aparición de comportamientos no-Newtonianos del lubricante y fenómenos de generación y disipación de calor, provocados por el movimiento relativo del fluido y las superficies en contacto. Para la validación del procedimiento, se han contrastado los resultados numéricos obtenidos con nuestro método, con los resultados numéricos y experimentales publicados por otros autores y con valores experimentales propios medidos en un equipo de ensayo de contacto puntual tipo MTM. El desarrollo de este programa ha dotado a la División de Ingeniería de Máquinas de una herramienta que ha permitido, y sobre todo va permitir, evaluar la importancia de cada uno de los parámetros reológicos en los diferentes problemas que va a tener que abordar, evaluación que hasta el momento se realizaba con métodos aproximados que describen la fenomenología con mucha menos precisión. A la hora de emplear nuestro procedimiento numérico para simular situaciones reales, nos hemos encontrado con el obstáculo de que es muy complicado encontrar, en la bibliografía y bases de datos, los parámetros que caracterizan el comportamiento reológico del lubricante en las condiciones de presión, temperatura y grado de cizallamiento habituales en las que se trabaja en este tipo de contactos, y las pocas medidas que existen para estas condiciones de funcionamiento son poco fiables. Por ello como complemento al objetivo principal de esta Tesis hemos desarrollado una metodología para caracterizar los lubricantes en estas condiciones extremas. Dicha metodología está basada en la buena descripción que hace nuestro programa del coeficiente de fricción, lo que nos ha permitido obtener los parámetros reológicos del III lubricante a partir de las medidas experimentales del coeficiente de fricción generado en un equipo MTM lubricado con el lubricante que se desea caracterizar. Madrid, Octubre de 2012 IV Improving energy efficiency and reducing the failure rate in lubricated contacts are issues that are of great interest in many sectors of industry, and right now posed operational difficulties and new challenges for the near future. Technological advances have increased the technical demands required to oils to fulfil its role by extending its operational variables to a wider range of applications, both operating conditions and to the wide variety of new materials which constitute the gear in which must be used. For this reason, has being encouraged currently to develop new procedures to understand the behaviour of this type of lubricated contacts, in order to achieve improvements in design techniques and the correct oil selection. In this Thesis we present a numerical methodology to simulate the puntual elastohydrodynamic contact behaviour (EHD), such as a roller bearing. The resolution of this problem presents various mathematical complexities and requires the development of an elaborate calculation procedure based on multilevel techniques. To make the procedure a valid tool in a wide range of operating conditions and types of lubricants, has been taken into account in calculating the possible occurrence of non-Newtonian behaviour of the lubricant and phenomena of generation and dissipation of heat, caused by the fluid relative motion and contacting surfaces. For the validation of the method, we have compared the numerical results obtained with our method with numerical and experimental results published by other authors also with own experimental values measured on point-contact test equipment MTM. The development of this program has provided the Machines Engineering Division of a tool that has allowed, and especially will allow to evaluate the importance of each of the rheological parameters on the various problems that will have to be addressed, evaluation performed hitherto described methods that phenomenology approximated with much less accuracy. When using our numerical procedure to simulate real situations we have encountered the obstacle that is very difficult to find, in the literature and database, parameters characterizing the rheological behaviour of the lubricant in the usual conditions of pressure, temperature and shear rate in which you work in this type of contact, and the few measures that exist for these operating conditions are unreliable. Thus in addition to the main objective of this thesis, we have developed a methodology to characterize the lubricants in these extreme conditions. This methodology is based on the good description, which makes our program, of the coefficient of friction, that allowed us to obtain the lubricant rheological parameters from experimental measurements of the friction coefficient generated on lubricated MTM equipment with the lubricant to be characterized.
Resumo:
Modern transmitters usually have to amplify and transmit signals with simultaneous envelope and phase modulation. Due to this property of the transmitted signal, linear power amplifiers (class A, B, or AB) are usually used as a solution for the power amplifier stage. These amplifiers have high linearity, but suffer from low efficiency when the transmitted signal has high peak-to-average power ratio. The Kahn envelope elimination and restoration technique is used to enhance the efficiency of RF transmitters, by combining highly efficient, nonlinear RF amplifier (class E) with a highly efficient envelope amplifier in order to obtain a linear and highly efficient RF amplifier. This paper presents a solution for the envelope amplifier based on a multilevel converter in series with a linear regulator. The multilevel converter is implemented by employing voltage dividers based on switching capacitors. The implemented envelope amplifier can reproduce any signal with a maximum spectral component of 2 MHz and give instantaneous maximum power of 50 W. The efficiency measurements show that when the signals with low average value are transmitted, the implemented prototypes have up to 20% higher efficiency than linear regulators used as a conventional solution.
Resumo:
This paper proposes a new method, oriented to crop row detection in images from maize fields with high weed pressure. The vision system is designed to be installed onboard a mobile agricultural vehicle, i.e. submitted to gyros, vibrations and undesired movements. The images are captured under image perspective, being affected by the above undesired effects. The image processing consists of three main processes: image segmentation, double thresholding, based on the Otsu’s method, and crop row detection. Image segmentation is based on the application of a vegetation index, the double thresholding achieves the separation between weeds and crops and the crop row detection applies least squares linear regression for line adjustment. Crop and weed separation becomes effective and the crop row detection can be favorably compared against the classical approach based on the Hough transform. Both gain effectiveness and accuracy thanks to the double thresholding that makes the main finding of the paper.
Resumo:
The type of signals obtained has conditioned chaos analysis tools. Almost in every case, they have analogue characteristics. But in certain cases, a chaotic digital signal is obtained and theses signals need a different approach than conventional analogue ones. The main objective of this paper will be to present some possible approaches to the study of this signals and how information about their characteristics may be obtained in the more straightforward possible way. We have obtained digital chaotic signals from an Optical Logic Cell with some feedback between output and one of the possible control gates. This chaos has been reported in several papers and its characteristics have been employed as a possible method to secure communications and as a way to encryption. In both cases, the influence of some perturbation in the transmission medium gave problems both for the synchronization of chaotic generators at emitter and receiver and for the recovering of information data. A proposed way to analyze the presence of some perturbation is to study the noise contents of transmitted signal and to implement a way to eliminate it. In our present case, the digital signal will be converted to a multilevel one by grouping bits in packets of 8 bits and applying conventional methods of time-frequency analysis to them. The results give information about the change in signals characteristics and hence some information about the noise or perturbations present. Equivalent representations to the phase and to the Feigenbaum diagrams for digital signals are employed in this case.