993 resultados para Systematic errors
Resumo:
Systematic trends in the properties of a linear split-gate heterojunction are studied by solving iteratively the Poisson and Schrödinger equations for different gate potentials and temperatures. A two-dimensional approximation is presented that is much simpler in the numerical implementation and that accurately reproduces all significant trends. In deriving this approximation, we provide a rigorous and quantitative basis for the formulation of models that assumes a two-dimensional character for the electron gas at the junction.
Resumo:
The results of an investigation on the limits of the random errors contained in the basic data of Physical Oceanography and their propagation through the computational procedures are presented in this thesis. It also suggest a method which increases the reliability of the derived results. The thesis is presented in eight chapters including the introductory chapter. Chapter 2 discusses the general theory of errors that are relevant in the context of the propagation of errors in Physical Oceanographic computations. The error components contained in the independent oceanographic variables namely, temperature, salinity and depth are deliniated and quantified in chapter 3. Chapter 4 discusses and derives the magnitude of errors in the computation of the dependent oceanographic variables, density in situ, gt, specific volume and specific volume anomaly, due to the propagation of errors contained in the independent oceanographic variables. The errors propagated into the computed values of the derived quantities namely, dynamic depth and relative currents, have been estimated and presented chapter 5. Chapter 6 reviews the existing methods for the identification of level of no motion and suggests a method for the identification of a reliable zero reference level. Chapter 7 discusses the available methods for the extension of the zero reference level into shallow regions of the oceans and suggests a new method which is more reliable. A procedure of graphical smoothening of dynamic topographies between the error limits to provide more reliable results is also suggested in this chapter. Chapter 8 deals with the computation of the geostrophic current from these smoothened values of dynamic heights, with reference to the selected zero reference level. The summary and conclusion are also presented in this chapter.
Resumo:
Quaternary stratigraphy of the Kerala coast and the genetic aspects of the sediments are discussed. The age of limeshells, and peaty sediments determined by radio carbon dating have been used for reconstruction of sea level changes. Evolution of red sands occurring in some parts of the coastal tract of Kerala is also discussed, based on textural parameters and quartz grain morphology.
Resumo:
Measurement is the act or the result of a quantitative comparison between a given quantity and a quantity of the same kind chosen as a unit. It is generally agreed that all measurements contain errors. In a measuring system where both a measuring instrument and a human being taking the measurement using a preset process, the measurement error could be due to the instrument, the process or the human being involved. The first part of the study is devoted to understanding the human errors in measurement. For that, selected person related and selected work related factors that could affect measurement errors have been identified. Though these are well known, the exact extent of the error and the extent of effect of different factors on human errors in measurement are less reported. Characterization of human errors in measurement is done by conducting an experimental study using different subjects, where the factors were changed one at a time and the measurements made by them recorded. From the pre‐experiment survey research studies, it is observed that the respondents could not give the correct answers to questions related to the correct values [extent] of human related measurement errors. This confirmed the fears expressed regarding lack of knowledge about the extent of human related measurement errors among professionals associated with quality. But in postexperiment phase of survey study, it is observed that the answers regarding the extent of human related measurement errors has improved significantly since the answer choices were provided based on the experimental study. It is hoped that this work will help users of measurement in practice to better understand and manage the phenomena of human related errors in measurement.
Resumo:
The present study is an attempt to highlight the problem of typographical errors in OPACS. The errors made while typing catalogue entries as well as importing bibliographical records from other libraries exist unnoticed by librarians resulting the non-retrieval of available records and affecting the quality of OPACs. This paper follows previous research on the topic mainly by Jeffrey Beall and Terry Ballard. The word “management” was chosen from the list of likely to be misspelled words identified by previous research. It was found that the word is wrongly entered in several forms in local, national and international OPACs justifying the observations of Ballard that typos occur in almost everywhere. Though there are lots of corrective measures proposed and are in use, the study asserts the fact that human effort is needed to get rid of the problem. The paper is also an invitation to the library professionals and system designers to construct a strategy to solve the issue
Resumo:
The present study is an attempt to highlight the problem of typographical errors in OPACS. The errors made while typing catalogue entries as well as importing bibliographical records from other libraries exist unnoticed by librarians resulting the non-retrieval of available records and affecting the quality of OPACs. This paper follows previous research on the topic mainly by Jeffrey Beall and Terry Ballard. The word “management” was chosen from the list of likely to be misspelled words identified by previous research. It was found that the word is wrongly entered in several forms in local, national and international OPACs justifying the observations of Ballard that typos occur in almost everywhere. Though there are lots of corrective measures proposed and are in use, the study asserts the fact that human effort is needed to get rid of the problem. The paper is also an invitation to the library professionals and system designers to construct a strategy to solve the issue
Resumo:
This paper introduces a simple and efficient method and its implementation in an FPGA for reducing the odometric localization errors caused by over count readings of an optical encoder based odometric system in a mobile robot due to wheel-slippage and terrain irregularities. The detection and correction is based on redundant encoder measurements. The method suggested relies on the fact that the wheel slippage or terrain irregularities cause more count readings from the encoder than what corresponds to the actual distance travelled by the vehicle. The standard quadrature technique is used to obtain four counts in each encoder period. In this work a three-wheeled mobile robot vehicle with one driving-steering wheel and two-fixed rear wheels in-axis, fitted with incremental optical encoders is considered. The CORDIC algorithm has been used for the computation of sine and cosine terms in the update equations. The results presented demonstrate the effectiveness of the technique
Resumo:
Die relativistische Multikonfigurations Dirac-Fock (MCDF) Methode ist gegenwärtig eines der am häufigsten benutzten Verfahren zur Berechnung der elektronischen Struktur und der Eigenschaften freier Atome. In diesem Verfahren werden die Wellenfunktionen ausgewählter atomarer Zustände als eine Linearkombination von sogenannten Konfigurationszuständen (CSF - Configuration State Functions) konstruiert, die in einem Teilraum des N-Elektronen Hilbert-Raumes eine (Vielteilchen-)Basis aufspannen. Die konkrete Konstruktion dieser Basis entscheidet letzlich über die Güte der Wellenfunktionen, die üblicherweise mit Hilfe einer Variation des Erwartungswertes zum no-pair Dirac-Coulomb Hamiltonoperators gewonnen werden. Mit Hilfe von MCDF Wellenfunktionen können die dominanten relativistischen und Korrelationseffekte in freien Atomen allgemein recht gut erfaßt und verstanden werden. Außer der instantanen Coulombabstoßung zwischen allen Elektronenpaaren werden dabei auch die relativistischen Korrekturen zur Elektron-Elektron Wechselwirkung, d.h. die magnetischen und Retardierungsbeiträge in der Wechselwirkung der Elektronen untereinander, die Ankopplung der Elektronen an das Strahlungsfeld sowie der Einfluß eines ausgedehnten Kernmodells erfaßt. Im Vergleich mit früheren MCDF Rechnungen werden in den in dieser Arbeit diskutierten Fallstudien Wellenfunktionsentwicklungen verwendet, die um 1-2 Größenordnungen aufwendiger sind und daher systematische Untersuchungen inzwischen auch an Atomen mit offenen d- und f-Schalen erlauben. Eine spontane Emission oder Absorption von Photonen kann bei freien Atomen theoretisch am einfachsten mit Hilfe von Übergangswahrscheinlichkeiten erfaßt werden. Solche Daten werden heute in vielen Forschungsbereichen benötigt, wobei neben den traditionellen Gebieten der Fusionsforschung und Astrophysik zunehmend auch neue Forschungsrichtungen (z.B. Nanostrukturforschung und Röntgenlithographie) zunehmend ins Blickfeld rücken. Um die Zuverlässigkeit unserer theoretischen Vorhersagen zu erhöhen, wurde in dieser Arbeit insbesondere die Relaxation der gebundenen Elektronendichte, die rechentechnisch einen deutlich größeren Aufwand erfordert, detailliert untersucht. Eine Berücksichtigung dieser Relaxationseffekte führt oftmals auch zu einer deutlich besseren Übereinstimmung mit experimentellen Werten, insbesondere für dn=1 Übergänge sowie für schwache und Interkombinationslinien, die innerhalb einer Hauptschale (dn=0) vorkommen. Unsere in den vergangenen Jahren verbesserten Rechnungen zu den Wellenfunktionen und Übergangswahrscheinlichkeiten zeigen deutlich den Fortschritt bei der Behandlung komplexer Atome. Gleichzeitig kann dieses neue Herangehen künftig aber auch auf (i) kompliziertere Schalensstrukturen, (ii) die Untersuchung von Zwei-Elektronen-ein-Photon (TEOP) Übergängen sowie (iii) auf eine Reihe weiterer atomarer Eigenschaften übertragen werden, die bekanntermaßen empflindlich von der Relaxation der Elektronendichte abhängen. Dies sind bspw. Augerzerfälle, die atomare Photoionisation oder auch strahlende und dielektronische Rekombinationsprozesse, die theoretisch bisher nur selten überhaupt in der Dirac-Fock Näherung betrachtet wurden.
Resumo:
A fully relativistic four-component Dirac-Fock-Slater program for diatomics, with numerically given AO's as basis functions is presented. We discuss the problem of the errors due to the finite basis-set, and due to the influence of the negative energy solutions of the Dirac Hamiltonian. The negative continuum contributions are found to be very small.
Resumo:
Resumen tomado de la publicaci??n. Resumen tambi??n en ingl??s
Resumo:
Resumen tomado de la publicaci??n
Resumo:
Addresses the problem of estimating the motion of an autonomous underwater vehicle (AUV), while it constructs a visual map ("mosaic" image) of the ocean floor. The vehicle is equipped with a down-looking camera which is used to compute its motion with respect to the seafloor. As the mosaic increases in size, a systematic bias is introduced in the alignment of the images which form the mosaic. Therefore, this accumulative error produces a drift in the estimation of the position of the vehicle. When the arbitrary trajectory of the AUV crosses over itself, it is possible to reduce this propagation of image alignment errors within the mosaic. A Kalman filter with augmented state is proposed to optimally estimate both the visual map and the vehicle position
Resumo:
Introducción: El TDAH tiene un componente genético importante; el gen de transportador de Dopamina (DAT1) se ha asociado con susceptibilidad al TDAH y con sus endofenotipos. El VNTR de 40pb en la región 3’UTR aumenta la expresión del DAT1. En Colombia no hay ningún estudio previo que indique evidencia de la asociación genética entre TDAH y el gen DAT1. Objetivo: Determinar asociación entre el VNTR del DAT1 y el fenotipo y/o endofenotipos del TDAH. Métodos: Se seleccionaron 73 pacientes con TDAH y 75 controles, se valoró en los casos inteligencia y funciones ejecutivas. Mediante (PCR) se amplificó el VNTR DAT1. Se establecieron estadísticos genético poblacionales, análisis de asociación y de regresión logística entre las pruebas neuropsicológicas y genotipo. Resultado: El polimorfismo del DAT1 no mostró asociación con TDAH, ni con alteraciones en las funciones ejecutivas. El genotipo 10/10 del VNTR DAT1 se encontró asociado con el índice de velocidad de procesamiento (p <0,05). En el subgrupo hiperactividad hubo asociación con algunas subpruebas de flexibilidad cognitiva, número de respuestas correctas, total de errores, número de respuestas perseverativas (p ≤ 0.01). En el subgrupo mixto se asoció con índice de comprensión verbal (p <0,05). Conclusiones: No hubo asociación entre el polimorfismo VNTR (DAT1) y el fenotipo de TDAH. Se encontraron asociaciones entre genotipo y algunos test de flexibilidad cognitiva e índice de comprensión verbal. Se establecieron los estadísticos genético poblacionales de este polimorfismo para la población analizada, el cual corresponde al primer reporte de una muestra de nuestro país.