13 resultados para Algebra of Errors

em Cochin University of Science


Relevância:

100.00% 100.00%

Publicador:

Resumo:

The results of an investigation on the limits of the random errors contained in the basic data of Physical Oceanography and their propagation through the computational procedures are presented in this thesis. It also suggest a method which increases the reliability of the derived results. The thesis is presented in eight chapters including the introductory chapter. Chapter 2 discusses the general theory of errors that are relevant in the context of the propagation of errors in Physical Oceanographic computations. The error components contained in the independent oceanographic variables namely, temperature, salinity and depth are deliniated and quantified in chapter 3. Chapter 4 discusses and derives the magnitude of errors in the computation of the dependent oceanographic variables, density in situ, gt, specific volume and specific volume anomaly, due to the propagation of errors contained in the independent oceanographic variables. The errors propagated into the computed values of the derived quantities namely, dynamic depth and relative currents, have been estimated and presented chapter 5. Chapter 6 reviews the existing methods for the identification of level of no motion and suggests a method for the identification of a reliable zero reference level. Chapter 7 discusses the available methods for the extension of the zero reference level into shallow regions of the oceans and suggests a new method which is more reliable. A procedure of graphical smoothening of dynamic topographies between the error limits to provide more reliable results is also suggested in this chapter. Chapter 8 deals with the computation of the geostrophic current from these smoothened values of dynamic heights, with reference to the selected zero reference level. The summary and conclusion are also presented in this chapter.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Solid phase extraction (SPE) is a powerful technique for preconcentration/removal or separation of trace and ultra trace amounts of toxic and nutrient elements. SPE effectively simplifies the labour intensive sample preparation, increase its reliability and eliminate the clean up step by using more selective extraction procedures. The synthesis of sorbents with a simplified procedure and diminution of the risks of errors shows the interest in the areas of environmental monitoring, geochemical exploration, food, agricultural, pharmaceutical, biochemical industry and high purity metal designing, etc. There is no universal SPE method because the sample pretreatment depends strongly on the analytical demand. But there is always an increasing demand for more sensitive, selective, rapid and reliable analytical procedures. Among the various materials, chelate modified naphthalene, activated carbon and chelate functionalized highly cross linked polymers are most important. In the biological and environmental field, large numbers of samples are to be analysed within a short span of time. Hence, online flow injection methods are preferred as they allow extraction, separation, identification and quantification of many numbers of analytes. The flow injection online preconcentration flame AAS procedure developed allows the determination of as low as 0.1 µg/l of nickel in soil and cobalt in human hair samples. The developed procedure is precise and rapid and allows the analysis of 30 samples per hour with a loading time of 60 s. The online FI manifold used in the present study permits high sampling, loading rates and thus resulting in higher preconcentration/enrichment factors of -725 and 600 for cobalt and nickel respectively with a 1 min preconcentration time compared to conventional FAAS signal. These enrichment factors are far superior to hitherto developed on line preconcentration procedures for inorganics. The instrumentation adopted in the present study allows much simpler equipment and low maintenance costs compared to costlier ICP-AES or ICP-MS instruments.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

In India, Food Security meant supply of food grains and the medium was Public Distribution System. Public Distribution System (PDS) is a rationing mechanism that entitles households to specified quantities of selected commodities at subsidized prices. The Objectives of PDS are maintaining Price Stability, rationing during times of scarcity, welfare of the poor, and keeping a check on private trade. Kerala has registered remarkable improvement in poverty reduction in general over the years among all social sections, including scheduled caste and scheduled tribe population. As part of the structural adjustment intended to reduce public expenditure, PDS has been modified as Revamped PDS (RPDS) during 1992 and later on as Targeted PDS (TPDS) in 1997, intended to target households on the basis of income criterion, classifying people as Below Poverty Line (BPL) and Above Poverty Line (APL). TPDS provides 25Kg. of food gra.ins through the Fair Price Shops per month @ Rs.3/- per Kg. of rice/ wheat to the BPL category and @Rs.8.90 and Rs.6.7O for rice and wheat respectively to the APL category of people. Since TPDS is intended to target the poor people, the subsidy spent by the government for the scheme should be beneficial to the poor people and naturally they should utilize the benefits by purchasing the food grains allotted under the scheme. Several studies have shown that there is underutilization of the allotments under TPDS. Therefore, the extent of utilization of TPDS in food grains, how and why remains as a major hurdle, in improving the structure and system of PDS. Livelihood of the tribal population being under threat due to increasing degradation of the resources, the targeting system ought to be effective among the tribal population. Therefore, performance of the TPDS in food grains, in terms of the utilization by the tribal population in Kerala, impact thereof and the factors, if any, affecting proper utilization were considered as the research problem in this study. The study concentrated on the pattern of consumption of food grains by the tribal people, whether their hunger needs are met by distribution of food grains through the TPDS, extent to which TPDS in food grains reduce their share of expenditure on food in the total household expenditure, and the factors affecting the utilization of the TPDS in food grains by the tribal population. Going through the literature, it has been noted that only few studies concentrated on the utilization of TPDS in food grains among the tribal population in Kerala.The Research Design used in this study is descriptive in nature, but exploratory in some aspects. Idukki, Palakkad and Wayanad have more than 60% of the population of the tribals in the state. Within the three districts mentioned above, 14 villages with scheduled tribe concentration were selected for the study. 95 tribal colonies were selected from among the various tribal settlements. Collection of primary data was made from 1231 households with in the above tribal colonies. Analysis of data on the socio-economic factors of the tribal people, pattern of food consumption, extent of reduction in the share of expenditure on food among the household expenditure of the tribal people and the impact of TPDS on the tribal families etc. and testing of hypotheses to find out the relation/association of each of the six variables, using the data on BPL and APL categories of households separately have resulted in findings such as six percent of the tribal families do not have Ration Cards, average per capita consumption of food grains by the tribal people utilizing TPDS meets 62% of their minimum requirement, whereas the per capita consumption of food grains by the tribal people is higher than the national average per capita consumption, 63% deficiency in food grains may be felt by tribal people in general, if TPDS is withdrawn, and the deficit for BPL tribal people may be 82%, TPDS facilitates a reduction of 9.71% in the food expenditure among the total household expenditure of the tribal people in general, share of food to non-food among BPL category of tribals is 55:45 and 40:60 among the APL, Variables, viz. household income, number of members in the family and distance of FPS from tribal settlements etc. have influence on the quantity of rice being purchased by the tribal people from the Fair Price Shops, and there is influence of household income and distance of FPS from tribal settlements on the quantity of rice being purchased by the tribal people from the open market. Rationing with differential pricing on phased allotments, rectification of errors in targeting, anomalies in norms and procedures for classifying tribal people as BPL/APL, exclusive Income Generation for tribal population, paddy cultivation in the landholdings possessed by the tribal people, special drive for allotment of Ration Cards to the tribal people, especially those belonging to the BPL category, Mobile Fair Price Shops in tribal settlements, ensure quality of the food grains distributed through the TPDS, distribution of wheat flour in packed condition instead of wheat through the Fair Price Shops are recommended to address the shortcomings and weaknesses of the TPDS vis-avis the tribal population in Kerala.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

A novel cavity perturbation technique using coaxial cavity resonators for the measurement of complex permittivity of liquids is presented. The method employs two types of resonators (Resonator I and Resonator II). Resonator I operates in the frequency range 600 MHz-7 GHz and resonator II operates in the frequency range 4 GHz-14 GHz. The introduction of the capillary tube filled with the sample liquid into the coaxial resonator causes shifts in the resonance frequency and loaded Q-factor of the resonator. The shifts in the resonance frequency and loaded Q-factor are used to determine the real and imaginary parts of the complex permittivity of the sample liquid, respectively. Using this technique, the dielectric parameters of water and nitrobenzene are measured. The results are compared with those obtained using other standard methods. The sources of errors are analyzed.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

In Safety critical software failure can have a high price. Such software should be free of errors before it is put into operation. Application of formal methods in the Software Development Life Cycle helps to ensure that the software for safety critical missions are ultra reliable. PVS theorem prover, a formal method tool, can be used for the formal verification of software in ADA Language for Flight Software Application (ALFA.). This paper describes the modeling of ALFA programs for PVS theorem prover. An ALFA2PVS translator is developed which automatically converts the software in ALFA to PVS specification. By this approach the software can be verified formally with respect to underflow/overflow errors and divide by zero conditions without the actual execution of the code.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

In Safety critical software failure can have a high price. Such software should be free of errors before it is put into operation. Application of formal methods in the Software Development Life Cycle helps to ensure that the software for safety critical missions are ultra reliable. PVS theorem prover, a formal method tool, can be used for the formal verification of software in ADA Language for Flight Software Application (ALFA.). This paper describes the modeling of ALFA programs for PVS theorem prover. An ALFA2PVS translator is developed which automatically converts the software in ALFA to PVS specification. By this approach the software can be verified formally with respect to underflow/overflow errors and divide by zero conditions without the actual execution of the code

Relevância:

80.00% 80.00%

Publicador:

Resumo:

RMS measuring device is a nonlinear device consisting of linear and nonlinear devices. The performance of rms measurement is influenced by a number of factors; i) signal characteristics, 2) the measurement technique used and 3) the device characteristics. RMS measurement is not simple, particularly when the signals are complex and unknown. The problem of rms measurement on high crest-factor signals is fully discussed and a solution to this problem is presented in this thesis. The problem of rms measurement is systematically analized and found to have mainly three types of errors: (1) amplitude or waveform error 2) Frequency error and (3) averaging error. Various rms measurement techniques are studied and compared. On the basis of this study the rms -measurement is reclassified three categories: (1) Wave-form-error-free measurement (2) High-frequncy-error measurement and (3) Low-frequency error-free measurement. In modern digital sampled-data systems the signals are complex and waveform-error-free rms measurement is highly appreciated. Among the three basic blocks of rms measuring device the squarer is the most important one. A squaring technique is selected, that permits shaping of the squarer error characteristic in such a way as to achieve waveform-errob free rms measurement. The squarer is designed, fabricated and tested. A hybrid rms measurement using an analog rms computing device and digital display combines the speed of analog techniques and the resolution and ease of measurement of digital techniques. An A/D converter is modified to perform the square-rooting operation. A 10-V rms voltmeter using the developed rms detector is fabricated and tested. The chapters two, three and four analyse the problems involved in rms measurement and present a comparative study of rms computing techniques and devices. The fifth chapter gives the details of the developed rms detector that permits wave-form-error free rms measurement. The sixth chapter, enumerates the the highlights of the thesis and suggests a list of future projects

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Measurement is the act or the result of a quantitative comparison between a given quantity and a quantity of the same kind chosen as a unit. It is generally agreed that all measurements contain errors. In a measuring system where both a measuring instrument and a human being taking the measurement using a preset process, the measurement error could be due to the instrument, the process or the human being involved. The first part of the study is devoted to understanding the human errors in measurement. For that, selected person related and selected work related factors that could affect measurement errors have been identified. Though these are well known, the exact extent of the error and the extent of effect of different factors on human errors in measurement are less reported. Characterization of human errors in measurement is done by conducting an experimental study using different subjects, where the factors were changed one at a time and the measurements made by them recorded. From the pre‐experiment survey research studies, it is observed that the respondents could not give the correct answers to questions related to the correct values [extent] of human related measurement errors. This confirmed the fears expressed regarding lack of knowledge about the extent of human related measurement errors among professionals associated with quality. But in postexperiment phase of survey study, it is observed that the answers regarding the extent of human related measurement errors has improved significantly since the answer choices were provided based on the experimental study. It is hoped that this work will help users of measurement in practice to better understand and manage the phenomena of human related errors in measurement.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This thesis is a study of abstract fuzzy convexity spaces and fuzzy topology fuzzy convexity spaces No attempt seems to have been made to develop a fuzzy convexity theoryin abstract situations. The purpose of this thesis is to introduce fuzzy convexity theory in abstract situations

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The present study is an attempt to highlight the problem of typographical errors in OPACS. The errors made while typing catalogue entries as well as importing bibliographical records from other libraries exist unnoticed by librarians resulting the non-retrieval of available records and affecting the quality of OPACs. This paper follows previous research on the topic mainly by Jeffrey Beall and Terry Ballard. The word “management” was chosen from the list of likely to be misspelled words identified by previous research. It was found that the word is wrongly entered in several forms in local, national and international OPACs justifying the observations of Ballard that typos occur in almost everywhere. Though there are lots of corrective measures proposed and are in use, the study asserts the fact that human effort is needed to get rid of the problem. The paper is also an invitation to the library professionals and system designers to construct a strategy to solve the issue

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The present study is an attempt to highlight the problem of typographical errors in OPACS. The errors made while typing catalogue entries as well as importing bibliographical records from other libraries exist unnoticed by librarians resulting the non-retrieval of available records and affecting the quality of OPACs. This paper follows previous research on the topic mainly by Jeffrey Beall and Terry Ballard. The word “management” was chosen from the list of likely to be misspelled words identified by previous research. It was found that the word is wrongly entered in several forms in local, national and international OPACs justifying the observations of Ballard that typos occur in almost everywhere. Though there are lots of corrective measures proposed and are in use, the study asserts the fact that human effort is needed to get rid of the problem. The paper is also an invitation to the library professionals and system designers to construct a strategy to solve the issue

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This paper introduces a simple and efficient method and its implementation in an FPGA for reducing the odometric localization errors caused by over count readings of an optical encoder based odometric system in a mobile robot due to wheel-slippage and terrain irregularities. The detection and correction is based on redundant encoder measurements. The method suggested relies on the fact that the wheel slippage or terrain irregularities cause more count readings from the encoder than what corresponds to the actual distance travelled by the vehicle. The standard quadrature technique is used to obtain four counts in each encoder period. In this work a three-wheeled mobile robot vehicle with one driving-steering wheel and two-fixed rear wheels in-axis, fitted with incremental optical encoders is considered. The CORDIC algorithm has been used for the computation of sine and cosine terms in the update equations. The results presented demonstrate the effectiveness of the technique