25 resultados para método de correção por equação matemática
em Universidade Federal do Rio Grande do Norte(UFRN)
Resumo:
The present study aims to check whether the use of activities mediated by the History of Mathematics can contribute to improve the understanding of resolution the 2nd degree equation for teachers and undergraduates that reproduce methods of solving such equations, uncritically, without domain of the justifications for their actions. For this, we adapted a didactic sequence with activities that aims to cause a rediscovery of resolutive formula of 2nd degree equation through the method known as cut and paste. Finally, we presented the activity module containing the didactic sequence used during the study, as suggestion for use in the classroom, by the math teacher
Resumo:
The usual programs for load flow calculation were in general developped aiming the simulation of electric energy transmission, subtransmission and distribution systems. However, the mathematical methods and algorithms used by the formulations were based, in majority, just on the characteristics of the transmittion systems, which were the main concern focus of engineers and researchers. Though, the physical characteristics of these systems are quite different from the distribution ones. In the transmission systems, the voltage levels are high and the lines are generally very long. These aspects contribute the capacitive and inductive effects that appear in the system to have a considerable influence in the values of the interest quantities, reason why they should be taken into consideration. Still in the transmission systems, the loads have a macro nature, as for example, cities, neiborhoods, or big industries. These loads are, generally, practically balanced, what reduces the necessity of utilization of three-phase methodology for the load flow calculation. Distribution systems, on the other hand, present different characteristics: the voltage levels are small in comparison to the transmission ones. This almost annul the capacitive effects of the lines. The loads are, in this case, transformers, in whose secondaries are connected small consumers, in a sort of times, mono-phase ones, so that the probability of finding an unbalanced circuit is high. This way, the utilization of three-phase methodologies assumes an important dimension. Besides, equipments like voltage regulators, that use simultaneously the concepts of phase and line voltage in their functioning, need a three-phase methodology, in order to allow the simulation of their real behavior. For the exposed reasons, initially was developped, in the scope of this work, a method for three-phase load flow calculation in order to simulate the steady-state behaviour of distribution systems. Aiming to achieve this goal, the Power Summation Algorithm was used, as a base for developping the three phase method. This algorithm was already widely tested and approved by researchers and engineers in the simulation of radial electric energy distribution systems, mainly for single-phase representation. By our formulation, lines are modeled in three-phase circuits, considering the magnetic coupling between the phases; but the earth effect is considered through the Carson reduction. Its important to point out that, in spite of the loads being normally connected to the transformers secondaries, was considered the hypothesis of existence of star or delta loads connected to the primary circuit. To perform the simulation of voltage regulators, a new model was utilized, allowing the simulation of various types of configurations, according to their real functioning. Finally, was considered the possibility of representation of switches with current measuring in various points of the feeder. The loads are adjusted during the iteractive process, in order to match the current in each switch, converging to the measured value specified by the input data. In a second stage of the work, sensibility parameters were derived taking as base the described load flow, with the objective of suporting further optimization processes. This parameters are found by calculating of the partial derivatives of a variable in respect to another, in general, voltages, losses and reactive powers. After describing the calculation of the sensibility parameters, the Gradient Method was presented, using these parameters to optimize an objective function, that will be defined for each type of study. The first one refers to the reduction of technical losses in a medium voltage feeder, through the installation of capacitor banks; the second one refers to the problem of correction of voltage profile, through the instalation of capacitor banks or voltage regulators. In case of the losses reduction will be considered, as objective function, the sum of the losses in all the parts of the system. To the correction of the voltage profile, the objective function will be the sum of the square voltage deviations in each node, in respect to the rated voltage. In the end of the work, results of application of the described methods in some feeders are presented, aiming to give insight about their performance and acuity
Resumo:
The management of water resources in the river basin level, as it defines the Law nº 9433/97, requires the effective knowledge of the processes of hydrological basin, resulting from studies based on consistent series of hydrological data that reflect the characteristics of the basin. In this context, the objective of this work was to develop the modeling of catchment basin of the river Jundiaí - RN and carry out the study of attenuation of a flood of the dam Tabatinga, by means of a monitoring project of hydrological data and climatology of the basin, with a view to promoting the development of research activities by applying methodologies unified and appropriate for the assessment of hydrological studies in the transition region of the semiarid and the forest zone on the coast of Rio Grande do Norte. For the study of the hydrological characteristics of the basin was conducted the automatic design of the basin of the river Jundiaí, with the aid of programs of geoprocessing, was adopted a hydrological model daily, the NRCS, which is a model determined and concentrated. For the use of this model was necessary to determine some parameters that are used in this model, as the Curve Number. Having in mind that this is the first study that is being conducted in the basin with the employment of this model, it was made sensitivity analysis of the results of this model from the adoption of different values of CN, situated within a range appropriate to the conditions of use, occupation and the nature of the soil of this basin. As the objective of this study was also developing a simulation model of the operation of the Tabatinga dam and with this flood control caused in the city of Macaíba, it was developed a mathematical model of fluid balance, developed to be used in Microsoft Excel. The simulation was conducted in two phases: the first step was promoted the water balance daily that allowed the analysis of the sensitivity of the model in relation to the volume of waiting, as well as the determination of the period of greatest discharges daily averages. From this point, it was assumed for the second stage, which was in the determination of the hydrograph of discharges effluent slots, that was determined by means of the fluid balance time, on the basis of the discharges effluents generated by a mathematical equation whose parameters were adjusted according to the hydrograph daily. Through the analyzes it was realized that the dam Tabatinga only has how to carry out the attenuation of floods through the regularization of the volume of waiting, with this there is a loss of approximately 56.5% on storage capacity of this dam, because for causing the attenuation effect of filled the shell of this dam has to remain more than 5m below the level of the sill, representing at least 50.582.927m3. The results obtained with the modeling represents a first step in the direction of improving the level of hydrological information about the behavior of the basins of the semiarid. In order to monitor quantitatively the hydrographic basin of the river Jundiaí will be necessary to install a rain gauge register, next to the Tabatinga dam and a pressure transducer, for regular measurements of flow in the reservoir of the dam. The climatological data will be collected in full automatic weather station installed in Agricultural School Jundiaí
Resumo:
One of the greatest challenges of demography, nowadays, is to obtain estimates of mortality, in a consistent manner, mainly in small areas. The lack of this information, hinders public health actions and leads to impairment of quality of classification of deaths, generating concern on the part of demographers and epidemiologists in obtaining reliable statistics of mortality in the country. In this context, the objective of this work is to obtain estimates of deaths adjustment factors for correction of adult mortality, by States, meso-regions and age groups in the northeastern region, in 2010. The proposal is based on two lines of observation: a demographic one and a statistical one, considering also two areas of coverage in the States of the Northeast region, the meso-regions, as larger areas and counties, as small areas. The methodological principle is to use the General Equation and Balancing demographic method or General Growth Balance to correct the observed deaths, in larger areas (meso-regions) of the states, since they are less prone to breakage of methodological assumptions. In the sequence, it will be applied the statistical empirical Bayesian estimator method, considering as sum of deaths in the meso-regions, the death value corrected by the demographic method, and as reference of observation of smaller area, the observed deaths in small areas (counties). As results of this combination, a smoothing effect on the degree of coverage of deaths is obtained, due to the association with the empirical Bayesian Estimator, and the possibility of evaluating the degree of coverage of deaths by age groups at counties, meso-regions and states levels, with the advantage of estimete adjustment factors, according to the desired level of aggregation. The results grouped by State, point to a significant improvement of the degree of coverage of deaths, according to the combination of the methods with values above 80%. Alagoas (0.88), Bahia (0.90), Ceará (0.90), Maranhão (0.84), Paraíba (0.88), Pernambuco (0.93), Piauí (0.85), Rio Grande do Norte (0.89) and Sergipe (0.92). Advances in the control of the registry information in the health system, linked to improvements in socioeconomic conditions and urbanization of the counties, in the last decade, provided a better quality of information registry of deaths in small areas
Resumo:
The development of oil wells drilling requires additional cares mainly if the drilling is in offshore ultra deep water with low overburden pressure gradients which cause low fracture gradients and, consequently, difficult the well drilling by the reduction of the operational window. To minimize, in the well planning phases, the difficulties faced by the drilling in those sceneries, indirect models are used to estimate fracture gradient that foresees approximate values for leakoff tests. These models generate curves of geopressures that allow detailed analysis of the pressure behavior for the whole well. Most of these models are based on the Terzaghi equation, just differentiating in the determination of the values of rock tension coefficient. This work proposes an alternative method for prediction of fracture pressure gradient based on a geometric correlation that relates the pressure gradients proportionally for a given depth and extrapolates it for the whole well depth, meaning that theses parameters vary in a fixed proportion. The model is based on the application of analytical proportion segments corresponding to the differential pressure related to the rock tension. The study shows that the proposed analytical proportion segments reaches values of fracture gradient with good agreement with those available for leakoff tests in the field area. The obtained results were compared with twelve different indirect models for fracture pressure gradient prediction based on the compacting effect. For this, a software was developed using Matlab language. The comparison was also made varying the water depth from zero (onshore wellbores) to 1500 meters. The leakoff tests are also used to compare the different methods including the one proposed in this work. The presented work gives good results for error analysis compared to other methods and, due to its simplicity, justify its possible application
Resumo:
The ethanol is the most overused psychoactive drug over the world; this fact makes it one of the main substances required in toxicological exams nowadays. The development of an analytical method, adaptation or implementation of a method known, involves a process of validation that estimates its efficiency in the laboratory routine and credibility of the method. The stability is defined as the ability of the sample of material to keep the initial value of a quantitative measure for a defined period within specific limits when stored under defined conditions. This study aimed to evaluate the method of Gas chromatography and study the stability of ethanol in blood samples, considering the variables time and temperature of storage, and the presence of preservative and, with that check if the conditions of conservation and storage used in this study maintain the quality of the sample and preserve the originally amount of analyte present. Blood samples were collected from 10 volunteers to evaluate the method and to study the stability of ethanol. For the evaluation of the method, part of the samples was added to known concentrations of ethanol. In the study of stability, the other side of the pool of blood was placed in two containers: one containing the preservative sodium fluoride 1% and the anticoagulant heparin and the other only heparin, was added ethanol at a concentration of 0.6 g/L, fractionated in two bottles, one being stored at 4ºC (refrigerator) and another at -20ºC (freezer), the tests were performed on the same day (time zero) and after 1, 3, 7, 14, 30 and 60 days of storage. The assessment found the difference in results during storage in relation to time zero. It used the technique of headspace associated with gas chromatography with the FID and capillary column with stationary phase of polyethylene. The best analysis of chromatographic conditions were: temperature of 50ºC (column), 150ºC (jet) and 250ºC (detector), with retention time for ethanol from 9.107 ± 0.026 and the tercbutanol (internal standard) of 8.170 ± 0.081 minutes, the ethanol being separated properly from acetaldehyde, acetone, methanol and 2-propanol, which are potential interfering in the determination of ethanol. The technique showed linearity in the concentration range of 0.01 and 3.2 g/L (0.8051 x + y = 0.6196; r2 = 0.999). The calibration curve showed the following equation of the line: y = x 0.7542 + 0.6545, with a linear correlation coefficient equal to 0.996. The average recovery was 100.2%, the coefficients of variation of accuracy and inter intra test showed values of up to 7.3%, the limit of detection and quantification was 0.01 g/L and showed coefficient of variation within the allowed. The analytical method evaluated in this study proved to be fast, efficient and practical, given the objective of this work satisfactorily. The study of stability has less than 20% difference in the response obtained under the conditions of storage and stipulated period, compared with the response obtained at time zero and at the significance level of 5%, no statistical difference in the concentration of ethanol was observed between analysis. The results reinforce the reliability of the method of gas chromatography and blood samples in search of ethanol, either in the toxicological, forensic, social or clinic
Resumo:
This work aims to describe and analyze the process of the mathematics teacher modernizing in Rio Grande do Norte, in the period from 1950 to 1980. For that, we use as theoretical foundation assumptions of Cultural History and memories of the researchers Maurice Halbwach, Ecléa Bosi and Paul Thompson. As methodological tools, we used bibliographical resources and semi-structured interviews, in order to do a historical reconstruct of the mathematics educational scene of institutions and people who taught mathematics in Rio Grande do Norte, or those who participated in the modernization of the teaching of this subject, recovering their training and its practices in teaching. For the analysis of the bibliographical resources, initially we organized in a systematic way the transcripts of the interviews and documents, which were accumulated during the research, so long our thoughts, returning to the theoretical basis of this research, through questioning of knowledge acquired and that guided the problem of our study. The analysis showed that, important moments to modernize the teaching of mathematics in Rio Grande do Norte happened such: (1) Training Course of Lay Teachers in Rio Grande do Norte, in 1965, (2) Course for Teachers in Normal Schools, in 1971 (3) Satelite Project on Interdisciplinary Advanced Communications (SPIAC) in 1973; (4) Lectures of the teacher Malba Tahan, at Natal, from the end of the 50 s, that could be analyzed through the lessons notes of the teacher Maria Nalva Xavier de Albuquerque and the narrative of teacher Evaldo Rodrigues de Carvalho and (5) Courses of the Campaign for Improvement of Secondary Education and Broadcasting (CISEB). Thereby, the modernization of the school s mathematics teaching in Rio Grande do Norte, in the period from 1950 to 1980, was given mainly by disclosure of the Discovery Method and by the Set Theory contents in Teacher Training Courses
Resumo:
The present work had as principal objective to analyze the, 9th grade students understanding about the solutions of an equation of the 2° degree, using geometric processes of the History of the Mathematics. To do so, the research had as base the elaboration and application of a group of teaching activities, based on Jean Piaget's construtivism. The research consisted of a methodological intervention, that has as subjects the students of a group of 9th grade of the State School José Martins de Vasconcelos, located in the municipal district of Mossoró, Rio Grande do Norte. The intervention was divided in three stages: application of an initial evaluation; development of activities‟ module with emphasis in constructive teaching; and the application of the final evaluation. The data presented in the initial evaluation revealed a low level of the students' understanding with relationship to the calculation of areas of rectangles, resolution of equations of the 1st and 2nd degrees, and they were to subsidize the elaboration of the teaching module. The data collected in the initial evaluation were commented and presented under descriptive statistics form. The results of the final evaluation were analyzed under the qualitative point of view, based on Richard Skemp's theory on the understanding of mathematical concepts. The general results showed a qualitative increase with relationship to the students' understanding on the mathematical concepts approached in the intervention. Such results indicate that a methodology using the previous student‟s knowledge and the development of teaching activities, learning in the construtivist theory, make possible an understanding on the part of the students concerning the thematic proposal
Resumo:
Most algorithms for state estimation based on the classical model are just adequate for use in transmission networks. Few algorithms were developed specifically for distribution systems, probably because of the little amount of data available in real time. Most overhead feeders possess just current and voltage measurements at the middle voltage bus-bar at the substation. In this way, classical algorithms are of difficult implementation, even considering off-line acquired data as pseudo-measurements. However, the necessity of automating the operation of distribution networks, mainly in regard to the selectivity of protection systems, as well to implement possibilities of load transfer maneuvers, is changing the network planning policy. In this way, some equipments incorporating telemetry and command modules have been installed in order to improve operational features, and so increasing the amount of measurement data available in real-time in the System Operation Center (SOC). This encourages the development of a state estimator model, involving real-time information and pseudo-measurements of loads, that are built from typical power factors and utilization factors (demand factors) of distribution transformers. This work reports about the development of a new state estimation method, specific for radial distribution systems. The main algorithm of the method is based on the power summation load flow. The estimation is carried out piecewise, section by section of the feeder, going from the substation to the terminal nodes. For each section, a measurement model is built, resulting in a nonlinear overdetermined equations set, whose solution is achieved by the Gaussian normal equation. The estimated variables of a section are used as pseudo-measurements for the next section. In general, a measurement set for a generic section consists of pseudo-measurements of power flows and nodal voltages obtained from the previous section or measurements in real-time, if they exist -, besides pseudomeasurements of injected powers for the power summations, whose functions are the load flow equations, assuming that the network can be represented by its single-phase equivalent. The great advantage of the algorithm is its simplicity and low computational effort. Moreover, the algorithm is very efficient, in regard to the accuracy of the estimated values. Besides the power summation state estimator, this work shows how other algorithms could be adapted to provide state estimation of middle voltage substations and networks, namely Schweppes method and an algorithm based on current proportionality, that is usually adopted for network planning tasks. Both estimators were implemented not only as alternatives for the proposed method, but also looking for getting results that give support for its validation. Once in most cases no power measurement is performed at beginning of the feeder and this is required for implementing the power summation estimations method, a new algorithm for estimating the network variables at the middle voltage bus-bar was also developed
Resumo:
Conselho Nacional de Desenvolvimento Científico e Tecnológico
Resumo:
This work consists on the theoretical and numerical analysis of some properties of circular microstrip patch antennas on isotropic and uniaxial anisotropic substrates. For this purpose, a full wave analysis is performed, using Hertz Vector Potentials method in the Hankel Transform domain. In the numerical analysis, the moment method is also used in order to determine some characteristics of the antenna, such as: resonant frequency and radiation pattern. The definition of Hertz potentials in the Hankel domain is used in association with Maxwell´s equations and the boundary conditions of the structures to obtain the Green´s functions, relating the components of the current density on the patch and the tangential electric field components. Then, the Galerkin method is used to generate a matrix equation whose nontrivial solution is the complex resonant frequency of the structure. In the analysis, a microstrip antenna with only one isotropic dielectric layer is initially considered. For this structure, the effect of using superconductor patches is also analyzed. An analysis of a circular microstrip antenna on an uniaxial anisotropic dielectric layer is performed, using the Hertz vector potentials oriented along the optical axis of the material, that is perpendicular to the microstrip ground plane. Afterwards, the circular microstrip antenna using two uniaxial anisotropic dielectric layers is investigated, considering the particular case in which the inferior layer is filled by air. In this study, numerical results for resonant frequency and radiation pattern for circular microstrip antennas on isotropic and uniaxial anisotropic substrates are presented and compared with measured and calculated results found in the literature
Resumo:
This work proposes a computer simulator for sucker rod pumped vertical wells. The simulator is able to represent the dynamic behavior of the systems and the computation of several important parameters, allowing the easy visualization of several pertinent phenomena. The use of the simulator allows the execution of several tests at lower costs and shorter times, than real wells experiments. The simulation uses a model based on the dynamic behavior of the rod string. This dynamic model is represented by a second order partial differencial equation. Through this model, several common field situations can be verified. Moreover, the simulation includes 3D animations, facilitating the physical understanding of the process, due to a better visual interpretation of the phenomena. Another important characteristic is the emulation of the main sensors used in sucker rod pumping automation. The emulation of the sensors is implemented through a microcontrolled interface between the simulator and the industrial controllers. By means of this interface, the controllers interpret the simulator as a real well. A "fault module" was included in the simulator. This module incorporates the six more important faults found in sucker rod pumping. Therefore, the analysis and verification of these problems through the simulator, allows the user to identify such situations that otherwise could be observed only in the field. The simulation of these faults receives a different treatment due to the different boundary conditions imposed to the numeric solution of the problem. Possible applications of the simulator are: the design and analysis of wells, training of technicians and engineers, execution of tests in controllers and supervisory systems, and validation of control algorithms
Resumo:
This work proposes a new technique for phasor estimation applied in microprocessor numerical relays for distance protection of transmission lines, based on the recursive least squares method and called least squares modified random walking. The phasor estimation methods have compromised their performance, mainly due to the DC exponential decaying component present in fault currents. In order to reduce the influence of the DC component, a Morphological Filter (FM) was added to the method of least squares and previously applied to the process of phasor estimation. The presented method is implemented in MATLABr and its performance is compared to one-cycle Fourier technique and conventional phasor estimation, which was also based on least squares algorithm. The methods based on least squares technique used for comparison with the proposed method were: forgetting factor recursive, covariance resetting and random walking. The techniques performance analysis were carried out by means of signals synthetic and signals provided of simulations on the Alternative Transient Program (ATP). When compared to other phasor estimation methods, the proposed method showed satisfactory results, when it comes to the estimation speed, the steady state oscillation and the overshoot. Then, the presented method performance was analyzed by means of variations in the fault parameters (resistance, distance, angle of incidence and type of fault). Through this study, the results did not showed significant variations in method performance. Besides, the apparent impedance trajectory and estimated distance of the fault were analysed, and the presented method showed better results in comparison to one-cycle Fourier algorithm
Resumo:
Ceramics with porous cellular structure, called ceramic foams, have a potential use in several applications, such as: thermal insulation, catalyst supports, filters, and others. Among these techniques to obtain porous ceramics the replication method is an important process. This method consists of impregnation of a sponge (usually polymer) with ceramic slurry, followed by a heat treatment, which will happen the decomposition of organic material and sintering the ceramic material, resulting in a ceramic structure which is a replica of impregnated sponge. Knowledge of the mechanical properties of these ceramics is important for these materials can be used commercially. Gibson and Ashby developed a mathematical model to describe the mechanical behavior of cellular solids. This model wasn´t for describing the ceramics behavior produced by the replica method, because it doesn´t consider the defects from this type of processing. In this study were researched mechanical behavior of porous alumina ceramics obtained by the replica method and proposed modifications to the model of Gibson and Ashby to accommodate this material. The polymer sponge used in processing was characterized by thermogravimetric analysis and scanning electron microscopy. The materials obtained after sintering were characterized by mechanical strength tests on 4-point bending and compression, density and porosity and by scanning electron microscopy. From these results it was evaluated the mechanical strength behavior compared to Gibson and Ashby model for solid cellular structure and was proposed a correction of this model through a factor related to struts integrity degree, which consider fissures present in the structure of these materials besides defects geometry within the struts
Resumo:
The topology optimization problem characterize and determine the optimum distribution of material into the domain. In other words, after the definition of the boundary conditions in a pre-established domain, the problem is how to distribute the material to solve the minimization problem. The objective of this work is to propose a competitive formulation for optimum structural topologies determination in 3D problems and able to provide high-resolution layouts. The procedure combines the Galerkin Finite Elements Method with the optimization method, looking for the best material distribution along the fixed domain of project. The layout topology optimization method is based on the material approach, proposed by Bendsoe & Kikuchi (1988), and considers a homogenized constitutive equation that depends only on the relative density of the material. The finite element used for the approach is a four nodes tetrahedron with a selective integration scheme, which interpolate not only the components of the displacement field but also the relative density field. The proposed procedure consists in the solution of a sequence of layout optimization problems applied to compliance minimization problems and mass minimization problems under local stress constraint. The microstructure used in this procedure was the SIMP (Solid Isotropic Material with Penalty). The approach reduces considerably the computational cost, showing to be efficient and robust. The results provided a well defined structural layout, with a sharpness distribution of the material and a boundary condition definition. The layout quality was proporcional to the medium size of the element and a considerable reduction of the project variables was observed due to the tetrahedrycal element