944 resultados para Real Root Isolation Methods


Relevância:

30.00% 30.00%

Publicador:

Resumo:

The aim of this thesis is to propose a novel control method for teleoperated electrohydraulic servo systems that implements a reliable haptic sense between the human and manipulator interaction, and an ideal position control between the manipulator and the task environment interaction. The proposed method has the characteristics of a universal technique independent of the actual control algorithm and it can be applied with other suitable control methods as a real-time control strategy. The motivation to develop this control method is the necessity for a reliable real-time controller for teleoperated electrohydraulic servo systems that provides highly accurate position control based on joystick inputs with haptic capabilities. The contribution of the research is that the proposed control method combines a directed random search method and a real-time simulation to develop an intelligent controller in which each generation of parameters is tested on-line by the real-time simulator before being applied to the real process. The controller was evaluated on a hydraulic position servo system. The simulator of the hydraulic system was built based on Markov chain Monte Carlo (MCMC) method. A Particle Swarm Optimization algorithm combined with the foraging behavior of E. coli bacteria was utilized as the directed random search engine. The control strategy allows the operator to be plugged into the work environment dynamically and kinetically. This helps to ensure the system has haptic sense with high stability, without abstracting away the dynamics of the hydraulic system. The new control algorithm provides asymptotically exact tracking of both, the position and the contact force. In addition, this research proposes a novel method for re-calibration of multi-axis force/torque sensors. The method makes several improvements to traditional methods. It can be used without dismantling the sensor from its application and it requires smaller number of standard loads for calibration. It is also more cost efficient and faster in comparison to traditional calibration methods. The proposed method was developed in response to re-calibration issues with the force sensors utilized in teleoperated systems. The new approach aimed to avoid dismantling of the sensors from their applications for applying calibration. A major complication with many manipulators is the difficulty accessing them when they operate inside a non-accessible environment; especially if those environments are harsh; such as in radioactive areas. The proposed technique is based on design of experiment methodology. It has been successfully applied to different force/torque sensors and this research presents experimental validation of use of the calibration method with one of the force sensors which method has been applied to.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The increasing presence of products derived from genetically modified (GM) plants in human and animal diets has led to the development of detection methods to distinguish biotechnology-derived foods from conventional ones. The conventional and real-time PCR have been used, respectively, to detect and quantify GM residues in highly processed foods. DNA extraction is a critical step during the analysis process. Some factors such as DNA degradation, matrix effects, and the presence of PCR inhibitors imply that a detection or quantification limit, established for a given method, is restricted to a matrix used during validation and cannot be projected to any other matrix outside the scope of the method. In Brazil, sausage samples were the main class of processed products in which Roundup Ready® (RR) soybean residues were detected. Thus, the validation of methodologies for the detection and quantification of those residues is absolutely necessary. Sausage samples were submitted to two different methods of DNA extraction: modified Wizard and the CTAB method. The yield and quality were compared for both methods. DNA samples were analyzed by conventional and real-time PCR for the detection and quantification of Roundup Ready® soybean in the samples. At least 200 ng of total sausage DNA was necessary for a reliable quantification. Reactions containing DNA amounts below this value led to large variations on the expected GM percentage value. In conventional PCR, the detection limit varied from 1.0 to 500 ng, depending on the GM soybean content in the sample. The precision, performance, and linearity were relatively high indicating that the method used for analysis was satisfactory.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Availability of analytical standards is a critical aspect in developing methods for quantitative analysis of anthocyanins. The anthocyanins cyanidin-3-O-glucoside and cyanidin-3-O-rutinoside were isolated from samples of freeze-dried açaí (Euterpe oleraceae Mart.), which is a round and purple well-known palm fruit in Brazil, and then used as standards. The isolation of the anthocyanins was performed by High Performance Liquid Chromatography (HPLC), using an adapted six-channel selection valve. The identification of anthocyanin pigments in açaí was based on mass spectrometric data for molecular ions and MS-MS product ions and on previous published data. After the collection procedure, standards with a high purity grade were obtained and an external standard curve of each anthocyanin was plotted.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Active magnetic bearing is a type of bearing which uses magnetic field to levitate the rotor. These bearings require continuous control of the currents in electromagnets and data from position of the rotor and the measured current from electromagnets. Because of this different identification methods can be implemented with no additional hardware. In this thesis the focus was to implement and test identification methods for active magnetic bearing system and to update the rotor model. Magnetic center calibration is a method used to locate the magnetic center of the rotor. Rotor model identification is an identification method used to identify the rotor model. Rotor model update is a method used to update the rotor model based on identification data. These methods were implemented and tested with a real machine where rotor was levitated with active magnetic bearings and the functionality of the methods was ensured. Methods were developed with further extension in mind and also with the possibility to apply them for different machines with ease.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The strongest wish of the customer concerning chemical pulp features is consistent, uniform quality. Variation may be controlled and reduced by using statistical methods. However, studies addressing the application and benefits of statistical methods in forest product sector are scarce. Thus, the customer wish is the root cause of the motivation behind this dissertation. The research problem addressed by this dissertation is that companies in the chemical forest product sector require new knowledge for improving their utilization of statistical methods. To gain this new knowledge, the research problem is studied from five complementary viewpoints – challenges and success factors, organizational learning, problem solving, economic benefit, and statistical methods as management tools. The five research questions generated on the basis of these viewpoints are answered in four research papers, which are case studies based on empirical data collection. This research as a whole complements the literature dealing with the use of statistical methods in the forest products industry. Practical examples of the application of statistical process control, case-based reasoning, the cross-industry standard process for data mining, and performance measurement methods in the context of chemical forest products manufacturing are brought to the public knowledge of the scientific community. The benefit of the application of these methods is estimated or demonstrated. The purpose of this dissertation is to find pragmatic ideas for companies in the chemical forest product sector in order for them to improve their utilization of statistical methods. The main practical implications of this doctoral dissertation can be summarized in four points: 1. It is beneficial to reduce variation in chemical forest product manufacturing processes 2. Statistical tools can be used to reduce this variation 3. Problem-solving in chemical forest product manufacturing processes can be intensified through the use of statistical methods 4. There are certain success factors and challenges that need to be addressed when implementing statistical methods

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Real exchange rate, exchange rate misalignment and economic growth in Brazil: 1994-2007. In this article we argue that the Brazilian economy presented in the period 1994-2007 a tendency of real exchange rate appreciation with respect to its equilibrium value, mainly from 2005. This exchange rate misalignment has worked to reduce the growth of Brazilian economy and is the root of the re-emergence of current account deficits.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This dissertation describes an approach for developing a real-time simulation for working mobile vehicles based on multibody modeling. The use of multibody modeling allows comprehensive description of the constrained motion of the mechanical systems involved and permits real-time solving of the equations of motion. By carefully selecting the multibody formulation method to be used, it is possible to increase the accuracy of the multibody model while at the same time solving equations of motion in real-time. In this study, a multibody procedure based on semi-recursive and augmented Lagrangian methods for real-time dynamic simulation application is studied in detail. In the semirecursive approach, a velocity transformation matrix is introduced to describe the dependent coordinates into relative (joint) coordinates, which reduces the size of the generalized coordinates. The augmented Lagrangian method is based on usage of global coordinates and, in that method, constraints are accounted using an iterative process. A multibody system can be modelled as either rigid or flexible bodies. When using flexible bodies, the system can be described using a floating frame of reference formulation. In this method, the deformation mode needed can be obtained from the finite element model. As the finite element model typically involves large number of degrees of freedom, reduced number of deformation modes can be obtained by employing model order reduction method such as Guyan reduction, Craig-Bampton method and Krylov subspace as shown in this study The constrained motion of the working mobile vehicles is actuated by the force from the hydraulic actuator. In this study, the hydraulic system is modeled using lumped fluid theory, in which the hydraulic circuit is divided into volumes. In this approach, the pressure wave propagation in the hoses and pipes is neglected. The contact modeling is divided into two stages: contact detection and contact response. Contact detection determines when and where the contact occurs, and contact response provides the force acting at the collision point. The friction between tire and ground is modelled using the LuGre friction model, which describes the frictional force between two surfaces. Typically, the equations of motion are solved in the full matrices format, where the sparsity of the matrices is not considered. Increasing the number of bodies and constraint equations leads to the system matrices becoming large and sparse in structure. To increase the computational efficiency, a technique for solution of sparse matrices is proposed in this dissertation and its implementation demonstrated. To assess the computing efficiency, augmented Lagrangian and semi-recursive methods are implemented employing a sparse matrix technique. From the numerical example, the results show that the proposed approach is applicable and produced appropriate results within the real-time period.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The paper analyzes the shortcomings of the Real Plan combining political science and economics, considering evidences on difficulties in implementing a complete plan of deindexation, namely: the distributional conflict, the corporatist relations between the State and society and the bureaucratic isolation of a highly specialized technocracy. The announcement of the plan triggered defensive reactions, fueling distributive tensions during the URV period, forcing the economic team to take measures that contradicted the overall guidelines followed. The persistence of indexation mechanisms indicate the resulting obstacles in seeking a reform of the State through a stabilization plan, even though the latter has the former as a precondition.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Currently, laser scribing is growing material processing method in the industry. Benefits of laser scribing technology are studied for example for improving an efficiency of solar cells. Due high-quality requirement of the fast scribing process, it is important to monitor the process in real time for detecting possible defects during the process. However, there is a lack of studies of laser scribing real time monitoring. Commonly used monitoring methods developed for other laser processes such a laser welding, are sufficient slow and existed applications cannot be implemented in fast laser scribing monitoring. The aim of this thesis is to find a method for laser scribing monitoring with a high-speed camera and evaluate reliability and performance of the developed monitoring system with experiments. The laser used in experiments is an IPG ytterbium pulsed fiber laser with 20 W maximum average power and Scan head optics used in the laser is Scanlab’s Hurryscan 14 II with an f100 tele-centric lens. The camera was connected to laser scanner using camera adapter to follow the laser process. A powerful fully programmable industrial computer was chosen for executing image processing and analysis. Algorithms for defect analysis, which are based on particle analysis, were developed using LabVIEW system design software. The performance of the algorithms was analyzed by analyzing a non-moving image from the scribing line with resolution 960x20 pixel. As a result, the maximum analysis speed was 560 frames per second. Reliability of the algorithm was evaluated by imaging scribing path with a variable number of defects 2000 mm/s when the laser was turned off and image analysis speed was 430 frames per second. The experiment was successful and as a result, the algorithms detected all defects from the scribing path. The final monitoring experiment was performed during a laser process. However, it was challenging to get active laser illumination work with the laser scanner due physical dimensions of the laser lens and the scanner. For reliable error detection, the illumination system is needed to be replaced.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A simple method was developed for treating corn seeds with oxamyl. It involved soaking the seeds to ensure oxamyl uptake, centrifugation to draw off excess solution, and drying under a stream of air to prevent the formation of fungus. The seeds were found to have an even distribution of oxamyl. Seeds remained fungus-free even 12 months after treatment. The highest nonphytotoxic treatment level was obtained by using a 4.00 mg/mL oxamyl solution. Extraction methods for the determination of oxamyl (methyl-N'N'-dimethyl-N-[(methylcarbamoyl)oxy]-l-thiooxamimidate), its oxime (methyl-N',N'-dimethyl-N-hydroxy-1-thiooxamimidate), and DMCF (N,N-dimethyl-1-cyanoformanade) in seed" root, and soil were developed. Seeds were processed by homogenizing, then shaking in methanol. Significantly more oxamyl was extracted from hydrated seeds as opposed to dry seeds. Soils were extracted by tumbling in methanol; recoveries range~ from 86 - 87% for oxamyl. Root was extracted to 93% efficiency for oxamyl by homogenizing the tissue in methanol. NucharAttaclay column cleanup afforded suitable extracts for analysis by RP-HPLC on a C18 column and UV detection at 254 nm. In the degradation study, oxamyl was found to dissipate from the seed down into the soil. It was also detected in the root. Oxime was detected in both the seed and soil, but not in the root. DMCF was detected in small amounts only in the seed.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We extend the class of M-tests for a unit root analyzed by Perron and Ng (1996) and Ng and Perron (1997) to the case where a change in the trend function is allowed to occur at an unknown time. These tests M(GLS) adopt the GLS detrending approach of Dufour and King (1991) and Elliott, Rothenberg and Stock (1996) (ERS). Following Perron (1989), we consider two models : one allowing for a change in slope and the other for both a change in intercept and slope. We derive the asymptotic distribution of the tests as well as that of the feasible point optimal tests PT(GLS) suggested by ERS. The asymptotic critical values of the tests are tabulated. Also, we compute the non-centrality parameter used for the local GLS detrending that permits the tests to have 50% asymptotic power at that value. We show that the M(GLS) and PT(GLS) tests have an asymptotic power function close to the power envelope. An extensive simulation study analyzes the size and power in finite samples under various methods to select the truncation lag for the autoregressive spectral density estimator. An empirical application is also provided.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper employs the one-sector Real Business Cycle model as a testing ground for four different procedures to estimate Dynamic Stochastic General Equilibrium (DSGE) models. The procedures are: 1 ) Maximum Likelihood, with and without measurement errors and incorporating Bayesian priors, 2) Generalized Method of Moments, 3) Simulated Method of Moments, and 4) Indirect Inference. Monte Carlo analysis indicates that all procedures deliver reasonably good estimates under the null hypothesis. However, there are substantial differences in statistical and computational efficiency in the small samples currently available to estimate DSGE models. GMM and SMM appear to be more robust to misspecification than the alternative procedures. The implications of the stochastic singularity of DSGE models for each estimation method are fully discussed.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Contexte. Les études cas-témoins sont très fréquemment utilisées par les épidémiologistes pour évaluer l’impact de certaines expositions sur une maladie particulière. Ces expositions peuvent être représentées par plusieurs variables dépendant du temps, et de nouvelles méthodes sont nécessaires pour estimer de manière précise leurs effets. En effet, la régression logistique qui est la méthode conventionnelle pour analyser les données cas-témoins ne tient pas directement compte des changements de valeurs des covariables au cours du temps. Par opposition, les méthodes d’analyse des données de survie telles que le modèle de Cox à risques instantanés proportionnels peuvent directement incorporer des covariables dépendant du temps représentant les histoires individuelles d’exposition. Cependant, cela nécessite de manipuler les ensembles de sujets à risque avec précaution à cause du sur-échantillonnage des cas, en comparaison avec les témoins, dans les études cas-témoins. Comme montré dans une étude de simulation précédente, la définition optimale des ensembles de sujets à risque pour l’analyse des données cas-témoins reste encore à être élucidée, et à être étudiée dans le cas des variables dépendant du temps. Objectif: L’objectif général est de proposer et d’étudier de nouvelles versions du modèle de Cox pour estimer l’impact d’expositions variant dans le temps dans les études cas-témoins, et de les appliquer à des données réelles cas-témoins sur le cancer du poumon et le tabac. Méthodes. J’ai identifié de nouvelles définitions d’ensemble de sujets à risque, potentiellement optimales (le Weighted Cox model and le Simple weighted Cox model), dans lesquelles différentes pondérations ont été affectées aux cas et aux témoins, afin de refléter les proportions de cas et de non cas dans la population source. Les propriétés des estimateurs des effets d’exposition ont été étudiées par simulation. Différents aspects d’exposition ont été générés (intensité, durée, valeur cumulée d’exposition). Les données cas-témoins générées ont été ensuite analysées avec différentes versions du modèle de Cox, incluant les définitions anciennes et nouvelles des ensembles de sujets à risque, ainsi qu’avec la régression logistique conventionnelle, à des fins de comparaison. Les différents modèles de régression ont ensuite été appliqués sur des données réelles cas-témoins sur le cancer du poumon. Les estimations des effets de différentes variables de tabac, obtenues avec les différentes méthodes, ont été comparées entre elles, et comparées aux résultats des simulations. Résultats. Les résultats des simulations montrent que les estimations des nouveaux modèles de Cox pondérés proposés, surtout celles du Weighted Cox model, sont bien moins biaisées que les estimations des modèles de Cox existants qui incluent ou excluent simplement les futurs cas de chaque ensemble de sujets à risque. De plus, les estimations du Weighted Cox model étaient légèrement, mais systématiquement, moins biaisées que celles de la régression logistique. L’application aux données réelles montre de plus grandes différences entre les estimations de la régression logistique et des modèles de Cox pondérés, pour quelques variables de tabac dépendant du temps. Conclusions. Les résultats suggèrent que le nouveau modèle de Cox pondéré propose pourrait être une alternative intéressante au modèle de régression logistique, pour estimer les effets d’expositions dépendant du temps dans les études cas-témoins

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Most panel unit root tests are designed to test the joint null hypothesis of a unit root for each individual series in a panel. After a rejection, it will often be of interest to identify which series can be deemed to be stationary and which series can be deemed nonstationary. Researchers will sometimes carry out this classification on the basis of n individual (univariate) unit root tests based on some ad hoc significance level. In this paper, we demonstrate how to use the false discovery rate (FDR) in evaluating I(1)=I(0) classifications based on individual unit root tests when the size of the cross section (n) and time series (T) dimensions are large. We report results from a simulation experiment and illustrate the methods on two data sets.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Les difficultés croissantes de l’enseignement supérieur en matière d’accessibilité aux formations et de disponibilité de filières de qualité, dans les pays en voie de développement de l’Afrique de l’Ouest notamment, conjuguées avec le développement vertigineux des technologies de l’information et de la communication (TIC), suscitent un grand espoir de faire de la formation à distance une solution alternative crédible des formations présentielles (OCDE, 2006). Or, si la littérature s’accorde à reconnaitre aux TIC et l’interactivité qu’elles procurent des facteurs favorisant l’apprentissage (Karsenti, 2006), la réalité du terrain éducatif lui impose de reconnaitre que non seulement la révolution de la formation ouverte et à distance (FOAD) n’est pas encore d’actualité (OCDE, 2006), mais qu’elle ne le sera que si, pour faire face à la distance transactionnelle, plus accrue en formation à distance, l’apprenant ne se contente plus d’apprendre, mais d’apprendre à apprendre, ce qui exige de lui des compétences d’autonomie. Or, malgré des décennies d’intérêt et d’investissement de la recherche, le développement de l’autonomie sur le terrain reste toujours marginal, les débats philosophiques ayant pris le pas sur la quête de solutions pratiques (Albero ,2003). La question de savoir comment les éducateurs de la FOAD utilisent les solutions existantes, censées favoriser l’autonomie de l’apprenant, telles certaines formes de tutorat et de travail de groupes, n’est pas sans intérêt, puisqu’elle permet de mieux comprendre le terrain cible et sa part de responsabilité dans cet insuccès de l’autonomie. S’inscrivant en droite ligne des travaux d’Albero (2003), la présente étude organise les principes d’autonomie suivant un cadre conceptuel privilégiant l’action et le développement, selon une dimension dynamique symbolisant l’importance du soutien à accorder à l’apprenant, une dimension topologique indiquant la nécessité pour ce soutien de prendre en compte les différents aspects sur lesquels l’apprenant peut exercer son autonomie et une dimension chronologique exprimant l’importance du désétayage. De façon pratique, cette étude, démarrée en 2009 dans le contexte de la FOAD du 2IE (Institut International des Ingénieurs de l’Eau et de l’Environnement), sis à Ouagadougou au Burkina Faso, s’organise en trois articles : le premier tente de comprendre si les conditions d’efficacité de la FOAD, selon les apprenants, formulent un besoin d’apprentissage en autonomie; le second tente de comprendre, à partir des propos des tuteurs, si l’encadrement des apprenants respecte les principes d’autonomie; enfin, le troisième article s’est penché, sur la base des intentions exprimées par les concepteurs, sur le respect des principes d’autonomie par les cours. L’éloignement des apprenants et l’objectif de comprendre leurs perceptions de la FOAD, nous ont fait préférer une approche de recherche de type mixte, à la fois qualitative pour mieux comprendre leur perception (Karsenti & Savoie-Zajc, 2004) et quantitative, pour tenir compte de leur éloignement. Pour la perception des éducateurs, nous avons opté pour une approche qualitative/descriptive plus appropriée dès que l’étude vise la compréhension du phénomène social abordé (Karsenti & Savoie-Zajc, 2004). Des résultats obtenus, la perception des 62 apprenants sur les 170 initialement invités par courriel à répondre au questionnaire, semble confirmer l’autonomie de l’apprenant comme une condition de leur réussite en FOAD. La demande des apprenants pour un soutien accru, malgré l’effort actuellement consenti par l’encadrement, effort reconnu par les apprenants eux-mêmes et qu’attestent les 10 tuteurs interviewés sur une quarantaine au 2IE, devrait, toutefois, inviter à la réflexion, puisque ne s’opposant pas réellement à un apprentissage en hétéronomie. À l’analyse, il apparait que l’insatisfaction des apprenants s’expliquerait par la non-prise en compte des différents aspects susceptibles d’influencer leur apprentissage. De plus, en nous référant aux entretiens avec les 11 concepteurs de cours sur un total de 30, il apparait que, bien que conscients de la nécessité d’adapter les cours au contexte de la FOAD, ni la modularité des contenus, ni la flexibilité de la structure des cours ne semblent être prises en compte. Au final, l’étude révèle l’urgence de lutter contre les habitudes acquises en formation présentielle et la nécessité d’employer des pédagogues professionnels formés pour une pédagogie d’autonomisation. Encore faudrait-il que l’autonomie soit véritablement consacrée par la littérature comme une praxis pour signifier qu’elle n’a d’autre fin qu’elle-même, et non comme une poiesis, pour dire que l’autonomie vise une production et cesserait dès que son objectif est atteint.