933 resultados para Visualization Of Interval Methods
Resumo:
Most soybean pathogens are seed transmitted, deserving emphasis the fungus Sclerotinia sclerotiorum, which has been presenting worrying levels of field incidence in some soybean cropping areas in several Brazilian states. The objective of this study was to verify the efficiency of different methods for detecting S. sclerotiorum on soybean seeds artificially infected in the laboratory and from field production areas with a historical disease incidence. Seed samples of seven different cultivars collected from naturally infested fields, and one seed sample artificially inoculated in the laboratory were used. The following detection methods recommended in the literature were compared: Blotter test at 7 ºC, 14 ºC, and 21 ºC; Rolled Paper; and Neon-S. Results demonstrated that these methods showed no repeatability and had a low sensitivity for detecting the pathogen in seeds from areas with disease incidence. They were effective, however, for its detection on artificially inoculated seeds. In the Blotter test method at 7 ºC, there was a lower incidence of other fungi considered undesirable during seed analysis.
Resumo:
Intelligence from a human source, that is falsely thought to be true, is potentially more harmful than a total lack of it. The veracity assessment of the gathered intelligence is one of the most important phases of the intelligence process. Lie detection and veracity assessment methods have been studied widely but a comprehensive analysis of these methods’ applicability is lacking. There are some problems related to the efficacy of lie detection and veracity assessment. According to a conventional belief an almighty lie detection method, that is almost 100% accurate and suitable for any social encounter, exists. However, scientific studies have shown that this is not the case, and popular approaches are often over simplified. The main research question of this study was: What is the applicability of veracity assessment methods, which are reliable and are based on scientific proof, in terms of the following criteria? o Accuracy, i.e. probability of detecting deception successfully o Ease of Use, i.e. easiness to apply the method correctly o Time Required to apply the method reliably o No Need for Special Equipment o Unobtrusiveness of the method In order to get an answer to the main research question, the following supporting research questions were answered first: What kinds of interviewing and interrogation techniques exist and how could they be used in the intelligence interview context, what kinds of lie detection and veracity assessment methods exist that are reliable and are based on scientific proof and what kind of uncertainty and other limitations are included in these methods? Two major databases, Google Scholar and Science Direct, were used to search and collect existing topic related studies and other papers. After the search phase, the understanding of the existing lie detection and veracity assessment methods was established through a meta-analysis. Multi Criteria Analysis utilizing Analytic Hierarchy Process was conducted to compare scientifically valid lie detection and veracity assessment methods in terms of the assessment criteria. In addition, a field study was arranged to get a firsthand experience of the applicability of different lie detection and veracity assessment methods. The Studied Features of Discourse and the Studied Features of Nonverbal Communication gained the highest ranking in overall applicability. They were assessed to be the easiest and fastest to apply, and to have required temporal and contextual sensitivity. The Plausibility and Inner Logic of the Statement, the Method for Assessing the Credibility of Evidence and the Criteria Based Content Analysis were also found to be useful, but with some limitations. The Discourse Analysis and the Polygraph were assessed to be the least applicable. Results from the field study support these findings. However, it was also discovered that the most applicable methods are not entirely troublefree either. In addition, this study highlighted that three channels of information, Content, Discourse and Nonverbal Communication, can be subjected to veracity assessment methods that are scientifically defensible. There is at least one reliable and applicable veracity assessment method for each of the three channels. All of the methods require disciplined application and a scientific working approach. There are no quick gains if high accuracy and reliability is desired. Since most of the current lie detection studies are concentrated around a scenario, where roughly half of the assessed people are totally truthful and the other half are liars who present a well prepared cover story, it is proposed that in future studies lie detection and veracity assessment methods are tested against partially truthful human sources. This kind of test setup would highlight new challenges and opportunities for the use of existing and widely studied lie detection methods, as well as for the modern ones that are still under development.
Resumo:
Soit $p_1 = 2, p_2 = 3, p_3 = 5,\ldots$ la suite des nombres premiers, et soient $q \ge 3$ et $a$ des entiers premiers entre eux. R\'ecemment, Daniel Shiu a d\'emontr\'e une ancienne conjecture de Sarvadaman Chowla. Ce dernier a conjectur\'e qu'il existe une infinit\'e de couples $p_n,p_{n+1}$ de premiers cons\'ecutifs tels que $p_n \equiv p_{n+1} \equiv a \bmod q$. Fixons $\epsilon > 0$. Une r\'ecente perc\'ee majeure, de Daniel Goldston, J\`anos Pintz et Cem Y{\i}ld{\i}r{\i}m, a \'et\'e de d\'emontrer qu'il existe une suite de nombres r\'eels $x$ tendant vers l'infini, tels que l'intervalle $(x,x+\epsilon\log x]$ contienne au moins deux nombres premiers $\equiv a \bmod q$. \'Etant donn\'e un couple de nombres premiers $\equiv a \bmod q$ dans un tel intervalle, il pourrait exister un nombre premier compris entre les deux qui n'est pas $\equiv a \bmod q$. On peut d\'eduire que soit il existe une suite de r\'eels $x$ tendant vers l'infini, telle que $(x,x+\epsilon\log x]$ contienne un triplet $p_n,p_{n+1},p_{n+2}$ de nombres premiers cons\'ecutifs, soit il existe une suite de r\'eels $x$, tendant vers l'infini telle que l'intervalle $(x,x+\epsilon\log x]$ contienne un couple $p_n,p_{n+1}$ de nombres premiers tel que $p_n \equiv p_{n+1} \equiv a \bmod q$. On pense que les deux \'enonc\'es sont vrais, toutefois on peut seulement d\'eduire que l'un d'entre eux est vrai, sans savoir lequel. Dans la premi\`ere partie de cette th\`ese, nous d\'emontrons que le deuxi\`eme \'enonc\'e est vrai, ce qui fournit une nouvelle d\'emonstration de la conjecture de Chowla. La preuve combine des id\'ees de Shiu et de Goldston-Pintz-Y{\i}ld{\i}r{\i}m, donc on peut consid\'erer que ce r\'esultat est une application de leurs m\'thodes. Ensuite, nous fournirons des bornes inf\'erieures pour le nombre de couples $p_n,p_{n+1}$ tels que $p_n \equiv p_{n+1} \equiv a \bmod q$, $p_{n+1} - p_n < \epsilon\log p_n$, avec $p_{n+1} \le Y$. Sous l'hypoth\`ese que $\theta$, le \og niveau de distribution \fg{} des nombres premiers, est plus grand que $1/2$, Goldston-Pintz-Y{\i}ld{\i}r{\i}m ont r\'eussi \`a d\'emontrer que $p_{n+1} - p_n \ll_{\theta} 1$ pour une infinit\'e de couples $p_n,p_{n+1}$. Sous la meme hypoth\`ese, nous d\'emontrerons que $p_{n+1} - p_n \ll_{q,\theta} 1$ et $p_n \equiv p_{n+1} \equiv a \bmod q$ pour une infinit\'e de couples $p_n,p_{n+1}$, et nous prouverons \'egalement un r\'esultat quantitatif. Dans la deuxi\`eme partie, nous allons utiliser les techniques de Goldston-Pintz-Y{\i}ld{\i}r{\i}m pour d\'emontrer qu'il existe une infinit\'e de couples de nombres premiers $p,p'$ tels que $(p-1)(p'-1)$ est une carr\'e parfait. Ce resultat est une version approximative d'une ancienne conjecture qui stipule qu'il existe une infinit\'e de nombres premiers $p$ tels que $p-1$ est une carr\'e parfait. En effet, nous d\'emontrerons une borne inf\'erieure sur le nombre d'entiers naturels $n \le Y$ tels que $n = \ell_1\cdots \ell_r$, avec $\ell_1,\ldots,\ell_r$ des premiers distincts, et tels que $(\ell_1-1)\cdots (\ell_r-1)$ est une puissance $r$-i\`eme, avec $r \ge 2$ quelconque. \'Egalement, nous d\'emontrerons une borne inf\'erieure sur le nombre d'entiers naturels $n = \ell_1\cdots \ell_r \le Y$ tels que $(\ell_1+1)\cdots (\ell_r+1)$ est une puissance $r$-i\`eme. Finalement, \'etant donn\'e $A$ un ensemble fini d'entiers non-nuls, nous d\'emontrerons une borne inf\'erieure sur le nombre d'entiers naturels $n \le Y$ tels que $\prod_{p \mid n} (p+a)$ est une puissance $r$-i\`eme, simultan\'ement pour chaque $a \in A$.
Resumo:
Combinational digital circuits can be evolved automatically using Genetic Algorithms (GA). Until recently this technique used linear chromosomes and and one dimensional crossover and mutation operators. In this paper, a new method for representing combinational digital circuits as 2 Dimensional (2D) chromosomes and suitable 2D crossover and mutation techniques has been proposed. By using this method, the convergence speed of GA can be increased significantly compared to the conventional methods. Moreover, the 2D representation and crossover operation provides the designer with better visualization of the evolved circuits. In addition to this, a technique to display automatically the evolved circuits has been developed with the help of MATLAB
Resumo:
This work presents Bayes invariant quadratic unbiased estimator, for short BAIQUE. Bayesian approach is used here to estimate the covariance functions of the regionalized variables which appear in the spatial covariance structure in mixed linear model. Firstly a brief review of spatial process, variance covariance components structure and Bayesian inference is given, since this project deals with these concepts. Then the linear equations model corresponding to BAIQUE in the general case is formulated. That Bayes estimator of variance components with too many unknown parameters is complicated to be solved analytically. Hence, in order to facilitate the handling with this system, BAIQUE of spatial covariance model with two parameters is considered. Bayesian estimation arises as a solution of a linear equations system which requires the linearity of the covariance functions in the parameters. Here the availability of prior information on the parameters is assumed. This information includes apriori distribution functions which enable to find the first and the second moments matrix. The Bayesian estimation suggested here depends only on the second moment of the prior distribution. The estimation appears as a quadratic form y'Ay , where y is the vector of filtered data observations. This quadratic estimator is used to estimate the linear function of unknown variance components. The matrix A of BAIQUE plays an important role. If such a symmetrical matrix exists, then Bayes risk becomes minimal and the unbiasedness conditions are fulfilled. Therefore, the symmetry of this matrix is elaborated in this work. Through dealing with the infinite series of matrices, a representation of the matrix A is obtained which shows the symmetry of A. In this context, the largest singular value of the decomposed matrix of the infinite series is considered to deal with the convergence condition and also it is connected with Gerschgorin Discs and Poincare theorem. Then the BAIQUE model for some experimental designs is computed and compared. The comparison deals with different aspects, such as the influence of the position of the design points in a fixed interval. The designs that are considered are those with their points distributed in the interval [0, 1]. These experimental structures are compared with respect to the Bayes risk and norms of the matrices corresponding to distances, covariance structures and matrices which have to satisfy the convergence condition. Also different types of the regression functions and distance measurements are handled. The influence of scaling on the design points is studied, moreover, the influence of the covariance structure on the best design is investigated and different covariance structures are considered. Finally, BAIQUE is applied for real data. The corresponding outcomes are compared with the results of other methods for the same data. Thereby, the special BAIQUE, which estimates the general variance of the data, achieves a very close result to the classical empirical variance.
Resumo:
The identification of chemical mechanism that can exhibit oscillatory phenomena in reaction networks are currently of intense interest. In particular, the parametric question of the existence of Hopf bifurcations has gained increasing popularity due to its relation to the oscillatory behavior around the fixed points. However, the detection of oscillations in high-dimensional systems and systems with constraints by the available symbolic methods has proven to be difficult. The development of new efficient methods are therefore required to tackle the complexity caused by the high-dimensionality and non-linearity of these systems. In this thesis, we mainly present efficient algorithmic methods to detect Hopf bifurcation fixed points in (bio)-chemical reaction networks with symbolic rate constants, thereby yielding information about their oscillatory behavior of the networks. The methods use the representations of the systems on convex coordinates that arise from stoichiometric network analysis. One of the methods called HoCoQ reduces the problem of determining the existence of Hopf bifurcation fixed points to a first-order formula over the ordered field of the reals that can then be solved using computational-logic packages. The second method called HoCaT uses ideas from tropical geometry to formulate a more efficient method that is incomplete in theory but worked very well for the attempted high-dimensional models involving more than 20 chemical species. The instability of reaction networks may lead to the oscillatory behaviour. Therefore, we investigate some criterions for their stability using convex coordinates and quantifier elimination techniques. We also study Muldowney's extension of the classical Bendixson-Dulac criterion for excluding periodic orbits to higher dimensions for polynomial vector fields and we discuss the use of simple conservation constraints and the use of parametric constraints for describing simple convex polytopes on which periodic orbits can be excluded by Muldowney's criteria. All developed algorithms have been integrated into a common software framework called PoCaB (platform to explore bio- chemical reaction networks by algebraic methods) allowing for automated computation workflows from the problem descriptions. PoCaB also contains a database for the algebraic entities computed from the models of chemical reaction networks.
Resumo:
In an earlier investigation (Burger et al., 2000) five sediment cores near the Rodrigues Triple Junction in the Indian Ocean were studied applying classical statistical methods (fuzzy c-means clustering, linear mixing model, principal component analysis) for the extraction of endmembers and evaluating the spatial and temporal variation of geochemical signals. Three main factors of sedimentation were expected by the marine geologists: a volcano-genetic, a hydro-hydrothermal and an ultra-basic factor. The display of fuzzy membership values and/or factor scores versus depth provided consistent results for two factors only; the ultra-basic component could not be identified. The reason for this may be that only traditional statistical methods were applied, i.e. the untransformed components were used and the cosine-theta coefficient as similarity measure. During the last decade considerable progress in compositional data analysis was made and many case studies were published using new tools for exploratory analysis of these data. Therefore it makes sense to check if the application of suitable data transformations, reduction of the D-part simplex to two or three factors and visual interpretation of the factor scores would lead to a revision of earlier results and to answers to open questions . In this paper we follow the lines of a paper of R. Tolosana- Delgado et al. (2005) starting with a problem-oriented interpretation of the biplot scattergram, extracting compositional factors, ilr-transformation of the components and visualization of the factor scores in a spatial context: The compositional factors will be plotted versus depth (time) of the core samples in order to facilitate the identification of the expected sources of the sedimentary process. Kew words: compositional data analysis, biplot, deep sea sediments
Resumo:
Many multivariate methods that are apparently distinct can be linked by introducing one or more parameters in their definition. Methods that can be linked in this way are correspondence analysis, unweighted or weighted logratio analysis (the latter also known as "spectral mapping"), nonsymmetric correspondence analysis, principal component analysis (with and without logarithmic transformation of the data) and multidimensional scaling. In this presentation I will show how several of these methods, which are frequently used in compositional data analysis, may be linked through parametrizations such as power transformations, linear transformations and convex linear combinations. Since the methods of interest here all lead to visual maps of data, a "movie" can be made where where the linking parameter is allowed to vary in small steps: the results are recalculated "frame by frame" and one can see the smooth change from one method to another. Several of these "movies" will be shown, giving a deeper insight into the similarities and differences between these methods
Resumo:
The performance of a model-based diagnosis system could be affected by several uncertainty sources, such as,model errors,uncertainty in measurements, and disturbances. This uncertainty can be handled by mean of interval models.The aim of this thesis is to propose a methodology for fault detection, isolation and identification based on interval models. The methodology includes some algorithms to obtain in an automatic way the symbolic expression of the residual generators enhancing the structural isolability of the faults, in order to design the fault detection tests. These algorithms are based on the structural model of the system. The stages of fault detection, isolation, and identification are stated as constraint satisfaction problems in continuous domains and solved by means of interval based consistency techniques. The qualitative fault isolation is enhanced by a reasoning in which the signs of the symptoms are derived from analytical redundancy relations or bond graph models of the system. An initial and empirical analysis regarding the differences between interval-based and statistical-based techniques is presented in this thesis. The performance and efficiency of the contributions are illustrated through several application examples, covering different levels of complexity.
Resumo:
El test de circuits és una fase del procés de producció que cada vegada pren més importància quan es desenvolupa un nou producte. Les tècniques de test i diagnosi per a circuits digitals han estat desenvolupades i automatitzades amb èxit, mentre que aquest no és encara el cas dels circuits analògics. D'entre tots els mètodes proposats per diagnosticar circuits analògics els més utilitzats són els diccionaris de falles. En aquesta tesi se'n descriuen alguns, tot analitzant-ne els seus avantatges i inconvenients. Durant aquests últims anys, les tècniques d'Intel·ligència Artificial han esdevingut un dels camps de recerca més importants per a la diagnosi de falles. Aquesta tesi desenvolupa dues d'aquestes tècniques per tal de cobrir algunes de les mancances que presenten els diccionaris de falles. La primera proposta es basa en construir un sistema fuzzy com a eina per identificar. Els resultats obtinguts son força bons, ja que s'aconsegueix localitzar la falla en un elevat tant percent dels casos. Per altra banda, el percentatge d'encerts no és prou bo quan a més a més s'intenta esbrinar la desviació. Com que els diccionaris de falles es poden veure com una aproximació simplificada al Raonament Basat en Casos (CBR), la segona proposta fa una extensió dels diccionaris de falles cap a un sistema CBR. El propòsit no és donar una solució general del problema sinó contribuir amb una nova metodologia. Aquesta consisteix en millorar la diagnosis dels diccionaris de falles mitjançant l'addició i l'adaptació dels nous casos per tal d'esdevenir un sistema de Raonament Basat en Casos. Es descriu l'estructura de la base de casos així com les tasques d'extracció, de reutilització, de revisió i de retenció, fent èmfasi al procés d'aprenentatge. En el transcurs del text s'utilitzen diversos circuits per mostrar exemples dels mètodes de test descrits, però en particular el filtre biquadràtic és l'utilitzat per provar les metodologies plantejades, ja que és un dels benchmarks proposats en el context dels circuits analògics. Les falles considerades son paramètriques, permanents, independents i simples, encara que la metodologia pot ser fàcilment extrapolable per a la diagnosi de falles múltiples i catastròfiques. El mètode es centra en el test dels components passius, encara que també es podria extendre per a falles en els actius.
Resumo:
Chemical methods to predict the bioavailable fraction of organic contaminants are usually validated in the literature by comparison with established bioassays. A soil spiked with polycyclic aromatic hydrocarbons (PAHs) was aged over six months and subjected to butanol, cyclodextrin and tenax extractions as well as an exhaustive extraction to determine total PAH concentrations at several time points. Earthworm (Eisenia fetida) and rye grass root (Lolium multiflorum) accumulation bioassays were conducted in parallel. Butanol extractions gave the best relationship with earthworm accumulation (r2 ≤ 0.54, p ≤ 0.01); cyclodextrin, butanol and acetone–hexane extractions all gave good predictions of accumulation in rye grass roots (r2 ≤ 0.86, p ≤ 0.01). However, the profile of the PAHs extracted by the different chemical methods was significantly different (p < 0.01) to that accumulated in the organisms. Biota accumulated a higher proportion of the heavier 4-ringed PAHs. It is concluded that bioaccumulation is a complex process that cannot be predicted by measuring the bioavailable fraction alone. The ability of chemical methods to predict PAH accumulation in Eisenia fetida and Lolium multiflorum was hindered by the varied metabolic fate of the different PAHs within the organisms.
Resumo:
Discussion of the numerical modeling of NDT methods based on the potential drop and the disruption of power lines to describe the nature, importance and application of modeling. La 1ère partie est consacrée aux applications aux contrôles par courants de Foucault. The first part is devoted to applications for inspection by eddy currents.
Resumo:
This paper examines the potential of using Participatory Farm Management methods to examine the suitability of a technology with farmers prior to on-farm trials. A study examining the suitability of green manuring as a technology for use with wet season tomato producers in Ghana is described. Findings from this case-study demonstrate that Participatory Budgeting can be used by farmers and researchers to analyse current cultivation practices, identify the options for including green manures into the system and explore the direct and wider resource implications of the technology. Scored-Causal Diagrams can be used to identify farmers' perceptions of the relative importance of the problem that the technology seeks to address. The use of the methods in this examine evaluation process appears to have the potential to improve the effectiveness and efficiency of the adaptive research process. This ensures that technologies subsequently examined in trials ate relevant to farmers' interests, existing systems and resources, thereby increasing the chances of farmer adoption. It is concluded that this process has potential for use-with other technologies and in other farming systems. (C) 2002 Elsevier Science Ltd. All rights reserved.
Resumo:
The present invention relates to haploid oil palm plants and homozygous doubled haploid oil palm plants. The invention also relates to methods for producing and selecting haploid and doubled haploid plants. More particularly, but not exclusively, the method may be used for selecting haploid and doubled haploid oil palm plants. Haploid and doubled haploid plants are selected by a large-scale screening based on a combination of the phenotype with the use of molecular methods combined with flow cytometry techniques to identify haploid and doubled haploid plants. More particularly, a method for selecting haploid and doubled haploid plants is described comprising: (a) germinating seeds; (b) selecting seedlings with atypical phenotype; (c) assessing heterozygosity using markers; (d) isolating cells from the seedlings and determining the DNA content of the cells; and (e) isolating and purifying the DNA and using defined molecular markers to characterise the genotype of the plant. The haploid oil palm plants may be used for producing homozygous doubled haploid oil palms: doubled haploids may be intercrossed to produce uniform F.sub.1 hybrids of superior properties.
Resumo:
With the rapid development of proteomics, a number of different methods appeared for the basic task of protein identification. We made a simple comparison between a common liquid chromatography-tandem mass spectrometry (LC-MS/MS) workflow using an ion trap mass spectrometer and a combined LC-MS and LC-MS/MS method using Fourier transform ion cyclotron resonance (FTICR) mass spectrometry and accurate peptide masses. To compare the two methods for protein identification, we grew and extracted proteins from E. coli using established protocols. Cystines were reduced and alkylated, and proteins digested by trypsin. The resulting peptide mixtures were separated by reversed-phase liquid chromatography using a 4 h gradient from 0 to 50% acetonitrile over a C18 reversed-phase column. The LC separation was coupled on-line to either a Bruker Esquire HCT ion trap or a Bruker 7 tesla APEX-Qe Qh-FTICR hybrid mass spectrometer. Data-dependent Qh-FTICR-MS/MS spectra were acquired using the quadrupole mass filter and collisionally induced dissociation into the external hexapole trap. Proteins were in both schemes identified by Mascot MS/MS ion searches and the peptides identified from these proteins in the FTICR MS/MS data were used for automatic internal calibration of the FTICR-MS data, together with ambient polydimethylcyclosiloxane ions.