980 resultados para Statistical testing
Resumo:
Não existe uma definição única de processo de memória de longo prazo. Esse processo é geralmente definido como uma série que possui um correlograma decaindo lentamente ou um espectro infinito de frequência zero. Também se refere que uma série com tal propriedade é caracterizada pela dependência a longo prazo e por não periódicos ciclos longos, ou que essa característica descreve a estrutura de correlação de uma série de longos desfasamentos ou que é convencionalmente expressa em termos do declínio da lei-potência da função auto-covariância. O interesse crescente da investigação internacional no aprofundamento do tema é justificado pela procura de um melhor entendimento da natureza dinâmica das séries temporais dos preços dos ativos financeiros. Em primeiro lugar, a falta de consistência entre os resultados reclama novos estudos e a utilização de várias metodologias complementares. Em segundo lugar, a confirmação de processos de memória longa tem implicações relevantes ao nível da (1) modelação teórica e econométrica (i.e., dos modelos martingale de preços e das regras técnicas de negociação), (2) dos testes estatísticos aos modelos de equilíbrio e avaliação, (3) das decisões ótimas de consumo / poupança e de portefólio e (4) da medição de eficiência e racionalidade. Em terceiro lugar, ainda permanecem questões científicas empíricas sobre a identificação do modelo geral teórico de mercado mais adequado para modelar a difusão das séries. Em quarto lugar, aos reguladores e gestores de risco importa saber se existem mercados persistentes e, por isso, ineficientes, que, portanto, possam produzir retornos anormais. O objetivo do trabalho de investigação da dissertação é duplo. Por um lado, pretende proporcionar conhecimento adicional para o debate da memória de longo prazo, debruçando-se sobre o comportamento das séries diárias de retornos dos principais índices acionistas da EURONEXT. Por outro lado, pretende contribuir para o aperfeiçoamento do capital asset pricing model CAPM, considerando uma medida de risco alternativa capaz de ultrapassar os constrangimentos da hipótese de mercado eficiente EMH na presença de séries financeiras com processos sem incrementos independentes e identicamente distribuídos (i.i.d.). O estudo empírico indica a possibilidade de utilização alternativa das obrigações do tesouro (OT’s) com maturidade de longo prazo no cálculo dos retornos do mercado, dado que o seu comportamento nos mercados de dívida soberana reflete a confiança dos investidores nas condições financeiras dos Estados e mede a forma como avaliam as respetiva economias com base no desempenho da generalidade dos seus ativos. Embora o modelo de difusão de preços definido pelo movimento Browniano geométrico gBm alegue proporcionar um bom ajustamento das séries temporais financeiras, os seus pressupostos de normalidade, estacionariedade e independência das inovações residuais são adulterados pelos dados empíricos analisados. Por isso, na procura de evidências sobre a propriedade de memória longa nos mercados recorre-se à rescaled-range analysis R/S e à detrended fluctuation analysis DFA, sob abordagem do movimento Browniano fracionário fBm, para estimar o expoente Hurst H em relação às séries de dados completas e para calcular o expoente Hurst “local” H t em janelas móveis. Complementarmente, são realizados testes estatísticos de hipóteses através do rescaled-range tests R/S , do modified rescaled-range test M - R/S e do fractional differencing test GPH. Em termos de uma conclusão única a partir de todos os métodos sobre a natureza da dependência para o mercado acionista em geral, os resultados empíricos são inconclusivos. Isso quer dizer que o grau de memória de longo prazo e, assim, qualquer classificação, depende de cada mercado particular. No entanto, os resultados gerais maioritariamente positivos suportam a presença de memória longa, sob a forma de persistência, nos retornos acionistas da Bélgica, Holanda e Portugal. Isto sugere que estes mercados estão mais sujeitos a maior previsibilidade (“efeito José”), mas também a tendências que podem ser inesperadamente interrompidas por descontinuidades (“efeito Noé”), e, por isso, tendem a ser mais arriscados para negociar. Apesar da evidência de dinâmica fractal ter suporte estatístico fraco, em sintonia com a maior parte dos estudos internacionais, refuta a hipótese de passeio aleatório com incrementos i.i.d., que é a base da EMH na sua forma fraca. Atendendo a isso, propõem-se contributos para aperfeiçoamento do CAPM, através da proposta de uma nova fractal capital market line FCML e de uma nova fractal security market line FSML. A nova proposta sugere que o elemento de risco (para o mercado e para um ativo) seja dado pelo expoente H de Hurst para desfasamentos de longo prazo dos retornos acionistas. O expoente H mede o grau de memória de longo prazo nos índices acionistas, quer quando as séries de retornos seguem um processo i.i.d. não correlacionado, descrito pelo gBm(em que H = 0,5 , confirmando- se a EMH e adequando-se o CAPM), quer quando seguem um processo com dependência estatística, descrito pelo fBm(em que H é diferente de 0,5, rejeitando-se a EMH e desadequando-se o CAPM). A vantagem da FCML e da FSML é que a medida de memória de longo prazo, definida por H, é a referência adequada para traduzir o risco em modelos que possam ser aplicados a séries de dados que sigam processos i.i.d. e processos com dependência não linear. Então, estas formulações contemplam a EMH como um caso particular possível.
Resumo:
The hidden-node problem has been shown to be a major source of Quality-of-Service (QoS) degradation in Wireless Sensor Networks (WSNs) due to factors such as the limited communication range of sensor nodes, link asymmetry and the characteristics of the physical environment. In wireless contention-based Medium Access Control protocols, if two nodes that are not visible to each other transmit to a third node that is visible to the formers, there will be a collision – usually called hidden-node or blind collision. This problem greatly affects network throughput, energy-efficiency and message transfer delays, which might be particularly dramatic in large-scale WSNs. This technical report tackles the hidden-node problem in WSNs and proposes HNAMe, a simple yet efficient distributed mechanism to overcome it. H-NAMe relies on a grouping strategy that splits each cluster of a WSN into disjoint groups of non-hidden nodes and then scales to multiple clusters via a cluster grouping strategy that guarantees no transmission interference between overlapping clusters. We also show that the H-NAMe mechanism can be easily applied to the IEEE 802.15.4/ZigBee protocols with only minor add-ons and ensuring backward compatibility with the standard specifications. We demonstrate the feasibility of H-NAMe via an experimental test-bed, showing that it increases network throughput and transmission success probability up to twice the values obtained without H-NAMe. We believe that the results in this technical report will be quite useful in efficiently enabling IEEE 802.15.4/ZigBee as a WSN protocol.
Resumo:
Introdução e objetivos: A síncope recorrente tem um impacto significativo na qualidade de vida. O desenvolvimento de escalas de medida de fácil aplicabilidade clínica para avaliar este impacto é fundamental. O objetivo do presente estudo é a validação preliminar da escala Impact of Syncope on Quality of Life, para a população portuguesa. Métodos: O instrumento foi submetido a um processo de tradução, validação, adequação cultural e cognitive debriefing. Participaram 39 doentes com história de síncopes recorrentes (>1 ano de evolução), submetidos a teste de inclinação em mesa basculante (teste de tilt), que constitui uma amostra de conveniência, com idade de 52,1±16,4 anos (21-83; 43,5% do sexo masculino), a maioria com uma situação profissional ativa (n=18) ou reformados (n=13). A versão portuguesa resultou numa versão semelhante unidimensional à original com 12 itens agregados num único somatório, tendo passado por validação estatística, com avaliação da fidelidade, validade e estabilidade no tempo. Resultados: Em relação à fidelidade, a consistência interna da escala é de 0,9. Avaliámos a validade convergente, tendo obtido resultados estatisticamente significativos (p<0,01). Avaliámos a validade divergente tendo obtido resultados estatisticamente significativos. Relativamente à estabilidade no tempo foi efetuado um teste-reteste do instrumento aos seis meses após o teste de inclinação com 22 doentes desta amostra não submetidos a intervenção clínica, que não mostrou alterações estatisticamente significativas da qualidade de vida. Conclusões: Os resultados obtidos indicam a pertinência da utilização deste instrumento em contexto português na avaliação da qualidade de vida de doentes com síncope recorrente.
Resumo:
This Thesis describes the application of automatic learning methods for a) the classification of organic and metabolic reactions, and b) the mapping of Potential Energy Surfaces(PES). The classification of reactions was approached with two distinct methodologies: a representation of chemical reactions based on NMR data, and a representation of chemical reactions from the reaction equation based on the physico-chemical and topological features of chemical bonds. NMR-based classification of photochemical and enzymatic reactions. Photochemical and metabolic reactions were classified by Kohonen Self-Organizing Maps (Kohonen SOMs) and Random Forests (RFs) taking as input the difference between the 1H NMR spectra of the products and the reactants. The development of such a representation can be applied in automatic analysis of changes in the 1H NMR spectrum of a mixture and their interpretation in terms of the chemical reactions taking place. Examples of possible applications are the monitoring of reaction processes, evaluation of the stability of chemicals, or even the interpretation of metabonomic data. A Kohonen SOM trained with a data set of metabolic reactions catalysed by transferases was able to correctly classify 75% of an independent test set in terms of the EC number subclass. Random Forests improved the correct predictions to 79%. With photochemical reactions classified into 7 groups, an independent test set was classified with 86-93% accuracy. The data set of photochemical reactions was also used to simulate mixtures with two reactions occurring simultaneously. Kohonen SOMs and Feed-Forward Neural Networks (FFNNs) were trained to classify the reactions occurring in a mixture based on the 1H NMR spectra of the products and reactants. Kohonen SOMs allowed the correct assignment of 53-63% of the mixtures (in a test set). Counter-Propagation Neural Networks (CPNNs) gave origin to similar results. The use of supervised learning techniques allowed an improvement in the results. They were improved to 77% of correct assignments when an ensemble of ten FFNNs were used and to 80% when Random Forests were used. This study was performed with NMR data simulated from the molecular structure by the SPINUS program. In the design of one test set, simulated data was combined with experimental data. The results support the proposal of linking databases of chemical reactions to experimental or simulated NMR data for automatic classification of reactions and mixtures of reactions. Genome-scale classification of enzymatic reactions from their reaction equation. The MOLMAP descriptor relies on a Kohonen SOM that defines types of bonds on the basis of their physico-chemical and topological properties. The MOLMAP descriptor of a molecule represents the types of bonds available in that molecule. The MOLMAP descriptor of a reaction is defined as the difference between the MOLMAPs of the products and the reactants, and numerically encodes the pattern of bonds that are broken, changed, and made during a chemical reaction. The automatic perception of chemical similarities between metabolic reactions is required for a variety of applications ranging from the computer validation of classification systems, genome-scale reconstruction (or comparison) of metabolic pathways, to the classification of enzymatic mechanisms. Catalytic functions of proteins are generally described by the EC numbers that are simultaneously employed as identifiers of reactions, enzymes, and enzyme genes, thus linking metabolic and genomic information. Different methods should be available to automatically compare metabolic reactions and for the automatic assignment of EC numbers to reactions still not officially classified. In this study, the genome-scale data set of enzymatic reactions available in the KEGG database was encoded by the MOLMAP descriptors, and was submitted to Kohonen SOMs to compare the resulting map with the official EC number classification, to explore the possibility of predicting EC numbers from the reaction equation, and to assess the internal consistency of the EC classification at the class level. A general agreement with the EC classification was observed, i.e. a relationship between the similarity of MOLMAPs and the similarity of EC numbers. At the same time, MOLMAPs were able to discriminate between EC sub-subclasses. EC numbers could be assigned at the class, subclass, and sub-subclass levels with accuracies up to 92%, 80%, and 70% for independent test sets. The correspondence between chemical similarity of metabolic reactions and their MOLMAP descriptors was applied to the identification of a number of reactions mapped into the same neuron but belonging to different EC classes, which demonstrated the ability of the MOLMAP/SOM approach to verify the internal consistency of classifications in databases of metabolic reactions. RFs were also used to assign the four levels of the EC hierarchy from the reaction equation. EC numbers were correctly assigned in 95%, 90%, 85% and 86% of the cases (for independent test sets) at the class, subclass, sub-subclass and full EC number level,respectively. Experiments for the classification of reactions from the main reactants and products were performed with RFs - EC numbers were assigned at the class, subclass and sub-subclass level with accuracies of 78%, 74% and 63%, respectively. In the course of the experiments with metabolic reactions we suggested that the MOLMAP / SOM concept could be extended to the representation of other levels of metabolic information such as metabolic pathways. Following the MOLMAP idea, the pattern of neurons activated by the reactions of a metabolic pathway is a representation of the reactions involved in that pathway - a descriptor of the metabolic pathway. This reasoning enabled the comparison of different pathways, the automatic classification of pathways, and a classification of organisms based on their biochemical machinery. The three levels of classification (from bonds to metabolic pathways) allowed to map and perceive chemical similarities between metabolic pathways even for pathways of different types of metabolism and pathways that do not share similarities in terms of EC numbers. Mapping of PES by neural networks (NNs). In a first series of experiments, ensembles of Feed-Forward NNs (EnsFFNNs) and Associative Neural Networks (ASNNs) were trained to reproduce PES represented by the Lennard-Jones (LJ) analytical potential function. The accuracy of the method was assessed by comparing the results of molecular dynamics simulations (thermal, structural, and dynamic properties) obtained from the NNs-PES and from the LJ function. The results indicated that for LJ-type potentials, NNs can be trained to generate accurate PES to be used in molecular simulations. EnsFFNNs and ASNNs gave better results than single FFNNs. A remarkable ability of the NNs models to interpolate between distant curves and accurately reproduce potentials to be used in molecular simulations is shown. The purpose of the first study was to systematically analyse the accuracy of different NNs. Our main motivation, however, is reflected in the next study: the mapping of multidimensional PES by NNs to simulate, by Molecular Dynamics or Monte Carlo, the adsorption and self-assembly of solvated organic molecules on noble-metal electrodes. Indeed, for such complex and heterogeneous systems the development of suitable analytical functions that fit quantum mechanical interaction energies is a non-trivial or even impossible task. The data consisted of energy values, from Density Functional Theory (DFT) calculations, at different distances, for several molecular orientations and three electrode adsorption sites. The results indicate that NNs require a data set large enough to cover well the diversity of possible interaction sites, distances, and orientations. NNs trained with such data sets can perform equally well or even better than analytical functions. Therefore, they can be used in molecular simulations, particularly for the ethanol/Au (111) interface which is the case studied in the present Thesis. Once properly trained, the networks are able to produce, as output, any required number of energy points for accurate interpolations.
Resumo:
OBJECTIVE To evaluate the cross-cultural validity of the Demand-Control Questionnaire, comparing the original Swedish questionnaire with the Brazilian version. METHODS We compared data from 362 Swedish and 399 Brazilian health workers. Confirmatory and exploratory factor analyses were performed to test structural validity, using the robust weighted least squares mean and variance-adjusted (WLSMV) estimator. Construct validity, using hypotheses testing, was evaluated through the inspection of the mean score distribution of the scale dimensions according to sociodemographic and social support at work variables. RESULTS The confirmatory and exploratory factor analyses supported the instrument in three dimensions (for Swedish and Brazilians): psychological demands, skill discretion and decision authority. The best-fit model was achieved by including an error correlation between work fast and work intensely (psychological demands) and removing the item repetitive work (skill discretion). Hypotheses testing showed that workers with university degree had higher scores on skill discretion and decision authority and those with high levels of Social Support at Work had lower scores on psychological demands and higher scores on decision authority. CONCLUSIONS The results supported the equivalent dimensional structures across the two culturally different work contexts. Skill discretion and decision authority formed two distinct dimensions and the item repetitive work should be removed.
Resumo:
In this work, an experimental study was performed on the influence of plug-filling, loading rate and temperature on the tensile strength of single-strap (SS) and double-strap (DS) repairs on aluminium structures. Whilst the main purpose of this work was to evaluate the feasibility of plug-filling for the strength improvement of these repairs, a parallel study was carried out to assess the sensitivity of the adhesive to external features that can affect the repairs performance, such as the rate of loading and environmental temperature. The experimental programme included repairs with different values of overlap length (L O = 10, 20 and 30 mm), and with and without plug-filling, whose results were interpreted in light of experimental evidence of the fracture modes and typical stress distributions for bonded repairs. The influence of the testing speed on the repairs strength was also addressed (considering 0.5, 5 and 25 mm/min). Accounting for the temperature effects, tests were carried out at room temperature (≈23°C), 50 and 80°C. This permitted a comparative evaluation of the adhesive tested below and above the glass transition temperature (T g), established by the manufacturer as 67°C. The combined influence of these two parameters on the repairs strength was also analysed. According to the results obtained from this work, design guidelines for repairing aluminium structures were
Resumo:
OBJECTIVE Determine the coverage rate of syphilis testing during prenatal care and the prevalence of syphilis in pregnant women in Brazil. METHODS This is a national hospital-based cohort study conducted in Brazil with 23,894 postpartum women between 2011 and 2012. Data were obtained using interviews with postpartum women, hospital records, and prenatal care cards. All postpartum women with a reactive serological test result recorded in the prenatal care card or syphilis diagnosis during hospitalization for childbirth were considered cases of syphilis in pregnancy. The Chi-square test was used for determining the disease prevalence and testing coverage rate by region of residence, self-reported skin color, maternal age, and type of prenatal and child delivery care units. RESULTS Prenatal care covered 98.7% postpartum women. Syphilis testing coverage rate was 89.1% (one test) and 41.2% (two tests), and syphilis prevalence in pregnancy was 1.02% (95%CI 0.84;1.25). A lower prenatal coverage rate was observed among women in the North region, indigenous women, those with less education, and those who received prenatal care in public health care units. A lower testing coverage rate was observed among residents in the North, Northeast, and Midwest regions, among younger and non-white skin-color women, among those with lower education, and those who received prenatal care in public health care units. An increased prevalence of syphilis was observed among women with < 8 years of education (1.74%), who self-reported as black (1.8%) or mixed (1.2%), those who did not receive prenatal care (2.5%), and those attending public (1.37%) or mixed (0.93%) health care units. CONCLUSIONS The estimated prevalence of syphilis in pregnancy was similar to that reported in the last sentinel surveillance study conducted in 2006. There was an improvement in prenatal care and testing coverage rate, and the goals suggested by the World Health Organization were achieved in two regions. Regional and social inequalities in access to health care units, coupled with other gaps in health assistance, have led to the persistence of congenital syphilis as a major public health problem in Brazil.
Resumo:
OBJECTIVE To analyze the clinical and laboratory characteristics of HIV-infected individuals upon admission to a reference health care center.METHODS This cross-sectional study was conducted between 1999 and 2010 on 527 individuals with confirmed serological diagnosis of HIV infection who were enrolled in an outpatient health care service in Santarém, PA, Northern Brazil. Data were collected from medical records and included the reason for HIV testing, clinical status, and count of peripheral CD4+ T lymphocytes upon enrollment. The data were divided into three groups, according to the patient’s year of admission – P1 (1999-2002), P2 (2003-2006), and P3 (2007-2010) – for comparative analysis of the variables of interest.RESULTS In the study group, 62.0% of the patients were assigned to the P3 group. The reason for undergoing HIV testing differed between genders. In the male population, most tests were conducted because of the presence of symptoms suggesting infection. Among women, tests were the result of knowledge of the partner’s seropositive status in groups P1 and P2. Higher proportion of women undergoing testing because of symptoms of HIV/AIDS infection abolished the difference between genders in the most recent period. A higher percentage of patients enrolling at a more advanced stage of the disease was observed in P3.CONCLUSIONS Despite the increased awareness of the number of HIV/AIDS cases, these patients have identified their serological status late and were admitted to health care units with active disease. The HIV/AIDS epidemic in Pará presents specificities in its progression that indicate the complex characteristics of the epidemic in the Northern region of Brazil and across the country.
Resumo:
Dissertação apresentada na Faculdade de Ciências e Tecnologia da Universidade Nova de Lisboa para a obtenção do grau de Mestre em Engenharia do Ambiente
Resumo:
Dissertação apresentada para obtenção do Grau de Doutor em Matemática, Estatística, pela Universidade Nova de Lisboa, faculdade de Ciências e Tecnologia
Resumo:
The Brazilian National Regulatory Agency for Private Health Insurance and Plans has recently published a technical note defining the criteria for the coverage of genetic testing to diagnose hereditary cancer. In this study we show the case of a patient with a breast lesion and an extensive history of cancer referred to a private service of genetic counseling. The patient met both criteria for hereditary breast and colorectal cancer syndrome screening. Her private insurance denied coverage for genetic testing because she lacks current or previous cancer diagnosis. After she appealed by lawsuit, the court was favorable and the test was performed using next-generation sequencing. A deletion of MLH1 exon 8 was found. We highlight the importance to offer genetic testing using multigene analysis for noncancer patients.
Resumo:
The main goals of the present work are the evaluation of the influence of several variables and test parameters on the melt flow index (MFI) of thermoplastics, and the determination of the uncertainty associated with the measurements. To evaluate the influence of test parameters on the measurement of MFI the design of experiments (DOE) approach has been used. The uncertainty has been calculated using a "bottom-up" approach given in the "Guide to the Expression of the Uncertainty of Measurement" (GUM). Since an analytical expression relating the output response (MFI) with input parameters does not exist, it has been necessary to build mathematical models by adjusting the experimental observations of the response variable in accordance with each input parameter. Subsequently, the determination of the uncertainty associated with the measurement of MFI has been performed by applying the law of propagation of uncertainty to the values of uncertainty of the input parameters. Finally, the activation energy (Ea) of the melt flow at around 200 degrees C and the respective uncertainty have also been determined.
Resumo:
Dissertation submitted in partial fulfilment of the requirements for the Degree of Master of Science in Geospatial Technologies
Resumo:
Beyond the classical statistical approaches (determination of basic statistics, regression analysis, ANOVA, etc.) a new set of applications of different statistical techniques has increasingly gained relevance in the analysis, processing and interpretation of data concerning the characteristics of forest soils. This is possible to be seen in some of the recent publications in the context of Multivariate Statistics. These new methods require additional care that is not always included or refered in some approaches. In the particular case of geostatistical data applications it is necessary, besides to geo-reference all the data acquisition, to collect the samples in regular grids and in sufficient quantity so that the variograms can reflect the spatial distribution of soil properties in a representative manner. In the case of the great majority of Multivariate Statistics techniques (Principal Component Analysis, Correspondence Analysis, Cluster Analysis, etc.) despite the fact they do not require in most cases the assumption of normal distribution, they however need a proper and rigorous strategy for its utilization. In this work, some reflections about these methodologies and, in particular, about the main constraints that often occur during the information collecting process and about the various linking possibilities of these different techniques will be presented. At the end, illustrations of some particular cases of the applications of these statistical methods will also be presented.
Resumo:
Second International Workshop on Analog and Mixed Signal Integrated Circuits for Space Applications (AMICSA 2008), Sintra, Portugal, Setembro de 2008