985 resultados para Relational Data Bases
Resumo:
The immobilization and electro-oxidation of guanine and adenine asDNA bases on glassy carbon electrode are evaluated by square wave voltammetric analysis. The influence of electrochemical pretreatments, nature of supporting electrolyte, pH, accumulation time and composition of DNA nucleotides on the immobilization effect and the electrochemical mechanism are discussed. Trace levels of either guanine or adenine can be readily detected following short accumulation time with detection limits of 35 and 40 ngmL−1 for guanine and adenine, respectively. The biosensors of guanine and adenine were employed for the voltammetric detection of antioxidant capacity in flavored water samples. The method relies on monitoring the changes of the intrinsic anodic response of the surface-confined guanine and adenine species, resulting from its interaction with free radicals from Fenton-type reaction in absence and presence of antioxidant. Ascorbic acid was used as standard to evaluate antioxidant capacities of samples. Analytical data was compared with that of FRAP method.
Resumo:
The aim of this paper is to develop models for experimental open-channel water delivery systems and assess the use of three data-driven modeling tools toward that end. Water delivery canals are nonlinear dynamical systems and thus should be modeled to meet given operational requirements while capturing all relevant dynamics, including transport delays. Typically, the derivation of first principle models for open-channel systems is based on the use of Saint-Venant equations for shallow water, which is a time-consuming task and demands for specific expertise. The present paper proposes and assesses the use of three data-driven modeling tools: artificial neural networks, composite local linear models and fuzzy systems. The canal from Hydraulics and Canal Control Nucleus (A parts per thousand vora University, Portugal) will be used as a benchmark: The models are identified using data collected from the experimental facility, and then their performances are assessed based on suitable validation criterion. The performance of all models is compared among each other and against the experimental data to show the effectiveness of such tools to capture all significant dynamics within the canal system and, therefore, provide accurate nonlinear models that can be used for simulation or control. The models are available upon request to the authors.
Resumo:
OBJETIVO: Avaliar a viabilidade de estratégia de relacionamento probabilístico de bases de dados na identificação de óbitos de pacientes submetidos a procedimentos de alta complexidade em cardiologia. MÉTODOS: O custo de processamento foi estimado com base em 1.672 registros de pacientes submetidos à cirurgia de revascularização do miocárdio, relacionados com todos os registros de óbito no Brasil em 2005. A acurácia do relacionamento baseou-se em linkage probabilístico entre 99 registros de autorização de internação hospitalar de pacientes submetidos a cirurgias cardíacas em instituto de referência em cardiologia, com status vital conhecido, e todos os registros de óbito do estado do Rio de Janeiro em 2005. O linkage foi realizado em quatro etapas: padronização das bases, blocagem, pareamento e classificação dos pares. Utilizou-se a blocagem em cinco passos, com chaves de blocagem com combinação de variáveis como soundex do primeiro e último nome, sexo e ano de nascimento. As variáveis utilizadas no pareamento foram "nome completo", com a utilização da distância de Levenshtein, e "data de nascimento". RESULTADOS: O segundo e o quinto passos de blocagem tiveram os maiores números de pares formados e os maiores tempos de processamento para o pareamento. O quarto passo demandou menor custo de processamento. No estudo de acurácia, após os cinco passos de blocagem, a sensibilidade do linkage foi de 90,6% e a especificidade foi de 100%. CONCLUSÕES: A estratégia de relacionamento probabilístico utilizada apresenta boa acurácia e poderá ser utilizada em estudos sobre a efetividade dos procedimentos de alta complexidade e alto custo em cardiologia.
Resumo:
Seismic data is difficult to analyze and classical mathematical tools reveal strong limitations in exposing hidden relationships between earthquakes. In this paper, we study earthquake phenomena in the perspective of complex systems. Global seismic data, covering the period from 1962 up to 2011 is analyzed. The events, characterized by their magnitude, geographic location and time of occurrence, are divided into groups, either according to the Flinn-Engdahl (F-E) seismic regions of Earth or using a rectangular grid based in latitude and longitude coordinates. Two methods of analysis are considered and compared in this study. In a first method, the distributions of magnitudes are approximated by Gutenberg-Richter (G-R) distributions and the parameters used to reveal the relationships among regions. In the second method, the mutual information is calculated and adopted as a measure of similarity between regions. In both cases, using clustering analysis, visualization maps are generated, providing an intuitive and useful representation of the complex relationships that are present among seismic data. Such relationships might not be perceived on classical geographic maps. Therefore, the generated charts are a valid alternative to other visualization tools, for understanding the global behavior of earthquakes.
Resumo:
Research on the problem of feature selection for clustering continues to develop. This is a challenging task, mainly due to the absence of class labels to guide the search for relevant features. Categorical feature selection for clustering has rarely been addressed in the literature, with most of the proposed approaches having focused on numerical data. In this work, we propose an approach to simultaneously cluster categorical data and select a subset of relevant features. Our approach is based on a modification of a finite mixture model (of multinomial distributions), where a set of latent variables indicate the relevance of each feature. To estimate the model parameters, we implement a variant of the expectation-maximization algorithm that simultaneously selects the subset of relevant features, using a minimum message length criterion. The proposed approach compares favourably with two baseline methods: a filter based on an entropy measure and a wrapper based on mutual information. The results obtained on synthetic data illustrate the ability of the proposed expectation-maximization method to recover ground truth. An application to real data, referred to official statistics, shows its usefulness.
Resumo:
Research on cluster analysis for categorical data continues to develop, new clustering algorithms being proposed. However, in this context, the determination of the number of clusters is rarely addressed. We propose a new approach in which clustering and the estimation of the number of clusters is done simultaneously for categorical data. We assume that the data originate from a finite mixture of multinomial distributions and use a minimum message length criterion (MML) to select the number of clusters (Wallace and Bolton, 1986). For this purpose, we implement an EM-type algorithm (Silvestre et al., 2008) based on the (Figueiredo and Jain, 2002) approach. The novelty of the approach rests on the integration of the model estimation and selection of the number of clusters in a single algorithm, rather than selecting this number based on a set of pre-estimated candidate models. The performance of our approach is compared with the use of Bayesian Information Criterion (BIC) (Schwarz, 1978) and Integrated Completed Likelihood (ICL) (Biernacki et al., 2000) using synthetic data. The obtained results illustrate the capacity of the proposed algorithm to attain the true number of cluster while outperforming BIC and ICL since it is faster, which is especially relevant when dealing with large data sets.
Resumo:
Cluster analysis for categorical data has been an active area of research. A well-known problem in this area is the determination of the number of clusters, which is unknown and must be inferred from the data. In order to estimate the number of clusters, one often resorts to information criteria, such as BIC (Bayesian information criterion), MML (minimum message length, proposed by Wallace and Boulton, 1968), and ICL (integrated classification likelihood). In this work, we adopt the approach developed by Figueiredo and Jain (2002) for clustering continuous data. They use an MML criterion to select the number of clusters and a variant of the EM algorithm to estimate the model parameters. This EM variant seamlessly integrates model estimation and selection in a single algorithm. For clustering categorical data, we assume a finite mixture of multinomial distributions and implement a new EM algorithm, following a previous version (Silvestre et al., 2008). Results obtained with synthetic datasets are encouraging. The main advantage of the proposed approach, when compared to the above referred criteria, is the speed of execution, which is especially relevant when dealing with large data sets.
Resumo:
Orientadora: Doutora Anabela Mesquita Teixeira Sarmento
Resumo:
Há mais de uma década começou a formar-se um movimento de reconversão nos estudos sobre as classes sociais, orientado pela ideia de que é preciso dar mais relevo aos processos culturais, à formação de identidades, ao gosto e à expressão de modos de vida e proceder ao exame direto da maneira como as posições de classe são vividas. Nesta perspetiva, as culturas de classe deixam de designar apenas tipos de comportamento próprios de grupos de desigualdade e passam a ser vistas também como modos de diferenciação, em que as categorias de classe operam através de distinções individualizadas e comparações relacionais com membros de outras classes. O que o conceito de identidade de classe quer agora indicar é um sentido das distâncias sociais dentro das hierarquias, que recorre às diferenças culturais para se exprimir. Este texto recenseia os desenvolvimentos teóricos e as bases empíricas desta linha de trabalho, emergente na sociologia britânica a partir do final da década de 90, com as propostas de autores como Savage, Skeggs, Reay, ou Bottero, entre outros.
Resumo:
Mestrado em Contabilidade
Resumo:
Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia Informática e de Computadores
Resumo:
Consider the problem of disseminating data from an arbitrary source node to all other nodes in a distributed computer system, like Wireless Sensor Networks (WSNs). We assume that wireless broadcast is used and nodes do not know the topology. We propose new protocols which disseminate data faster and use fewer broadcasts than the simple broadcast protocol.
Resumo:
Nowadays, due to the incredible grow of the mobile devices market, when we want to implement a client-server applications we must consider mobile devices limitations. In this paper we discuss which can be the more reliable and fast way to exchange information between a server and an Android mobile application. This is an important issue because with a responsive application the user experience is more enjoyable. In this paper we present a study that test and evaluate two data transfer protocols, socket and HTTP, and three data serialization formats (XML, JSON and Protocol Buffers) using different environments and mobile devices to realize which is the most practical and fast to use.
Resumo:
Dissertação de Natureza Científica para obtenção do grau de Mestre em Engenharia Civil na Área de Especialização de Edificações
Resumo:
Dissertation elaborated for the partial fulfilment of the requirements of the Master Degree in Civil Engineering in the Speciality Area of Hydarulics