725 resultados para CLUSTER VALIDITY
Resumo:
The present research paper presents five different clustering methods to identify typical load profiles of medium voltage (MV) electricity consumers. These methods are intended to be used in a smart grid environment to extract useful knowledge about customer’s behaviour. The obtained knowledge can be used to support a decision tool, not only for utilities but also for consumers. Load profiles can be used by the utilities to identify the aspects that cause system load peaks and enable the development of specific contracts with their customers. The framework presented throughout the paper consists in several steps, namely the pre-processing data phase, clustering algorithms application and the evaluation of the quality of the partition, which is supported by cluster validity indices. The process ends with the analysis of the discovered knowledge. To validate the proposed framework, a case study with a real database of 208 MV consumers is used.
Resumo:
This paper consists in the characterization of medium voltage (MV) electric power consumers based on a data clustering approach. It is intended to identify typical load profiles by selecting the best partition of a power consumption database among a pool of data partitions produced by several clustering algorithms. The best partition is selected using several cluster validity indices. These methods are intended to be used in a smart grid environment to extract useful knowledge about customers’ behavior. The data-mining-based methodology presented throughout the paper consists in several steps, namely the pre-processing data phase, clustering algorithms application and the evaluation of the quality of the partitions. To validate our approach, a case study with a real database of 1.022 MV consumers was used.
Resumo:
This paper presents the characterization of high voltage (HV) electric power consumers based on a data clustering approach. The typical load profiles (TLP) are obtained selecting the best partition of a power consumption database among a pool of data partitions produced by several clustering algorithms. The choice of the best partition is supported using several cluster validity indices. The proposed data-mining (DM) based methodology, that includes all steps presented in the process of knowledge discovery in databases (KDD), presents an automatic data treatment application in order to preprocess the initial database in an automatic way, allowing time saving and better accuracy during this phase. These methods are intended to be used in a smart grid environment to extract useful knowledge about customers’ consumption behavior. To validate our approach, a case study with a real database of 185 HV consumers was used.
Resumo:
One of the top ten most influential data mining algorithms, k-means, is known for being simple and scalable. However, it is sensitive to initialization of prototypes and requires that the number of clusters be specified in advance. This paper shows that evolutionary techniques conceived to guide the application of k-means can be more computationally efficient than systematic (i.e., repetitive) approaches that try to get around the above-mentioned drawbacks by repeatedly running the algorithm from different configurations for the number of clusters and initial positions of prototypes. To do so, a modified version of a (k-means based) fast evolutionary algorithm for clustering is employed. Theoretical complexity analyses for the systematic and evolutionary algorithms under interest are provided. Computational experiments and statistical analyses of the results are presented for artificial and text mining data sets. (C) 2010 Elsevier B.V. All rights reserved.
Resumo:
This paper tackles the problem of showing that evolutionary algorithms for fuzzy clustering can be more efficient than systematic (i.e. repetitive) approaches when the number of clusters in a data set is unknown. To do so, a fuzzy version of an Evolutionary Algorithm for Clustering (EAC) is introduced. A fuzzy cluster validity criterion and a fuzzy local search algorithm are used instead of their hard counterparts employed by EAC. Theoretical complexity analyses for both the systematic and evolutionary algorithms under interest are provided. Examples with computational experiments and statistical analyses are also presented.
Resumo:
This journal provides immediate open access to its content on the principle that making research freely available to the public supports a greater global exchange of knowledge.
Resumo:
Modeling the fundamental performance limits of Wireless Sensor Networks (WSNs) is of paramount importance to understand their behavior under the worst-case conditions and to make the appropriate design choices. This is particular relevant for time-sensitive WSN applications, where the timing behavior of the network protocols (message transmission must respect deadlines) impacts on the correct operation of these applications. In that direction this paper contributes with a methodology based on Network Calculus, which enables quick and efficient worst-case dimensioning of static or even dynamically changing cluster-tree WSNs where the data sink can either be static or mobile. We propose closed-form recurrent expressions for computing the worst-case end-to-end delays, buffering and bandwidth requirements across any source-destination path in a cluster-tree WSN. We show how to apply our methodology to the case of IEEE 802.15.4/ZigBee cluster-tree WSNs. Finally, we demonstrate the validity and analyze the accuracy of our methodology through a comprehensive experimental study using commercially available technology, namely TelosB motes running TinyOS.
Resumo:
Modeling the fundamental performance limits of Wireless Sensor Networks (WSNs) is of paramount importance to understand their behavior under worst-case conditions and to make the appropriate design choices. In that direction this paper contributes with an analytical methodology for modeling cluster-tree WSNs where the data sink can either be static or mobile. We assess the validity and pessimism of analytical model by comparing the worst-case results with the values measured through an experimental test-bed based on Commercial-Off- The-Shelf (COTS) technologies, namely TelosB motes running TinyOS.
Resumo:
Modelling the fundamental performance limits of wireless sensor networks (WSNs) is of paramount importance to understand the behaviour of WSN under worst case conditions and to make the appropriate design choices. In that direction, this paper contributes with a methodology for modelling cluster tree WSNs with a mobile sink. We propose closed form recurrent expressions for computing the worst case end to end delays, buffering and bandwidth requirements across any source-destination path in the cluster tree assuming error free channel. We show how to apply our theoretical results to the specific case of IEEE 802.15.4/ZigBee WSNs. Finally, we demonstrate the validity and analyze the accuracy of our methodology through a comprehensive experimental study, therefore validating the theoretical results through experimentation.
Resumo:
While Cluster-Tree network topologies look promising for WSN applications with timeliness and energy-efficiency requirements, we are yet to witness its adoption in commercial and academic solutions. One of the arguments that hinder the use of these topologies concerns the lack of flexibility in adapting to changes in the network, such as in traffic flows. This paper presents a solution to enable these networks with the ability to self-adapt their clusters’ duty-cycle and scheduling, to provide increased quality of service to multiple traffic flows. Importantly, our approach enables a network to change its cluster scheduling without requiring long inaccessibility times or the re-association of the nodes. We show how to apply our methodology to the case of IEEE 802.15.4/ZigBee cluster-tree WSNs without significant changes to the protocol. Finally, we analyze and demonstrate the validity of our methodology through a comprehensive simulation and experimental validation using commercially available technology on a Structural Health Monitoring application scenario.
Resumo:
Background: The validity of ensemble averaging on event-related potential (ERP) data has been questioned, due to its assumption that the ERP is identical across trials. Thus, there is a need for preliminary testing for cluster structure in the data. New method: We propose a complete pipeline for the cluster analysis of ERP data. To increase the signalto-noise (SNR) ratio of the raw single-trials, we used a denoising method based on Empirical Mode Decomposition (EMD). Next, we used a bootstrap-based method to determine the number of clusters, through a measure called the Stability Index (SI). We then used a clustering algorithm based on a Genetic Algorithm (GA)to define initial cluster centroids for subsequent k-means clustering. Finally, we visualised the clustering results through a scheme based on Principal Component Analysis (PCA). Results: After validating the pipeline on simulated data, we tested it on data from two experiments – a P300 speller paradigm on a single subject and a language processing study on 25 subjects. Results revealed evidence for the existence of 6 clusters in one experimental condition from the language processing study. Further, a two-way chi-square test revealed an influence of subject on cluster membership.
Resumo:
The aim of the present study was to determine whether under-reporting rates vary between dietary pattern Clusters. Subjects were sixty-five Brazilian women. During 3 weeks, anthropometric data were collected. total energy expenditure (TEE) was determined by the doubly labelled water method and diet Was Measured. Energy intake (El) and the daily frequency of consumption per 1000 kJ of twenty-two food groups were obtained from a FFQ. These frequencies were entered into a Cluster analysis procedure in order to obtain dietary patterns. Under-reporters were defined Lis those who did not lose more than 1 kg of body weight during the study and presented EI:TEE less than 0.82. Three dietary pattern clusters were identified and named according to their most recurrent food groups: sweet foods (SW). starchy foods (ST) and health), (H). Subjects from the healthy cluster had the lowest mean EI:TEE (SW = 0.86, ST = 0.71 and H = 0.58: P = 0.003) and EI - TEE (SW = -0.49 MJ, ST = - 3.20 MJ and H = -5.09 MJ; P = 0.008). The proportion of Under-reporters was 45.2 (95 % CI 35.5, 55.0) % in the SW Cluster: 58.3 (95 % CI 48.6, 68.0) % in the ST Cluster and 70.0 (95 % CI 61.0, 79) % in the H cluster (P=0.34). Thus, in Brazilian women, Under-reporting of El is not uniformly distributed among, dietary pattern clusters and tends to be more severe among subjects from the healthy cluster. This cluster is more consistent with both dietary guidelines and with what lay individuals usually consider `healthy eating`.
Resumo:
The new social panorama resulting from aging of the Brazilian population is leading to significant transformations within healthcare. Through the cluster analysis strategy, it was sought to describe the specific care demands of the elderly population, using frailty components. Cross-sectional study based on reviewing medical records, conducted in the geriatric outpatient clinic, Hospital de Clínicas, Universidade Estadual de Campinas (Unicamp). Ninety-eight elderly users of this clinic were evaluated using cluster analysis and instruments for assessing their overall geriatric status and frailty characteristics. The variables that most strongly influenced the formation of clusters were age, functional capacities, cognitive capacity, presence of comorbidities and number of medications used. Three main groups of elderly people could be identified: one with good cognitive and functional performance but with high prevalence of comorbidities (mean age 77.9 years, cognitive impairment in 28.6% and mean of 7.4 comorbidities); a second with more advanced age, greater cognitive impairment and greater dependence (mean age 88.5 years old, cognitive impairment in 84.6% and mean of 7.1 comorbidities); and a third younger group with poor cognitive performance and greater number of comorbidities but functionally independent (mean age 78.5 years old, cognitive impairment in 89.6% and mean of 7.4 comorbidities). These data characterize the profile of this population and can be used as the basis for developing efficient strategies aimed at diminishing functional dependence, poor self-rated health and impaired quality of life.
Resumo:
to assess the construct validity and reliability of the Pediatric Patient Classification Instrument. correlation study developed at a teaching hospital. The classification involved 227 patients, using the pediatric patient classification instrument. The construct validity was assessed through the factor analysis approach and reliability through internal consistency. the Exploratory Factor Analysis identified three constructs with 67.5% of variance explanation and, in the reliability assessment, the following Cronbach's alpha coefficients were found: 0.92 for the instrument as a whole; 0.88 for the Patient domain; 0.81 for the Family domain; 0.44 for the Therapeutic procedures domain. the instrument evidenced its construct validity and reliability, and these analyses indicate the feasibility of the instrument. The validation of the Pediatric Patient Classification Instrument still represents a challenge, due to its relevance for a closer look at pediatric nursing care and management. Further research should be considered to explore its dimensionality and content validity.
Resumo:
Introductions: In the care of hypertension, it is important that health professionals possess available tools that allow evaluating the impairment of the health-related quality of life, according to the severity of hypertension and the risk for cardiovascular events. Among the instruments developed for the assessment of health-related quality of life, there is the Mini-Cuestionario of Calidad de Vida en la Hipertensión Arterial (MINICHAL) recently adapted to the Brazilian culture. Objective: To estimate the validity of known groups of the Brazilian version of the MINICHAL regarding the classification of risk for cardiovascular events, symptoms, severity of dyspnea and target-organ damage. Methods: Data of 200 hypertensive outpatients concerning sociodemographic and clinical information and health-related quality of life were gathered by consulting the medical charts and the application of the Brazilian version of MINICHAL. The Mann-Whitney test was used to compare health-related quality of life in relation to symptoms and target-organ damage. The Kruskal-Wallis test and ANOVA with ranks transformation were used to compare health-related quality of life in relation to the classification of risk for cardiovascular events and intensity of dyspnea, respectively. Results: The MINICHAL was able to discriminate health-related quality of life in relation to symptoms and kidney damage, but did not discriminate health-related quality of life in relation to the classification of risk for cardiovascular events. Conclusion: The Brazilian version of the MINICHAL is a questionnaire capable of discriminating differences on the health‑related quality of life regarding dyspnea, chest pain, palpitation, lipothymy, cephalea and renal damage.Fundamento: No cuidado ao hipertenso, é importante que o profissional de saúde disponha de ferramentas que possibilitem avaliar o comprometimento da qualidade de vida relacionada à saúde, de acordo com a gravidade da hipertensão e o risco para eventos cardiovasculares. Dentre os instrumentos criados para avaliação da qualidade de vida relacionada à saúde, destaca-se o Mini-Cuestionario de Calidad de Vida en la Hipertensión Arterial (MINICHAL), recentemente adaptado para a cultura brasileira. Objetivo: Estimar a validade de grupos conhecidos da versão brasileira do MINICHAL em relação à classificação de risco para eventos cardiovasculares, sintomas, intensidade da dispneia e lesões de órgãos-alvo. Métodos: Foram investigados 200 hipertensos em seguimento ambulatorial, cujos dados sociodemográficos, clínicos e de qualidade de vida relacionada à saúde foram obtidos por meio de consulta ao prontuário e da aplicação da versão brasileira do MINICHAL. O teste de Mann-Whitney foi utilizado para comparar qualidade de vida relacionada à saúde em relação aos sintomas e às lesões de órgãos-alvo. Teste de Kruskal-Wallis e ANOVA com transformação nos ranks foram empregados para comparar qualidade de vida relacionada à saúde em relação à classificação de risco para eventos cardiovasculares e intensidade da dispneia, respectivamente. Resultados: O MINICHAL discriminou qualidade de vida relacionada à saúde em relação aos sintomas e dano renal (lesões de órgãos-alvo), porém não discriminou qualidade de vida relacionada à saúde em relação à classificação de risco para eventos cardiovasculares. Conclusão: A versão brasileira do MINICHAL é um instrumento capaz de discriminar diferenças na qualidade de vida relacionada à saúde em relação aos sintomas de dispneia, precordialgia, palpitação, lipotímia, cefaleia e presença de dano renal.