919 resultados para Model-based Categorical Sequence Clustering
Resumo:
Virtual machines (VMs) are powerful platforms for building agile datacenters and emerging cloud systems. However, resource management for a VM-based system is still a challenging task. First, the complexity of application workloads as well as the interference among competing workloads makes it difficult to understand their VMs’ resource demands for meeting their Quality of Service (QoS) targets; Second, the dynamics in the applications and system makes it also difficult to maintain the desired QoS target while the environment changes; Third, the transparency of virtualization presents a hurdle for guest-layer application and host-layer VM scheduler to cooperate and improve application QoS and system efficiency. This dissertation proposes to address the above challenges through fuzzy modeling and control theory based VM resource management. First, a fuzzy-logic-based nonlinear modeling approach is proposed to accurately capture a VM’s complex demands of multiple types of resources automatically online based on the observed workload and resource usages. Second, to enable fast adaption for resource management, the fuzzy modeling approach is integrated with a predictive-control-based controller to form a new Fuzzy Modeling Predictive Control (FMPC) approach which can quickly track the applications’ QoS targets and optimize the resource allocations under dynamic changes in the system. Finally, to address the limitations of black-box-based resource management solutions, a cross-layer optimization approach is proposed to enable cooperation between a VM’s host and guest layers and further improve the application QoS and resource usage efficiency. The above proposed approaches are prototyped and evaluated on a Xen-based virtualized system and evaluated with representative benchmarks including TPC-H, RUBiS, and TerraFly. The results demonstrate that the fuzzy-modeling-based approach improves the accuracy in resource prediction by up to 31.4% compared to conventional regression approaches. The FMPC approach substantially outperforms the traditional linear-model-based predictive control approach in meeting application QoS targets for an oversubscribed system. It is able to manage dynamic VM resource allocations and migrations for over 100 concurrent VMs across multiple hosts with good efficiency. Finally, the cross-layer optimization approach further improves the performance of a virtualized application by up to 40% when the resources are contended by dynamic workloads.
Resumo:
Software development guidelines are a set of rules which can help improve the quality of software. These rules are defined on the basis of experience gained by the software development community over time. This paper discusses a set of design guidelines for model-based development of complex real-time embedded software systems. To be precise, we propose nine design conventions, three design patterns and thirteen antipatterns for developing UML-RT models. These guidelines have been identified based on our analysis of around 100 UML-RT models from industry and academia. Most of the guidelines are explained with the help of examples, and standard templates from the current state of the art are used for documenting the design rules.
Resumo:
Emerging cybersecurity vulnerabilities in supervisory control and data acquisition (SCADA) systems are becoming urgent engineering issues for modern substations. This paper proposes a novel intrusion detection system (IDS) tailored for cybersecurity of IEC 61850 based substations. The proposed IDS integrates physical knowledge, protocol specifications and logical behaviours to provide a comprehensive and effective solution that is able to mitigate various cyberattacks. The proposed approach comprises access control detection, protocol whitelisting, model-based detection, and multi-parameter based detection. This SCADA-specific IDS is implemented and validated using a comprehensive and realistic cyber-physical test-bed and data from a real 500kV smart substation.
Resumo:
Thesis (Ph.D.)--University of Washington, 2016-08
Resumo:
Thesis (Ph.D.)--University of Washington, 2016-08
Resumo:
According to law number 12.715/2012, Brazilian government instituted guidelines for a program named Inovar-Auto. In this context, energy efficiency is a survival requirement for Brazilian automotive industry from September 2016. As proposed by law, energy efficiency is not going to be calculated by models only. It is going to be calculated by the whole universe of new vehicles registered. In this scenario, the composition of vehicles sold in market will be a key factor on profits of each automaker. Energy efficiency and its consequences should be taken into consideration in all of its aspects. In this scenario, emerges the following question: which is the efficiency curve of one automaker for long term, allowing them to adequate to rules, keep balancing on investment in technologies, increasing energy efficiency without affecting competitiveness of product lineup? Among several variables to be considered, one can highlight the analysis of manufacturing costs, customer value perception and market share, which characterizes this problem as a multi-criteria decision-making. To tackle the energy efficiency problem required by legislation, this paper proposes a framework of multi-criteria decision-making. The proposed framework combines Delphi group and Analytic Hierarchy Process to identify suitable alternatives for automakers to incorporate in main Brazilian vehicle segments. A forecast model based on artificial neural networks was used to estimate vehicle sales demand to validate expected results. This approach is demonstrated with a real case study using public vehicles sales data of Brazilian automakers and public energy efficiency data.
Resumo:
A variety of physical and biomedical imaging techniques, such as digital holography, interferometric synthetic aperture radar (InSAR), or magnetic resonance imaging (MRI) enable measurement of the phase of a physical quantity additionally to its amplitude. However, the phase can commonly only be measured modulo 2π, as a so called wrapped phase map. Phase unwrapping is the process of obtaining the underlying physical phase map from the wrapped phase. Tile-based phase unwrapping algorithms operate by first tessellating the phase map, then unwrapping individual tiles, and finally merging them to a continuous phase map. They can be implemented computationally efficiently and are robust to noise. However, they are prone to failure in the presence of phase residues or erroneous unwraps of single tiles. We tried to overcome these shortcomings by creating novel tile unwrapping and merging algorithms as well as creating a framework that allows to combine them in modular fashion. To increase the robustness of the tile unwrapping step, we implemented a model-based algorithm that makes efficient use of linear algebra to unwrap individual tiles. Furthermore, we adapted an established pixel-based unwrapping algorithm to create a quality guided tile merger. These original algorithms as well as previously existing ones were implemented in a modular phase unwrapping C++ framework. By examining different combinations of unwrapping and merging algorithms we compared our method to existing approaches. We could show that the appropriate choice of unwrapping and merging algorithms can significantly improve the unwrapped result in the presence of phase residues and noise. Beyond that, our modular framework allows for efficient design and test of new tile-based phase unwrapping algorithms. The software developed in this study is freely available.
Resumo:
The recent advent of new technologies has led to huge amounts of genomic data. With these data come new opportunities to understand biological cellular processes underlying hidden regulation mechanisms and to identify disease related biomarkers for informative diagnostics. However, extracting biological insights from the immense amounts of genomic data is a challenging task. Therefore, effective and efficient computational techniques are needed to analyze and interpret genomic data. In this thesis, novel computational methods are proposed to address such challenges: a Bayesian mixture model, an extended Bayesian mixture model, and an Eigen-brain approach. The Bayesian mixture framework involves integration of the Bayesian network and the Gaussian mixture model. Based on the proposed framework and its conjunction with K-means clustering and principal component analysis (PCA), biological insights are derived such as context specific/dependent relationships and nested structures within microarray where biological replicates are encapsulated. The Bayesian mixture framework is then extended to explore posterior distributions of network space by incorporating a Markov chain Monte Carlo (MCMC) model. The extended Bayesian mixture model summarizes the sampled network structures by extracting biologically meaningful features. Finally, an Eigen-brain approach is proposed to analyze in situ hybridization data for the identification of the cell-type specific genes, which can be useful for informative blood diagnostics. Computational results with region-based clustering reveals the critical evidence for the consistency with brain anatomical structure.
Resumo:
This thesis considers a three- dimensional numerical model based on 3-D Navier— Stokes and continuity equations involving various wind speeds (North west), water surface levels, horizontal shier stresses, eddy viscosity, densities of oil and gas condensate- water mixture flows. The model is used to simulate the prediction of the surface movement of oil and gas condensate slicks from spill accident in the north coasts of Persian Gulf.
Resumo:
A simple model based on, the maximum energy that an athlete can produce in a small time interval is used to describe the high and long jump. Conservation of angular momentum is used to explain why an athlete should, run horizontally to perform a vertical jump. Our results agree with world records. (c) 2005 American Association of Physics Teachers.
Resumo:
The primary goal of systems biology is to integrate complex omics data, and data obtained from traditional experimental studies in order to provide a holistic understanding of organismal function. One way of achieving this aim is to generate genome-scale metabolic models (GEMs), which contain information on all metabolites, enzyme-coding genes, and biochemical reactions in a biological system. Drosophila melanogaster GEM has not been reconstructed to date. Constraint-free genome-wide metabolic model of the fruit fly has been reconstructed in our lab, identifying gaps, where no enzyme was identified and metabolites were either only produced or consume. The main focus of the work presented in this thesis was to develop a pipeline for efficient gap filling using metabolomics approaches combined with standard reverse genetics methods, using 5-hydroxyisourate hydrolase (5-HIUH) as an example. 5-HIUH plays a role in urate degradation pathway. Inability to degrade urate can lead to inborn errors of metabolism (IEMs) in humans, including hyperuricemia. Based on sequence analysis Drosophila CG30016 gene was hypothesised to encode 5- HIUH. CG30016 knockout flies were examined to identify Malpighian tubules phenotype, and shortened lifespan might reflect kidney disorders in hyperuricemia in humans. Moreover, LC-MS analysis of mutant tubules revealed that CG30016 is involved in purine metabolism, and specifically urate degradation pathway. However, the exact role of the gene has not been identified, and the complete method for gap filling has not been developed. Nevertheless, thanks to the work presented here, we are a step closer towards the development of a gap-filling pipeline in Drosophila melanogaster GEM. Importantly, the areas that require further optimisation were identified and are the focus of future research. Moreover, LC-MS analysis confirmed that tubules rather than the whole fly were more suitable for metabolomics analysis of purine metabolism. Previously, Dow/Davies lab has generated the most complete tissue-specific transcriptomic atlas for Drosophila – FlyAtlas.org, which provides data on gene expression across multiple tissues of adult fly and larva. FlyAtlas revealed that transcripts of many genes are enriched in specific Drosophila tissues, and that it is possible to deduce the functions of individual tissues within the fly. Based on FlyAtlas data, it has become clear that the fly (like other metazoan species) must be considered as a set of tissues, each 2 with its own distinct transcriptional and functional profile. Moreover, it revealed that for about 30% of the genome, reverse genetic methods (i.e. mutation in an unknown gene followed by observation of phenotype) are only useful if specific tissues are investigated. Based on the FlyAtlas findings, we aimed to build a primary tissue-specific metabolome of the fruit fly, in order to establish whether different Drosophila tissues have different metabolomes and if they correspond to tissue-specific transcriptome of the fruit fly (FlyAtlas.org). Different fly tissues have been dissected and their metabolome elucidated using LC-MS. The results confirmed that tissue metabolomes differ significantly from each other and from the whole fly, and that some of these differences can be correlated to the tissue function. The results illustrate the need to study individual tissues as well as the whole organism. It is clear that some metabolites that play an important role in a given tissue might not be detected in the whole fly sample because their abundance is much lower in comparison to other metabolites present in all tissues, which prevent the detection of the tissue-specific compound.
Resumo:
The main aim of this study was to determine the impact of innovation on productivity in service sector companies — especially those in the hospitality sector — that value the reduction of environmental impact as relevant to the innovation process. We used a structural analysis model based on the one developed by Crépon, Duguet, and Mairesse (1998). This model is known as the CDM model (an acronym of the authors’ surnames). These authors developed seminal studies in the field of the relationships between innovation and productivity (see Griliches 1979; Pakes and Grilliches 1980). The main advantage of the CDM model is its ability to integrate the process of innovation and business productivity from an empirical perspective.
Resumo:
Ionic liquids (ILs) have attracted great attention, from both industry and academia, as alternative fluids for very different types of applications. The large number of cations and anions allow a wide range of physical and chemical characteristics to be designed. However, the exhaustive measurement of all these systems is impractical, thus requiring the use of a predictive model for their study. In this work, the predictive capability of the conductor-like screening model for real solvents (COSMO-RS), a model based on unimolecular quantum chemistry calculations, was evaluated for the prediction water activity coefficient at infinite dilution, gamma(infinity)(w), in several classes of ILs. A critical evaluation of the experimental and predicted data using COSMO-RS was carried out. The global average relative deviation was found to be 27.2%, indicating that the model presents a satisfactory prediction ability to estimate gamma(infinity)(w) in a broad range of ILs. The results also showed that the basicity of the ILs anions plays an important role in their interaction with water, and it considerably determines the enthalpic behavior of the binary mixtures composed by Its and water. Concerning the cation effect, it is possible to state that generally gamma(infinity)(w) increases with the cation size, but it is shown that the cation-anion interaction strength is also important and is strongly correlated to the anion ability to interact with water. The results here reported are relevant in the understanding of ILs-water interactions and the impact of the various structural features of its on the gamma(infinity)(w) as these allow the development of guidelines for the choice of the most suitable lLs with enhanced interaction with water.
Resumo:
Inter-subject parcellation of functional Magnetic Resonance Imaging (fMRI) data based on a standard General Linear Model (GLM) and spectral clustering was recently proposed as a means to alleviate the issues associated with spatial normalization in fMRI. However, for all its appeal, a GLM-based parcellation approach introduces its own biases, in the form of a priori knowledge about the shape of Hemodynamic Response Function (HRF) and task-related signal changes, or about the subject behaviour during the task. In this paper, we introduce a data-driven version of the spectral clustering parcellation, based on Independent Component Analysis (ICA) and Partial Least Squares (PLS) instead of the GLM. First, a number of independent components are automatically selected. Seed voxels are then obtained from the associated ICA maps and we compute the PLS latent variables between the fMRI signal of the seed voxels (which covers regional variations of the HRF) and the principal components of the signal across all voxels. Finally, we parcellate all subjects data with a spectral clustering of the PLS latent variables. We present results of the application of the proposed method on both single-subject and multi-subject fMRI datasets. Preliminary experimental results, evaluated with intra-parcel variance of GLM t-values and PLS derived t-values, indicate that this data-driven approach offers improvement in terms of parcellation accuracy over GLM based techniques.
Resumo:
In the first part of this thesis we search for beyond the Standard Model physics through the search for anomalous production of the Higgs boson using the razor kinematic variables. We search for anomalous Higgs boson production using proton-proton collisions at center of mass energy √s=8 TeV collected by the Compact Muon Solenoid experiment at the Large Hadron Collider corresponding to an integrated luminosity of 19.8 fb-1.
In the second part we present a novel method for using a quantum annealer to train a classifier to recognize events containing a Higgs boson decaying to two photons. We train that classifier using simulated proton-proton collisions at √s=8 TeV producing either a Standard Model Higgs boson decaying to two photons or a non-resonant Standard Model process that produces a two photon final state.
The production mechanisms of the Higgs boson are precisely predicted by the Standard Model based on its association with the mechanism of electroweak symmetry breaking. We measure the yield of Higgs bosons decaying to two photons in kinematic regions predicted to have very little contribution from a Standard Model Higgs boson and search for an excess of events, which would be evidence of either non-standard production or non-standard properties of the Higgs boson. We divide the events into disjoint categories based on kinematic properties and the presence of additional b-quarks produced in the collisions. In each of these disjoint categories, we use the razor kinematic variables to characterize events with topological configurations incompatible with typical configurations found from standard model production of the Higgs boson.
We observe an excess of events with di-photon invariant mass compatible with the Higgs boson mass and localized in a small region of the razor plane. We observe 5 events with a predicted background of 0.54 ± 0.28, which observation has a p-value of 10-3 and a local significance of 3.35σ. This background prediction comes from 0.48 predicted non-resonant background events and 0.07 predicted SM higgs boson events. We proceed to investigate the properties of this excess, finding that it provides a very compelling peak in the di-photon invariant mass distribution and is physically separated in the razor plane from predicted background. Using another method of measuring the background and significance of the excess, we find a 2.5σ deviation from the Standard Model hypothesis over a broader range of the razor plane.
In the second part of the thesis we transform the problem of training a classifier to distinguish events with a Higgs boson decaying to two photons from events with other sources of photon pairs into the Hamiltonian of a spin system, the ground state of which is the best classifier. We then use a quantum annealer to find the ground state of this Hamiltonian and train the classifier. We find that we are able to do this successfully in less than 400 annealing runs for a problem of median difficulty at the largest problem size considered. The networks trained in this manner exhibit good classification performance, competitive with the more complicated machine learning techniques, and are highly resistant to overtraining. We also find that the nature of the training gives access to additional solutions that can be used to improve the classification performance by up to 1.2% in some regions.