9 resultados para Chen-Burer algorithm

em Deakin Research Online - Australia


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Inducing general functions from specific training examples is a central problem in the machine learning. Using sets of If-then rules is the most expressive and readable manner. To find If-then rules, many induction algorithms such as ID3, AQ, CN2 and their variants, were proposed. Sequential covering is the kernel technique of them. To avoid testing all possible selectors, Entropy gain is used to select the best attribute in ID3. Constraint of the size of star was introduced in AQ and beam search was adopted in CN2. These methods speed up their induction algorithms but many good selectors are filtered out. In this work, we introduce a new induction algorithm that is based on enumeration of all possible selectors. Contrary to the previous works, we use pruning power to reduce irrelative selectors. But we can guarantee that no good selectors are filtered out. Comparing with other techniques, the experiment results demonstrate
that the rules produced by our induction algorithm have high consistency and simplicity.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Accurate prediction of protein structures is very important for many applications such as drug discovery and biotechnology. Building side chains is an essential to get any reliable prediction of the protein structure for any given a protein main chain conformation. Most of the methods that predict side chain conformations use statistically generated data from known protein structures. It is a computationally intractable problem to search suitable side chains from all possible rotamers simultaneously using information of known protein structures. Reducing the number of possibility is a main issue to predict side chain conformation. This paper proposes an enumeration based similarity search algorithm to predict side chain conformations. By introducing “beam search” technique, a significant number of unrelated side chain rotamers can easily be eliminated. As a result, we can search for suitable residue side chains from all possible side chain conformations.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Visualization is one of the most effective methods for analyzing how high-dimensional data are distributed. Dimensionality reduction techniques, such as PCA, can be used to map high dimensional data to a two- or three-dimensional space. In this paper, we propose an algorithm called HyperMap that can be effectively applied to visualization. Our algorithm can be seen as a generalization of FastMap. It preserves its linear computation complexity, and overcomes several main shortcomings, especially in visualization. Since there are more than two pivot objects in each axis of a target space, more distance information needs to be preserved in each dimension. Then in visualization, the number of pivot objects can go beyond the limitation of six (2-pivot objects × 3-dimensions). Our HyperMap algorithm also gives more flexibility to the target space, such that the data distribution can be observed from various viewpoints. Its effectiveness is confirmed by empirical evaluations on both real and synthetic datasets.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The radial return mapping algorithm within the computational context of a hybrid Finite Element and Particle-In-Cell (FE/PIC) method is constructed to allow a fluid flow FE/PIC code to be applied solid mechanic problems with large displacements and large deformations. The FE/PIC method retains the robustness of an Eulerian mesh and enables tracking of material deformation by a set of Lagrangian particles or material points. In the FE/PIC approach the particle velocities are interpolated from nodal velocities and then the particle position is updated using a suitable integration scheme, such as the 4th order Runge-Kutta scheme[1]. The strain increments are obtained from gradients of the nodal velocities at the material point positions, which are then used to evaluate the stress increment and update history variables. To obtain the stress increment from the strain increment, the nonlinear constitutive equations are solved in an incremental iterative integration scheme based on a radial return mapping algorithm[2]. A plane stress extension of a rectangular shape J2 elastoplastic material with isotropic, kinematic and combined hardening is performed as an example and for validation of the enhanced FE/PIC method. It is shown that the method is suitable for analysis of problems in crystal plasticity and metal forming. The method is specifically suitable for simulation of neighbouring microstructural phases with different constitutive equations in a multiscale material modelling framework.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Sequential minimal optimization (SMO) is quite an efficient algorithm for training the support vector machine. The most important step of this algorithm is the selection of the working set, which greatly affects the training speed. The feasible direction strategy for the working set selection can decrease the objective function, however, may augment to the total calculation for selecting the working set in each of the iteration. In this paper, a new candidate working set (CWS) Strategy is presented considering the cost on the working set selection and cache performance. This new strategy can select several greatest violating samples from Cache as the iterative working sets for the next several optimizing steps, which can improve the efficiency of the kernel cache usage and reduce the computational cost related to the working set selection. The results of the theory analysis and experiments demonstrate that the proposed method can reduce the training time, especially on the large-scale datasets.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Hybrid storage systems that consist of flash-based solid state drives (SSDs) and traditional disks are now widely used. In hybrid storage systems, there exists a two-level cache hierarchy that regard dynamic random access memory (DRAM) as the first level cache and SSD as the second level cache for disk storage. However, this two-level cache hierarchy typically uses independent cache replacement policies for each level, which makes cache resource management inefficient and reduces system performance. In this paper, we propose a novel adaptive multi-level cache (AMC) replacement algorithm in hybrid storage systems. The AMC algorithm adaptively adjusts cache blocks between DRAM and SSD cache levels using an integrated solution. AMC uses combined selective promote and demote operations to dynamically determine the level in which the blocks are to be cached. In this manner, the AMC algorithm achieves multi-level cache exclusiveness and makes cache resource management more efficient. By using real-life storage traces, our evaluation shows the proposed algorithm improves hybrid multi-level cache performance and also increases the SSD lifetime compared with traditional multi-level cache replacement algorithms.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The 0/1 Knapsack Problem (KP), which is a classical NP-complete problem, has been widely applied to solving many real world problems. Ant system (AS), as one of the earliest ant colony optimization (ACO) algorithms, provides approximate solutions to 0/1 KPs. However, there are some shortcomings such as low efficiency and premature convergence in most AS algorithms. In order to overcome the shortcomings of AS, this paper proposes a rank-based AS algorithm, denoted as RAS to solve 0/1 KP. Taking advantages of the ranked ants with a higher profit, the pheromone of items will be updated with better solutions in RAS. Experimental results in different datasets show that this new kind of AS algorithm can obtain a higher efficiency and robustness when solving 0/1 KP.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

As a typical NP-complete problem, 0/1 Knapsack Problem (KP), has been widely applied in many domains for solving practical problems. Although ant colony optimization (ACO) algorithms can obtain approximate solutions to 0/1 KP, there exist some shortcomings such as the low convergence rate, premature convergence and weak robustness. In order to get rid of the above-mentioned shortcomings, this paper proposes a new kind of Physarum-based hybrid optimization algorithm, denoted as PM-ACO, based on the critical paths reserved by Physarum-inspired mathematical (PM) model. By releasing additional pheromone to items that are on the important pipelines of PM model, PM-ACO algorithms can enhance item pheromone matrix and realize a positive feedback process of updating item pheromone. The experimental results in two different datasets show that PM-ACO algorithms have a stronger robustness and a higher convergence rate compared with traditional ACO algorithms.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

With the emergence of the big data age, the issue of how to obtain valuable knowledge from a dataset efficiently and accurately has attracted increasingly attention from both academia and industry. This paper presents a Parallel Random Forest (PRF) algorithm for big data on the Apache Spark platform. The PRF algorithm is optimized based on a hybrid approach combining data-parallel and task-parallel optimization. From the perspective of data-parallel optimization, a vertical data-partitioning method is performed to reduce the data communication cost effectively, and a data-multiplexing method is performed is performed to allow the training dataset to be reused and diminish the volume of data. From the perspective of task-parallel optimization, a dual parallel approach is carried out in the training process of RF, and a task Directed Acyclic Graph (DAG) is created according to the parallel training process of PRF and the dependence of the Resilient Distributed Datasets (RDD) objects. Then, different task schedulers are invoked for the tasks in the DAG. Moreover, to improve the algorithm's accuracy for large, high-dimensional, and noisy data, we perform a dimension-reduction approach in the training process and a weighted voting approach in the prediction process prior to parallelization. Extensive experimental results indicate the superiority and notable advantages of the PRF algorithm over the relevant algorithms implemented by Spark MLlib and other studies in terms of the classification accuracy, performance, and scalability.