1000 resultados para computational algebra


Relevância:

70.00% 70.00%

Publicador:

Resumo:

We construct five new elements of degree 6 in the nucleus of the free alternative algebra. We use the representation theory of the symmetric group to locate the elements. We use the computer algebra system ALBERT and an extension of ALBERT to express the elements in compact form and to show that these new elements are not a consequence of the known clegree-5 elements in the nucleus. We prove that these five new elements and four known elements form a basis for the subspace of nuclear elements of degree 6. Our calculations are done using modular arithmetic to save memory and time. The calculations can be done in characteristic zero or any prime greater than 6, and similar results are expected. We generated the nuclear elements using prime 103. We check our answer using five other primes.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

We consider boundary layer flow of a micropolar fluid driven by a porous stretching sheet. A similarity solution is defined, and numerical solutions using Runge-Kutta and quasilinearisation schemes are obtained. A perturbation analysis is also used to derive analytic solutions to first order in the perturbing parameter. The resulting closed form solutions involve relatively complex expressions, and the analysis is made more tractable by a combination of offline and online work using a computational algebra system (CAS). For this combined numerical and analytic approach, the perturbation analysis yields a number of benefits with regard to the numerical work. The existence of a closed form solution helps to discriminate between acceptable and spurious numerical solutions. Also, the expressions obtained from the perturbation work can provide an accurate description of the solution for ranges of parameters where the numerical approaches considered here prove computationally more difficult.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The minimum distance of linear block codes is one of the important parameter that indicates the error performance of the code. When the code rate is less than 1/2, efficient algorithms are available for finding minimum distance using the concept of information sets. When the code rate is greater than 1/2, only one information set is available and efficiency suffers. In this paper, we investigate and propose a novel algorithm to find the minimum distance of linear block codes with the code rate greater than 1/2. We propose to reverse the roles of information set and parity set to get virtually another information set to improve the efficiency. This method is 67.7 times faster than the minimum distance algorithm implemented in MAGMA Computational Algebra System for a (80, 45) linear block code.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

We consider polynomial identities satisfied by nonhomogeneous subalgebras of Lie and special Jordan superalgebras: we ignore the grading and regard the superalgebra as an ordinary algebra. The Lie case has been studied by Volichenko and Baranov: they found identities in degrees 3, 4 and 5 which imply all the identities in degrees <= 6. We simplify their identities in degree 5, and show that there are no new identities in degree 7. The Jordan case has not previously been studied: we find identities in degrees 3, 4, 5 and 6 which imply all the identities in degrees <= 6, and demonstrate the existence of further new identities in degree 7. our proofs depend on computer algebra: we use the representation theory of the symmetric group, the Hermite normal form of an integer matrix, the LLL algorithm for lattice basis reduction, and the Chinese remainder theorem. (C) 2009 Elsevier Inc. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Linear algebra provides theory and technology that are the cornerstones of a range of cutting edge mathematical applications, from designing computer games to complex industrial problems, as well as more traditional applications in statistics and mathematical modelling. Once past introductions to matrices and vectors, the challenges of balancing theory, applications and computational work across mathematical and statistical topics and problems are considerable, particularly given the diversity of abilities and interests in typical cohorts. This paper considers two such cohorts in a second level linear algebra course in different years. The course objectives and materials were almost the same, but some changes were made in the assessment package. In addition to considering effects of these changes, the links with achievement in first year courses are analysed, together with achievement in a following computational mathematics course. Some results that may initially appear surprising provide insight into the components of student learning in linear algebra.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The R statistical environment and language has demonstrated particular strengths for interactive development of statistical algorithms, as well as data modelling and visualisation. Its current implementation has an interpreter at its core which may result in a performance penalty in comparison to directly executing user algorithms in the native machine code of the host CPU. In contrast, the C++ language has no built-in visualisation capabilities, handling of linear algebra or even basic statistical algorithms; however, user programs are converted to high-performance machine code, ahead of execution. A new method avoids possible speed penalties in R by using the Rcpp extension package in conjunction with the Armadillo C++ matrix library. In addition to the inherent performance advantages of compiled code, Armadillo provides an easy-to-use template-based meta-programming framework, allowing the automatic pooling of several linear algebra operations into one, which in turn can lead to further speedups. With the aid of Rcpp and Armadillo, conversion of linear algebra centered algorithms from R to C++ becomes straightforward. The algorithms retains the overall structure as well as readability, all while maintaining a bidirectional link with the host R environment. Empirical timing comparisons of R and C++ implementations of a Kalman filtering algorithm indicate a speedup of several orders of magnitude.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Matrix decompositions, where a given matrix is represented as a product of two other matrices, are regularly used in data mining. Most matrix decompositions have their roots in linear algebra, but the needs of data mining are not always those of linear algebra. In data mining one needs to have results that are interpretable -- and what is considered interpretable in data mining can be very different to what is considered interpretable in linear algebra. --- The purpose of this thesis is to study matrix decompositions that directly address the issue of interpretability. An example is a decomposition of binary matrices where the factor matrices are assumed to be binary and the matrix multiplication is Boolean. The restriction to binary factor matrices increases interpretability -- factor matrices are of the same type as the original matrix -- and allows the use of Boolean matrix multiplication, which is often more intuitive than normal matrix multiplication with binary matrices. Also several other decomposition methods are described, and the computational complexity of computing them is studied together with the hardness of approximating the related optimization problems. Based on these studies, algorithms for constructing the decompositions are proposed. Constructing the decompositions turns out to be computationally hard, and the proposed algorithms are mostly based on various heuristics. Nevertheless, the algorithms are shown to be capable of finding good results in empirical experiments conducted with both synthetic and real-world data.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Numerical Linear Algebra (NLA) kernels are at the heart of all computational problems. These kernels require hardware acceleration for increased throughput. NLA Solvers for dense and sparse matrices differ in the way the matrices are stored and operated upon although they exhibit similar computational properties. While ASIC solutions for NLA Solvers can deliver high performance, they are not scalable, and hence are not commercially viable. In this paper, we show how NLA kernels can be accelerated on REDEFINE, a scalable runtime reconfigurable hardware platform. Compared to a software implementation, Direct Solver (Modified Faddeev's algorithm) on REDEFINE shows a 29X improvement on an average and Iterative Solver (Conjugate Gradient algorithm) shows a 15-20% improvement. We further show that solution on REDEFINE is scalable over larger problem sizes without any notable degradation in performance.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper we have developed methods to compute maps from differential equations. We take two examples. First is the case of the harmonic oscillator and the second is the case of Duffing's equation. First we convert these equations to a canonical form. This is slightly nontrivial for the Duffing's equation. Then we show a method to extend these differential equations. In the second case, symbolic algebra needs to be used. Once the extensions are accomplished, various maps are generated. The Poincare sections are seen as a special case of such generated maps. Other applications are also discussed.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This thesis studies three classes of randomized numerical linear algebra algorithms, namely: (i) randomized matrix sparsification algorithms, (ii) low-rank approximation algorithms that use randomized unitary transformations, and (iii) low-rank approximation algorithms for positive-semidefinite (PSD) matrices.

Randomized matrix sparsification algorithms set randomly chosen entries of the input matrix to zero. When the approximant is substituted for the original matrix in computations, its sparsity allows one to employ faster sparsity-exploiting algorithms. This thesis contributes bounds on the approximation error of nonuniform randomized sparsification schemes, measured in the spectral norm and two NP-hard norms that are of interest in computational graph theory and subset selection applications.

Low-rank approximations based on randomized unitary transformations have several desirable properties: they have low communication costs, are amenable to parallel implementation, and exploit the existence of fast transform algorithms. This thesis investigates the tradeoff between the accuracy and cost of generating such approximations. State-of-the-art spectral and Frobenius-norm error bounds are provided.

The last class of algorithms considered are SPSD "sketching" algorithms. Such sketches can be computed faster than approximations based on projecting onto mixtures of the columns of the matrix. The performance of several such sketching schemes is empirically evaluated using a suite of canonical matrices drawn from machine learning and data analysis applications, and a framework is developed for establishing theoretical error bounds.

In addition to studying these algorithms, this thesis extends the Matrix Laplace Transform framework to derive Chernoff and Bernstein inequalities that apply to all the eigenvalues of certain classes of random matrices. These inequalities are used to investigate the behavior of the singular values of a matrix under random sampling, and to derive convergence rates for each individual eigenvalue of a sample covariance matrix.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

South Central University

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper describes the deployment on GPUs of PROP, a program of the 2DRMP suite which models electron collisions with H-like atoms and ions. Because performance on GPUs is better in single precision than in double precision, the numerical stability of the PROP program in single precision has been studied. The numerical quality of PROP results computed in single precision and their impact on the next program of the 2DRMP suite has been analyzed. Successive versions of the PROP program on GPUs have been developed in order to improve its performance. Particular attention has been paid to the optimization of data transfers and of linear algebra operations. Performance obtained on several architectures (including NVIDIA Fermi) are presented.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Radiation induced bystander effects are secondary effects caused by the production of chemical signals by cells in response to radiation. We present a Bio-PEPA model which builds on previous modelling work in this field to predict: the surviving fraction of cells in response to radiation, the relative proportion of cell death caused by bystander signalling, the risk of non-lethal damage and the probability of observing bystander signalling for a given dose. This work provides the foundation for modelling bystander effects caused by biologically realistic dose distributions, with implications for cancer therapies.