Currently displaying 1 – 20 of 20

Showing per page

Order by Relevance | Title | Year of publication

Combining the preconditioned conjugate gradient method and a matrix iterative method

Jan Zítko — 1996

Applications of Mathematics

The preconditioned conjugate gradient method for solving the system of linear algebraic equations with a positive definite matrix is investigated. The initial approximation for conjugate gradient is constructed as a result of a matrix iteration method after m steps. The behaviour of the error vector for such a combined method is studied and special numerical tests and conclusions are made.

Using successive approximations for improving the convergence of GMRES method

Jan Zítko — 1998

Applications of Mathematics

In this paper, our attention is concentrated on the GMRES method for the solution of the system ( I - T ) x = b of linear algebraic equations with a nonsymmetric matrix. We perform m pre-iterations y l + 1 = T y l + b before starting GMRES and put y m for the initial approximation in GMRES. We derive an upper estimate for the norm of the error vector in dependence on the m th powers of eigenvalues of the matrix T . Further we study under what eigenvalues lay-out this upper estimate is the best one. The estimate shows and numerical...

Convergence of extrapolation coefficients

Jan Zítko — 1984

Aplikace matematiky

Let x k + 1 = T x k + b be an iterative process for solving the operator equation x = T x + b in Hilbert space X . Let the sequence { x k } k = o formed by the above described iterative process be convergent for some initial approximation x o with a limit x * = T x * + b . For given l > 1 , m 0 , m 1 , , m l let us define a new sequence { y k } k = m 1 by the formula y k = α 0 ( k ) x k + α 1 ( k ) x k - m 1 + ... + α l ( k ) x k - m l , where α i ( k ) are obtained by solving a minimization problem for a given functional. In this paper convergence properties of α i ( k ) are investigated and on the basis of the results thus obtainded it is proved that lim k x * - y k / x * - x k p = 0 for some p 1 .

Two step extrapolation and optimum choice of relaxation factor of the extrapolated S.O.R. method

Jan Zítko — 1988

Aplikace matematiky

Limits of the extrapolation coefficients are rational functions of several poles with the largest moduli of the resolvent operator R ( λ , T ) = ( λ I - T ) - 1 and therefore good estimates of these poles could be calculated from these coefficients. The calculation is very easy for the case of two coefficients and its practical effect in finite dimensional space is considerable. The results are used for acceleration of S.O.R. method.

Derivation of BiCG from the conditions defining Lanczos' method for solving a system of linear equations

Petr TichýJan Zítko — 1998

Applications of Mathematics

Lanczos’ method for solving the system of linear algebraic equations A x = b consists in constructing a sequence of vectors x k in such a way that r k = b - A x k r 0 + A 𝒦 k ( A , r 0 ) and r k 𝒦 k ( A T , r ˜ 0 ) . This sequence of vectors can be computed by the BiCG (BiOMin) algorithm. In this paper is shown how to obtain the recurrences of BiCG (BiOMin) directly from this conditions.

Poznámka k superrelaxační metodě

Emil HumhalJan Zítko — 1967

Aplikace matematiky

The paper investigates the dependence of the number of iterations on the change of a relaxation factor by the over relaxation method (SOR). This dependence has been formulated in two theorems and explained in tables and graphs enclosed to the paper.

Automatic differentiation and its program realization

Jan HartmanLadislav LukšanJan Zítko — 2009


Automatic differentiation is an effective method for evaluating derivatives of function, which is defined by a formula or a program. Program for evaluating of value of function is by automatic differentiation modified to program, which also evaluates values of derivatives. Computed values are exact up to computer precision and their evaluation is very quick. In this article, we describe a program realization of automatic differentiation. This implementation is prepared in the system UFO, but its...

Calculation of the greatest common divisor of perturbed polynomials

Zítko, JanEliaš, Ján — 2013

Programs and Algorithms of Numerical Mathematics

The coefficients of the greatest common divisor of two polynomials f and g (GCD ( f , g ) ) can be obtained from the Sylvester subresultant matrix S j ( f , g ) transformed to lower triangular form, where 1 j d and d = deg(GCD ( f , g ) ) needs to be computed. Firstly, it is supposed that the coefficients of polynomials are given exactly. Transformations of S j ( f , g ) for an arbitrary allowable j are in details described and an algorithm for the calculation of the GCD ( f , g ) is formulated. If inexact polynomials are given, then an approximate greatest...

Approximate polynomial GCD

Eliaš, JánZítko, Jan — 2013

Programs and Algorithms of Numerical Mathematics

The computation of polynomial greatest common divisor (GCD) ranks among basic algebraic problems with many applications, for example, in image processing and control theory. The problem of the GCD computing of two exact polynomials is well defined and can be solved symbolically, for example, by the oldest and commonly used Euclid’s algorithm. However, this is an ill-posed problem, particularly when some unknown noise is applied to the polynomial coefficients. Hence, new methods for the GCD computation...

An improvement of Euclid's algorithm

Zítko, JanKuřátko, Jan — 2010

Programs and Algorithms of Numerical Mathematics

The paper introduces the calculation of a greatest common divisor of two univariate polynomials. Euclid’s algorithm can be easily simulated by the reduction of the Sylvester matrix to an upper triangular form. This is performed by using c - s transformation and Q R -factorization methods. Both procedures are described and numerically compared. Computations are performed in the floating point environment.

Comparison of algorithms for calculation of the greatest common divisor of several polynomials

Eckstein, JiříZítko, Jan — 2015

Programs and Algorithms of Numerical Mathematics

The computation of the greatest common divisor (GCD) has many applications in several disciplines including computer graphics, image deblurring problem or computing multiple roots of inexact polynomials. In this paper, Sylvester and Bézout matrices are considered for this purpose. The computation is divided into three stages. A rank revealing method is shortly mentioned in the first one and then the algorithms for calculation of an approximation of GCD are formulated. In the final stage the coefficients...

Alternating iterative scheme for the solution of block-structured systems

Šípek, JanZítko, Jan — 2004

Programs and Algorithms of Numerical Mathematics

We consider the solution of linear system with a block-structured matrix of saddle point type. The solution technique is based on the idea of the classical alternating-direction implicit iterative method where symmetric-antisymmetric splitting of the coefficient matrix is used. To find an optimal parameter for solving the system with a symmetric matrix, the polynomial filters are considered. The CGW method is used for systems with skew-symmetric matrix. The numerical tests compare the results obtained...

Page 1 Next

Download Results (CSV)