登入選單
返回Google圖書搜尋
註釋Optimization algorithms typically require the solution of many systems of linear equations B sub Y sub = b sub. When large numbers of variables or constraints are present, these linear systems could account for much of the total computation time. Both direct and iterative equation solvers are needed in practice. Unfortunately, most of the off-the shelf solvers are designed for single systems, whereas optimization problems give rise to hundreds or thousands of systems. To avoid refactorization, or to speed the convergence of an iterative method, it is essential to note that B sub is related to B sub - 1. The authors review various sparse matrices that arise in optimization, and discuss compromises that are currently being made in dealing with them. Since significant advances continue to be made with single-system solvers they give special attention to methods that allow such solvers to be used repeatedly on a sequence of modified systems (e.g., the product-form update; use of the Schur complement). The speed of factorizing a matrix then becomes relatively less important than the efficiency of subsequent solves with very many right-hand sides. At the same time it is hoped that future improvements to linear-equation software will be oriented more specifically to the case of related matrices B sub k. (Author).