Matrix-vector products (mat-vecs) form the core of iterative methods used for solving dense linear systems. Often, these systems arise in the solution of integral equations used in electromagnetics, heat transfer, and wave propagation. In this paper, we present a parallel approximate method for computing mat-vecs used in the solution of integral equations. We use this method to compute dense mat-vecs of hundreds of thousands of elements. The combined speedups obtained from the use of approximate methods and parallel processing represent an improvement of several orders of magnitude over exact mat-vecs on uniprocessors. We demonstrate that our parallel formulation incurs minimal parallel processing overhead and scales up to a large number of processors. We study the impact of varying the accuracy of the approximate mat-vec on overall time and on parallel efficiency. Experimental results are presented for 256 processor Cray T3D and Thinking Machines CM5 parallel computers. We have achieved computation rates in excess of 5 GFLOPS on the T3D.
|Original language||English (US)|
|Number of pages||20|
|Journal||Proceedings of the ACM/IEEE Supercomputing Conference|
|State||Published - Dec 1 1995|
|Event||Proceedings of the 1995 ACM/IEEE Supercomputing Conference. Part 2 (of 2) - San Diego, CA, USA|
Duration: Dec 3 1995 → Dec 8 1995