Communication and matrix computations on large message passing systems (Q750124): Difference between revisions
From MaRDI portal
Created a new Item |
Added link to MaRDI item. |
||
links / mardi / name | links / mardi / name | ||
Revision as of 10:24, 30 January 2024
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | Communication and matrix computations on large message passing systems |
scientific article |
Statements
Communication and matrix computations on large message passing systems (English)
0 references
1990
0 references
This paper studies the consequences for matrix computations of working with a large number of general purpose processors, say from 10000 to 20000, connected through a topology that allows any processor to communicate directly only to its immediate neighbors, and assuming a rate of a million arithmetic operations per second for each processor. (No such a system has been built so far.) One consequence of these hypotheses is that problems can be posed that are too large to be solved in a reasonable amount of time. This suggests that the next generation of parallel computers must be designed to solve smaller matrix problems faster, rather than to solve the largest problems they could handle. To be computationally intensive, parallel algorithms must communicate a lot. However, it is known that the rate of computation to communication decreases with problem size. But if a problem is large enough to be efficiently solved, is it also so large that its solving can only be reached in an unreasonable time span? The author partitions an arbitrary matrix into square submatrices and assigns these to individual processors. Some typical communication tasks are then considered and their implementation done in two types of networks: a grid and a hypercube. The communication times for problems of different sizes are then computed and compared with the computation costs, so as to determine break-even points. The main result assures that the efficiency of certain important matrix algorithms at fine granularity is restricted by the granularity of the communication. A remedy, called simple streaming, is then proposed.
0 references
concurrent calculations
0 references
message passing systems
0 references
timing analysis
0 references
matrix computations
0 references
parallel computers
0 references
parallel algorithms
0 references
hypercube
0 references
communication times
0 references
computation costs
0 references
efficiency
0 references