TY - GEN
T1 - Efficient parallel algorithms on restartable fail-stop processors
AU - Kanellakis, Paris C.
AU - Shvartsman, Alex A.
N1 - Funding Information:
The model of parallel computation known as the Parallel Random Access Machine or PRAM [FW 78] has attracted much attention in recent years. Many e@-cient and optimal algorithms have been designed for it, see the surveys [EG 88, KR 90]. The PRAM is a convenient abstraction that combines the power of parallelism with the simplicity of a RAM, but it has several unre- ‘ Computer Science Dept., Brown University, PO Box 1910, Providence, RI 02912, USA. pck@cs.brown.edu. The research of thks author was supported by NSF grant IRI-861 7344 and ONR grant NOO014-91-J-1613. t ComputeT Science Dept., Brown University, PO Box 1910, Providence, RI 02912, USA, and Digital Equipment Corporation, LI<G2-2/T2, 55o King Street, Littleton, MA 01460, USA. aas@cs.brown.edu.
Publisher Copyright:
© 1991 ACM
PY - 1991/7/1
Y1 - 1991/7/1
N2 - We study efficient deterministic executions of parallel algorithms on restartable fail-stop CRCW PRAMs. We allow the PRAM processors to be subject to arbitrary stop failures and restarts, that are determined by an on-line adversary, and that result in loss of private memory but do not affect shared memory. For this model, we define and justify the complexity measures of: completed work, where processors are charged for completed fixed-size update cycles, and overhead ratio, which amortizes the work over necessary work and failures. This framework is a nontrivial extension of the fail-stop no-restart model of [KS 89]. We present a simulation strategy for any N processor PRAM on a restartable fail-stop P processor CRCW PRAM such that: it guarantees a terminating execution of each simulated N processor step, with 0(log2 N) overhead ratio, and 0(min{A'' + Plog2 N + Mlog JV, N P0'8}) (sub-quadratic) completed work, where M is the number of failures during this step's simulation. This strategy is work-optimal when the number of simulating processors is P < N/log2 N and the total number of failures per each simulated N processor step is 0(N/logN). These results are based on a new algorithm for the Write-All problem "P processors write l's in an array of size N", together with a modification of the main algorithm of [KS 89] and with the techniques in [KPS 90, Shv 89]. We observe that, on P = N restartable fail-stop processors, the Write-All problem requires Q(N/ogN) completed work, and this lower bound holds even under the additional assumption that processors can read and locally process the entire shared memory at unit cost. Under this unrealistic assumption we have a matching upper bound. The lower bound also applies to the expected completed work of randomized algorithms that are subject to on-line adversaries. Finally, we desribe a simple on-line adversary that causes inefficiency in many randomized algorithms.
AB - We study efficient deterministic executions of parallel algorithms on restartable fail-stop CRCW PRAMs. We allow the PRAM processors to be subject to arbitrary stop failures and restarts, that are determined by an on-line adversary, and that result in loss of private memory but do not affect shared memory. For this model, we define and justify the complexity measures of: completed work, where processors are charged for completed fixed-size update cycles, and overhead ratio, which amortizes the work over necessary work and failures. This framework is a nontrivial extension of the fail-stop no-restart model of [KS 89]. We present a simulation strategy for any N processor PRAM on a restartable fail-stop P processor CRCW PRAM such that: it guarantees a terminating execution of each simulated N processor step, with 0(log2 N) overhead ratio, and 0(min{A'' + Plog2 N + Mlog JV, N P0'8}) (sub-quadratic) completed work, where M is the number of failures during this step's simulation. This strategy is work-optimal when the number of simulating processors is P < N/log2 N and the total number of failures per each simulated N processor step is 0(N/logN). These results are based on a new algorithm for the Write-All problem "P processors write l's in an array of size N", together with a modification of the main algorithm of [KS 89] and with the techniques in [KPS 90, Shv 89]. We observe that, on P = N restartable fail-stop processors, the Write-All problem requires Q(N/ogN) completed work, and this lower bound holds even under the additional assumption that processors can read and locally process the entire shared memory at unit cost. Under this unrealistic assumption we have a matching upper bound. The lower bound also applies to the expected completed work of randomized algorithms that are subject to on-line adversaries. Finally, we desribe a simple on-line adversary that causes inefficiency in many randomized algorithms.
UR - http://www.scopus.com/inward/record.url?scp=0013024626&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=0013024626&partnerID=8YFLogxK
U2 - 10.1145/112600.112603
DO - 10.1145/112600.112603
M3 - Conference contribution
AN - SCOPUS:0013024626
SN - 0897914392
T3 - Proceedings of the Annual ACM Symposium on Principles of Distributed Computing
SP - 23
EP - 36
BT - Proceedings of the Annual ACM Symposium on Principles of Distributed Computing
PB - Association for Computing Machinery
T2 - 10th Annual ACM Symposium on Principles of Distributed Computing, PODC 1991
Y2 - 19 August 1991 through 21 August 1991
ER -