Redundancy and Parallelism Tradeoffs for Reliable, High-Performance Architectures
Open Access
Author:
Parashar, Angshuman
Graduate Program:
Computer Science and Engineering
Degree:
Doctor of Philosophy
Document Type:
Dissertation
Date of Defense:
February 08, 2007
Committee Members:
Anand Sivasubramaniam, Committee Chair/Co-Chair Mary Jane Irwin, Committee Member Chitaranjan Das, Committee Member Vijaykrishnan Narayanan, Committee Member Kenneth Jenkins, Committee Member Sudhanva Gurumurthi, Committee Member
For decades, high performance processors have provided architectural and microarchitectural abstractions that enable applications to exploit parallelism as a means to fruitfully utilize the exponentially increasing on-chip transistor counts. With shrinking device sizes, reduced supply voltages and growing integration densities leading to logic elements becoming increasingly susceptible to transient faults, research works have established techniques to leverage existing on-chip parallelism-targeted abstractions to provide redundancy for processor pipelines, thereby increasing their resilience to transient faults.
While prior works have viewed the impact of provisioning such redundancy as performance and/or implementation costs, this thesis attempts to: (a) establish the view that there exists a fundamental tradeoff between parallelism and redundancy, (b) propose efficient mechanisms that can be built to enable processor cores to operate at multiple points in this tradeoff space and argue the utility of these mechanisms, and (c) demonstrate how locality-based dataflow characteristics can be exploited in novel ways to shift this tradeoff space to more efficient regions.