High Performance Compilers for Parallel Computing / Edition 1

High Performance Compilers for Parallel Computing / Edition 1

by Michael Wolfe
     
 

ISBN-10: 0805327304

ISBN-13: 9780805327304

Pub. Date: 06/06/1995

Publisher: Pearson

High Performance Compilers for Parallel Computing provides a clear understanding of the analysis and optimization methods used in modern commercial research compilers for parallel systems. By the author of the classic 1989 monograph Optimizing Supercompilers for Supercomputers, this book covers the knowledge and skills necessary to build a

…  See more details below

Overview

High Performance Compilers for Parallel Computing provides a clear understanding of the analysis and optimization methods used in modern commercial research compilers for parallel systems. By the author of the classic 1989 monograph Optimizing Supercompilers for Supercomputers, this book covers the knowledge and skills necessary to build a competitive, advanced compiler for parallel or high-performance computers. Starting with a review of basic terms and algorithms used - such as graphs, trees, and matrix algebra - Wolfe shares the lessons of his 20 years experience developing compiler products. He provides a complete catalog of program restructuring methods that have proven useful in the discovery of parallelism or performance optimization and discusses compiling details for each type of parallel system described, from simple code generation, through basic and aggressive optimizations. A wide variety of parallel systems are presented, from bus-based cache-coherent shared memory multiprocessors and vector computers, to message-passing multicomputers and large-scale shared memory systems.

0805327304B04062001

Read More

Product Details

ISBN-13:
9780805327304
Publisher:
Pearson
Publication date:
06/06/1995
Edition description:
New Edition
Pages:
500
Product dimensions:
6.90(w) x 9.10(h) x 1.50(d)

Table of Contents

1. High Performance Systems.

An Example Program: Matrix Multiplication.

Structure of a Compiler.

2. Programming Language Features.

Languages for High Performance.

Sequential and Parallel Loops.

Roundoff Error.

3. Basic Graph Concepts.

Sets, Tuples, Logic.

Graphs.

Control Dependence.

4. Review of Linear Algebra.

Real Vectors and Matrices.

Integer Matrices and Lattices.

Linear System of Equations.

System of Integer Equations.

Systems of Linear Inequalities.

Systems of Integer Linear Inequalities.

Extreme Values of Affine Functions.

5. Data Dependence.

Data Dependence in Loops.

Data Dependence in Conditionals.

Data Dependence in Parallel Loops.

Program Dependence Graph.

6. Scalar Analysis with Factored Use-Def Chains.

Constructing Factored Use-Def Chains.

FUD Chains for Arrays.

Finding All Reaching Definitions.

Implicit References in FUD Chains.

InductionVariables Using FUD Chains.

Constant Propagation with FUD Chains.

Data Dependence for Scalars.

7. Data Dependence Analysis for Arrays.

Building the Dependence System.

Dependence System Solvers.

General Solver.

Summary of Solvers.

Complications.

Run-time Dependence Testing.

8. Other Dependence Problems.

Array Region Analysis.

Pointer Analysis.

I/O Dependence.

Procedure Calls.

Interprocedural Analysis.

9. Loop Restructuring.

Simpile Transformations.

Loop Fusion.

Loop Fission.

Loop Reversal.

Loop Interchanging.

Loop Skewing.

Linear Loop Transformations.

Strip-Mining.

Loop Tiling.

Other Loop Transformations.

Interprocedural Transformations.

10. Optimizing for Locality.

Single Reference to Each Array.

Multiple References.

General Tiling.

Fission and Fusion for Locality.

11. Concurrency Analysis.

Code for Concurrent Loops.

Concurrency from Sequential Loops.

Concurrency from Parallel Loops.

Nested Loops.

Roundoff Error.

Exceptions and Debuggers.

12. Vector Analysis.

Vector Code.

Vector Code from Sequential Loops.

Vector Code from Forall Loops.

Nested Loops.

Roundoff Error, Exceptions, and Debuggers.

Multivector Computers.

13. Message-Passing Machines.

SIMD Machines.

MIMD Machines.

Data Layout.

Parallel Code for Array Assignment.

Remote Data Access.

Automatic Data Layout.

Multiple Array Assignments.

Other Topics.

14. Scalable Shared-Memory Machines.

Global Cache Coherence.

Local Cache Coherence.

Latency Tolerant Machines.

Glossary.

References.

Author Index.

Index. 0805327304T04062001

Read More

Customer Reviews

Average Review:

Write a Review

and post it to your social network

     

Most Helpful Customer Reviews

See all customer reviews >