Gordon Bell Prize Winners

Home  /   Gordon Bell Prize Winners
Year Category Award Application(s) Performance
1987 General Purpose Computer First Place Beam Stress Analysis

Surface Wave Simulation

Unstable fluid flow model

400 – 600 speedup on a 1,024 node N-CUBE
Honorable Mention Global Ocean Model 450 Mflops on a Cray X/MP48
Honorable Mention QCD and

Circuit Simulation

Speedups ranging from 39-458 on three applications run on hypercubes
Honorable Mention Nonlinear network optimization 1.5 sec. Execution time on a Connection Machine
1988 Performance First Place Static finite element analysis 1 Gflop on 8-proc. Cray Y-MP

Running time reduced from 15 min. to 30 sec.

Price-Performance Honorable Mention Fluid flow problem using the spectral methd 800 speedup on a 1,024 node N-CUBE
Compiler Parallelization Honorable Mention Automatic parallelization of a financial application 350 times speedup on a 1,024 N-cube and 50 speedup on a 64 node Intel iPSC-2.


Year Category Award Application(s) Performance
1989 Performance First Place Seismic data processing 6 Gflops on a CM-2

(also, 500 Mflops/$ 1M)

Price-Performance First Place Oil reservoir modeling 400 Mflops/ $ 1 M on a CM-2
Performance Honorable Mention Parallel search for VLSI design 1,100 speedup on a 1,024 processor CM
Price-Performance Honorable Mention DNA sequence matching 77k MIPs/$ 1 M
1990 Price-Performance First Place Electronic structure of a high-temperature superconductor 800 Mflops/ $ 1 M on a 128-node Intel iPSC/860
Compiler Parallelization Second Place Grid generation program used to solve partial differential equations 1,900 speedup on a 2,048 node CM-2 (2.3 Gflops)
Performance Honorable Mention Seismic data processing 14 Gflops on a CM-2
Compiler Parallelization Honorable Mention Parallelizing Pascal Compiler 25x on a 25 node Sequent Symmetry


Year Category Award Application(s) Performance
1992 Price-Performance First Place Simulation of polymer chains parallelized over a heterogeneous collection of distributed machines 1 Gips / $ 1 M
Speedup First Place Large, sparse linear system solver that enabled the solution of vortex configurations in superconductors and the modeling of the vibration of piezo-electric crystals. 4 Gflops on an Intel Touchstone Delta. Speedups between 350 and 500.
Performance First Place Simulation of 9 million gravitating stars by parallelizing a tree code 5 Gflops on an Intel Touchstone Delta
1993 Performance First Place Modeling of a shock front using the Boltzmann Equation 60 Gflops on a 1,024 processor CM-5
Price-Performance First Place Image analysis using the bispectrum analysis algorithm 6.5 Gflops/ $ 1 M on a custom-built machine called SNAP
Performance Honorable Mention Simulating the micro-structure of grain boundaries in solids. 50 Gflops on a 1,024 processor CM-5


Year Category Award Application(s) Performance
1994 Performance First Place Structural mechanics modeling using the boundary element method 140 Gflops on a 1,904 node Intel Paragon
Price-Performance First Place Quantum mechanical interacitons among 216 silicon atoms 3 Gflops / $ 1 M on a cluster of eight HP workstations
Performance Honorable Mention Isotropic Turbulence and other CFD codes 120 Gflops on a 140 processor Numerical Wind Tunnel
1995 Price-Performance First Place Modeling of air flow in flue pipes 3.6 Gflops/ $ 1 M on a cluster of 20 HP workstations
Performance First Place Quantum chromodynamics simulation 179 Gflops on 128 processors of the Numerical Wind Tunnel
Special-Purpose Machines First Place Simulation of the motion of 100,000 stars 112 Gflops using the Grape-4 machine with 288 processors


Year Category Award Application(s) Performance
1996 Price-Performance First Place Electronic structures calculations 6.3 Gflops/ $ 1 M on an SGI Power Challenge with 6 MIPS R8000 processors
Performance First Place Fluid dynamics problem 111 Gflops on 166 processor Numerical Wind Tunnel
Performance Honorable Mention Simulation of the motion of 780,000 stars 333 Gflops using the Grape-4 machine w/ 1,269 processors
1997 Price-Performance First Prize Modeling suspensions 10.8 Gflops/ $ 1 M on 28 DEC Alpha machines
Performance First Prize-

Part 1

Simulating the motion of 322,000,000 self-gravitating particles 430 Gflops on ASCI Red using 4,096 processors
Price-Performance First Prize-

Part 2

Two problems: vortex fluid flow modeled with 360,000 particles; galaxy formation following 10,000,000 self-gravitating particles 18 Gflops/ $ 1 M on a cluster of 16 Intel Pentium Pros (200 Mhz.)


Year Category Award Application(s) Performance
1998 Performance First Prize First principles calculation, of a unit cell (512 atoms) model of non-collinear magnetic arrangements for metallic magnets using a variation of the locally self-consistent multiple scattering method. 657 Gflops on a 1024-PE Cray T3E system (600 Mhz)
Performance Second Prize Electronic structures: a silicon bulk periodic unit cell of 3072 atoms, and an aluminum oxide surface unit cell of 2160 atoms, using a complete dense generalized Hermitian eigenvalue-eigenvector calculation 605 Gflops on the ASCI Red machine with 9200 processors (200 Mhz.)
Price-Performance First Prize 3 lattice quantum chromodynamics computations 79.7 Gflops / $ 1 M on a custom system with 2,048 PE’s using a Texas Instruments chip (32-bit floating point ops.)
Price-Performance Second Prize Simulation of a shock wave propagating through a structure of 61 million atoms 64.9 Gflops/ $ 1 M using a 70 PE system of DEC Alpha’s (533 Mhz.)



Year Category Award Application(s) Performance
1999 Performance First Prize Very high resolution simulation of fluid turbulence in compressible flows 1.18 Tflop/s on short run on 5832 CPU’s on ASCI Blue Pacific, 1.04 Tflop/s sustained on one-hour run, 600 Gflop/s on one-week run on 3840 CPU’s
Price Performance First Prize Astrophysical n-body simulation 144 Glops / $ 1 M on custom-built GRAPE-5 32-processor system
Special First Prize, Shared Unstructured tetrahedral mesh fluid dynamics using PETSc library 156 Gflop/s on 2048 nodes of ASCI Red, using one CPU per node for computation
Special First Prize, Shared Spectral element calculation using a sparse system solver 319 Gflop/s on 2048 nodes of ASCI Red, using two CPU’s per node for computation