Year | Category | Award | Application(s) | Performance | |
1987 | General Purpose Computer | First Place | Beam Stress Analysis
Surface Wave Simulation Unstable fluid flow model |
400 – 600 speedup on a 1,024 node N-CUBE | |
Honorable Mention | Global Ocean Model | 450 Mflops on a Cray X/MP48 | |||
Honorable Mention | QCD and
Circuit Simulation |
Speedups ranging from 39-458 on three applications run on hypercubes | |||
Honorable Mention | Nonlinear network optimization | 1.5 sec. Execution time on a Connection Machine | |||
1988 | Performance | First Place | Static finite element analysis | 1 Gflop on 8-proc. Cray Y-MP
Running time reduced from 15 min. to 30 sec. |
|
Price-Performance | Honorable Mention | Fluid flow problem using the spectral methd | 800 speedup on a 1,024 node N-CUBE | ||
Compiler Parallelization | Honorable Mention | Automatic parallelization of a financial application | 350 times speedup on a 1,024 N-cube and 50 speedup on a 64 node Intel iPSC-2. |
Year | Category | Award | Application(s) | Performance | |
1989 | Performance | First Place | Seismic data processing | 6 Gflops on a CM-2
(also, 500 Mflops/$ 1M) |
|
Price-Performance | First Place | Oil reservoir modeling | 400 Mflops/ $ 1 M on a CM-2 | ||
Performance | Honorable Mention | Parallel search for VLSI design | 1,100 speedup on a 1,024 processor CM | ||
Price-Performance | Honorable Mention | DNA sequence matching | 77k MIPs/$ 1 M | ||
1990 | Price-Performance | First Place | Electronic structure of a high-temperature superconductor | 800 Mflops/ $ 1 M on a 128-node Intel iPSC/860 | |
Compiler Parallelization | Second Place | Grid generation program used to solve partial differential equations | 1,900 speedup on a 2,048 node CM-2 (2.3 Gflops) | ||
Performance | Honorable Mention | Seismic data processing | 14 Gflops on a CM-2 | ||
Compiler Parallelization | Honorable Mention | Parallelizing Pascal Compiler | 25x on a 25 node Sequent Symmetry |
Year | Category | Award | Application(s) | Performance | |
1992 | Price-Performance | First Place | Simulation of polymer chains parallelized over a heterogeneous collection of distributed machines | 1 Gips / $ 1 M | |
Speedup | First Place | Large, sparse linear system solver that enabled the solution of vortex configurations in superconductors and the modeling of the vibration of piezo-electric crystals. | 4 Gflops on an Intel Touchstone Delta. Speedups between 350 and 500. | ||
Performance | First Place | Simulation of 9 million gravitating stars by parallelizing a tree code | 5 Gflops on an Intel Touchstone Delta | ||
1993 | Performance | First Place | Modeling of a shock front using the Boltzmann Equation | 60 Gflops on a 1,024 processor CM-5 | |
Price-Performance | First Place | Image analysis using the bispectrum analysis algorithm | 6.5 Gflops/ $ 1 M on a custom-built machine called SNAP | ||
Performance | Honorable Mention | Simulating the micro-structure of grain boundaries in solids. | 50 Gflops on a 1,024 processor CM-5 |
Year | Category | Award | Application(s) | Performance | |
1994 | Performance | First Place | Structural mechanics modeling using the boundary element method | 140 Gflops on a 1,904 node Intel Paragon | |
Price-Performance | First Place | Quantum mechanical interacitons among 216 silicon atoms | 3 Gflops / $ 1 M on a cluster of eight HP workstations | ||
Performance | Honorable Mention | Isotropic Turbulence and other CFD codes | 120 Gflops on a 140 processor Numerical Wind Tunnel | ||
1995 | Price-Performance | First Place | Modeling of air flow in flue pipes | 3.6 Gflops/ $ 1 M on a cluster of 20 HP workstations | |
Performance | First Place | Quantum chromodynamics simulation | 179 Gflops on 128 processors of the Numerical Wind Tunnel | ||
Special-Purpose Machines | First Place | Simulation of the motion of 100,000 stars | 112 Gflops using the Grape-4 machine with 288 processors |
Year | Category | Award | Application(s) | Performance | |
1996 | Price-Performance | First Place | Electronic structures calculations | 6.3 Gflops/ $ 1 M on an SGI Power Challenge with 6 MIPS R8000 processors | |
Performance | First Place | Fluid dynamics problem | 111 Gflops on 166 processor Numerical Wind Tunnel | ||
Performance | Honorable Mention | Simulation of the motion of 780,000 stars | 333 Gflops using the Grape-4 machine w/ 1,269 processors | ||
1997 | Price-Performance | First Prize | Modeling suspensions | 10.8 Gflops/ $ 1 M on 28 DEC Alpha machines | |
Performance | First Prize-
Part 1 |
Simulating the motion of 322,000,000 self-gravitating particles | 430 Gflops on ASCI Red using 4,096 processors | ||
Price-Performance | First Prize-
Part 2 |
Two problems: vortex fluid flow modeled with 360,000 particles; galaxy formation following 10,000,000 self-gravitating particles | 18 Gflops/ $ 1 M on a cluster of 16 Intel Pentium Pros (200 Mhz.) |
Year | Category | Award | Application(s) | Performance | |
1998 | Performance | First Prize | First principles calculation, of a unit cell (512 atoms) model of non-collinear magnetic arrangements for metallic magnets using a variation of the locally self-consistent multiple scattering method. | 657 Gflops on a 1024-PE Cray T3E system (600 Mhz) | |
Performance | Second Prize | Electronic structures: a silicon bulk periodic unit cell of 3072 atoms, and an aluminum oxide surface unit cell of 2160 atoms, using a complete dense generalized Hermitian eigenvalue-eigenvector calculation | 605 Gflops on the ASCI Red machine with 9200 processors (200 Mhz.) | ||
Price-Performance | First Prize | 3 lattice quantum chromodynamics computations | 79.7 Gflops / $ 1 M on a custom system with 2,048 PE’s using a Texas Instruments chip (32-bit floating point ops.) | ||
Price-Performance | Second Prize | Simulation of a shock wave propagating through a structure of 61 million atoms | 64.9 Gflops/ $ 1 M using a 70 PE system of DEC Alpha’s (533 Mhz.) |
Year | Category | Award | Application(s) | Performance | |
1999 | Performance | First Prize | Very high resolution simulation of fluid turbulence in compressible flows | 1.18 Tflop/s on short run on 5832 CPU’s on ASCI Blue Pacific, 1.04 Tflop/s sustained on one-hour run, 600 Gflop/s on one-week run on 3840 CPU’s | |
Price Performance | First Prize | Astrophysical n-body simulation | 144 Glops / $ 1 M on custom-built GRAPE-5 32-processor system | ||
Special | First Prize, Shared | Unstructured tetrahedral mesh fluid dynamics using PETSc library | 156 Gflop/s on 2048 nodes of ASCI Red, using one CPU per node for computation | ||
Special | First Prize, Shared | Spectral element calculation using a sparse system solver | 319 Gflop/s on 2048 nodes of ASCI Red, using two CPU’s per node for computation |