HELP! * YELLOW=global GREY=local Full HTML for

GLOBAL foilset HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure

Given by Geoffrey Fox at Trip to China on July 12-28,96. Foils prepared July 6 1996
Abstract * Foil Index for this file See also color IMAGE

We describe the structure of seven talks making up this review of HPCC from today to the Web and Petaflop performance in future
Here we describe current status with HPCC in some sense both a failure and a great success
This requires looking at hardware, software and the critical lack of commercial adoption of this technology
We discuss COTS and trickle up and down technology strategies
We describe education and interdisciplinary computational science in both simulation and information arenas

Table of Contents for full HTML of HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure


1 Status of "Classic" HPCC -- June1996
Overall Status: Success or Failure?

2 Abstract of HPCC Current Status 1996
3 NPAC Home Page
4 Components of This HPCC Presentation
5 Superficial Observations on High Performance Computing-I
6 Superficial Observations on High Performance Computing-II
7 Superficial Observations on High Performance Communication
8 Some Implications of HPCC Observations
9 Advances in Parallel Computer and High Speed Network (HPCC) Technology
10 Some Hardware/Software Trends over next 5 years
11 What is Status of HPCC Applications?
12 Two Major Parallel Programming Paradigms
13 When will Parallel Computing Take Over ?
14 The Federal High Performance Computing and Communication Initiative (HPCCI)
15 Performance of High End Machines Years 1940-2000
16 Performance of High End Machines Years 1980-2000
17 Peak Supercomputer Performance
18 What Happens now that HPCC Initiative is no longer in place?
19 Some Important Trends -- COTS is King!
20 Comments on COTS for Hardware
21 Performance Per Transistor
22 However we need more than fast enough machines
We also need a large enough market to sustain technology (systems and software)

23 NII Compute & Communications Capability in Year 2000 --> 2005
24 Returning to Today - I
25 Returning to Today - II
26 Software Issues/Choices - I
27 The Sad Story of HPF and Some Applications
28 Software Issues/Choices - II
29 Software Issues/Choices - III
30 Need to Educate People to take advantage of HPCC technologies
31 Educational and (Re)training Challenges
32 What is Computational Science?
33 Program in Computational Science
Implemented within current academic framework

34 Program in Information Age Computational Science Implemented Within Current Academic Program
35 Parallel Processing and Society
36 Concurrent Construction of a Wall
Using N = 8 Bricklayers
Decomposition by Vertical Sections

37 Quantitative Speed-Up Analysis for Construction of Hadrian's Wall
38 Amdahl's law for Real World Parallel Processing
39 Pipelining --Another Parallel Processing Strategy for Hadrian's Wall
40 Hadrian's Wall Illustrates that the Topology of Processor Must Include Topology of Problem
41 General Speed Up Analysis
42 Nature's Concurrent Computers
43 Comparison of Concurrent Processing in Society and Computing
44 Data Parallelism is a Universal Source of Scaling Parallelism
45 We have learnt that Parallel Computing Works !
46 Methodology of Parallel Computing
47 Concurrent Computation as a Mapping Problem -I
48 Concurrent Computation as a Mapping Problem - II
49 Concurrent Computation as a Mapping Problem - III

This table of Contents Abstract



HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 1 Status of "Classic" HPCC -- June1996
Overall Status: Success or Failure?

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
http://www.npac.syr.edu/users/gcf/hpcc96status/index.html
Presented during Trip to China July 12-28,1996
Geoffrey Fox
NPAC
Syracuse University
111 College Place
Syracuse NY 13244-4100

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 2 Abstract of HPCC Current Status 1996

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
We describe the structure of seven talks making up this review of HPCC from today to the Web and Petaflop performance in future
Here we describe current status with HPCC in some sense both a failure and a great success
This requires looking at hardware, software and the critical lack of commercial adoption of this technology
We discuss COTS and trickle up and down technology strategies
We describe education and interdisciplinary computational science in both simulation and information arenas

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 3 NPAC Home Page

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Full HTML Index

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 4 Components of This HPCC Presentation

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
There are seven talks in this series:
HPCC Status -- this talk -- Overall Technical and Political Status
HPCC Today I -- MPP Hardware Architectures and Machines
HPCC Today II -- Software
HPCC Today III -- Applications -- Grand Challenges Industry
HPCC Tomorrow I -- Problem Solving Environments
HPCC Tomorrow II -- Petaflop (10^15 Operations per second) in the year 2007?
HPCC Tomorrow III -- The use of Web Technology in HPCC

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 5 Superficial Observations on High Performance Computing-I

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Secs 87 Full HTML Index
Parallel Computing Works!
Technology well understood for Science and Engineering
  • Good parallel algorithms, several examples of major applications in many fields exploring range of issues
  • Data and Message Parallel programming models developed
Supercomputing market small (few percent at best) and probably decreasing in size
  • Essential to have good common software infrastructure
  • Productivity tools -- Software Engineering -- Programming Support tools POOR
  • The parallel software "industry" is very small

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 6 Superficial Observations on High Performance Computing-II

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Secs 141 Full HTML Index
No silver programming bullet -- I doubt if new language will revolutionize parallel programmimng and make much easier
  • Hardware (shared memory) could be helpful
Social forces are tending to hinder adoption of parallel computing as most applications are areas where large scale computing already common
  • Parallelizing existing applications (porting sequential software) very hard
  • Opportunities offered by use of MPP's often require major organizational changes

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 7 Superficial Observations on High Performance Communication

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Secs 86 Full HTML Index
ATM ISDN Wireless Satellite advancing rapidly in commercial arena which is adopting research rapidly
Social forces (deregulation in the U.S.A.) are tending to accelerate adoption of digital communication technologies
  • These are often NEW applications (porting of POTS relatively easy!) such as interactive TV/Shopping
  • Tremendous competition between different telecommunication sectors encourages new technology now to ensure future success
Not clear how to make money on Web(Internet) but growing interest/acceptance by general public
  • huge sales in home multimedia PC's -- comparable to TV's in volume
Integration of Communities and Opportunities
  • Computing and Communication and Information Industries merging -- similar impact on academic departments will(should) happen

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 8 Some Implications of HPCC Observations

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Secs 63 Full HTML Index
Technology Opportunities in Integration of High Performance Computing and Communication Systems
  • Merging of networking, parallel computing, distributed comouting communities
  • This SOLVES previous difficulties observed for high performance computing as implies a much larger distributed (world-wide metacomputing) computing base
New Business opportunities linking Enterprise Information Systems to Community networks to current cable/network TV journalism
New educational needs at interface of computer science and communications/information applications
Major implications for education -- the Virtual University

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 9 Advances in Parallel Computer and High Speed Network (HPCC) Technology

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Secs 99 Full HTML Index
Performance of both communication networks and computers will increase by a factor of 1000 during the 1990's
  • New uses of Computers to design new drugs, search terabyte databases etc.
  • National Information Infrastructure will see pervasive deployment of upgraded Internet to give megagabit/second interactive links to homes and offices allowing interactive realtime video.
  • Greater utility of computers in "Old Applications"
Competitive advantage to industries that can use either or both High Performance Computers and Communication Networks. (United States clearly ahead of Japan and Europe in these technologies.)

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 10 Some Hardware/Software Trends over next 5 years

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Secs 76 Full HTML Index
ATM networks have rapidly transitioned from research Gigabit networks to commercial deployment
  • ATM likely to be a major force in local area as well as wide area networks
Computer Hardware trends imply that all computers (PC's ---> Supercomputers) will be parallel by the year 2000
  • Up to 1993, parallel computers are from small start-up companies (except Intel Supercomputer Division)
  • Now Cray, Convex (HP), Digital, IBM have massively parallel computing systems and Silicon Graphics is becoming a powerful high performance computing vendor
  • Several architectures but only one : Distributed memory MIMD multicomputer is known to scale from one to very many processors
Software is challenge and could prevent/delay hardware trend that suggests parallelism will be a mainline computer architecture
  • We must get systems software correct
  • Simultaneously develop applications software in gradually improving parallel programming environment

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 11 What is Status of HPCC Applications?

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
Parallel Computing Works in Nearly all scientific and engineering applications

As described in Book Parallel Computing Works! (Fox,Messina,Williams)
The necessary algorithms are in general understood in most cases
The implementation of -- especially adaptive irregular -- algorithms is not easy because:
The software tools are immature and do not usually offer direct help for say:
  • Adaptive (un)structured meshs
  • Fast multipole method for particle dynamics
There are several different approachs and not clear what will "win" and what will actually "work" when
Need abstractions of the "hard" problem (component)s and toolkits to tackle them
  • Templates will describe promising approach

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 12 Two Major Parallel Programming Paradigms

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
Data Parallel and Message Parallel

These are Message Parallel and Data Parallel Resources
  • Suggest Message Parallel Fortran as description of Fortran plus Message passing (Per Brinch Hansen) to be consistent with HPF or CMFortran as Data Parallel Fortran
There are trade-offs in Ease of Programming (not same for each user!), Portability, Maturity of Software, Generality of Problem class
Message Parallel is most mature, somewhat less portable in principle but not necessarily in practice, tackles all problems and some consider painful to program
  • Most succesful parallel applications have been Message Parallel
  • Fortran-M and MPI will describe this approach

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 13 When will Parallel Computing Take Over ?

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Secs 72 Full HTML Index
Switch from conventional to new types of technology is a phase transition
Needs headroom (Carver Mead) which is large (factor of 10 ?) due to large new software investment
Machines such as the nCUBE-1 and CM-2 were comparable in cost performance to conventional supercomputers
  • Enough to show that "Parallel Computing Works"
  • Not enough to take over!
Cray T3D, Intel Paragon, CM-5, DECmpp (Maspar MP-2), IBM SP-2, nCUBE-3 have enough headroom to take over from traditional computers ?

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 14 The Federal High Performance Computing and Communication Initiative (HPCCI)

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
Originally $2.9 billion over 5 years starting in 1992 and
  • Rapidly growing Information technology component starting in 1994 and total budget became over $1 billion per year
This drove race to Teraflop performance and is now OVER!
The Grand Challenges
  • Enabled by teraflop computers and important to economy or fundamental research
    • Global warming - NOAA
    • Oil reservoir and environmental simulation - DOE
    • Structural and aerodynamic calculations - NASA
    • Earth observing satellite - data analysis - NASA
    • Human genome - NIH, DOE
    • Quantum chromodynamics - Fundamental Physics
    • Gravitational waves from black holes - Fundamental Physics
    • Molecular modeling - Fundamental Chemistry

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 15 Performance of High End Machines Years 1940-2000

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Secs 72 Full HTML Index

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 16 Performance of High End Machines Years 1980-2000

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Secs 33 Full HTML Index

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 17 Peak Supercomputer Performance

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
For "Convential" MPP/Distributed Shared Memory Architecture
Now(1996) Peak is 0.1 to 0.2 Teraflops in Production Centers
  • Note both SGI and IBM are changing architectures:
  • IBM Distributed Memory to Distributed Shared Memory
  • SGI Shared Memory to Distributed Shared Memory
In 1999, one will see production 1 Teraflop systems
In 2003, one will see production 10 Teraflop Systems
In 2007, one will see production 50-100 Teraflop Systems
Memory is Roughly 0.25 to 1 Terabyte per 1 Teraflop
If you are lucky/work hard: Realized performance is 30% of Peak

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 18 What Happens now that HPCC Initiative is no longer in place?

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
Most of the activities it started are ongoing!
It achieved goal of Teraflop performance -- see Intel P6 machine at Sandia
But it failed to develop a viable commercial base
And although hardware peak performs at advertised rate, the software environment is poor
  • This could be due to poor hardware as well as lack of sufficient resources to sustain software effort
Academic Activities -- NSF Supercomputer centers -- are very healthy as much easier to put such codes on MPP as short in lifetime and lines of code
Next initiatives -- based on PetaFlop goal -- will include a federal development as well as research component as can't assume "brilliant research" will be picked up by industry

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 19 Some Important Trends -- COTS is King!

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
Everybody now believes in COTS -- Consumer On the Shelf Technology -- one must use commercial building blocks for any specialized system whether it be a DoD weopens program or high end Supercomputer
  • These are both Niche Markets!
COTS for hardware can be applied to a greater or less extent
  • Gordon Bell's SNAP system says we will only have ATM networks of PC's running WindowsNT
  • SGI HP and IBM will take commodity processor nodes but link with custom switches (with different versions of distributed shared memory support)
COTS for Software is less common but (I expect) to become much more common
  • HPF producing HTTP not MPI with Java visulalization is an example of Software COTS

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 20 Comments on COTS for Hardware

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
Currently MPP's have COTS processors and specialized networks but this could reverse
  • Pervasive ATM will indeed lead to COTS Networks BUT
  • Current microprocessors are roughly near optimal in terms of megaflops per square meter of silicon BUT
  • As (explicit) parallelism shunned by modern microprocessor, silicon is used for wasteful speculative execution with expectation that future systems will move to 8 way functional parallelism.
Thus estimate that 250,000 transistors (excluding on chip cache) is optimal for performance per square mm of silicon
  • Modern microprocessor is around ten times this size
Again simplicity is optimal but this requires parallelism
Contrary trend is that memory dominates use of silicon and so performance per square mm of silicon is often not relevant

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 21 Performance Per Transistor

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Full HTML Index
Performance data from uP vendors
Transistor count excludes on-chip caches
Performance normalized by clock rate
Conclusion: Simplest is best! (250K Transistor CPU)
Millions of Transistors (CPU)
Millions of Transistors (CPU)
Normalized SPECINTS
Normalized SPECFLTS

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 22 However we need more than fast enough machines
We also need a large enough market to sustain technology (systems and software)

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Secs 89 Full HTML Index
This is both Grand Challenges augmented by National Challenges but also
Build HPCC technologies on a broad not niche base starting at bottom not top of computing pyramid

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 23 NII Compute & Communications Capability in Year 2000 --> 2005

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Full HTML Index
Each of three components (network connections, clients, servers) has capital value of order $10 to $100 Billion

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 24 Returning to Today - I

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
Tightly Coupled MPP's were (SP2,Paragon,CM5 etc) distributed memory but at least at the low end they are becoming hardware assisted shared memory
  • Unclear how well compilers will support this in a scaling fashion -- we will see how SGI/Cray systems based ideas pioneered at Stanford fair!
Note this is an example of COTS at work -- SGI/Sun/.. Symmetric Multiprocessors (Power Challenge from SGI) attractive as bus will support upto 16 processors in elegant shared memory software world.
  • Previously such systems were not pwerful enough to be interesting
Clustering such SGI Power Challenge like systems produces a powerful but difficult to program (as both distributed and shared memory) heterogeneous system
Meanwhile Tera Computer will offer a true Uniform Memory access shared memory using ingenious multi threaded software/hardware to hide latency
  • Unclear if this competitive in cost/performance with (scruffy) COTS approach

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 25 Returning to Today - II

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
Trend I -- Hardware Assisted Tightly Coupled Shared Memory MPP's are replacing pure distributed memory systems
Trend II -- The World Wide Web and increasing power of individual workstations is making geographically distributed heterogeneous distributed memory systems more attractive
Trend III -- To confuse the issue, the technology trends in next ten years suggest yet different architecture(s) such as PIM
Better use Scalable Portable Software with conflicting technology/architecture trends BUT must address latency agenda which isn't clearly portable!

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 26 Software Issues/Choices - I

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
Functional Parallelism supported by sequential compiler intrachip in modern superscalar processor at very small grain size
Multithreading as in Java supports object or small grain size functional parallelism within a Java Applet -- this can be implemented with message passing not shared memory
Data Parallelism is source of Most scaling parallelism and needed for more than a few way speedup --
  • HPF emerging slowly as hardware/technology/politics (grand versus national challenges) changes faster than one can write good complete compilers
  • Unfortunately most languages find it hard to express data parallelism in natural automatic way even though is "obvious in problem"
  • Must use domain decomposition with target architecture dependent "objects" and so inevitably harder than pure object based parallelism where objects come from problem!

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 27 The Sad Story of HPF and Some Applications

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
The Good News -- In 1993, NPAC selected to Partner with two Grand Challenge Groups as part of their "computer science" support
Both Applications were rather suitable for HPF as regular grids
The Bad News -- Both Application groups have abandoned HPF as couldn't wait for working compilers with appropriate features
  • But BOTH have switched to using Fortran90!!
Numerical Relativity needed adaptive mesh and this supplied by DAGH from Texas which is like HPF but simpler in overall capability but can handle adaptive meshes
  • But need to wait for new features in DAGH to handle say conjugate gradient which is automatic in HPF
  • Need to parallelize sophisticated data types (tensors) in Fortran90
In Data Assimilation, a NASA Goddard application, they are recoding in Fortran90 plus MPI as MUST meet specific performance goals in a year

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 28 Software Issues/Choices - II

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
MetaProblems exhibit large grain size functional parallelism as in
  • Multidisciplinary Optimization
Coordination or Integration Software Addresses this.
See SCCS 757 (available from online NPAC reports) which discusses applications and software approaches for all types of parallelism
Java Applet Model is natural approach which can be illustrated by new VRML 2.0 which will build on
  • Success of SIMNET/DSI -- Distributed Simulation for DoD
  • with totally distributed set of Applets invoked by handlers (interrupts = active messages) loosely integrated over world wide Web -- a worldwide object parallelism
  • e.g. Pressing door knob invokes Java applet as does missile hitting tank.

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 29 Software Issues/Choices - III

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
Architectures such as PIM emphasis opportunities if we could develop software models/compilers which could effectively use substantial parallelism
Current partyline microprocessors assume that substantial parallelism cannot be easily extracted from programs designed for single Chips
One area we are exploring is how to extract from/build in parallelism for Java
  • We call this the HPJava project
A major new direction in Computer Science is "Problem Solving Environments" which are domain specific systems at a "higher level" than compilers which are a "toolkit" of components such as:
  • Linear Algebra Solvers
  • Elliptic or Hyperbolic PDE Solvers
  • Particular (adaptive) data structures such as those for multigrid
  • Visualization etc.

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 30 Need to Educate People to take advantage of HPCC technologies

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Secs 20 Full HTML Index

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 31 Educational and (Re)training Challenges

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Secs 72 Full HTML Index
Computation joins theory and experiment as the three complementary approachs to study of science and engineering
Current industries such as Media and Telecommunications which have been dominated by analog technologies will need to adjust to growing use of digital (computer) technologies
Need for new educational approachs such as Computational Science centered on interdiciplinary border between computer science and applications with both a
  • Science and Engineering and
  • Information (Communications ) Track

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 32 What is Computational Science?

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
Computational Science is an interdisciplinary field that integrates computer science and applied mathematics with a wide variety of application areas that use significant computation to solve their problems
Includes the study of computational techniques
  • Science and Engineering - Grand Challenges
  • Society and Business - National Challenge
Includes the study of new algorithms, languages and models in computer science and applied mathematics required by the use of high performance computing and communications in any (?) important application
  • At interface of (applied) computer science and applications
Includes computation of complex systems using physical analogies such as neural networks and genetic optimization.

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 33 Program in Computational Science
Implemented within current academic framework

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Full HTML Index

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 34 Program in Information Age Computational Science Implemented Within Current Academic Program

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Full HTML Index

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 35 Parallel Processing and Society

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
The fundamental principles behind the use of concurrent computers are identical to those used in society - in fact they are partly why society exists.
If a problem is too large for one person, one does not hire a SUPERman, but rather puts together a team of ordinary people...
cf. Construction of Hadrians Wall

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 36 Concurrent Construction of a Wall
Using N = 8 Bricklayers
Decomposition by Vertical Sections

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Full HTML Index
Domain Decomposition is Key to Parallelism
Need "Large" Subdomains l >> l overlap

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 37 Quantitative Speed-Up Analysis for Construction of Hadrian's Wall

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Full HTML Index

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 38 Amdahl's law for Real World Parallel Processing

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Important Information in IMAGE
Full HTML Index
AMDAHL"s LAW or
Too many cooks spoil the broth
Says that
Speedup S is small if efficiency e small
or for Hadrian's wall
equivalently S is small if length l small
But this is irrelevant as we do not need parallel processing unless problem big!

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 39 Pipelining --Another Parallel Processing Strategy for Hadrian's Wall

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Full HTML Index
"Pipelining" or decomposition by horizontal section is:
  • In general less effective
  • and leads to less parallelism
  • (N = Number of bricklayers must be < number of layers of bricks)

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 40 Hadrian's Wall Illustrates that the Topology of Processor Must Include Topology of Problem

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Full HTML Index
Hadrian's Wall is one dimensional
Humans represent a flexible processor node that can be arranged in different ways for different problems
The lesson for computing is:
Original MIMD machines used a hypercube topology. The hypercube includes several topologies including all meshes. It is a flexible concurrent computer that can tackle a broad range of problems. Current machines use different interconnect structure from hypercube but preserve this capability.

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 41 General Speed Up Analysis

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Full HTML Index
Comparing Computer and Hadrian's Wall Cases

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 42 Nature's Concurrent Computers

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Full HTML Index
At the finest resolution, collection of neurons sending and receiving messages by axons and dendrites
At a coarser resolution
Society is a collection of brains sending and receiving messages by sight and sound
Ant Hill is a collection of ants (smaller brains) sending and receiving messages by chemical signals
Lesson: All Nature's Computers Use Message Passing
With several different Architectures

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 43 Comparison of Concurrent Processing in Society and Computing

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Important Information in IMAGE
Full HTML Index
Problems are large - use domain decomposition Overheads are edge effects
Topology of processor matches that of domain - processor with rich flexible node/topology matches most domains
Regular homogeneous problems easiest but
irregular or
Inhomogeneous
Can use local and global parallelism
Can handle concurrent calculation and I/O
Nature always uses message passing as in parallel computers (at lowest level)

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 44 Data Parallelism is a Universal Source of Scaling Parallelism

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 45 We have learnt that Parallel Computing Works !

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
Data Parallelism - universal form of scaling parallelism
Functional Parallelism - Important but typically modest speedup. - Critical in multidisciplinary applications.
On any machine architecture
  • Distributed Memory MIMD
  • Distributed Memory SIMD
  • Shared memory - this affects programming model
  • This affects generality
  • SIMD ~ 50% academic problems
  • but < 50% commercial

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 46 Methodology of Parallel Computing

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * See also color IMAGE
Full HTML Index
Simple, but general and extensible to many more nodes is domain decomposition
All successful concurrent machines with
  • Many nodes
  • High performance (this excludes Dataflow)
Have obtained parallelism from "Data Parallelism" or "Domain Decomposition"
Problem is an algorithm applied to data set
  • Obtain concurrency by acting on data concurrently.
The three architectures considered here differ as follows:
  • MIMD Distributed Memory -- Processing and Data Distributed
  • MIMD Shared Memory -- Processing Distributed but memory shared
  • SIMD Distributed Memory -- Synchronous Processing on Distributed Data

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 47 Concurrent Computation as a Mapping Problem -I

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Full HTML Index
2 Different types of Mappings in Physical Spaces
Both are static
  • a) Seismic Migration with domain decomposition on 4 nodes
  • b)Universe simulation with irregular data but static 16 node decomposition
  • but this problem would be best with dynamic irregular decomposition

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 48 Concurrent Computation as a Mapping Problem - II

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Full HTML Index
Different types of Mappings -- A very dynamic case without any underlying Physical Space
c)Computer Chess with dynamic game tree decomposed onto 4 nodes

HELP! * YELLOW=global GREY=local HTML version of GLOBAL Foils prepared July 6 1996

Foil 49 Concurrent Computation as a Mapping Problem - III

From HPCC Status -- TeraFlop to Web and Petaflops -- Success and Failure Trip to China -- July 12-28,96. * Critical Information in IMAGE
Full HTML Index

Northeast Parallel Architectures Center, Syracuse University, npac@npac.syr.edu

If you have any comments about this server, send e-mail to webmaster@npac.syr.edu.

Page produced by wwwfoil on Tue Feb 18 1997