--------------------------------
Notes, GEM Meeting, 12/5/98, Argent Hotel, San Francisco, CA
Notes taken by S. McGinnis, University of Colorado
---------------------------------
Roscoe Giles:
Computing and networking power is increasing at a very fast pace.
PACI focuses on communities of scientists interacting, rather than
single centers doing everything.
The Big Idea: to develop "The Grid": a system of dependable,
consistent, pervasive access to (high-end) computing resources. This
would transform the Internet into a utility, giving access to
computation in the same way that we have access to electricity,
without the end-user needing to care about the details. The Grid
would possess fractal structure, (highly) parallel components, and
variable-strength nodes and links.
The Goal: is to make a powerful computation grid available and
*accessible* to scientific users.
---------------------------------
Geoffrey Fox:
The GEM problem can be broken up into six components:
1) user interface
2) fault scale (non-local) equation solver (for Green's functions)
3) modules specifying local physics and friction
4) model evaluation, data analysis, and data visualization
5) data storage, indexing, and access
6) investigation of complexity & pattern dynamics
Because it is a young program, with little legacy code, GEM should be
able to take advantage of Distributed Object Web Technologies. [Using
the Web to link together ephemeral code objects, rather than persistent
documents.]
GEM requires HPCC, on the order of 1-100 Teraflops. GEM should also
take advantage of tree codes, which are *much* more efficient than N^2
codes for large N; N is huge in geophysics problems.
GEM's User Intrerface goal is to create seamless access to distributed
resources.
GEM can also get lots of leverage off inter-organization collaboration.
For the integration of GEM into a Problem-Solving Environment, we should
begin creating an object web.
The Pacific Disaster Center (PDC) is a good opportunity for these
approaches.
The goal is not necessarily to *do* these things now, but to be *aware*
of them and to make decisions so as to take advantage of them in the
future.
-----------------------------
Jay Parker
Improving the Earth Science / Computer Science (ES/CS) link
Note that Earth Scientists don't actually need to know much CS jargon.
New computing paradigms can lead to better collaboration.
Web-Distributed Object (currently a "rapidly-evolving" technology) could
affect: model elements, coding and integration, geometry and rheology
specification, data analysis and visualization, how the work is done,
and image and idea communication.
One dream:
Work can become a remote access jigsaw puzzle.
We could pool existing model codes into a toybox for all; to do this,
you would need to:
--Extend the web browser
--Wrap (small) existing legacy codes in java
--Add geometry specification, results viewing modules
--Define I/O parameters in wrappers
--Implement security mechanisms
One experience:
MIDAS for radar simulation
Allowed limited mix-and-match processing. A specialist was able to get
the entire system running in 2 weeks (with all the applications in place
through ftp, telnet, etc). Construction was easier than expected, but
it added layers to maintain that were not kept up.
GEM work would involve early core code and later flexible modular
plug-in physics.
Eventually, GEM should encompass: distrubution of web tools, access to a
"big model", and a remote collaborative environment
Meanwhile, we should: cooperate in the usual ways and share code, data,
and papers via ftp and the web.
----------------------------------
Bill Klein
Crossover scientists (formerly in Earth Science, now in CS) are a
valuable and needed resource for GEM.
[Much general discussion]
----------------------------------
Bernard Minster
We want to couple together many different models at many scales, but we
must make sure that we maintain modularity of the models.
We should change our idea of scale, using *LOTS* of computer resources.
GEM must THINK BIG.
Earth Science is a very data-intensive field. We need to learn how to
deal with that.
Useful Tools:
Mathematicians are working on new methods for stable simulation of
complex systems: multi-symplectic integrators.
Code exists at Argonne National Labs that will take any modeling code
and produce from it the *adjoint* code. The adjoint code runs the
physical evolution of the system *backwards* in time....allowing
data to be assimilated into the original code.
Minster also discussed the new NSF Earth System Collaboratory idea
now in the planning stages:
420M$/yr x 5 yrs = 1.4B$
THINK BIG!
---------------------------------
Charlie Sammis
[Discussion about the reasons why a large simulation is needed. Is GEM
proposing a large tool, and not a large science problem?
Some arguments for science (not tool) include use of GEM as a tool to
analyze scientific assumptoins about the problem, and the question of
why the simple models are viewed as valid for such a complex process.]
---------------------------------------
Terry Tullis
What an earthquake fault looks like depends on the scale at which you
view it. The question: as we scale up, what do we throw out? (Since
we must throw out something.) How do we properly parameterize
sub-gridscale processes at the next hierarchical level?
The answer may depend on the question being asked, and we should
catalog the questions.
---------------------------
[Attempt by scribe to summarize the general mindset of the GEM group,
with regard to the question of what GEM attempts to study:
Earthquakes are a complex system which consists of many interacting
components, both literally (faults) and symbolically / mathematically
(friction vs geometry). What we do not understand is the relative
importance of these interactions on difference scales. GEM proposes
generating a computational infrastructure that will allow comparison
of the elements in order to understand what is important and
whether a given interaction can be parameterized.]
---------------------------------
Lisa Grant
The US nuclear testing program (in the 80's) was a very successful
program with many similarities to the GEM problem. The process this
program used was to loop on:
prediction
test/observation
code calibration
Proposed Observation / Calibration component for the earthquake problem
is to study the rupture potential of the San Andreas - Parkfield -
Cholame - Carrizo faults.
This is a well-constrained system
--single fault, simple geometry
--large data set
--data at various scales
with a potentially large payoff in science and hazard mitigation.
Testing and prediction will be most successful if we focus not on when
an event happens, but on where and in what manner.
Geology can provide some boundary conditions, and slip distribution can
also help to calibrate models.
-------------------------
Kerry Sieh
The critical question is, how regular are earthquakes?
One fault is unlikely to have a G-R distribution.
Characteristic events do exist for some faults.
A possibility: displacement tends to be similar, but recurrence time is
irregular.
Do we need more complicated and expensive models?
Modellers and geologists need to communicate more. Modellers should ask
the geologists to:
1) establish general behavior of faults and fault systems
2) determine quantitative histories for particular systems
----------------------------
Mike Watkins
Most of GEM is about improving forward models; we want to work towards
doing inverse modelling.
A useful tool is the "data assimilation" method. Using variational
objective analysis, you can tune model parameters. This model also
allows you to determine whether or not parameters are significant.
-------------------------------
Tom Jordan
If we integrate all these themes, we see that we are talking about the
interaction between physics, hypotheses, models, and data. For this
reason, we shouldn't even use terms like "model validation" and "model
verification".
Need to decide the focus of GEM: theoretical function (numerical
laboratory) or observational function (data reconcilliation,
assimilation, and extrapolation)?
Two proposed tasks/foci: Establishing the scaling boundaries of
earthquake processes as a geosystem, and data assimilation.
[Dissenting opinion: we should break the grand endeavor into multiple
smaller (but still big) projects.]
We have, basically, a new way of doing science: model inference. (Note
that the model need not have equations in it. Example: avalanche
problems.)
The study of a driven, non-equlibrium, threshold system is applicable to
many fields.
GEM should focus on a *mission goal*, which is one element of a
strategic research plan to "understand earthquakes"; an SRP will
generate lots of successes, as it did in astrophysics.
The systems approach is the most attractive approach to the NSF.
-----------------------------------
End of Meeting