Chapter 18: Scientific Computing

The EECS Research Summary for 2003


Automatic Performance Tuning of Sparse Matrix Kernels

Richard Vuduc, Shoaib Kamil1, Rajesh Nishtala2, Benjamin Lee3, and Attila Gyulassy4
(Professors James W. Demmel and Katherine A. Yelick)
(DOE) DE-FG03-94ER25219, (DOE) DE-FC03-98ER25351, and (LLNL/DOE) W-7405-ENG-48

The overall performance in a variety of scientific computing and information retrieval applications is dominated by a few computational kernels. One important class of such operations are sparse matrix kernels--computations with matrices having relatively few non-zero entries. Performance tuning of sparse kernels is a particularly tedious and time-consuming task because performance is a complicated function of the kernel, machine, and non-zero structure of the matrix: for every machine, a user must carefully choose a data structure and implementation (code and transformations) that minimize the matrix storage while achieving the best possible performance.

The goal of this research project is to generate implementations of particular sparse kernels tuned to a given matrix and machine. Our work builds on Sparsity [1], an early successful prototype for sparse matrix-vector multiply (y = A*x, where A is a sparse matrix and x, y, are dense vectors), in the following ways:

This research is being conducted by members of the Berkeley Benchmarking and OPtimization (BeBOP) project [5].

[1]
E.-J. Im, "Optimizing the Performance of Sparse Matrix-Vector Multiplication," PhD thesis, UC Berkeley, May 2000.
[2]
R. Vuduc, J. Demmel, K. Yelick, et al., "Performance Optimizations and Bounds for Sparse-Matrix Vector Multiply," Supercomputing, November 2002.
[3]
R. Vuduc, S. Kamil, et al., "Automatic Performance Tuning and Analysis of Sparse Triangular Solve," Int. Conf. Supercomputing, Workshop on Performance Optimization of High-level Languages and Libraries, June 2002.
[4]
BLAS Technical Forum: http://www.netlib.org/ blas/blast-forum.
[5]
The BeBOP Homepage: http://www.cs.berkeley. edu/~richie/bebop.
1Undergraduate (EECS)
2Undergraduate (EECS)
3Undergraduate (EECS)
4Undergraduate (EECS)

More information (http://www.cs.berkeley.edu/~richie/bebop) or

Send mail to the author : (richie@cs.berkeley.edu)

Simulation of Microelectromechanical Systems

David Bindel, David Garmire, Shyam Lakshmin, and Jiawang Nie
(Professor James W. Demmel)
NTONC 29823

SUGAR is a simulation tool for MEMS that inherits its name and philosophy from the IC simulator SPICE. SUGAR is used by researchers at the Berkeley Sensor and Actuator Center, and also at outside sites. Currently, about 500 users have filled out the registration form, and there are many unregistered users as well.

We are working to extend SUGAR in several ways:

We have also deployed a version of SUGAR as a web service running on the Millennium cluster. Also, the SUGAR software was used in a graduate introduction to MEMS course.

There now exists an open source SUGAR repository under the BSD software license on sourceforge.net at http://sourceforge.net/projects/mems. Other relevant information may be found through this website.

SUGAR is a collaborative effort of many graduate students and professors. Affiliated graduate students include David Bindel, Jason V. Clark, David Garmire, Babak Jamshidi, Rishi Kant, Shyam Lakshmin, Jiawang Nie. Affiliated professors include Alice Agogino, Zhaojun Bai, James Demmel, Sanjay Govindjee, Ming Gu, and Kristofer S. J. Pister.


More information (http://www.cs.berkeley.edu/~strive/sugar) or

Send mail to the author : (strive@eecs.berkeley.edu)

A Computer Algebra Integration Web Server


(Professor Richard J. Fateman)
(NSF) CCR-9901933

This work is part of the continuing activity at Berkeley in computer systems for symbolic mathematics (computer algebra). This research involves the establishment and extension of TILU, our online table-look-up symbolic integration service. This table-look-up service currently includes only about 700 forms, although we have data to extend it to the majority of about 13,000 forms in current tables, including many definite integrals involving special functions.

Why bother with table look-up, when there are alleged to be algorithms for symbolic integration? There are several reasons. The best of the algorithms are for indefinite integration only; even so, they somewhat ignore special functions, and often provide a useless result: that no answer exists in closed form in terms of elementary functions. The algebraic algorithms generally ignore the domain of validity of the answer. The results of some of the more powerful techniques are, in many cases, woefully unsimplified. Our table look-up can do the task in about 6.5 milliseconds CPU time on an HP-9000/715 at http://torte.cs.berkeley.edu:8010/tilu. If we can't find it, you will know soon. By comparison, typical algorithms can be orders of magnitude slower.

TILU serves an average of 125 requests a day, and is now tied to a button on the new release of the "Graphing Calculator" shipped with every Apple Macintosh computer.

In addition to studying extensions of this system to more integrals, we are also considering applications to other domains such as symbolic solution of ordinary differential equations.


More information (http://http.cs.berkeley.edu/~fateman) or

Send mail to the author : (fateman@cs.berkeley.edu)

Adaptive Approximations and Exact Penalization for the Solution of Generalized Semi-Infinite Min-Max Problems

Johannes O. Royset1 and Armen Der Kiureghian2
(Professor Elijah Polak)
Norwegian Research Council, (NSF) ECS-9900985, and Space Sciences Laboratory-Lockheed Martin Mini-grant Program

We develop an implementable algorithm for the solution of a class of generalized semi-infinite min-max problems. To this end, first we use exact penalties to convert a generalized semi-infinite min-max problem into a finite family of semi-infinite min-max-min problems. Second, the inner min-function is smoothed and the semi-infinite max part is approximated, using discretization, to obtain a three-parameter family of finite min-max problems. Under a calmness assumption, we show that when the penalty is sufficiently large the semi-infinite min-max-min problems have the same solutions as the original problem, and that when the smoothing and discretization parameters go to infinity the solutions of the finite min-max problems converge to solutions of the original problem, provided the penalty parameter is sufficiently large.

Our algorithm combines tests for adjusting the penalty, the smoothing, and the discretization parameters and makes use of a min-max algorithm as a subroutine. In effect, the min-max algorithm is applied to a sequence of gradually better-approximating min-max problems, with the penalty parameter eventually ceasing to increase, but the smoothing and discretization parameters driven to infinity. A numerical example demonstrates the viability of the algorithm.

1Graduate Student (non-EECS)
2Outside Adviser (non-EECS)

Send mail to the author : (joroyset@ce.berkeley.edu)

Optimization of Cost Functions that Are Defined on the Solution of Differential Algebraic Equations

Michael Wetter1, Van P. Carey2, and Frederick C. Winkelmann3
(Professor Elijah Polak)
(DOE) DE-AC03-76SF00098 and (NSF) ECS-9900985

We are interested in the optimization of a cost function that is evaluated by a thermal building simulation program. Such programs solve a complex system of differential algebraic equations, and can nowadays only be optimized heuristically.

We developed a generalized pattern search method with adaptive precision function evaluations that uses coarse approximations to the cost function in the early iterations, with the approximation precision controlled by a test. Such an approach leads to substantial time savings, and the precision control guarantees that the optimization converges to a stationary point of the infinite precision cost function.

However, implementing the precision control requires knowledge of an error bound function which is hard to obtain for thermal building simulation programs. Therefore, we currently develop simulation models that allow us to obtain such an error bound function, and that reduce computation time for the optimization significantly.

In parallel to the above optimization method, we develop a smoothing method that will allow using existing simulation programs for optimization with guaranteed convergence properties.

1Graduate Student (non-EECS), LBNL, EETD, Simulation Research Group
2Outside Adviser (non-EECS)
3Outside Adviser (non-EECS), LBNL, EETD, Simulation Research Group

More information (http://www.me.berkeley.edu/~mwetter) or

Send mail to the author : (mwetter@lbl.gov)