[Crib-list] Speaker: NADYA BLISS -- Computational Research in Boston Seminar -- Friday, May 2, 2008 -- TIME: 12:30 PM -- Room 32-144 (fwd)
Shirley Entzminger
daisymae at math.mit.edu
Thu May 1 13:37:24 EDT 2008
COMPUTATIONAL RESEARCH in BOSTON SEMINAR
DATE: Friday, MAY 2, 2008
TIME: 12:30 PM
LOCATION: Building 32, Room 144 (Stata Center)
(Pizza and beverages will be provided at 12:15 PM outside room.)
TITLE: SMaRT: Sparse Mapping and Routing Toolbox
SPEAKER: NADYA BLISS (MIT - Lincoln Lab)
ABSTRACT:
Over the last few years, MIT Lincoln Laboratory has been developing techniques
for automatic algorithm-to-architecture mapping of signal processing
applications. Due to the nature of signal processing operations, modeling,
program analysis, and mapping techniques have been limited to dense matrix
computations. As more post-processing algorithms move to the sensor front end,
efficient parallel implementations of those algorithms for sparse data are
becoming necessary. Post-processing algorithms are typically represented in
graphical form, but can also be formulated using sparse matrix notation. The
efficiency of the graph algorithms is often only a small fraction (< 0.01) of
the peak throughput of a conventional architecture, and the efficiency
deteriorates as the problem size increases. Furthermore, graph algorithms are
not well suited for parallelization. Sparse matrix formulations expose
intrinsic algorithm parallelism; however, efficient mappings are still needed
to exploit the parallelism and deliver desirable efficiency.
This talk will motivate the need for efficient sparse matrix processing with an
example from the graph algorithm community and highlight a key computational
kernel: sparse matrix-matrix multiply. We will first introduce a high-level
overview of our dense mapping framework and highlight its limitations with
regards to parallelization of sparse computations. We will then show how SMaRT
overcomes these limitations using fine-grain architecture modeling, dependency
analysis, mapping and routing techniques. Finally, we will present results for
a number of implementations of the matrix multiply kernel. Our results show
over an order of magnitude performance improvement over traditional sparse 2D
block-cyclic mappings. This performance improvement can be attributed to
detailed runtime analysis of the computation, sensitivity to the sparsity of
the data, co-optimized map and route determination, and the detailed hardware
modeling capability.
This is joint work with Una-May O'Reilly at CSAIL and Sanjeev Mohindra at
Lincoln Laboratory.
This work is sponsored by the Department of the Air Force under Air Force
contract FA8721-05-C-0002. Opinions, interpretations, conclusions and
recommendations are those of the author and are not necessarily endorsed by the
United States Government
*****************************************************************************
Massachusetts Institute of Technology
Cambridge, MA 02139
http://www-math. mit.edu/crib
For information on CRiB, contact:
Alan Edelman: edelman at math.mit.edu
Steven G. Johnson: stevenj at math.mit.edu
Jeremy Kepner: kepner at ll.mit.edu
Patrick Dreher: dreher at mit.edu
More information about the CRiB-list
mailing list