Mapping tensor products onto VLSI networks with reduced I/O

A. Elnaggar*, H. M. Alnuweiri, M. R. Ito

*Corresponding author for this work

Research output: Contribution to journalConference articlepeer-review

1 Citation (Scopus)

Abstract

This paper presents a methodology for designing folded VLSI networks for implementing tensor-product forms. Using tensor-products leads to very efficient expressions for a large number of computations in digital signal processing and matrix arithmetic. The resulting networks can trade-off total time delay with I/O bandwidth and chip area. The main goal is to parametrize the VLSI architecture so that it can be implemented under various packaging constraints including the available number of I/O pins, available chip-area, and certain restrictions on maximum wire length. Our methods result in folded VLSI networks with optimal AT2 trade-off for digital filtering and multidimensional transforms, where A is the total area of the VLSI circuit (or chip) and T is its total time delay.

Original languageEnglish
Pages (from-to)150-155
Number of pages6
JournalProceedings of the IEEE Great Lakes Symposium on VLSI
Publication statusPublished - 1994
Externally publishedYes
EventProceedings of the 4th Great Lakes Symposium on VLSI - Notre Dame, IN, USA
Duration: 4 Mar 19945 Mar 1994

Fingerprint

Dive into the research topics of 'Mapping tensor products onto VLSI networks with reduced I/O'. Together they form a unique fingerprint.

Cite this