TangramFP

Date:

TangramFP: Energy-Efficient, Bit-Parallel Multiply-Accumulate for Deep Neural Networks

Code repository for the paper
TangramFP: Energy-Efficient, Bit-Parallel, Multiply-Accumulate for Deep Neural Networks
Published in IEEE 36th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD), 2024
🎉 Best Paper Candidate!

🔗 GitHub Repository


Repository Contents

This repository includes the full TangramFP implementation across multiple levels:

1. PyTorch Wrapper for TangramFP

A PyTorch-based implementation supporting:

  • Matrix Multiplication
  • Convolution
  • Linear Layer
  • Depth-wise Convolution

2. Backend C Code for TangramFP

  • Supports 16-bit and [NEW!] 32-bit floating-point precision.

3. Verilog Code for TangramFP MAC Units

  • Implementations for 16-bit and 32-bit MAC units.

How to Use

Follow the README in the repository for setup, build, and usage instructions.