Chisel Pipelined GEMM (#30)

* Reset to 644 file permissions

* Add json files to src/test/resources for testing

* Add new TensorGemmPipelinedSplit module and rename existing TensorGemm to TensorGemmOrig

* Tests for TensorGemmPipelinedSplit, TensorGemmOrig, and associated submodules

* Add jackson plugin dependency and stricter Scala checks

* Remove debug prints

* Rename x.json and y.json to gemm_1uop_overflow_offset.json and gemm_2uop_overflow_cascaded.json respectively

* All occurrences of '\( ' replaced with '\('

* Add linting rule to flag spaces after lparen characters

* Remove comment

* Rename TensorGemmOrig to TensorGemmSimple
11 files changed
tree: dee39349a9bea3082da0c8f27ebeed4d83986cab
  1. .asf.yaml
  2. .gitignore
  3. Jenkinsfile
  7. apps/
  8. config/
  9. hardware/
  10. include/
  11. src/
  12. tests/

VTA Hardware Design Stack

Build Status

VTA (versatile tensor accelerator) is an open-source deep learning accelerator complemented with an end-to-end TVM-based compiler stack.

The key features of VTA include:

  • Generic, modular, open-source hardware
    • Streamlined workflow to deploy to FPGAs.
    • Simulator support to prototype compilation passes on regular workstations.
  • Driver and JIT runtime for both simulator and FPGA hardware back-end.
  • End-to-end TVM stack integration
    • Direct optimization and deployment of models from deep learning frameworks via TVM.
    • Customized and extensible TVM compiler back-end.
    • Flexible RPC support to ease deployment, and program FPGAs with the convenience of Python.