|  | ..  Licensed to the Apache Software Foundation (ASF) under one | 
|  | or more contributor license agreements.  See the NOTICE file | 
|  | distributed with this work for additional information | 
|  | regarding copyright ownership.  The ASF licenses this file | 
|  | to you under the Apache License, Version 2.0 (the | 
|  | "License"); you may not use this file except in compliance | 
|  | with the License.  You may obtain a copy of the License at | 
|  |  | 
|  | ..    http://www.apache.org/licenses/LICENSE-2.0 | 
|  |  | 
|  | ..  Unless required by applicable law or agreed to in writing, | 
|  | software distributed under the License is distributed on an | 
|  | "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY | 
|  | KIND, either express or implied.  See the License for the | 
|  | specific language governing permissions and limitations | 
|  | under the License. | 
|  |  | 
|  | Publications | 
|  | ============ | 
|  |  | 
|  | TVM is developed as part of peer-reviewed research in machine learning compiler | 
|  | framework for CPUs, GPUs, and machine learning accelerators. | 
|  |  | 
|  | This document includes references to publications describing the research, | 
|  | results, and design that use or built on top of TVM. | 
|  |  | 
|  | 2018 | 
|  |  | 
|  | * `TVM: An Automated End-to-End Optimizing Compiler for Deep Learning`__, [Slides_] | 
|  |  | 
|  | .. __: https://arxiv.org/abs/1802.04799 | 
|  | .. _Slides: https://www.usenix.org/system/files/osdi18-chen.pdf | 
|  |  | 
|  | * `Learning to Optimize Tensor Programs`__, [Slides] | 
|  |  | 
|  | .. __: https://arxiv.org/pdf/1805.08166.pdf | 
|  |  | 
|  | 2020 | 
|  |  | 
|  | * `Ansor: Generating High-Performance Tensor Programs for Deep Learning`__, [Slides__] [Tutorial__] | 
|  |  | 
|  | .. __: https://arxiv.org/abs/2006.06762 | 
|  | .. __: https://www.usenix.org/sites/default/files/conference/protected-files/osdi20_slides_zheng.pdf | 
|  | .. __: https://tvm.apache.org/2021/03/03/intro-auto-scheduler | 
|  |  | 
|  | 2021 | 
|  |  | 
|  | * `Nimble: Efficiently Compiling Dynamic Neural Networks for Model Inference`__, [Slides__] | 
|  |  | 
|  | .. __: https://arxiv.org/abs/2006.03031 | 
|  | .. __: https://shenhaichen.com/slides/nimble_mlsys.pdf | 
|  |  | 
|  | * `Cortex: A Compiler for Recursive Deep Learning Models`__, [Slides__] | 
|  |  | 
|  | .. __: https://arxiv.org/pdf/2011.01383.pdf | 
|  | .. __: https://mlsys.org/media/mlsys-2021/Slides/1507.pdf | 
|  |  | 
|  | * `UNIT: Unifying Tensorized Instruction Compilation`__, [Slides] | 
|  |  | 
|  | .. __: https://arxiv.org/abs/2101.08458 | 
|  |  | 
|  | * `Lorien: Efficient Deep Learning Workloads Delivery`__, [Slides] | 
|  |  | 
|  | .. __: https://assets.amazon.science/c2/46/2481c9064a8bbaebcf389dd5ad75/lorien-efficient-deep-learning-workloads-delivery.pdf | 
|  |  | 
|  |  | 
|  | * `Bring Your Own Codegen to Deep Learning Compiler`__, [Slides] [Tutorial__] | 
|  |  | 
|  | .. __: https://arxiv.org/abs/2105.03215 | 
|  | .. __: https://tvm.apache.org/2020/07/15/how-to-bring-your-own-codegen-to-tvm | 
|  |  | 
|  | 2022 | 
|  |  | 
|  | * `DietCode: Automatic optimization for dynamic tensor program`__, [Slides] | 
|  |  | 
|  | .. __: https://proceedings.mlsys.org/paper/2022/file/fa7cdfad1a5aaf8370ebeda47a1ff1c3-Paper.pdf | 
|  |  | 
|  | * `Bolt: Bridging the Gap between Auto-tuners and Hardware-native Performance`__, [Slides] | 
|  |  | 
|  | .. __: https://proceedings.mlsys.org/paper/2022/file/38b3eff8baf56627478ec76a704e9b52-Paper.pdf | 
|  |  | 
|  | * `The CoRa Tensor Compiler: Compilation for Ragged Tensors with Minimal Padding`__, [Slides] | 
|  |  | 
|  | .. __: https://arxiv.org/abs/2110.10221 |