PET: Optimizing Tensor Programs with Partially Equivalent Transformations and Automated Corrections
https://www.usenix.org/conference/osdi21/presentation/wang
- Current system considers only full equivalent transformations - Pro: preserve functionality 
- Con: miss optimization opportunities 
 
- Partial Equivalent Transformation - Pro: better performance - Faster ML operators 
- More efficient tensor layouts 
- HW-specific optimizations 
 
- Con: potential accuracy loss 
 
- Key: benefits while preserving equivalence? - Correction preserves equivalence 
 

- Super optimization 
- Multi-linearity of DNN computations - DNN computation = MLTP + non-linear activations 
- Reduce complexity from O(m*n) to O(1) 
 
Corrector:
- Re-compute the incorrect outputs using the original program 
- opportunistically fuse correction kernels with other operators 
- Eval: less than 1% overhead 
Program optimizer:
- Search-based program optimizer - Beam search 
 
Last updated
Was this helpful?