PET: Optimizing Tensor Programs with Partially Equivalent Transformations and Automated Corrections
https://www.usenix.org/conference/osdi21/presentation/wang
Last updated
Was this helpful?
https://www.usenix.org/conference/osdi21/presentation/wang
Last updated
Was this helpful?
Current system considers only full equivalent transformations
Pro: preserve functionality
Con: miss optimization opportunities
Partial Equivalent Transformation
Pro: better performance
Faster ML operators
More efficient tensor layouts
HW-specific optimizations
Con: potential accuracy loss
Key: benefits while preserving equivalence?
Correction preserves equivalence
Super optimization
Multi-linearity of DNN computations
DNN computation = MLTP + non-linear activations
Reduce complexity from O(m*n) to O(1)
Corrector:
Re-compute the incorrect outputs using the original program
opportunistically fuse correction kernels with other operators
Eval: less than 1% overhead
Program optimizer:
Search-based program optimizer
Beam search