V0.7.0 Release Plan
cp5555 opened this issue · comments
Release Manager
Endgame
- Code freeze: Jan. 3rd, 2023
- Bug Bash date: Jan 13th, 2023
- Release date: Jan 20th, 2023
Main Features
SuperBench Improvement
-
- Update version to include revision hash and date (#427)
-
- Fix Transformers version to avoid Tensorrt failure (#441)
-
- Add CUDA11.8 Docker image for Nvidia arch90 GPUs (#449)
-
- Support
sb deploy
without docker pulling (#466)
- Support
Micro-benchmark Improvement
-
- Support list of custom config string in cudnn-functions and cublas-functions (#414)
-
- Support GEMM-FLOPS for Nvidia arch90 GPUs (#456)
-
- Add wait time option to resolve mem-bw unstable issue (#438)
-
- Fix bug for incorrect datatype judgement in cublas-function source code. (#462)
Model-benchmark Improvement
Distributed Benchmark Improvement
-
- Support pair-wise pattern in IB validation benchmark. (#453)
-
- Support topo-aware, pair-wise, and K-batch pattern in nccl-bw benchmark. (#454)
Backlog
Inference Benchmark Improvement
- Support VGG, LSTM, and GPT-2 small in TensorRT Inference Backend
- Support VGG, LSTM, and GPT-2 small in ORT Inference Backend
- Support more TensorRT parameters (Related to #366)
Document
- Metric Reasoning Doc