ITU AI/ML in 5G Challenge: DNN Inference Optimization Challenge
The talk will describe the problem ITU-ML5G-PS-018 DNN Inference Optimization. This problem is about how to optimize inference efficiency of deep learning models since computing efficiency, memory footprint and inference latency tends to be the bottleneck when deploying large deep learning models. The problem will be described in details in the presentation, including why we propose this problem, status quo of related research, what are expected from the competitors/participants and some suggestions will be given. Hopefully more competitors will get intested in this problem and propose their innovative solutions.