8 JUL 2020: Facebook
In this paper the authors experiment with 51 languages, first of its kind in the multilingual ASR family. They train different models:
They use seq-2-seq architecture
They train using sentence pieces model for the output representations.
Sampling frequency helps when there is just a balance between uniform and natural frequency. neither of the extreme.
They use curriculum learning as explained in section 3.3.1