"With 23.8 PPL on WikiText-103, WaveletLM beats both GPT-2, which was trained on 80× more data, and Transformer-XL Standard, which..."