PLATO-XL: Exploring the Large-scale Pre-training of Dialogue Generation
pdf: https://arxiv.org/pdf/2109.09519.pdf
abs: https://arxiv.org/abs/2109.09519
present the models of PLATO-XL with up to 11 billion parameters, trained on both Chinese and English social media conversations