microsoft/unilm
here, LayoutReader, microsoft/unilm official, unilm, на сайте с 18 декабря 2022, 15:50
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities.
Fundamental research to improve modeling generality and capability, as well as training stability and efficiency for Transformers at any scale.