Large-scale Multi-Modal Pre-trained Models: A Comprehensive Survey
[article]
Xiao Wang, Guangyao Chen, Guangwu Qian, Pengcheng Gao, Xiao-Yong Wei, Yaowei Wang, Yonghong Tian, Wen Gao
2024
arXiv
pre-print
In this work, we give a comprehensive survey of these models and hope this paper could provide new insights and helps fresh researchers to track the most cutting-edge works. ...
Then, we introduce the task definition, key challenges, and advantages of multi-modal pre-training models (MM-PTMs), and discuss the MM-PTMs with a focus on data, objectives, network architectures, and ...
arXiv NLP SC, 34 An Empirical Survey of the Effectiveness of Debiasing Techniques for Pre-Trained Language Models [45] 2021 arXiv NLP DC, 21 A multi-layer bidirectional transformer encoder for pre-trained ...
arXiv:2302.10035v3
fatcat:j2y4ti2f65aqbnap4ln436c7ci