PLUS
PLUS
Home
People
Research
Publications
Internal
Contact
Light
Dark
Automatic
Vision-Language Pretraining
KD-VLP: Improving End-to-End Vision-and-Language Pretraining with Object Knowledge Distillation
Self-supervised vision-and-language pretraining (VLP) aims to learn transferable multi-modal representations from large-scale image-text data and to achieve strong performances on a broad scope of vision-language tasks after finetuning. Previous …
Cite
×