MMPT'21: International Joint Workshop on Multi-Modal Pre-Training for Multimedia Understanding
Abstract
References
Index Terms
- MMPT'21: International Joint Workshop on Multi-Modal Pre-Training for Multimedia Understanding
Recommendations
Vision-language pre-training via modal interaction
AbstractExisting vision-language pre-training models typically extract region features and conduct fine-grained local alignment based on masked image/text completion or object detection methods. However, these models often design independent subtasks for ...
Highlights- In the field of cross-modal pre-training, a unified pre-training framework for multi-task collaboration based on modal interaction is proposed.
- Two new subtasks including image filling and text filling are designed to complete missing ...
Poster: Boosting Adversarial Robustness by Adversarial Pre-training
CCS '23: Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications SecurityVision Transformer (ViT) shows superior performance on various tasks, but, similar to other deep learning techniques, it is vulnerable to adversarial attacks. Due to the differences between ViT and traditional CNNs, previous works designed new ...
Comments
Information & Contributors
Information
Published In
![cover image ACM Conferences](/cms/asset/faaca846-c856-44a9-a676-c316f1a51cb4/3460426.cover.jpg)
- General Chairs:
- Wen-Huang Cheng,
- Mohan Kankanhalli,
- Meng Wang,
- Program Chairs:
- Wei-Ta Chu,
- Jiaying Liu,
- Marcel Worring
Sponsors
Publisher
Association for Computing Machinery
New York, NY, United States
Publication History
Check for updates
Author Tags
Qualifiers
- Abstract
Conference
Acceptance Rates
Contributors
Other Metrics
Bibliometrics & Citations
Bibliometrics
Article Metrics
- 0Total Citations
- 74Total Downloads
- Downloads (Last 12 months)2
- Downloads (Last 6 weeks)0
Other Metrics
Citations
View Options
Login options
Check if you have access through your login credentials or your institution to get full access on this article.
Sign in