GA-SATIC: Semi-autoregressive Transformer Image Captioning Based on Geometric Attention
Abstract
References
Index Terms
- GA-SATIC: Semi-autoregressive Transformer Image Captioning Based on Geometric Attention
Recommendations
Semi-Autoregressive Image Captioning
MM '21: Proceedings of the 29th ACM International Conference on MultimediaCurrent state-of-the-art approaches for image captioning typically adopt an autoregressive manner, i.e., generating descriptions word by word, which suffers from slow decoding issue and becomes a bottleneck in real-time applications. Non-autoregressive ...
Neural attention for image captioning: review of outstanding methods
AbstractImage captioning is the task of automatically generating sentences that describe an input image in the best way possible. The most successful techniques for automatically generating image captions have recently used attentive deep learning models. ...
Image captioning using transformer-based double attention network
AbstractImage captioning generates a human-like description for a query image, which has attracted considerable attention recently. The most broadly utilized model for image description is an encoder–decoder structure, where the encoder ...
Graphical abstractDisplay Omitted
Comments
Information & Contributors
Information
Published In

Publisher
Association for Computing Machinery
New York, NY, United States
Publication History
Check for updates
Author Tags
Qualifiers
- Research-article
- Research
- Refereed limited
Conference
Contributors
Other Metrics
Bibliometrics & Citations
Bibliometrics
Article Metrics
- 0Total Citations
- 28Total Downloads
- Downloads (Last 12 months)8
- Downloads (Last 6 weeks)2
Other Metrics
Citations
Cited By
View allView Options
Login options
Check if you have access through your login credentials or your institution to get full access on this article.
Sign inFull Access
View options
View or Download as a PDF file.
PDFeReader
View online with eReader.
eReaderHTML Format
View this article in HTML Format.
HTML Format