Skip to content

Latest commit

 

History

History
43 lines (28 loc) · 1.11 KB

README.md

File metadata and controls

43 lines (28 loc) · 1.11 KB

Borrowed some code from AdaptiveAttention.

Overview

Overview

Summarization

  • We propose a compositional captioning model, where caption generation follows a recursive structure, which naturally fits the properties of human language. Specifically, we could generate diverse captions based on the same visual semantics by changing syntax, or attentions.

Good

Codes

  • nlp: scripts for processing sentences using a nlp toolkit

  • data: scripts for processing data

  • misc: models, dataloaders, and auxiliary codes

  • image_models: pretrained image models

  • test: scripts for studies

  • scripts: bashs

Citation

@inproceedings{dai2018aneural,
title = {A Neural Compositional Paradigm for Image Captioning},
author = {Dai, Bo and Fidler, Sanja and Lin, Dahua},
booktitle = {Advances in Neural Information Processing Systems 31},
pages = {658--668},
year = {2018}
}