Latent Memory-augmented Graph Transformer for Visual Storytelling

Document Type

Conference Proceeding

Publication Title

MM 2021 - Proceedings of the 29th ACM International Conference on Multimedia

Abstract

Visual storytelling aims to automatically generate a human-like short story given an image stream. Most existing works utilize either scene-level or object-level representations, neglecting the interaction among objects in each image and the sequential dependency between consecutive images. In this paper, we present a novel Latent Memory-augmented Graph Transformer∼(LMGT ), a Transformer based framework for visual story generation. LMGT directly inherits the merits from the Transformer, which is further enhanced with two carefully designed components, i.e., a graph encoding module and a latent memory unit. Specifically, the graph encoding module exploits the semantic relationships among image regions and attentively aggregates critical visual features based on the parsed scene graphs. Furthermore, to better preserve inter-sentence coherence and topic consistency, we introduce an augmented latent memory unit that learns and records highly summarized latent information as the story line from the image stream and the sentence history. Experimental results on three widely-used datasets demonstrate the superior performance of LMGT over the state-of-the-art methods.

First Page

4892

Last Page

4901

DOI

10.1145/3474085.3475236

Publication Date

10-17-2021

Keywords

memory network, scene graph, transformer, visual storytelling

Comments

IR conditions: non-described

Share

COinS