pith. machine review for the scientific record. sign in

arxiv: 1810.11954 · v1 · submitted 2018-10-20 · 💻 cs.CL · cs.AI

Recognition: unknown

A Knowledge-Grounded Multimodal Search-Based Conversational Agent

Authors on Pith no claims yet
classification 💻 cs.CL cs.AI
keywords multimodalconversationaldialogueknowledge-groundedmodelsearch-basedaccessadditional
0
0 comments X
read the original abstract

Multimodal search-based dialogue is a challenging new task: It extends visually grounded question answering systems into multi-turn conversations with access to an external database. We address this new challenge by learning a neural response generation system from the recently released Multimodal Dialogue (MMD) dataset (Saha et al., 2017). We introduce a knowledge-grounded multimodal conversational model where an encoded knowledge base (KB) representation is appended to the decoder input. Our model substantially outperforms strong baselines in terms of text-based similarity measures (over 9 BLEU points, 3 of which are solely due to the use of additional information from the KB.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.