Dialmat: Dialogue-enabled Transformer With Moment-based Adversarial Training · The Large Language Model Bible Contribute to LLM-Bible

Dialmat: Dialogue-enabled Transformer With Moment-based Adversarial Training

Kaneda Kanta, Korekata Ryosuke, Wada Yuiga, Nagashima Shunya, Kambara Motonari, Iioka Yui, Matsuo Haruka, Imai Yuto, Nishimura Takayuki, Sugiura Komei. Arxiv 2023

[Paper]    
Agentic Model Architecture Pretraining Methods Security Training Techniques Transformer

This paper focuses on the DialFRED task, which is the task of embodied instruction following in a setting where an agent can actively ask questions about the task. To address this task, we propose DialMAT. DialMAT introduces Moment-based Adversarial Training, which incorporates adversarial perturbations into the latent space of language, image, and action. Additionally, it introduces a crossmodal parallel feature extraction mechanism that applies foundation models to both language and image. We evaluated our model using a dataset constructed from the DialFRED dataset and demonstrated superior performance compared to the baseline method in terms of success rate and path weighted success rate. The model secured the top position in the DialFRED Challenge, which took place at the CVPR 2023 Embodied AI workshop.

Similar Work