Emptransfo: A Multi-head Transformer Architecture For Creating Empathetic Dialog Systems · The Large Language Model Bible Contribute to LLM-Bible

Emptransfo: A Multi-head Transformer Architecture For Creating Empathetic Dialog Systems

Zandie Rohola, Mahoor Mohammad H.. Arxiv 2020

[Paper]    
GPT Model Architecture Pretraining Methods Transformer

Understanding emotions and responding accordingly is one of the biggest challenges of dialog systems. This paper presents EmpTransfo, a multi-head Transformer architecture for creating an empathetic dialog system. EmpTransfo utilizes state-of-the-art pre-trained models (e.g., OpenAI-GPT) for language generation, though models with different sizes can be used. We show that utilizing the history of emotions and other metadata can improve the quality of generated conversations by the dialog system. Our experimental results using a challenging language corpus show that the proposed approach outperforms other models in terms of Hit@1 and PPL (Perplexity).

Similar Work