Loading Now

Summary of Large Language Models As Zero-shot Dialogue State Tracker Through Function Calling, by Zekun Li et al.


Large Language Models as Zero-shot Dialogue State Tracker through Function Calling

by Zekun Li, Zhiyu Zoey Chen, Mike Ross, Patrick Huber, Seungwhan Moon, Zhaojiang Lin, Xin Luna Dong, Adithya Sagar, Xifeng Yan, Paul A. Crook

First submitted to arxiv on: 16 Feb 2024

Categories

  • Main: Computation and Language (cs.CL)
  • Secondary: Artificial Intelligence (cs.AI)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The proposed FnCTOD approach utilizes function calling to improve zero-shot dialogue state tracking (DST) in large language models (LLMs). This method enables adaptation to diverse domains without extensive data collection or model tuning. The experimental results demonstrate exceptional performance, surpassing the previous state-of-the-art achieved by ChatGPT and improving its average joint goal accuracy (JGA) by 5.6%. The individual model results for GPT-3.5 and GPT-4 are boosted by 4.8% and 14%, respectively. Additionally, fine-tuning on a small collection of diverse task-oriented dialogues enables modestly sized models to achieve DST performance comparable to ChatGPT while maintaining their chat capabilities.
Low GrooveSquid.com (original content) Low Difficulty Summary
FnCTOD is a new way to make large language models better at understanding what’s going on in conversations that are working towards specific goals. It does this by using “function calling” which helps the model keep track of the conversation’s state without needing lots of extra data or fine-tuning. The results show that this approach works really well, beating the previous best result and improving how well the models can understand what’s going on in conversations.

Keywords

» Artificial intelligence  » Fine tuning  » Gpt  » Tracking  » Zero shot