Loading Now

Summary of Retrieval & Fine-tuning For In-context Tabular Models, by Valentin Thomas et al.


Retrieval & Fine-Tuning for In-Context Tabular Models

by Valentin Thomas, Junwei Ma, Rasa Hosseinzadeh, Keyvan Golestan, Guangwei Yu, Maksims Volkovs, Anthony Caterini

First submitted to arxiv on: 7 Jun 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: None

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The paper addresses the challenge of applying transformer-based deep learning to large and complex tabular datasets. Recent advancements in in-context learning have shown promise on smaller datasets, but struggle to scale to larger ones. The proposed approach combines retrieval and fine-tuning: nearest neighbours are collected from a local subset of the data, and then task-specific fine-tuning is performed using this retrieved set. This locally-calibrated PFN (LoCalPFN) model outperforms existing in-context models on 95 datasets curated by TabZilla from OpenML, achieving state-of-the-art results even with respect to tuned tree-based models.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper tries to make deep learning work better for really big and complicated tables of data. Right now, it’s hard to use this kind of AI on very large datasets because they’re too complex. The solution is to take a smaller part of the dataset that looks similar to what you’re trying to predict, and then train the model using those examples. This new approach, called LoCalPFN, works really well even compared to other methods that are specifically designed for tables.

Keywords

* Artificial intelligence  * Deep learning  * Fine tuning  * Transformer