Loading Now

Summary of Not Another Imputation Method: a Transformer-based Model For Missing Values in Tabular Datasets, by Camillo Maria Caruso et al.


Not Another Imputation Method: A Transformer-based Model for Missing Values in Tabular Datasets

by Camillo Maria Caruso, Paolo Soda, Valerio Guarrasi

First submitted to arxiv on: 16 Jul 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: None

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The paper introduces “Not Another Imputation Method” (NAIM), a transformer-based model designed to handle missing values in tabular datasets without traditional imputation techniques. NAIM employs feature-specific embeddings and a masked self-attention mechanism that learns from available data, avoiding the need for imputation. The model is evaluated on 5 public datasets against 10 state-of-the-art models, demonstrating superior performance with various imputation techniques. The results highlight NAIM’s predictive capabilities and resilience in missing data scenarios.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper is about a new way to deal with missing information in big tables of data that computers use to learn from experience. Usually, we fill in the missing bits before training our models. But this new method, called NAIM, can skip this step altogether! It uses special words and attention techniques to figure out what’s going on even when some data is missing. The authors tested it with lots of real-world datasets and showed that it works better than other ways people are trying to solve this problem.

Keywords

» Artificial intelligence  » Attention  » Self attention  » Transformer