Loading Now

Summary of Comparative Study on the Performance Of Categorical Variable Encoders in Classification and Regression Tasks, by Wenbin Zhu et al.


Comparative Study on the Performance of Categorical Variable Encoders in Classification and Regression Tasks

by Wenbin Zhu, Runwen Qiu, Ying Fu

First submitted to arxiv on: 18 Jan 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: None

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
Machine learning practitioners often struggle to choose the right encoder for categorical variables in their datasets. This is because different encoders can have a significant impact on model performance. In this study, researchers categorize machine learning models into three types: those that perform affine transformations (ATI), tree-based models, and others. They theoretically prove that the one-hot encoder is the best choice for ATI models, while the target encoder is most suitable for tree-based models. The study also conducts comprehensive experiments to evaluate 14 encoders on 28 datasets with eight machine-learning models. The results agree with the theoretical analysis, providing valuable insights for data scientists in fields like fraud detection and disease diagnosis.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper helps us choose the right way to turn words into numbers for our machine learning model. It shows that different methods can work better or worse depending on what kind of model we’re using. The researchers tested 14 different ways to do this on lots of datasets and showed which ones work best with certain types of models. This can help us make better choices when working with data.

Keywords

* Artificial intelligence  * Encoder  * Machine learning  * One hot