Loading Now

Summary of Reducing and Exploiting Data Augmentation Noise Through Meta Reweighting Contrastive Learning For Text Classification, by Guanyi Mou et al.


Reducing and Exploiting Data Augmentation Noise through Meta Reweighting Contrastive Learning for Text Classification

by Guanyi Mou, Yichuan Li, Kyumin Lee

First submitted to arxiv on: 26 Sep 2024

Categories

  • Main: Computation and Language (cs.CL)
  • Secondary: Machine Learning (cs.LG)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The proposed novel framework leverages meta learning and contrastive learning techniques to reweight augmented samples in text classification tasks, refining their feature representations based on quality. The framework incorporates novel weight-dependent enqueue and dequeue algorithms, demonstrating the ability to cooperate with existing deep learning models like RoBERTa-base and Text-CNN, as well as augmentation techniques such as Wordnet and Easydata. Experimental results show significant improvements of up to 4.3% absolute improvement on seven GLUE benchmark datasets compared to the best baseline.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper proposes a new way to improve deep learning models’ performance when using augmented data in text classification tasks. The idea is to use special algorithms that look at how good or bad each piece of augmented data is, and adjust the model’s behavior based on this information. This helps the model learn more effectively from the high-quality data and ignore the low-quality data. The results show that this approach can lead to significant improvements in performance, with an average improvement of 1.6-4.3% compared to the best previous methods.

Keywords

» Artificial intelligence  » Cnn  » Deep learning  » Meta learning  » Text classification