Loading Now

Summary of Eda-dm: Enhanced Distribution Alignment For Post-training Quantization Of Diffusion Models, by Xuewen Liu et al.


EDA-DM: Enhanced Distribution Alignment for Post-Training Quantization of Diffusion Models

by Xuewen Liu, Zhikai Li, Junrui Xiao, Qingyi Gu

First submitted to arxiv on: 9 Jan 2024

Categories

  • Main: Computer Vision and Pattern Recognition (cs.CV)
  • Secondary: Machine Learning (cs.LG)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The proposed Enhanced Distribution Alignment for Post-Training Quantization of Diffusion Models (EDA-DM) aims to address the distribution mismatch issues that plague existing post-training quantization (PTQ) methods for diffusion models. By selecting calibration samples based on density and variety in the latent space, EDA-DM improves alignment at the sample level. Additionally, the modified loss function used during block reconstruction aligns output distributions between the quantized model and its full-precision counterpart at different network granularity levels.
Low GrooveSquid.com (original content) Low Difficulty Summary
Diffusion models are really good at generating images, but they’re not very fast because they need to do a lot of complicated calculations. To make them faster, scientists tried shrinking the model’s size by using something called post-training quantization (PTQ). However, this didn’t work well for diffusion models because their “thoughts” or activations are super dynamic and hard to compress. The new method, EDA-DM, fixes this problem by choosing the right samples to use during training and adjusting how the model is evaluated during reconstruction.

Keywords

* Artificial intelligence  * Alignment  * Diffusion  * Latent space  * Loss function  * Precision  * Quantization