Summary of Ebft: Effective and Block-wise Fine-tuning For Sparse Llms, by Song Guo et al.
EBFT: Effective and Block-Wise Fine-Tuning for Sparse LLMsby Song Guo, Fan Wu, Lei Zhang, Xiawu…
EBFT: Effective and Block-Wise Fine-Tuning for Sparse LLMsby Song Guo, Fan Wu, Lei Zhang, Xiawu…
QuRating: Selecting High-Quality Data for Training Language Modelsby Alexander Wettig, Aatmik Gupta, Saumya Malik, Danqi…
SLEB: Streamlining LLMs through Redundancy Verification and Elimination of Transformer Blocksby Jiwon Song, Kyungseok Oh,…
Permute-and-Flip: An optimally stable and watermarkable decoder for LLMsby Xuandong Zhao, Lei Li, Yu-Xiang WangFirst…
Nevermind: Instruction Override and Moderation in Large Language Modelsby Edward KimFirst submitted to arxiv on:…
DenseFormer: Enhancing Information Flow in Transformers via Depth Weighted Averagingby Matteo Pagliardini, Amirkeivan Mohtashami, Francois…
Comparing Template-based and Template-free Language Model Probingby Sagi Shaier, Kevin Bennett, Lawrence E Hunter, Katharina…
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantizationby Coleman Hooper, Sehoon…
Convergence analysis of t-SNE as a gradient flow for point cloud on a manifoldby Seonghyeon…
Dynamic Layer Tying for Parameter-Efficient Transformersby Tamir David Hay, Lior WolfFirst submitted to arxiv on:…