Summary of Framequant: Flexible Low-bit Quantization For Transformers, by Harshavardhan Adepu et al.
FrameQuant: Flexible Low-Bit Quantization for Transformersby Harshavardhan Adepu, Zhanpeng Zeng, Li Zhang, Vikas SinghFirst submitted…
FrameQuant: Flexible Low-Bit Quantization for Transformersby Harshavardhan Adepu, Zhanpeng Zeng, Li Zhang, Vikas SinghFirst submitted…
Merging Text Transformer Models from Different Initializationsby Neha Verma, Maha ElbayadFirst submitted to arxiv on:…
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bitsby Shuming Ma,…
How Can LLM Guide RL? A Value-Based Approachby Shenao Zhang, Sirui Zheng, Shuqi Ke, Zhihan…
Deep Learning Approaches for Improving Question Answering Systems in Hepatocellular Carcinoma Researchby Shuning Huo, Yafei…
Dynamic Evaluation of Large Language Models by Meta Probing Agentsby Kaijie Zhu, Jindong Wang, Qinlin…
Efficient data selection employing Semantic Similarity-based Graph Structures for model trainingby Roxana Petcu, Subhadeep MajiFirst…
Improving Language Understanding from Screenshotsby Tianyu Gao, Zirui Wang, Adithya Bhaskar, Danqi ChenFirst submitted to…
UMBCLU at SemEval-2024 Task 1A and 1C: Semantic Textual Relatedness with and without machine translationby…
Symbolic Autoencoding for Self-Supervised Sequence Learningby Mohammad Hossein Amani, Nicolas Mario Baldwin, Amin Mansouri, Martin…