Summary of Llm-neo: Parameter Efficient Knowledge Distillation For Large Language Models, by Runming Yang et al.
LLM-NEO: Parameter Efficient Knowledge Distillation for Large Language Modelsby Runming Yang, Taiqiang Wu, Jiahao Wang,…
LLM-NEO: Parameter Efficient Knowledge Distillation for Large Language Modelsby Runming Yang, Taiqiang Wu, Jiahao Wang,…
Accelerating Large Language Model Training with 4D Parallelism and Memory Consumption Estimatorby Kazuki Fujii, Kohei…
Robust and Efficient Fine-tuning of LLMs with Bayesian Reparameterization of Low-Rank Adaptationby Ayan Sengupta, Vaibhav…
Variational Low-Rank Adaptation Using IVONby Bai Cong, Nico Daheim, Yuesong Shen, Daniel Cremers, Rio Yokota,…
Can Custom Models Learn In-Context? An Exploration of Hybrid Architecture Performance on In-Context Learning Tasksby…
Self-Consistency Preference Optimizationby Archiki Prasad, Weizhe Yuan, Richard Yuanzhe Pang, Jing Xu, Maryam Fazel-Zarandi, Mohit…
A Comprehensive Survey of Small Language Models in the Era of Large Language Models: Techniques,…
Stochastic Monkeys at Play: Random Augmentations Cheaply Break LLM Safety Alignmentby Jason Vega, Junsheng Huang,…
“Give Me BF16 or Give Me Death”? Accuracy-Performance Trade-Offs in LLM Quantizationby Eldar Kurtic, Alexandre…
TableGPT2: A Large Multimodal Model with Tabular Data Integrationby Aofeng Su, Aowen Wang, Chao Ye,…