Summary of A Post-training Enhanced Optimization Approach For Small Language Models, by Keke Zhai
A Post-Training Enhanced Optimization Approach for Small Language Modelsby Keke ZhaiFirst submitted to arxiv on:…
A Post-Training Enhanced Optimization Approach for Small Language Modelsby Keke ZhaiFirst submitted to arxiv on:…
A Mamba Foundation Model for Time Series Forecastingby Haoyu Ma, Yushu Chen, Wenlai Zhao, Jinzhe…
ATM: Improving Model Merging by Alternating Tuning and Mergingby Luca Zhou, Daniele Solombrino, Donato Crisostomi,…
Stochastic Monkeys at Play: Random Augmentations Cheaply Break LLM Safety Alignmentby Jason Vega, Junsheng Huang,…
Specialized Foundation Models Struggle to Beat Supervised Baselinesby Zongzhe Xu, Ritvik Gupta, Wenduo Cheng, Alexander…
On the Loss of Context-awareness in General Instruction Fine-tuningby Yihan Wang, Andrew Bai, Nanyun Peng,…
MM-Embed: Universal Multimodal Retrieval with Multimodal LLMsby Sheng-Chieh Lin, Chankyu Lee, Mohammad Shoeybi, Jimmy Lin,…
TeleOracle: Fine-Tuned Retrieval-Augmented Generation with Long-Context Support for Networkby Nouf Alabbasi, Omar Erak, Omar Alhussein,…
Improving Steering Vectors by Targeting Sparse Autoencoder Featuresby Sviatoslav Chalnev, Matthew Siu, Arthur ConmyFirst submitted…
Provably Transformers Harness Multi-Concept Word Semantics for Efficient In-Context Learningby Dake Bu, Wei Huang, Andi…