Summary of A Post-training Enhanced Optimization Approach For Small Language Models, by Keke Zhai
A Post-Training Enhanced Optimization Approach for Small Language Modelsby Keke ZhaiFirst submitted to arxiv on:…
A Post-Training Enhanced Optimization Approach for Small Language Modelsby Keke ZhaiFirst submitted to arxiv on:…
A Mamba Foundation Model for Time Series Forecastingby Haoyu Ma, Yushu Chen, Wenlai Zhao, Jinzhe…
Stochastic Monkeys at Play: Random Augmentations Cheaply Break LLM Safety Alignmentby Jason Vega, Junsheng Huang,…
Specialized Foundation Models Struggle to Beat Supervised Baselinesby Zongzhe Xu, Ritvik Gupta, Wenduo Cheng, Alexander…
TeleOracle: Fine-Tuned Retrieval-Augmented Generation with Long-Context Support for Networkby Nouf Alabbasi, Omar Erak, Omar Alhussein,…
On the Loss of Context-awareness in General Instruction Fine-tuningby Yihan Wang, Andrew Bai, Nanyun Peng,…
MM-Embed: Universal Multimodal Retrieval with Multimodal LLMsby Sheng-Chieh Lin, Chankyu Lee, Mohammad Shoeybi, Jimmy Lin,…
Provably Transformers Harness Multi-Concept Word Semantics for Efficient In-Context Learningby Dake Bu, Wei Huang, Andi…
Improving Steering Vectors by Targeting Sparse Autoencoder Featuresby Sviatoslav Chalnev, Matthew Siu, Arthur ConmyFirst submitted…
Training on test proteins improves fitness, structure, and function predictionby Anton Bushuiev, Roman Bushuiev, Nikola…