Summary of From Bytes to Borsch: Fine-tuning Gemma and Mistral For the Ukrainian Language Representation, by Artur Kiulian et al.
From Bytes to Borsch: Fine-Tuning Gemma and Mistral for the Ukrainian Language Representationby Artur Kiulian,…
From Bytes to Borsch: Fine-Tuning Gemma and Mistral for the Ukrainian Language Representationby Artur Kiulian,…
Exploring and Improving Drafts in Blockwise Parallel Decodingby Taehyeon Kim, Ananda Theertha Suresh, Kishore Papineni,…
Knowledgeable Agents by Offline Reinforcement Learning from Large Language Model Rolloutsby Jing-Cheng Pang, Si-Hang Yang,…
On Speculative Decoding for Multimodal Large Language Modelsby Mukul Gagrani, Raghavv Goel, Wonseok Jeon, Junyoung…
ChimpVLM: Ethogram-Enhanced Chimpanzee Behaviour Recognitionby Otto Brookes, Majid Mirmehdi, Hjalmar Kuhl, Tilo BurghardtFirst submitted to…
Training a Vision Language Model as Smartphone Assistantby Nicolai Dorka, Janusz Marecki, Ammar AnwarFirst submitted…
PoliTune: Analyzing the Impact of Data Selection and Fine-Tuning on Economic and Political Biases in…
Toward a Theory of Tokenization in LLMsby Nived Rajaraman, Jiantao Jiao, Kannan RamchandranFirst submitted to…
Variance-reduced Zeroth-Order Methods for Fine-Tuning Language Modelsby Tanmay Gautam, Youngsuk Park, Hao Zhou, Parameswaran Raman,…
RecurrentGemma: Moving Past Transformers for Efficient Open Language Modelsby Aleksandar Botev, Soham De, Samuel L…