Summary of Hydra: Sequentially-dependent Draft Heads For Medusa Decoding, by Zachary Ankner et al.
Hydra: Sequentially-Dependent Draft Heads for Medusa Decodingby Zachary Ankner, Rishab Parthasarathy, Aniruddha Nrusimha, Christopher Rinard,…
Hydra: Sequentially-Dependent Draft Heads for Medusa Decodingby Zachary Ankner, Rishab Parthasarathy, Aniruddha Nrusimha, Christopher Rinard,…
Beyond Answers: Transferring Reasoning Capabilities to Smaller LLMs Using Multi-Teacher Knowledge Distillationby Yijun Tian, Yikun…
Fine-Tuned Language Models Generate Stable Inorganic Materials as Textby Nate Gruver, Anuroop Sriram, Andrea Madotto,…
Retrieve to Explain: Evidence-driven Predictions with Language Modelsby Ravi Patel, Angus Brayne, Rogier Hintzen, Daniel…
Distinguishing the Knowable from the Unknowable with Language Modelsby Gustaf Ahdritz, Tian Qin, Nikhil Vyas,…
Make Every Move Count: LLM-based High-Quality RTL Code Generation Using MCTSby Matthew DeLorenzo, Animesh Basak…
DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Modelsby Zhihong Shao, Peiyi Wang,…
Nevermind: Instruction Override and Moderation in Large Language Modelsby Edward KimFirst submitted to arxiv on:…
Applying Unsupervised Semantic Segmentation to High-Resolution UAV Imagery for Enhanced Road Scene Parsingby Zihan Ma,…
Evading Data Contamination Detection for Language Models is (too) Easyby Jasper Dekoninck, Mark Niklas Müller,…