Summary of Learning with Less: Knowledge Distillation From Large Language Models Via Unlabeled Data, by Juanhui Li et al.
Learning with Less: Knowledge Distillation from Large Language Models via Unlabeled Databy Juanhui Li, Sreyashi…
Learning with Less: Knowledge Distillation from Large Language Models via Unlabeled Databy Juanhui Li, Sreyashi…
Chain Association-based Attacking and Shielding Natural Language Processing Systemsby Jiacheng Huang, Long ChenFirst submitted to…
Fair Summarization: Bridging Quality and Diversity in Extractive Summariesby Sina Bagheri Nezhad, Sayan Bandyapadhyay, Ameeta…
Conditional [MASK] Discrete Diffusion Language Modelby Hyukhun Koh, Minha Jhang, Dohyung Kim, Sangmook Lee, Kyomin…
Qwen2.5-32B: Leveraging Self-Consistent Tool-Integrated Reasoning for Bengali Mathematical Olympiad Problem Solvingby Saad Tahmid, Sourav SarkerFirst…
AI Multi-Agent Interoperability Extension for Managing Multiparty Conversationsby Diego Gosmar, Deborah A. Dahl, Emmett Coin,…
GPTKB: Comprehensively Materializing Factual LLM Knowledgeby Yujia Hu, Tuan-Phong Nguyen, Shrestha Ghosh, Simon RazniewskiFirst submitted…
M3SciQA: A Multi-Modal Multi-Document Scientific QA Benchmark for Evaluating Foundation Modelsby Chuhan Li, Ziyao Shangguan,…
Exploring the Benefits of Domain-Pretraining of Generative Large Language Models for Chemistryby Anurag Acharya, Shivam…
Larger models yield better results? Streamlined severity classification of ADHD-related concerns using BERT-based knowledge distillationby…