Summary of Core Tokensets For Data-efficient Sequential Training Of Transformers, by Subarnaduti Paul et al.
Core Tokensets for Data-efficient Sequential Training of Transformersby Subarnaduti Paul, Manuel Brack, Patrick Schramowski, Kristian…
Core Tokensets for Data-efficient Sequential Training of Transformersby Subarnaduti Paul, Manuel Brack, Patrick Schramowski, Kristian…
TextHawk2: A Large Vision-Language Model Excels in Bilingual OCR and Grounding with 16x Fewer Tokensby…
PAD: Personalized Alignment of LLMs at Decoding-Timeby Ruizhe Chen, Xiaotian Zhang, Meng Luo, Wenhao Chai,…
Correlation-Aware Select and Merge Attention for Efficient Fine-Tuning and Context Length Extensionby Ning Wang, Zekun…
LANTERN: Accelerating Visual Autoregressive Models with Relaxed Speculative Decodingby Doohyuk Jang, Sihwan Park, June Yong…
Unveiling Language Skills via Path-Level Circuit Discoveryby Hang Chen, Jiaying Zhu, Xinyu Yang, Wenya WangFirst…
1 Trillion Token (1TT) Platform: A Novel Framework for Efficient Data Sharing and Compensation in…
Contrastive Token Learning with Similarity Decay for Repetition Suppression in Machine Translationby Huangyu Dai, Ben…
See then Tell: Enhancing Key Information Extraction with Vision Groundingby Shuhang Liu, Zhenrong Zhang, Pengfei…
Inference-Time Language Model Alignment via Integrated Value Guidanceby Zhixuan Liu, Zhanhui Zhou, Yuanfu Wang, Chao…