Summary of Adam-mini: Use Fewer Learning Rates to Gain More, by Yushun Zhang et al.
Adam-mini: Use Fewer Learning Rates To Gain Moreby Yushun Zhang, Congliang Chen, Ziniu Li, Tian…
Adam-mini: Use Fewer Learning Rates To Gain Moreby Yushun Zhang, Congliang Chen, Ziniu Li, Tian…
Understanding and Mitigating Tokenization Bias in Language Modelsby Buu Phan, Marton Havasi, Matthew Muckley, Karen…
Segment Any Text: A Universal Approach for Robust, Efficient and Adaptable Sentence Segmentationby Markus Frohmann,…
Anomaly Detection of Tabular Data Using LLMsby Aodong Li, Yunhan Zhao, Chen Qiu, Marius Kloft,…
Does Cross-Cultural Alignment Change the Commonsense Morality of Language Models?by Yuu JinnaiFirst submitted to arxiv…
Memory-Enhanced Neural Solvers for Efficient Adaptation in Combinatorial Optimizationby Felix Chalumeau, Refiloe Shabe, Noah De…
Reducing Fine-Tuning Memory Overhead by Approximate and Memory-Sharing Backpropagationby Yuchen Yang, Yingdong Shi, Cheems Wang,…
Towards Scalable Exact Machine Unlearning Using Parameter-Efficient Fine-Tuningby Somnath Basu Roy Chowdhury, Krzysztof Choromanski, Arijit…
Crosslingual Capabilities and Knowledge Barriers in Multilingual Large Language Modelsby Lynn Chua, Badih Ghazi, Yangsibo…
RuleR: Improving LLM Controllability by Rule-based Data Recyclingby Ming Li, Han Chen, Chenguang Wang, Dang…