Summary of Layer-wise Regularized Dropout For Neural Language Models, by Shiwen Ni et al.
Layer-wise Regularized Dropout for Neural Language Modelsby Shiwen Ni, Min Yang, Ruifeng Xu, Chengming Li,…
Layer-wise Regularized Dropout for Neural Language Modelsby Shiwen Ni, Min Yang, Ruifeng Xu, Chengming Li,…
Mitigating the Linguistic Gap with Phonemic Representations for Robust Cross-lingual Transferby Haeji Jung, Changdae Oh,…
An Empirical Study on Cross-lingual Vocabulary Adaptation for Efficient Language Model Inferenceby Atsuki Yamaguchi, Aline…
Evaluating and Improving Continual Learning in Spoken Language Understandingby Muqiao Yang, Xiang Li, Umberto Cappellazzo,…
Large Language Models: A Surveyby Shervin Minaee, Tomas Mikolov, Narjes Nikzad, Meysam Chenaghlu, Richard Socher,…
LB-KBQA: Large-language-model and BERT based Knowledge-Based Question and Answering Systemby Yan Zhao, Zhongyun Li, Yushan…
“What’s my model inside of?”: Exploring the role of environments for grounded natural language understandingby…
Paramanu: A Family of Novel Efficient Generative Foundation Language Models for Indian Languagesby Mitodru Niyogi,…
BPDec: Unveiling the Potential of Masked Language Modeling Decoder in BERT pretrainingby Wen Liang, Youzhi…
UniMS-RAG: A Unified Multi-source Retrieval-Augmented Generation for Personalized Dialogue Systemsby Hongru Wang, Wenyu Huang, Yang…