Summary of Task-adaptive Pretrained Language Models Via Clustered-importance Sampling, by David Grangier et al.
Task-Adaptive Pretrained Language Models via Clustered-Importance Samplingby David Grangier, Simin Fan, Skyler Seto, Pierre AblinFirst…
Task-Adaptive Pretrained Language Models via Clustered-Importance Samplingby David Grangier, Simin Fan, Skyler Seto, Pierre AblinFirst…
No Need to Talk: Asynchronous Mixture of Language Modelsby Anastasiia Filippova, Angelos Katharopoulos, David Grangier,…
Investigating on RLHF methodologyby Alexey Kutalev, Sergei MarkoffFirst submitted to arxiv on: 2 Oct 2024CategoriesMain:…
AutoTM 2.0: Automatic Topic Modeling Framework for Documents Analysisby Maria Khodorchenko, Nikolay Butakov, Maxim Zuev,…
Rotated Runtime Smooth: Training-Free Activation Smoother for accurate INT4 inferenceby Ke Yi, Zengke Liu, Jianwei…
MaskLLM: Learnable Semi-Structured Sparsity for Large Language Modelsby Gongfan Fang, Hongxu Yin, Saurav Muralidharan, Greg…
Context-Aware Membership Inference Attacks against Pre-trained Large Language Modelsby Hongyan Chang, Ali Shahin Shamsabadi, Kleomenis…
A Controlled Study on Long Context Extension and Generalization in LLMsby Yi Lu, Jing Nathan…
Understanding Knowledge Drift in LLMs through Misinformationby Alina Fastowski, Gjergji KasneciFirst submitted to arxiv on:…
Improving Pretraining Data Using Perplexity Correlationsby Tristan Thrush, Christopher Potts, Tatsunori HashimotoFirst submitted to arxiv…