Summary of Serverlessllm: Low-latency Serverless Inference For Large Language Models, by Yao Fu et al.
ServerlessLLM: Low-Latency Serverless Inference for Large Language Modelsby Yao Fu, Leyang Xue, Yeqi Huang, Andrei-Octavian…
ServerlessLLM: Low-Latency Serverless Inference for Large Language Modelsby Yao Fu, Leyang Xue, Yeqi Huang, Andrei-Octavian…
True Knowledge Comes from Practice: Aligning LLMs with Embodied Environments via Reinforcement Learningby Weihao Tan,…
How Can Large Language Models Understand Spatial-Temporal Data?by Lei Liu, Shuo Yu, Runze Wang, Zhenxun…
Assessing the Portability of Parameter Matrices Trained by Parameter-Efficient Finetuning Methodsby Mohammed Sabry, Anya BelzFirst…
Towards 3D Molecule-Text Interpretation in Language Modelsby Sihang Li, Zhiyuan Liu, Yanchen Luo, Xiang Wang,…
Knowledge Distillation from Language-Oriented to Emergent Communication for Multi-Agent Remote Controlby Yongjun Kim, Sejin Seo,…
Freely Long-Thinking Transformer (FraiLT)by Akbay TabakFirst submitted to arxiv on: 21 Jan 2024CategoriesMain: Machine Learning…
MolTailor: Tailoring Chemical Molecular Representation to Specific Tasks via Text Promptsby Haoqiang Guo, Sendong Zhao,…
Medusa: Simple LLM Inference Acceleration Framework with Multiple Decoding Headsby Tianle Cai, Yuhong Li, Zhengyang…
Critical Data Size of Language Models from a Grokking Perspectiveby Xuekai Zhu, Yao Fu, Bowen…