Summary of Wkvquant: Quantizing Weight and Key/value Cache For Large Language Models Gains More, by Yuxuan Yue et al.
WKVQuant: Quantizing Weight and Key/Value Cache for Large Language Models Gains Moreby Yuxuan Yue, Zhihang…
WKVQuant: Quantizing Weight and Key/Value Cache for Large Language Models Gains Moreby Yuxuan Yue, Zhihang…
AICAttack: Adversarial Image Captioning Attack with Attention-Based Optimizationby Jiyao Li, Mingze Ni, Yifei Dong, Tianqing…
What Evidence Do Language Models Find Convincing?by Alexander Wan, Eric Wallace, Dan KleinFirst submitted to…
In-Context Learning with Transformers: Softmax Attention Adapts to Function Lipschitznessby Liam Collins, Advait Parulekar, Aryan…
Measuring and Controlling Instruction (In)Stability in Language Model Dialogsby Kenneth Li, Tianle Liu, Naomi Bashkansky,…
EdgeQAT: Entropy and Distribution Guided Quantization-Aware Training for the Acceleration of Lightweight LLMs on the…
An end-to-end attention-based approach for learning on graphsby David Buterez, Jon Paul Janet, Dino Oglic,…
Can Transformers Predict Vibrations?by Fusataka Kuniyoshi, Yoshihide SawadaFirst submitted to arxiv on: 16 Feb 2024CategoriesMain:…
Graph-based Forecasting with Missing Data through Spatiotemporal Downsamplingby Ivan Marisca, Cesare Alippi, Filippo Maria BianchiFirst…
SAMformer: Unlocking the Potential of Transformers in Time Series Forecasting with Sharpness-Aware Minimization and Channel-Wise…