Summary of Q-value Regularized Transformer For Offline Reinforcement Learning, by Shengchao Hu et al.
Q-value Regularized Transformer for Offline Reinforcement Learningby Shengchao Hu, Ziqing Fan, Chaoqin Huang, Li Shen,…
Q-value Regularized Transformer for Offline Reinforcement Learningby Shengchao Hu, Ziqing Fan, Chaoqin Huang, Li Shen,…
Demystifying amortized causal discovery with transformersby Francesco Montagna, Max Cairney-Leeming, Dhanya Sridhar, Francesco LocatelloFirst submitted…
Automatic Domain Adaptation by Transformers in In-Context Learningby Ryuichiro Hataya, Kota Matsui, Masaaki ImaizumiFirst submitted…
On Mesa-Optimization in Autoregressively Trained Transformers: Emergence and Capabilityby Chenyu Zheng, Wei Huang, Rongzhen Wang,…
Are Self-Attentions Effective for Time Series Forecasting?by Dongbin Kim, Jinseong Park, Jaewook Lee, Hoki KimFirst…
Amortized Active Causal Induction with Deep Reinforcement Learningby Yashas Annadani, Panagiotis Tigas, Stefan Bauer, Adam…
Disentangling and Integrating Relational and Sensory Information in Transformer Architecturesby Awni Altabaa, John LaffertyFirst submitted…
Understanding Linear Probing then Fine-tuning Language Models from NTK Perspectiveby Akiyoshi Tomihari, Issei SatoFirst submitted…
Acceleration of Grokking in Learning Arithmetic Operations via Kolmogorov-Arnold Representationby Yeachan Park, Minseok Kim, Yeoneung…
Zamba: A Compact 7B SSM Hybrid Modelby Paolo Glorioso, Quentin Anthony, Yury Tokpanov, James Whittington,…