Summary of Teaching Mlp More Graph Information: a Three-stage Multitask Knowledge Distillation Framework, by Junxian Li et al.
Teaching MLP More Graph Information: A Three-stage Multitask Knowledge Distillation Frameworkby Junxian Li, Bin Shi,…
Teaching MLP More Graph Information: A Three-stage Multitask Knowledge Distillation Frameworkby Junxian Li, Bin Shi,…
Comparing Graph Transformers via Positional Encodingsby Mitchell Black, Zhengchao Wan, Gal Mishne, Amir Nayyeri, Yusu…
From Self-Attention to Markov Models: Unveiling the Dynamics of Generative Transformersby M. Emrullah Ildiz, Yixiao…
Polyhedral Complex Derivation from Piecewise Trilinear Networksby Jin-Hwa KimFirst submitted to arxiv on: 16 Feb…
Subgraphormer: Unifying Subgraph GNNs and Graph Transformers via Graph Productsby Guy Bar-Shalom, Beatrice Bevilacqua, Haggai…
Todyformer: Towards Holistic Dynamic Graph Transformers with Structure-Aware Tokenizationby Mahdi Biparva, Raika Karimi, Faezeh Faez,…
How do Transformers perform In-Context Autoregressive Learning?by Michael E. Sander, Raja Giryes, Taiji Suzuki, Mathieu…
XTSFormer: Cross-Temporal-Scale Transformer for Irregular-Time Event Prediction in Clinical Applicationsby Tingsong Xiao, Zelin Xu, Wenchong…
Beyond the Limits: A Survey of Techniques to Extend the Context Length in Large Language…
Theoretical Understanding of In-Context Learning in Shallow Transformers with Unstructured Databy Yue Xing, Xiaofeng Lin,…