Summary of Hierarchical Attention Models For Multi-relational Graphs, by Roshni G. Iyer et al.
Hierarchical Attention Models for Multi-Relational Graphsby Roshni G. Iyer, Wei Wang, Yizhou SunFirst submitted to…
Hierarchical Attention Models for Multi-Relational Graphsby Roshni G. Iyer, Wei Wang, Yizhou SunFirst submitted to…
Wasserstein Wormhole: Scalable Optimal Transport Distance with Transformersby Doron Haviv, Russell Zhang Kunes, Thomas Dougherty,…
RF-Diffusion: Radio Signal Generation via Time-Frequency Diffusionby Guoxuan Chi, Zheng Yang, Chenshu Wu, Jingao Xu,…
TransformerFAM: Feedback attention is working memoryby Dongseong Hwang, Weiran Wang, Zhuoyuan Huo, Khe Chai Sim,…
Foundational GPT Model for MEGby Richard Csaky, Mats W.J. van Es, Oiwi Parker Jones, Mark…
Megalodon: Efficient LLM Pretraining and Inference with Unlimited Context Lengthby Xuezhe Ma, Xiaomeng Yang, Wenhan…
The Illusion of State in State-Space Modelsby William Merrill, Jackson Petty, Ashish SabharwalFirst submitted to…
Inheritune: Training Smaller Yet More Attentive Language Modelsby Sunny Sanyal, Ravid Shwartz-Ziv, Alexandros G. Dimakis,…
Transformer-based Joint Modelling for Automatic Essay Scoring and Off-Topic Detectionby Sourya Dipta Das, Yash Vadi,…
Revealing Trends in Datasets from the 2022 ACL and EMNLP Conferencesby Jesse Atuhurra, Hidetaka KamigaitoFirst…