Summary of State Space Model For New-generation Network Alternative to Transformers: a Survey, by Xiao Wang et al.
State Space Model for New-Generation Network Alternative to Transformers: A Surveyby Xiao Wang, Shiao Wang,…
State Space Model for New-Generation Network Alternative to Transformers: A Surveyby Xiao Wang, Shiao Wang,…
Hierarchical Attention Models for Multi-Relational Graphsby Roshni G. Iyer, Wei Wang, Yizhou SunFirst submitted to…
TransformerFAM: Feedback attention is working memoryby Dongseong Hwang, Weiran Wang, Zhuoyuan Huo, Khe Chai Sim,…
`Eyes of a Hawk and Ears of a Fox’: Part Prototype Network for Generalized Zero-Shot…
Megalodon: Efficient LLM Pretraining and Inference with Unlimited Context Lengthby Xuezhe Ma, Xiaomeng Yang, Wenhan…
BERT-LSH: Reducing Absolute Compute For Attentionby Zezheng Li, Kingston YipFirst submitted to arxiv on: 12…
Inheritune: Training Smaller Yet More Attentive Language Modelsby Sunny Sanyal, Ravid Shwartz-Ziv, Alexandros G. Dimakis,…
Neural Sequence-to-Sequence Modeling with Attention by Leveraging Deep Learning Architectures for Enhanced Contextual Understanding in…
VeTraSS: Vehicle Trajectory Similarity Search Through Graph Modeling and Representation Learningby Ming Cheng, Bowen Zhang,…
DisorderUnetLM: Validating ProteinUnet for efficient protein intrinsic disorder predictionby Krzysztof Kotowski, Irena Roterman, Katarzyna StaporFirst…