Summary of Local-global Attention: An Adaptive Mechanism For Multi-scale Feature Integration, by Yifan Shao
Local-Global Attention: An Adaptive Mechanism for Multi-Scale Feature Integrationby Yifan ShaoFirst submitted to arxiv on:…
Local-Global Attention: An Adaptive Mechanism for Multi-Scale Feature Integrationby Yifan ShaoFirst submitted to arxiv on:…
Tackling the Abstraction and Reasoning Corpus with Vision Transformers: the Importance of 2D Representation, Positions,…
Dependency Transformer Grammars: Integrating Dependency Structures into Transformer Language Modelsby Yida Zhao, Chao Lou, Kewei…
Learning Image Priors through Patch-based Diffusion Models for Solving Inverse Problemsby Jason Hu, Bowen Song,…
A Structure-Aware Lane Graph Transformer Model for Vehicle Trajectory Predictionby Sun Zhanbo, Dong Caiyin, Ji…
Exploring the Role of Token in Transformer-based Time Series Forecastingby Jianqi Zhang, Jingyao Wang, Chuxiong…
LongHeads: Multi-Head Attention is Secretly a Long Context Processorby Yi Lu, Xin Zhou, Wei He,…
VSFormer: Value and Shape-Aware Transformer with Prior-Enhanced Self-Attention for Multivariate Time Series Classificationby Wenjie Xi,…
PCA-Featured Transformer for Jamming Detection in 5G UAV Networksby Joseanne Viana, Hamed Farkhari, Pedro Sebastiao,…
Hansel: Output Length Controlling Framework for Large Language Modelsby Seoha Song, Junhyun Lee, Hyeonmok KoFirst…