Decentralized Attention Fails Centralized Signals: Rethinking Transformers for Medical Time Series

This research introduces CoTAR, a novel Transformer architecture that replaces decentralized attention with a centralized mechanism to solve synchronization ...

Level: advanced

By Guoqi Yu, Juncheng Wang, Chen Yang, Jing Qin, Angelica I. Aviles-Rivero, Shujun Wang

Category: research