Efficient High-Resolution Time Series Classification via Attention Kronecker Decomposition
CoRR(2024)
摘要
The high-resolution time series classification problem is essential due to
the increasing availability of detailed temporal data in various domains. To
tackle this challenge effectively, it is imperative that the state-of-the-art
attention model is scalable to accommodate the growing sequence lengths
typically encountered in high-resolution time series data, while also
demonstrating robustness in handling the inherent noise prevalent in such
datasets. To address this, we propose to hierarchically encode the long time
series into multiple levels based on the interaction ranges. By capturing
relationships at different levels, we can build more robust, expressive, and
efficient models that are capable of capturing both short-term fluctuations and
long-term trends in the data. We then propose a new time series transformer
backbone (KronTime) by introducing Kronecker-decomposed attention to process
such multi-level time series, which sequentially calculates attention from the
lower level to the upper level. Experiments on four long time series datasets
demonstrate superior classification results with improved efficiency compared
to baseline methods.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要