Normalization techniques are crucial for enhancing Transformer models’ performance and stability in time series analysis tasks, yet traditional methods like batch and layer normalization often lead to issues such as token shift, attention shift, and …