All you need to know about the state of the art Transformer Neural Network Architecture, adapted to Time Series Tasks. Keras code included.
Table of Contents
- Learnable Time Representation (Time 2 Vec)
- Bag Of Tricks (things to consider when training Transformers)
Attention Is All You Need they said. Is it a more robust convolution? Is it just a hack to squeeze more learning capacity out of fewer parameters? Is it supposed to be…