Skip to content

Attention Network

Model description

In the realm of sequence-to-sequence tasks, Attention Networks have garnered attention (pun intended) for their ability to focus on specific parts of the input when generating the output, much like how humans pay attention to specific portions of input when performing tasks like reading or listening. This is particularly useful in time-series prediction tasks like CGM data forecasting, where not all past data points are equally informative for predicting future glucose levels. By weighing the importance of different input data points, Attention Networks aim for more accurate and interpretable predictions.

The best hyperparameters after optimization are the following:

Hyperparameter Value
head_size 448
num_heads 1
ff_dim 480
num_transformer_blocks 6
mlp_units [352, 352]
mlp_dropout 0.0
dropout 0.0

More results

Aside from the scores reported in the overview, we show here the Confusion Matrix and a regression example for our network

The cm The ts The loss