This is a fascinating read on how the attention function (kind of like it's attention span while learning) of an AI works, and the off-by-one bug that may be present in the maths/implementaton that underpins it: https://www.evanmiller.org/attention-is-off-by-one.html