That’s because they don’t see the letters, but tokens instead. A token can be one letter, but is usually bigger. So what the llm sees might be something like
st
raw
be
r
r
y
When seeing it like that it’s more obvious why the llm’s are struggling with it
I think the strawberry problem is to ask it how many R’s are in strawberry. Current AI gets it wrong almost every time.
That’s because they don’t see the letters, but tokens instead. A token can be one letter, but is usually bigger. So what the llm sees might be something like
When seeing it like that it’s more obvious why the llm’s are struggling with it