, Keuntae Park), and KAIST (professors Kimin Lee
and Seungwon Shin)!
You can find our paper here: arxiv.org/abs/2410.23684 (11/11)
, Keuntae Park), and KAIST (professors Kimin Lee
and Seungwon Shin)!
You can find our paper here: arxiv.org/abs/2410.23684 (11/11)
We think so! When we tokenize the same phrase differently to *avoid* incomplete tokens, the models generally performed much better (including a 93% reduction in Llama3.1). (7/11)
We think so! When we tokenize the same phrase differently to *avoid* incomplete tokens, the models generally performed much better (including a 93% reduction in Llama3.1). (7/11)
Improbable bigrams were significantly higher to hallucinations.
(For this, we only used trained tokens to remove influence of glitch tokens.) (6/11)
Improbable bigrams were significantly higher to hallucinations.
(For this, we only used trained tokens to remove influence of glitch tokens.) (6/11)
A target phrase is considered hallucinatory only if the model fails to repeat the phrase in all 3 prompts. (5/11)
A target phrase is considered hallucinatory only if the model fails to repeat the phrase in all 3 prompts. (5/11)
If the pair is re-encodable to the incomplete tokens, it is a legal incomplete bigram. (4/11)
If the pair is re-encodable to the incomplete tokens, it is a legal incomplete bigram. (4/11)
Such tokens with stray bytes rely on adjacent tokens' stray bytes to resolve as a character.
If two such tokens combine into an "improbable bigram" like ट能, we get a phrase that causes model errors. (3/11)
Such tokens with stray bytes rely on adjacent tokens' stray bytes to resolve as a character.
If two such tokens combine into an "improbable bigram" like ट能, we get a phrase that causes model errors. (3/11)
Have you ever wondered what "ट能" means?
Probably not, since it's not a meaningful phrase.
But if you ever did, any well-trained LLM should be able to tell you that. Right?
Not quite! We discover phrases like "ट能" trigger vulnerabilities in Byte-Level BPE Tokenizers. (1/11)
Have you ever wondered what "ट能" means?
Probably not, since it's not a meaningful phrase.
But if you ever did, any well-trained LLM should be able to tell you that. Right?
Not quite! We discover phrases like "ट能" trigger vulnerabilities in Byte-Level BPE Tokenizers. (1/11)