Comment by fooofw
1 day ago
The tokenization can represent uncommon words with multiple tokens. Inputting your example on https://platform.openai.com/tokenizer (GPT-4o) gives me (tokens separated by "|"):
lower|case|un|se|parated|name
1 day ago
The tokenization can represent uncommon words with multiple tokens. Inputting your example on https://platform.openai.com/tokenizer (GPT-4o) gives me (tokens separated by "|"):
lower|case|un|se|parated|name
No comments yet
Contribute on Hacker News ↗