Efficiency Formalization
Why Tokenizer Differences Matter
You'll notice ZON performs differently across models. Here's why:
- GPT-4o (o200k): Highly optimized for code/JSON.
- Result: ZON is neck-and-neck with TSV, but wins on structure.
- Llama 3: Loves explicit tokens.
- Result: ZON wins big (-10.6% vs TOON) because Llama tokenizes ZON's structure very efficiently.
- Claude 3.5: Balanced tokenizer.
- Result: ZON provides consistent savings (-4.4% vs TOON).
Takeaway: ZON is the safest bet for multi-model deployments. It's efficient everywhere, unlike JSON which is inefficient everywhere.
