Tokens Aren't Meaning — They're Compression Hacks

Tokens Aren’t Meaning — They’re Compression Hacks Everyone assumes tokens ≈ words. Wrong. They’re byte substrings glued by frequency, and this fundamental misunderstanding costs companies millions in inference costs and model failures. TL;DR: Your tokenizer doesn’t understand language, it’s just compressing frequent byte sequences. A typo can cost you 33% more tokens. Your Arabic users pay 7x more than English users. And “Be accurate” works better than “Do not hallucinate” for both cost AND quality reasons. ...

August 30, 2025 · 6 min · Lakshay Chhabra

Why Your Model Can't Learn New Concepts (Even with Perfect Data)

Why Your Model Can’t Learn New Concepts (Even with Perfect Data) You just spent months annotating 50,000 examples of your proprietary concept “TurboIN” (your new indexing architecture for Indian markets). Your model still thinks it’s about turbochargers in Indiana. Not a data quality issue. Not a quantity issue. Your model literally cannot learn concepts that don’t exist in its tokenizer embedding space. You’re trying to teach calculus to someone who doesn’t have numbers. ...

August 5, 2025 · 9 min · Lakshay Chhabra

10 Ways Tokenization Screws With Your Model (and Wallet)

Hidden GEMS of Tokenization: The Secrets Nobody Tells You Your model just confused “therapist” with “the rapist” because someone added an invisible Unicode character. Your French bread neurons are firing when processing English medical “pain” terms. Your carefully tuned model got worse at processing currency because fine-tuning on “$AAPL” accidentally shifted what “$” means globally. Welcome to the tokenization secrets that aren’t in any documentation. TL;DR: Beyond the obvious tokenization problems, there’s a shadow world of hidden disasters. Positional encodings break differently for fragmented tokens. Attention heads specialize wrong. Gradients flow differently. Your tokenizer might be fighting invisible Unicode duplicates. These aren’t edge cases, they’re actively destroying your model’s performance right now. ...

August 4, 2025 · 7 min · Lakshay Chhabra