Developer Teaches Claude Caveman Speech to Cut Tokens
Alexander Huso taught Claude to speak like a caveman to reduce output tokens, Business Insider reports. Huso told Business Insider the experiment saved some tokens but produced poor-quality output; he is quoted saying, "I wouldn't trust it to write any serious code." After trying the prompt hack, Huso posted the result on Reddit and said he learned a lesson about virality, according to the as-told-to essay. The anecdote illustrates a practical tradeoff between aggressive token-saving prompt styles and usable model output when working with paid LLM plans.
What happened
Alexander Huso experimented with getting Claude to respond in a terse "caveman" style to reduce output token usage, Business Insider reports. Business Insider quotes Huso saying the compressed style saved some tokens but produced poor output quality: "I wouldn't trust it to write any serious code." Business Insider also reports Huso posted the experiment on Reddit and described learning a lesson about how content can go viral.
Editorial analysis - technical context
Industry-pattern observations: Prompt engineers and power users routinely trade verbosity for token cost. For transformer-based models, every additional output token increases billed usage, so users attempt stylistic constraints or token-compression tactics to lower cost. Those tactics commonly reduce semantic detail, error-correcting phrasing, and explicitness that models use to produce reliable code or precise instructions. In practice, forcing a model into an unnatural register (for example, extreme terseness or deliberate grammatical degradation) reduces the model's effective signal for tasks that require structure, such as coding, debugging, or technical writing.
Context and significance
Editorial analysis: The Business Insider anecdote is a concrete example of a broader tradeoff in paid LLM usage: token-economics versus output quality. For practitioners, the story underscores that micro-optimizations to reduce billing can materially harm downstream utility, particularly for correctness-sensitive outputs like code. It also shows how playful or experimental prompts can attract attention on social platforms even when the technical payoff is limited.
What to watch
For practitioners: monitor these indicators when evaluating token-saving tactics:
- •changes in functional correctness or test-passing rates for model-generated code
- •token-cost savings versus time spent prompting and validating outputs
- •community-shared prompt patterns and whether they generalize beyond novelty
Business Insider is the source for the reported experiment and quotes; Huso provided the as-told-to account to Business Insider.
Scoring Rationale
Anecdotal but relevant: the story highlights a common practitioner tradeoff between token costs and output quality when using paid LLM plans. It is useful for prompt engineers but not a frontier technical advance.
Practice interview problems based on real data
1,500+ SQL & Python problems across 15 industry datasets — the exact type of data you work with.
Try 250 free problems