Eduards Ruzga explores how the AI community debates local LLMs versus cloud APIs versus subscriptions – but almost entirely on vibes. He attempted to address this by building an open-source tool that calculates quality-adjusted tokens per dollar across all three options.
He shares surprising findings: running GLM-4.7 Flash locally on a $3,500 MacBook delivers 22×worse value than calling the same model via API. ChatGPT Pro provides 6.4× more tokens than Claude Max at the same price point.
The methodology – combining benchmark scores, hardware costs, subscription limits, and API pricing into one comparable metric – is open for scrutiny and contribution. Attendees leave with a live tool, a framework for evaluating AI cost-effectiveness, and concrete data to replace opinions with math.
Tool and source: https://wonderwhy-er.github.io/llm-value-comparison/
Full analysis: https://wonderwhy-er.medium.com/51fc0ad0dbd7

