Skip to content

[Article] The Token Cost of Beautiful AI: OpenUI Lang vs AI SDK vs JSON#12

Open
manja316 wants to merge 1 commit intothesysdev:mainfrom
manja316:article/token-cost-of-beautiful-ai
Open

[Article] The Token Cost of Beautiful AI: OpenUI Lang vs AI SDK vs JSON#12
manja316 wants to merge 1 commit intothesysdev:mainfrom
manja316:article/token-cost-of-beautiful-ai

Conversation

@manja316
Copy link
Copy Markdown

@manja316 manja316 commented Apr 9, 2026

Closes #4

Summary

Independent benchmark analysis comparing token costs across three generative UI approaches:

  • OpenUI Lang vs Vercel json-render (RFC 6902 patches) vs Thesys C1 JSON (nested component tree)
  • 7 real UI scenarios using OpenUI's own benchmark suite
  • Token counts via tiktoken with gpt-5 encoder
  • Cost projections at 10K and 100K daily users across GPT-4o, Claude 3.5 Sonnet, and GPT-4o-mini

Key findings

  • OpenUI Lang averages 52.8% fewer tokens than json-render and 51.7% fewer than C1 JSON
  • Savings range from 45% (data-heavy dashboards) to 67% (component-dense forms)
  • Annualized savings on GPT-4o at 10K users/day: ~$982/year
  • Balanced tradeoff analysis covering streaming, error recovery, ecosystem, and learning curve

Article structure

  1. The Setup — side-by-side format comparison
  2. The Numbers — full 7-scenario benchmark table
  3. What This Costs at Scale — dollar projections
  4. The Methodology Question — honest assessment of benchmark approach
  5. What the Numbers Don't Tell You — learning curve, streaming, error recovery, ecosystem
  6. When Each Approach Wins — decision framework
  7. Running the Benchmark Yourself — reproducible instructions

Independent benchmark analysis comparing token costs across three generative UI
approaches using OpenUI's own benchmark suite. Includes 7 scenarios, cost
projections at scale, and balanced tradeoff analysis.

Closes thesysdev#4

Co-Authored-By: Paperclip <noreply@paperclip.ing>
@entelligence-ai-pr-reviews
Copy link
Copy Markdown

entelligence-ai-pr-reviews Bot commented Apr 9, 2026

EntelligenceAI PR Summary

Introduces a new technical article analyzing the token cost and tradeoffs of three generative UI rendering formats across seven UI scenarios.

  • Benchmarks token consumption for OpenUI Lang, Vercel json-render/RFC 6902 patches, and Thesys C1 JSON
  • Provides quantitative token counts and cost projections at scale for each format
  • Covers qualitative tradeoffs including streaming behavior, error recovery, ecosystem maturity, and DSL learning curve

Confidence Score: 5/5 - Safe to Merge

Safe to merge — this PR introduces a new technical article benchmarking token costs across OpenUI Lang, Vercel json-render/RFC 6902 patches, and Thesys C1 JSON, and no issues were identified during automated review. The content is purely additive (a new article file) with no runtime logic, security surfaces, or functional code changes that could introduce regressions. The analysis appears well-scoped, covering both quantitative token counts and qualitative tradeoffs across seven UI scenarios.

Key Findings:

  • No review comments were generated and zero issues were flagged across any severity tier, including no critical, significant, or medium-severity findings.
  • The PR is purely content/documentation — it adds a technical article with no code changes, meaning there is no risk of runtime errors, logic bugs, or security vulnerabilities being introduced.
  • Coverage was noted as 0/1 changed files formally reviewed by heuristics, but given the non-code nature of the change (article content), this does not represent a meaningful gap in confidence.
  • The PR achieves a clear educational goal by providing quantitative cost projections and streaming/error-recovery tradeoff analysis, which adds genuine value to the repository's content.

@entelligence-ai-pr-reviews
Copy link
Copy Markdown

Walkthrough

Adds a new technical article benchmarking token consumption and cost across three generative UI formats — OpenUI Lang, Vercel json-render/RFC 6902 patches, and Thesys C1 JSON — over seven UI scenarios. The article presents quantitative token counts, cost projections at scale, and qualitative tradeoff analysis covering streaming behavior, error recovery, ecosystem maturity, and DSL learning curve.

Changes

File(s) Summary
articles/token-cost-of-beautiful-ai.md Adds a new technical article benchmarking token usage and cost for three generative UI formats (OpenUI Lang, Vercel json-render/RFC 6902 patches, Thesys C1 JSON) across seven UI scenarios, including quantitative token counts, at-scale cost projections, and qualitative tradeoff analysis.

Sequence Diagram

This diagram shows the interactions between components:

sequenceDiagram
    participant Dev as Developer
    participant GenCLI as "pnpm generate"
    participant BenchCLI as "pnpm bench"
    participant OpenAI as OpenAI API
    participant AST as AST Parser
    participant Conv as Format Converter
    participant Tiktoken as Tiktoken Counter

    Dev->>GenCLI: run with OPENAI_API_KEY
    loop for each of 7 UI scenarios
        GenCLI->>OpenAI: prompt: generate UI (temp=0)
        OpenAI-->>GenCLI: OpenUI Lang response
        GenCLI->>AST: parse OpenUI Lang into AST
        AST-->>GenCLI: structured AST
        GenCLI->>Conv: convert AST to json-render (RFC 6902 patches)
        Conv-->>GenCLI: json-render sample
        GenCLI->>Conv: convert AST to C1 JSON (nested tree)
        Conv-->>GenCLI: C1 JSON sample
        GenCLI->>Conv: convert AST to YAML
        Conv-->>GenCLI: YAML sample
        GenCLI->>GenCLI: save all 4 format samples to disk
    end

    Dev->>BenchCLI: run token benchmark
    loop for each scenario x each format
        BenchCLI->>Tiktoken: count tokens (gpt-5 encoder)
        Tiktoken-->>BenchCLI: token count
    end
    BenchCLI->>BenchCLI: compute % savings vs json-render and C1 JSON
    BenchCLI-->>Dev: print comparison table

    Note over Dev, Tiktoken: OpenUI Lang saves ~52% tokens vs JSON formats on average
    Note over Conv, Tiktoken: All formats represent identical UI (same AST source)
Loading

🔗 Cross-Repository Impact Analysis

Enable automatic detection of breaking changes across your dependent repositories. → Set up now

Learn more about Cross-Repository Analysis

What It Does

  • Automatically identifies repositories that depend on this code
  • Analyzes potential breaking changes across your entire codebase
  • Provides risk assessment before merging to prevent cross-repo issues

How to Enable

  1. Visit Settings → Code Management
  2. Configure repository dependencies
  3. Future PRs will automatically include cross-repo impact analysis!

Benefits

  • 🛡️ Prevent breaking changes across repositories
  • 🔍 Catch integration issues before they reach production
  • 📊 Better visibility into your multi-repo architecture

@entelligence-ai-pr-reviews
Copy link
Copy Markdown

LGTM 👍 No issues found.


One thing the benchmark README is transparent about: the model generates OpenUI Lang only, then the AST is mechanically converted to the other formats. This is fair for measuring _representation efficiency_ — same information, fewer bytes.

But it sidesteps a real question: would a model prompted to generate JSON directly produce as many tokens as the mechanical conversion? LLMs can be terse or verbose depending on their system prompt. A well-tuned JSON generation prompt might produce leaner output than a literal AST-to-JSON mapping.
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

How can the response be terse/verbose for the same request? The reason we chose to use AST is to ensure the same components are used in all 3 formats.

### Learning curve
OpenUI Lang is a new DSL. JSON is universal. Every developer on your team already reads JSON. OpenUI Lang requires learning positional argument semantics (`FormControl("Label", Input(...))` — which arg is which?), reference-based composition (`nameField = ...` then use `nameField`), and a schema that maps positions to names.

If your team is two people building fast, learning a DSL is 30 minutes. If you're integrating into a 50-person org with existing JSON tooling, it's a migration project.
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

OpenUI lang is not for developers

For production systems, the partial-failure behavior of line-oriented and patch-oriented formats is materially better than monolithic JSON.

### Ecosystem and tooling
json-render (Vercel) has 13K+ stars and renderers for React, Vue, Svelte, Solid, and React Native. OpenUI currently supports React. If you're building for multiple frameworks, json-render's ecosystem is broader today.
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Both projects are framework agnostic. OpenUI has first party support for React but community can bring in their own frameworks if needed.

### Ecosystem and tooling
json-render (Vercel) has 13K+ stars and renderers for React, Vue, Svelte, Solid, and React Native. OpenUI currently supports React. If you're building for multiple frameworks, json-render's ecosystem is broader today.

## When Each Approach Wins
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

please rewrite this section

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Written Content: The Token Cost of Beautiful AI: OpenUI Lang vs. AI SDK vs. JSON: What You're Actually Paying For

2 participants