Skip to content

Commit a2a417e

Browse files
🤖 revert: ORPC migration and related commits (#777)
Reverts: - 470e4eb perf: fix streaming content delay from ORPC schema validation (#774) - b437977 feat: add backend support for soft-interrupts (#767) - df30cbc fix: use ResultSchema for sendMessage output to prevent field stripping (#773) - 41c77ef fix: testUtils formatting (#771) - 3ee7288 refactor: migrate IPC layer to ORPC for type-safe RPC (#763) Due to a huge number of regressions. _Generated with `mux`_
1 parent 470e4eb commit a2a417e

File tree

223 files changed

+13622
-13645
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

223 files changed

+13622
-13645
lines changed

.github/actions/setup-mux/action.yml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -35,3 +35,4 @@ runs:
3535
if: steps.cache-node-modules.outputs.cache-hit != 'true'
3636
shell: bash
3737
run: bun install --frozen-lockfile
38+

.github/workflows/release.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ on:
66
workflow_dispatch:
77
inputs:
88
tag:
9-
description: "Tag to release (e.g., v1.2.3). If provided, will checkout and release this tag regardless of current branch."
9+
description: 'Tag to release (e.g., v1.2.3). If provided, will checkout and release this tag regardless of current branch.'
1010
required: false
1111
type: string
1212

.github/workflows/terminal-bench.yml

Lines changed: 19 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -4,34 +4,34 @@ on:
44
workflow_call:
55
inputs:
66
model_name:
7-
description: "Model to use (e.g., anthropic:claude-sonnet-4-5)"
7+
description: 'Model to use (e.g., anthropic:claude-sonnet-4-5)'
88
required: false
99
type: string
1010
thinking_level:
11-
description: "Thinking level (off, low, medium, high)"
11+
description: 'Thinking level (off, low, medium, high)'
1212
required: false
1313
type: string
1414
dataset:
15-
description: "Terminal-Bench dataset to use"
15+
description: 'Terminal-Bench dataset to use'
1616
required: false
1717
type: string
18-
default: "terminal-bench-core==0.1.1"
18+
default: 'terminal-bench-core==0.1.1'
1919
concurrency:
20-
description: "Number of concurrent tasks (--n-concurrent)"
20+
description: 'Number of concurrent tasks (--n-concurrent)'
2121
required: false
2222
type: string
23-
default: "4"
23+
default: '4'
2424
livestream:
25-
description: "Enable livestream mode (verbose output to console)"
25+
description: 'Enable livestream mode (verbose output to console)'
2626
required: false
2727
type: boolean
2828
default: false
2929
sample_size:
30-
description: "Number of random tasks to run (empty = all tasks)"
30+
description: 'Number of random tasks to run (empty = all tasks)'
3131
required: false
3232
type: string
3333
extra_args:
34-
description: "Additional arguments to pass to terminal-bench"
34+
description: 'Additional arguments to pass to terminal-bench'
3535
required: false
3636
type: string
3737
secrets:
@@ -42,34 +42,34 @@ on:
4242
workflow_dispatch:
4343
inputs:
4444
dataset:
45-
description: "Terminal-Bench dataset to use"
45+
description: 'Terminal-Bench dataset to use'
4646
required: false
47-
default: "terminal-bench-core==0.1.1"
47+
default: 'terminal-bench-core==0.1.1'
4848
type: string
4949
concurrency:
50-
description: "Number of concurrent tasks (--n-concurrent)"
50+
description: 'Number of concurrent tasks (--n-concurrent)'
5151
required: false
52-
default: "4"
52+
default: '4'
5353
type: string
5454
livestream:
55-
description: "Enable livestream mode (verbose output to console)"
55+
description: 'Enable livestream mode (verbose output to console)'
5656
required: false
5757
default: false
5858
type: boolean
5959
sample_size:
60-
description: "Number of random tasks to run (empty = all tasks)"
60+
description: 'Number of random tasks to run (empty = all tasks)'
6161
required: false
6262
type: string
6363
model_name:
64-
description: "Model to use (e.g., anthropic:claude-sonnet-4-5, openai:gpt-5.1-codex)"
64+
description: 'Model to use (e.g., anthropic:claude-sonnet-4-5, openai:gpt-5.1-codex)'
6565
required: false
6666
type: string
6767
thinking_level:
68-
description: "Thinking level (off, low, medium, high)"
68+
description: 'Thinking level (off, low, medium, high)'
6969
required: false
7070
type: string
7171
extra_args:
72-
description: "Additional arguments to pass to terminal-bench"
72+
description: 'Additional arguments to pass to terminal-bench'
7373
required: false
7474
type: string
7575

@@ -147,3 +147,4 @@ jobs:
147147
benchmark.log
148148
if-no-files-found: warn
149149
retention-days: 30
150+

.storybook/mocks/orpc.ts

Lines changed: 0 additions & 217 deletions
This file was deleted.

.storybook/preview.tsx

Lines changed: 1 addition & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,6 @@
1-
import React, { useMemo } from "react";
1+
import React from "react";
22
import type { Preview } from "@storybook/react-vite";
33
import { ThemeProvider, type ThemeMode } from "../src/browser/contexts/ThemeContext";
4-
import { ORPCProvider } from "../src/browser/orpc/react";
5-
import { createMockORPCClient } from "./mocks/orpc";
64
import "../src/browser/styles/globals.css";
75

86
const preview: Preview = {
@@ -24,16 +22,6 @@ const preview: Preview = {
2422
theme: "dark",
2523
},
2624
decorators: [
27-
// Global ORPC provider - ensures useORPC works in all stories
28-
(Story) => {
29-
const client = useMemo(() => createMockORPCClient(), []);
30-
return (
31-
<ORPCProvider client={client}>
32-
<Story />
33-
</ORPCProvider>
34-
);
35-
},
36-
// Theme provider
3725
(Story, context) => {
3826
// Default to dark if mode not set (e.g., Chromatic headless browser defaults to light)
3927
const mode = (context.globals.theme as ThemeMode | undefined) ?? "dark";

babel.config.js

Lines changed: 0 additions & 19 deletions
This file was deleted.

0 commit comments

Comments
 (0)