Testing CLIs
Testing a CLI is harder than testing a library. Libraries are functions — you call them and check the return value. CLIs are processes — they read from argv, env vars, config files, and stdin, then write to stdout, stderr, and the filesystem. Lots of moving parts.
Why It's Hard
The naive way to test a CLI:
# run the command
output=$(mycli greet Alice --loud)
# check the output
[ "$output" = "HELLO, ALICE!" ] || echo "FAIL"This works, technically. But it's slow (spawns a new process each time), fragile (depends on exact output formatting), and limited (how do you test env vars? config files? interactive prompts? error messages?).
Real-world CLIs have tests like:
- "If
--regionis missing butDEPLOY_REGIONis set, use the env var" - "If both flag and config file provide a value, the flag wins"
- "If the prompt is cancelled, exit with code 1"
- "In JSON mode, errors should be structured JSON on stderr"
Good luck doing that with shell scripts.
Two Approaches
1. Subprocess Testing (Black-Box)
Run the actual compiled binary as a child process:
import { execFile } from 'child_process';
const { stdout, stderr, exitCode } = await execFile('./mycli', ['greet', 'Alice']);
expect(stdout).toBe('Hello, Alice!\n');
expect(exitCode).toBe(0);Pros: Tests the real thing. Catches packaging issues. Cons: Slow. Hard to mock env/config. Can't test prompts easily. Platform-dependent.
2. In-Process Testing (White-Box)
Run the command handler as a function, injecting all inputs:
const result = await runCommand(greet, ['Alice', '--loud']);
expect(result.stdout).toEqual(['HELLO, ALICE!\n']);
expect(result.exitCode).toBe(0);Pros: Fast. Full control. Can inject env, config, prompt answers, output capture. Cons: Doesn't test the actual binary entry point.
Most CLI frameworks don't give you option 2. You're stuck shelling out and parsing text. This is a solved problem — the test harness just needs to exist as a first-class feature.
The examples below use dreamcli's test harness, but the patterns apply to any framework that offers in-process testing.
What to Test
Happy Paths
The command works with valid input:
const result = await runCommand(greet, ['Alice']);
expect(result.stdout).toEqual(['Hello, Alice!\n']);
expect(result.exitCode).toBe(0);Flag Resolution
Flags resolve from the right source:
// env var provides the value
const result = await runCommand(cmd, [], {
env: { MY_REGION: 'eu' },
});
expect(result.stdout).toContain('eu');Error Cases
Bad input produces helpful errors:
const result = await runCommand(cmd, ['--unknown']);
expect(result.exitCode).toBe(2);
expect(result.stderr.join('')).toContain('Unknown flag');Missing Required Values
Required flags that aren't provided fail clearly:
const result = await runCommand(cmd, []);
expect(result.exitCode).not.toBe(0);
expect(result.stderr.join('')).toContain('Missing required');JSON Mode
Structured output is valid JSON:
const result = await runCommand(cmd, ['list'], { json: true });
const data = JSON.parse(result.stdout.join(''));
expect(data).toBeInstanceOf(Array);Interactive Prompts
Prompt answers resolve correctly:
const result = await runCommand(cmd, [], {
answers: ['eu', true],
});
expect(result.exitCode).toBe(0);Prompt Cancellation
Ctrl+C during a prompt exits gracefully:
const result = await runCommand(cmd, [], {
answers: [CANCEL],
});
expect(result.exitCode).not.toBe(0);Isolation
Good CLI tests don't touch real state:
- No
process.argvmutation — pass argv as a parameter - No real env vars — inject env as an object
- No real filesystem — inject config as an object
- No real TTY — capture output to arrays
- No real prompts — provide answers programmatically
Each test runs in isolation. No beforeEach cleanup, no shared state, no order dependencies.
What's Next?
- Getting Started — build your first dreamcli command
- Testing Commands guide — dreamcli's in-process test harness