Fix Flaky Tests
Quick Start
You'll typically receive a Tuist test case URL or identifier. Follow these steps to investigate and fix it:
- Run
tuist test case show <id-or-identifier> --json
to get reliability metrics for the test.
- Run
tuist test case run list Module/Suite/TestCase --flaky --json
to see flaky run patterns.
- Run
tuist test case run show <run-id> --json
on failing flaky runs to get failure messages and file paths.
- Read the test source at the reported path and line, identify the flaky pattern, and fix it.
- Verify by running the test multiple times to confirm it passes consistently.
If no specific test is provided, start with the Discovery section below.
Discovery
When no specific test case is provided, find all flaky tests in the project:
bash
tuist test case list --flaky --json --page-size 50
This returns all test cases currently flagged as flaky. Key fields:
- / / — the test identifier
- — helps prioritize (fix fast unit tests first)
- — whether the test is already quarantined
Triage strategy:
- Group tests by suite — multiple flaky tests in the same suite often share a root cause.
- Check if failures share a — tests that all failed in the same run may have been killed by a process crash, not individual test bugs.
- Look at failure messages to categorize: test logic bugs vs infrastructure issues (network errors, server 502s, conflicts on retry).
Investigation
1. Get test case metrics
You can pass either the UUID or the
identifier:
bash
tuist test case show <id> --json
tuist test case show Module/Suite/TestCase --json
Key fields:
- — percentage of successful runs (higher is better)
- — percentage of runs marked flaky in the last 30 days
- / — volume context
- — current state
2. View flaky run history
bash
tuist test case run list Module/Suite/TestCase --flaky --json
The identifier uses the format
ModuleName/SuiteName/TestCaseName
or
when there is no suite. This returns only runs that were detected as flaky.
3. View full run history
bash
tuist test case run list Module/Suite/TestCase --json --page-size 20
Look for patterns:
- Does it fail on specific branches?
- Does it fail only on CI () or also locally?
- Are failures clustered around specific commits?
4. Get failure details
bash
tuist test case run show <run-id> --json
Key fields:
- — the assertion or error message
- — source file path
- — exact line of failure
- — type of issue (assertion_failure, etc.)
- — if present, shows retry behavior (pass/fail sequence)
- — the broader test run this execution belongs to
Code Analysis
- Open the file at and go to .
- Read the full test function and its setup/teardown.
- Identify which of the common flaky patterns below applies.
- Check if the test shares state with other tests in the same suite.
Common Flaky Patterns
Timing and async issues
- Missing waits: Test checks a result before an async operation completes. Fix: use , expectations with timeouts, or polling.
- Race conditions: Multiple concurrent operations access shared state. Fix: synchronize access or use serial queues.
- Hardcoded timeouts: or fixed delays that are too short on CI. Fix: use condition-based waits instead of fixed delays.
Shared state
- Test pollution: One test modifies global/static state that another test depends on. Fix: reset state in setUp/tearDown or use unique instances per test.
- Singleton contamination: Shared singletons carry state between tests. Fix: inject dependencies or reset singletons.
- File system leftovers: Tests leave files that affect subsequent runs. Fix: use temporary directories and clean up.
Environment dependencies
- Network calls: Tests hit real services that may be slow or unavailable. Fix: mock network calls.
- Date/time sensitivity: Tests depend on current time or timezone. Fix: inject a clock or freeze time.
- File system paths: Hardcoded paths that differ between environments. Fix: use relative paths or temp directories.
Order dependence
- Implicit ordering: Test passes only when run after another test that sets up required state. Fix: make each test self-contained.
- Parallel execution conflicts: Tests that work in isolation but fail when run concurrently. Fix: use unique resources per test.
Fix Implementation
After identifying the pattern:
- Apply the smallest fix that addresses the root cause.
- Do not refactor unrelated code.
- If the fix requires a test utility (like a mock or helper), check if one already exists before creating a new one.
Verification
Running tests repeatedly
Run the specific test repeatedly until failure using
's built-in repetition support:
bash
xcodebuild test -workspace <workspace> -scheme <scheme> -only-testing <module>/<suite>/<test> -test-iterations <count> -run-tests-until-failure
This runs the test up to
times and stops at the first failure. Choose the iteration count based on how long the test takes — for fast unit tests use 50–100, for slower integration or acceptance tests use 2–5.
Reproducing before fixing
Before applying a fix, try to reproduce the flaky failure locally. A successful reproduction confirms your root cause analysis and lets you verify the fix directly. Use the "Running tests repeatedly" approach above, or the race condition strategies below if concurrency is suspected.
Some flaky scenarios — especially race conditions, CI-specific timing issues, or environment-dependent failures — may be difficult or impossible to reproduce locally. If you cannot reproduce after reasonable effort, proceed with fixing based on code analysis and failure logs. A fix backed by clear evidence of a bug (e.g. unsynchronized shared state, TOCTOU pattern) is valid even without local reproduction.
Reproducing race conditions
Race conditions and concurrency bugs often only manifest under CI-level parallelism and are hard to reproduce locally. Try these strategies in order:
- Increase parallelism: Add
-parallel-testing-enabled YES
to run test suites concurrently.
- Run broader test suites: Instead of running a single test, run the entire module (e.g.
-only-testing ModuleTests
) to increase contention on shared resources.
- Thread Sanitizer: Run with TSan enabled to detect data races deterministically. Note: TSan adds overhead which can change timing, so some races may not trigger under TSan.
bash
xcodebuild test -workspace <workspace> -scheme <scheme> -only-testing <module> -enableThreadSanitizer YES
- High iteration count with broad scope: Combine all the above — run the full module with parallelism and many iterations.
If a race condition cannot be reproduced locally but the code is provably thread-unsafe (e.g. unsynchronized mutation of shared state), the fix is still valid. Verify the fix by confirming the tests pass with the same reproduction strategies above. Document in the commit message that the fix addresses a CI-only race condition identified through code analysis and failure logs.
Done Checklist
- Identified the root cause of flakiness
- Applied a targeted fix
- Verified the test passes consistently (multiple runs)
- Did not introduce new test dependencies or shared state
- Committed the fix with a descriptive message