| name | cts-triage |
| description | Run CTS test suites and investigate failures |
Triage Process
When working on a category of CTS tests, follow this systematic process to identify issues, prioritize fixes, and document findings.
Step 0: Divide Into Manageable Chunks
List all the tests matching a selector:
cargo xtask cts -- --list 'webgpu:api,validation,*' 2>&1 | wc -l
If there is a reasonable number of tests matching the selector (less than a couple hundred or so, but this isn't a hard cutoff), then you can proceed with triage. Otherwise, make a list of more detailed wildcards that match fewer tests, verify that each wildcard matches a reasonable number of tests, and triage each wildcard separately.
Step 1: Get Overall Statistics
Run the full test suite for the category to understand the scope:
cargo xtask cts 'webgpu:api,validation,category:*' 2>&1 | grep -E "(Summary|Passed|Failed)" | tail -5
This gives you the pass rate and number of failures. Document this as your baseline.
Step 2: Identify Test Subcategories
Review the output from the running the CTS (with or without --list) to
identify any subcategories that may exist within the suite being analyzed.
Subcategories typically have an additional :- or ,-delimited word in the
test name. Running tests by subcategory may be more manageable than running
with the entire suite at once or running individual tests.
Step 3: Run Each Subcategory
Test each subcategory individually to identify which ones are passing vs failing:
cargo xtask cts 'webgpu:api,validation,category:subcategory:*' 2>&1 | grep -E "(Passed|Failed|Skipped)" | tail -3
Track the pass rate for each subcategory. This helps you identify:
- What's already working (don't break it!)
- Where the failures are concentrated
- Which issues affect multiple categories
Step 4: Analyze Failure Patterns
For failing subcategories, look at what tests are failing:
cargo xtask cts 'webgpu:api,validation,category:subcategory:*' 2>&1 | grep "\[fail\]" | head -20
Look for patterns:
- Format-specific failures: May indicate missing format capability checks
- Parameter-specific failures: Validation missing for specific parameter combinations
Step 5: Examine Specific Failures
Pick a representative failing test and run it individually to see the error:
cargo xtask cts 'webgpu:api,validation,category:subcategory:specific_test' 2>&1 | tail -30
Look for:
- "EXPECTATION FAILED: DID NOT REJECT": wgpu is accepting invalid input (validation gap)
- "Validation succeeded unexpectedly": Similar to above
- "Unexpected validation error occurred": wgpu is rejecting valid input
- Error message content: Tells you what validation is triggering or missing
Step 6: Check CTS Test Source
To understand what the test expects, read the TypeScript source:
grep -A 40 "test_name" cts/src/webgpu/api/validation/path/file.spec.ts
The test source shows:
- What configurations are being tested
- What the expected behavior is (pass/fail)
- The validation rules from the WebGPU spec
- Comments explaining the rationale
Step 7: Categorize Issues
Group failures into categories:
High Priority - Validation Gaps:
- wgpu accepts invalid configurations that should fail
- Security or correctness implications
- Example: Accepting wrong texture formats, missing aspect checks
Medium Priority - Spec Compliance:
- Edge cases not handled correctly
- Optional field validation issues
- Example: depthCompare optional field handling
Low Priority - Minor Gaps:
- Less common scenarios
- Limited real-world impact
- Example: Depth bias with non-triangle topologies
Known Issues - Skip:
- Known failure patterns (documented in AGENTS.md)
- Track count but don't try to fix
Step 8: Identify Root Causes
For validation gaps, find where validation should happen:
Search for existing validation:
grep -n "relevant_keyword" wgpu-core/src/device/resource.rsLook for render/compute pipeline creation:
- Render pipeline:
wgpu-core/src/device/resource.rsaroundcreate_render_pipeline - Compute pipeline: Similar location
- Look for existing validation patterns you can follow
- Render pipeline:
Check for helper functions:
grep "fn is_" wgpu-types/src/texture/format.rsFind error enums:
grep "pub enum.*Error" wgpu-core/src/pipeline.rs
Step 9: Implement Fixes
When implementing fixes:
- Add error variants if needed (in
wgpu-core/src/pipeline.rs) - Add helper methods (in
wgpu-typesif checking properties) - Add validation checks (in
wgpu-core/src/device/resource.rs) - Test the fix with specific failing tests
- Run full subcategory to verify all related tests pass
- Check you didn't break passing tests
Step 10: Document Findings
Create or update a triage document (e.g., category_triage.md).
Do not write information about changes you have made to the triage document. Only capture the state of the tests and any investigation into open issues.
# Category CTS Tests - Triage Report
**Overall Status:** XP/YF/ZS (%/%/%)
## Passing Sub-suites ✅
[List sub-suites that have no failures (all pass or skip)]
## Remaining Issues ⚠️
[List sub-suites that have failures and if it can be stated concisely, a summary of the issue]
## Issue Detail
[List detail of any investigation into failures. Do not go into detail about passed suites, just list the failures.]
### 1. title, e.g. a distinguishing word from the test selector
**Test selector:** `webgpu:api,validation,render_pipeline,depth_stencil_state:format:*`
**What it tests:** [Description]
**Example failure:**
[a selector for a single failing test, e.g.:]
```
webgpu:api,validation,render_pipeline,depth_stencil_state:depthCompare_optional:isAsync=false;format="stencil8"
```
**Error:**
[error message from the failing tests, e.g.:]
```
Unexpected validation error occurred: Depth/stencil state is invalid:
Format Stencil8 does not have a depth aspect, but depth test/write is enabled
```
**Root cause:**
[Your analysis of the root cause. Do not speculate. Only include the results of specific investigation you have done.]
The validation is triggering incorrectly. When `depthCompare` is undefined/missing in JavaScript, it's getting translated to a default value that makes `is_depth_enabled()` return true, even for stencil-only formats.
**Fix needed:**
[Your proposed fix. Again, do not speculate. Only state the fix if it is obvious from the root cause analysis, or if you have done specific investigation into how to fix it.]
### 2. title
[repeat as needed for additional issues]
Step 11: Update test.lst
For fixed tests that are now passing, add them to cts_runner/test.lst:
Use wildcards to minimize lines:
webgpu:api,validation,category:subcategory:isAsync=false;*Group related tests when possible:
- If multiple subcategories all pass, use a higher-level wildcard
- Example:
webgpu:api,validation,category:*if all subcategories pass
Maintain alphabetical order roughly in the file
Step 12: Verify and Build
Before finishing:
# Format code
cargo fmt
# Check for errors
cargo clippy --tests
# Build to ensure no compilation errors
cargo build
# Run the tests you added to test.lst
cargo xtask cts 'webgpu:api,validation,category:subcategory:isAsync=false;*'
Common Patterns
If the user asked to investigate a failure, and the cause of the failure is noted here with "do not attempt to fix", then stop and ask the user before attempting to fix.
Pattern: Format-specific failures
- Check if format validation is missing
- Look for
is_depth_stencil_format(),is_color_format()etc. - May need to add format capability checks
Pattern: Aspect-related failures
- Check if code validates format aspects (DEPTH, STENCIL, COLOR)
- Use
hal::FormatAspects::from(format)to check - Validate operations match available aspects
Pattern: Optional field failures
- May be WebGPU optional field semantics issue
- Check if undefined in JS becomes a default value in Rust
- May need to distinguish "not set" from "set to default"
Pattern: Atomics accepted incorrectly
- Naga allows referencing an atomic directly in an expression
- Should only allow accessing via
atomicLoad,atomicStore, etc. - Only investigate as necessary to confirm this is the issue. Do not attempt to fix. Refer user to https://github.com/gfx-rs/wgpu/issues/5474.
Pattern: Error reporting for destroyed resources
- Tests that check for validation errors when a destroyed resource is used.
wgpuoften reports these errors later than WebGPU requires, causing the tests to fail. wgpumay report these errors earlier than it should, causing the test to fail with an unexpected validation error.- Look for:
- Tests with
state="destroyed"parameter - Tests checking that operations on destroyed buffers or textures should fail
- Tests with
- Example failing tests:
webgpu:api,validation,encoding,cmds,compute_pass:indirect_dispatch_buffer_state:withstate="destroyed"subcases
- Only investigate as necessary to confirm this is the issue. Do not attempt to fix. Refer user to https://github.com/gfx-rs/wgpu/issues/7881.
Tips
- Start with high-impact fixes: Validation gaps with security implications
- Look for existing patterns: Other validation code shows the style
- Test incrementally: Fix one category at a time, verify it works
- Document as you go: Don't wait until the end to write the triage
- Ask for clarification: If test expectations are unclear, check the WebGPU spec or test source
- Track your progress: Update pass rates as you fix issues