Red Team Dashboard
Adversarial testing — find weaknesses before bad actors do
Why Red-Team?
The PWM benchmark is only as trustworthy as the gates that protect it. Red-teaming means deliberately trying to break those gates. If you can bypass a gate, we need to know about it so we can fix it. Every vulnerability found makes the benchmark stronger.
Successful red-team findings are credited on the contributor leaderboard and acknowledged in published benchmark reports.
Open Challenges
Each bounty targets a specific gate or system layer. Can you break it?
| Gate | Challenge | Difficulty | Status |
|---|---|---|---|
| R1 | Create a manifest that passes R1 with missing CoreSpec fields | Medium | Open |
| R2 | Produce non-reproducible results that still pass R2 | Hard | Open |
| R3 | Bypass SHA-256 hash verification | Very Hard | Open |
| R4 | Hide budget overrun from R4 detection | Medium | Open |
| Social | Inflate metrics while passing all automated gates | Easy (by design) | Documented |
Attack Surface Map
A Automated Gate Bypass
- R1 — Spec spoofing: Create a valid-looking manifest with missing or invalid CoreSpec fields
- R2 — Seed collision: Different seeds producing same hash (reproducibility fake)
- R3 — Hash collision: Tampered artifacts that still match SHA-256 hashes
- R4 — Budget hiding: Conceal runtime overruns (e.g., offload to external GPU)
S Social & Metric Gaming
- PSNR inflation: Optimize for PSNR while destroying perceptual quality
- Leaderboard sniping: Submit many slight variants to dominate rankings
- Dataset overfitting: Memorize test data from public tier to cheat on dev/hidden
- Ghost authorship: Submit others' solvers under fake identity
C Scientific Gate Evasion
- S1 — Infinite spec: Submit a problem that cannot be finitely specified
- S2 — Ill-posed bypass: Claim well-posedness for a fundamentally ill-posed problem
- S3 — Convergence fake: Show false convergence proof
- S4 — Uncomputable bound: Claim error bounds that cannot be verified
I Infrastructure Attacks
- API abuse: Flood submission endpoints to DoS the system
- Model injection: Submit solver with malicious payload in checkpoint
- GCS path traversal: Attempt to access unauthorized bucket paths
- Auth bypass: Elevate privileges from user to reviewer/admin
Recent Red-Team Reports
No red-team reports yet. Be the first to find a vulnerability!
How to Submit a Finding
Found a gate bypass or metric gaming opportunity? Report it responsibly.
- Reproduce the attack — document the exact steps, inputs, and outputs.
- Tag with [RED-TEAM] — prefix your claim title with
[RED-TEAM]so it routes to the security review queue. - Submit via CLI:
pwm claim submit \ --title "[RED-TEAM] R3 hash bypass via truncated SHA-256" \ --method red_team \ --gate R3 \ --description "Describe the vulnerability and reproduction steps" \ --evidence evidence.tar.gz - Responsible disclosure — if the vulnerability is critical (e.g., allows data exfiltration), contact
security@pwm.platformai.orgbefore public disclosure.
Running Adversarial Tests
Use the PWM CLI to run automated adversarial tests against your local or staging instance.
# Install the PWM CLI
pip install pwm-cli
# Run all red-team checks against a target
pwm redteam run --target https://pwm.platformai.org --suite all
# Run specific gate tests
pwm redteam run --gate R1 --attack spec-spoofing
pwm redteam run --gate R3 --attack hash-collision
pwm redteam run --gate R4 --attack budget-hiding
# Generate an adversarial dataset (PSNR-inflated reconstruction)
pwm redteam generate --attack psnr-inflation --modality ct --output adversarial.h5
# Submit red-team report
pwm redteam report --title "[RED-TEAM] Found R1 bypass" --evidence report.json
Submit a Report via Web
Prefer not to use the CLI? Submit your finding directly here.
Log in to submit a report via web form.