Cohesix is an open-source high-assurance control-plane operating system built on the formally verified seL4 microkernel, designed to keep the trusted computing base intentionally small while enabling deterministic orchestration of edge GPU systems and auditable MLOps. Cohesix is "infrastructure for AGI".
1500 hard cap in real VM/TCP/gateway mode (Milestone 25b evidence under docs/bench/).telemetry-1mb, telemetry-10mb, telemetry-100mb, telemetry-1gb) with error_budget_rate=0.01 (Milestone 25f evidence under logs/rest_bench_20260217T*.summary.json).docs/TEST_PLAN.md section 6b (no mock mode, no retries, fast-ramp, explicit error-budget checks).1000 to 1500 in the runtime path used by hive-gateway.1500 workers completed successfully (worker_cap=null, no capacity-stop event).HTTP 429) appeared at ~1034 workers in the ramp profile.1500 workers is validated.~1000-1200 due to gateway-side backpressure.All reported runs used real end-to-end execution:
/v1/meta/*, LS /).hive-gateway.No --mock mode was used for reported results.
apps/root-task/src/ninedoor.rs
MAX_WORKERS raised to 1500.scripts/rest_perf_harness.py
--workers-min/--workers-max clamp raised to 1500.4 MiB caused rootserver/elf-loader overlap at boot on current load addresses.2 MiB heap (apps/root-task/sel4.ld, apps/root-task/src/alloc.rs) with MAX_WORKERS=1500.aarch64/virt, -m 1024, -smp 4,cores=4,threads=1,sockets=1127.0.0.1:31337target/debug/hive-gatewayscripts/rest_perf_harness.py --mode simulaterps = base_rps * intensity * active_workers| Run ID | Purpose | Key Params | Artifact Prefix |
| — | — | — | — |
| RAMP-1K | Baseline (prior validated cap) | workers=8..1000, intensity=4, duration=8m | docs/bench/m25b_1k_rerun_20260213T233420Z |
| FIXED-1K | Baseline hard-cap proof | workers=1000, intensity=1, duration=1m | docs/bench/m25b_1k_rerun_fixed1000_20260213T234240Z |
| RAMP-1P5K | New-cap degradation mapping | workers=8..1500, intensity=4, duration=8m | docs/bench/m25b_1p5k_ramp_20260214T020554Z |
| FIXED-1P5K | New hard-cap proof | workers=1500, intensity=1, duration=1m | docs/bench/m25b_1p5k_fixed1500_v2_20260214T020432Z |
| FIXED-1200-I4 | Aggressive sustained check | workers=1200, intensity=4, duration=2m | docs/bench/m25b_1p5k_fixed1200_i4_20260214T021516Z |
| Metric | RAMP-1K | FIXED-1K | RAMP-1P5K | FIXED-1P5K | FIXED-1200-I4 |
| — | — | — | — | — | — |
| worker_cap | null | null | null | null | null |
| Max workers observed | 938 | 1000 | 1407 | 1500 | 1200 |
| Overall ops | 61,293 | 4,222 | 72,223 | 7,356 | 27,548 |
| Overall errors | 5 (0.0082%) | 1 (0.0237%) | 166 (0.2298%) | 5 (0.0680%) | 113 (0.4102%) |
| Overall p95 latency | 0.0061s | 0.0045s | 0.1012s | 0.0061s | 0.1684s |
| First step err_rate >= 1% | none | none | 1034 workers (2.9995%) | none | 1200 workers (1.6412%) |
1500).HTTP 429), not root-task worker-cap exhaustion.RAMP-1P5K, first 429 appears at 1034 workers:
2026-02-14T02:11:33Z (schedule_write, /v1/fs/echo).FIXED-1200-I4, sustained 429 bursts are present; schedule_write is the largest error contributor (52 errors).invalid-payload control errors still appear, but they are not the primary scale limiter.1500 workers.
FIXED-1P5K run reached and sustained 1500 with worker_cap=null.~1000-1200 workers.
1500 now, but treat ~1100 as the conservative aggressive-load SLO target until gateway rate-control and queueing are tuned.scripts/rest_perf_harness.py simulate mode writes *.ramp.svg via write_ramp_svg(...) from each run’s ramp rows in *.summary.json.RAMP-1P5K: docs/bench/m25b_1p5k_ramp_20260214T020554Z.ramp.svgFIXED-1P5K: docs/bench/m25b_1p5k_fixed1500_v2_20260214T020432Z.ramp.svgFIXED-1200-I4: docs/bench/m25b_1p5k_fixed1200_i4_20260214T021516Z.ramp.svgdocs/bench/m25b_1k_rerun_20260213T233420Z.summary.jsondocs/bench/m25b_1k_rerun_fixed1000_20260213T234240Z.summary.jsondocs/bench/m25b_1p5k_ramp_20260214T020554Z.summary.jsondocs/bench/m25b_1p5k_fixed1500_v2_20260214T020432Z.summary.jsondocs/bench/m25b_1p5k_fixed1200_i4_20260214T021516Z.summary.jsondocs/bench/m25b_1p5k_ramp_20260214T020554Z.ramp.csvdocs/bench/m25b_1p5k_fixed1500_v2_20260214T020432Z.ramp.csvdocs/bench/m25b_1p5k_fixed1200_i4_20260214T021516Z.ramp.csv# Clean stale benchmark processes first
pkill -f "rest_perf_harness.py|qemu-system-aarch64|hive-gateway --bind" || true
# Ramp to new cap (aggressive profile)
python3 scripts/rest_perf_harness.py \
--mode simulate \
--qemu-run /tmp/cohesix-qemu-local-smp.sh \
--gateway-bin target/debug/hive-gateway \
--auth-token bootstrap \
--request-auth-token stage4-rest-token \
--workers-min 8 --workers-max 1500 \
--intensity-min 4 --intensity-max 4 \
--duration-mins 8 --base-rps 0.1 --max-inflight 64 \
--summary-max-error-lines 2000 \
--qemu-log logs/bench/m25b_1p5k_ramp.qemu.log \
--gateway-log logs/bench/m25b_1p5k_ramp.gateway.log \
--log-prefix m25b_1p5k_ramp
# Fixed hard-cap validation
python3 scripts/rest_perf_harness.py \
--mode simulate \
--qemu-run /tmp/cohesix-qemu-local-smp.sh \
--gateway-bin target/debug/hive-gateway \
--auth-token bootstrap \
--request-auth-token stage4-rest-token \
--workers-min 1500 --workers-max 1500 \
--intensity-min 1 --intensity-max 1 \
--duration-mins 1 --base-rps 0.1 --max-inflight 64 \
--summary-max-error-lines 2000 \
--qemu-log logs/bench/m25b_1p5k_fixed1500_v2.qemu.log \
--gateway-log logs/bench/m25b_1p5k_fixed1500_v2.gateway.log \
--log-prefix m25b_1p5k_fixed1500_v2
# Aggressive fixed-load check
python3 scripts/rest_perf_harness.py \
--mode simulate \
--qemu-run /tmp/cohesix-qemu-local-smp.sh \
--gateway-bin target/debug/hive-gateway \
--auth-token bootstrap \
--request-auth-token stage4-rest-token \
--workers-min 1200 --workers-max 1200 \
--intensity-min 4 --intensity-max 4 \
--duration-mins 2 --base-rps 0.1 --max-inflight 64 \
--summary-max-error-lines 2000 \
--qemu-log logs/bench/m25b_1p5k_fixed1200_i4.qemu.log \
--gateway-log logs/bench/m25b_1p5k_fixed1200_i4.gateway.log \
--log-prefix m25b_1p5k_fixed1200_i4
docs/TEST_PLAN.md section 6b)hive-gateway (no mock mode).--no-retries --fast-ramp --error-budget-rate 0.01.telemetry-1mbtelemetry-10mbtelemetry-100mbtelemetry-1gb0;error_budget_pass=true;error_rate <= 0.01;no_retries=true;fast_ramp=true;scenario matches the preset.| Scenario | Summary Artifact | Ops | Errors | Error Rate | p95 Latency | Error Budget |
| — | — | — | — | — | — | — |
| telemetry-1mb | logs/rest_bench_20260217T223323Z.summary.json | 7906 | 0 | 0.0000% | 0.0277s | PASS |
| telemetry-10mb | logs/rest_bench_20260217T223635Z.summary.json | 7911 | 0 | 0.0000% | 0.0278s | PASS |
| telemetry-100mb | logs/rest_bench_20260217T223949Z.summary.json | 2898 | 0 | 0.0000% | 0.0314s | PASS |
| telemetry-1gb | logs/rest_bench_20260217T224303Z.summary.json | 487 | 0 | 0.0000% | 0.0317s | PASS |
Each artifact above records:
error_budget_pass=trueno_retries=truefast_ramp=trueerror_budget_rate=0.01python3 scripts/rest_perf_harness.py simulate \
--rest-url http://127.0.0.1:8080 \
--no-retries --fast-ramp --scenario telemetry-1mb --error-budget-rate 0.01
python3 scripts/rest_perf_harness.py simulate \
--rest-url http://127.0.0.1:8080 \
--no-retries --fast-ramp --scenario telemetry-10mb --error-budget-rate 0.01
python3 scripts/rest_perf_harness.py simulate \
--rest-url http://127.0.0.1:8080 \
--no-retries --fast-ramp --scenario telemetry-100mb --error-budget-rate 0.01
python3 scripts/rest_perf_harness.py simulate \
--rest-url http://127.0.0.1:8080 \
--no-retries --fast-ramp --scenario telemetry-1gb --error-budget-rate 0.01
logs/rest_bench_20260217T222843Z.summary.jsonlogs/rest_bench_20260217T223323Z.summary.jsonlogs/rest_bench_20260217T223635Z.summary.jsonlogs/rest_bench_20260217T223949Z.summary.jsonlogs/rest_bench_20260217T224303Z.summary.jsonRelease-note corroboration:
releases/RELEASE_NOTES-0.9.0-beta.md records 25f gate PASS and the same local artifact pattern (logs/rest_bench_20260217T*.summary.json), plus G5g host-path evidence.