[WIP/Draft] Testnet Phase #2

Phase #2 advances Cortensor from “system-wide regression” into trust-layer hardening — tightening validator logic, expanding redundancy, and beginning formal gas + runtime optimization. Building on Testnet Phase #1’s end-to-end stability baselines, this phase focuses on making validation more accurate, more resilient, and cheaper to operate at scale across both Testnet-0 (Arbitrum Sepolia) and Testnet-1 (L3 COR Rollup).


Phase 2 — Validation Refinement & Infrastructure Hardening

Timeline: Early Q1 → Mid Q1 2026 Goal: Strengthen validator logic, expand redundancy, and begin formal gas optimization. Focus: Accuracy, trust stability, and infrastructure robustness.


Contest Rules

Stake Requirement

  • All participants must stake 10,000 $COR per node (Pool 3) to qualify for Testnet Phase #2.

  • Nodes must remain staked and active throughout the testing window to receive rewards.

Node Uptime & Performance

Operators must:

  • Maintain continuous uptime and responsiveness to assigned sessions.

  • Support redundant validation flows (more multi-run / multi-miner validation paths).

  • Handle higher validator throughput, including scoring + reporting without bottlenecks.

  • Ensure correct registration + telemetry sync across both Testnet-0 and Testnet-1 when participating in both.

Participation & Reporting

Participants are expected to:

  • Report bugs, validator anomalies, scoring drift, or reliability issues.

  • Help validate improvements to Validator v2/v3, redundancy policies, and router reliability.

  • Share structured logs + findings through Discord reporting channels.

Disqualification Criteria

Cortensor reserves the right to disqualify any operator for:

  • Misreporting uptime or falsifying telemetry data.

  • Exploiting reward, routing, or validation mechanisms.

  • Any manipulation that undermines fairness, accuracy, or network stability.


Rewards

Base Reward

  • Base rewards is $20 per node.

  • Qualification Requirements:

    • Active uptime during the entire test window

    • Precommit Score ≥ 60 %

    • Node Level Success Rate

Role-Based Bonus

  • +1 % per verified Discord role (capped at 10 %).

Node Level Bonus

  • Level-based bonus scales linearly (+$0.80 per level, 0–10):

    • Level 0 = $0 → Level 10 = $5 extra per node per month.

Performance Prize

Multi-Node Bonus (New Ops Only)

  • 3 + nodes → +10 % bonus

  • 5 + nodes → +20 % bonus


Testing Focus

1) Validator v2/v3 Refinement (PoI + PoUW)

Phase #2 is centered on tightening trust calibration and reducing scoring drift:

  • Deep PoI (quantitative) validation refinement: improve numerical consistency, dispersion bounds, and convergence under redundancy.

  • Deep PoUW (qualitative) validation refinement: strengthen rubric adherence (accuracy, relevance, completeness, clarity, safety) and reduce judge variance.

  • Improve validator determinism under real load: fewer “flaky” outcomes, better repeatability, clearer failure modes.

2) Validator Engine Experiments: Ollama Models + Embedding Variants

We will explicitly experiment with Ollama-based model stacks on validator nodes to improve hardware compatibility and verification quality:

  • Validate Ollama model variants as LLM-verifier options (judge models / verifier models), comparing stability, latency, and scoring consistency.

  • Test multiple embedding model variants to address known issues where certain validator GPU setups produce inconsistent/unsupported embedding behavior with the current engine.

  • Target outcome: better hardware adaptation on validator nodes, fewer environment-specific failures, and more predictable validator throughput.

3) QuantitativeStats / QualitativeStats Expansion

To tune thresholds and reduce false positives/negatives:

  • Expand datasets for calibration of PoI thresholds and PoUW scoring.

  • Improve LLM-verifier scoring alignment across task domains (general QA, code, reasoning, extraction, etc.).

  • Track distributions over time to detect drift, regressions, and “boundary-case” instability.

4) Redundancy Expansion & Trust Stability

Strengthen reliability through broader redundancy policies:

  • Increase multi-run / multi-miner redundancy where needed.

  • Validate aggregation logic (median/majority/weighted) and ensure it behaves correctly under partial failures.

  • Measure trust stability: score variance, disagreement rate, and disagreement handling behavior.

5) Router Surface Experiments: /delegate + /validate

Phase #2 expands structured testing of the router’s agent-facing surfaces:

  • /delegate experimentation:

    • Policy-driven delegation (cost/risk hints), routing behavior, retries, and fallbacks

    • Multi-run delegation and “same task, different miners/models” variance testing

    • Response structure stability for downstream agent consumption

  • /validate experimentation:

    • Validator tier selection, scoring stability under different verifier policies

    • Regression testing across domains and edge cases (hallucination, partial answers, ambiguous prompts)

    • Throughput/latency profiling for validation-heavy workloads

6) Oracle & Internal Service Reliability

Harden the supporting pipeline that feeds validation and observability:

  • Router APIs and internal services reliability under sustained load.

  • Indexer consistency + freshness, including recovery from missed blocks/events.

  • Dashboard feed integrity (no stale/partial views, accurate merging across networks).

7) Infrastructure Redundancy: RPC, Failover, Backups

Make “survives failures” a first-class requirement:

  • Multiple RPC endpoints per network + automatic failover behavior.

  • Validator backups / fallback paths to prevent single points of failure.

  • Validate ops playbooks: restart recovery, resync behavior, and degraded-mode handling.

8) Gas Optimization & Runtime Tuning (Formal Start)

Begin structured profiling and cost reduction:

  • Profile Cortensord runtime parameters and on-chain execution paths.

  • Optimize key L2 contracts/modules:

    • SessionQueueValidation

    • SessionPayment

    • Validator storage layout

  • Reduce unnecessary log events and nested calls to minimize gas and execution overhead.

9) Dashboard Expansion: Trust Observability

Expand instrumentation so Phase #2 improvements are measurable:

  • Real-time PoI / PoUW analytics

  • Validator health views (availability, queue latency, drift indicators)

  • Endpoint health indicators (RPC/indexer/router feed status) and alert-style visibility

10) L3 Rollup Comparison: Self-Managed vs RaaS (Apple-to-Apple)

A dedicated Phase #2 focus to validate tradeoffs before scaling Testnet-1:

  • Benchmark Testnet-1a (self-managed L3) vs Testnet-1 (RaaS L3) under the same workload and validation requirements.

  • Measure:

    • Indexer/validator freshness under load

    • RPC reliability + failover behavior

    • Throughput/latency + telemetry completeness

    • Operational overhead and where cost/runtime tuning matters most

  • Goal: a data-backed decision on which L3 approach best supports Testnet → Mainnet progression.

11) Light MCP Integration: Component Assignment Experiments (Nice To Have)

Phase #2 includes early “agent-ready wiring” work via light MCP integration:

  • Experiment with MCP component assignment (mapping tasks to tools/components cleanly)

  • Validate minimal MCP compatibility for task execution + validation loops (without over-scoping full protocol rollout)

  • Goal: prove the router can begin acting as an agent execution surface where tools/components are assigned deterministically and observed end-to-end.


Key Outcomes

By the end of Phase #2, Cortensor should achieve:

  • Higher-accuracy validation with more stable PoI/PoUW scoring.

  • Expanded redundancy with measurable trust improvement and lower variance.

  • Infrastructure robustness via failover, backups, and reduced service fragility.

  • Early gas + runtime optimizations reducing cost and execution overhead on L2.

  • A fully instrumented trust layer, observable in real time through dashboard analytics.


Outcome Statement

Testnet Phase #2 is about making Cortensor’s trust layer redundant, cost-efficient, and fully observable — a foundation where validation isn’t just “working,” but stable under load, resilient under failure, and optimized for scale.

Join Cortensor Discord: https://discord.gg/cortensor

Last updated