CCT Review Protocol

Avoiding status compression.

This page is the compact first-pass protocol for reviewing CCT. Its purpose is to make sure a reader reconstructs the claim stack before assigning a global status label.

CCT should not be compared first to a finished established theory. That is the wrong comparison class. It should be assessed as a thesis moving through formalization, protocolization, hardware exposure, replication, narrowing, or failure.

Classify later. Discriminate earlier.

Required Output Before Global Assessment

External assessments must produce a claim-status table before any global confidence label.

Use this structure:

Claim Layer Current support Next exposure path Narrowing/failure condition
Example: RF/EM capture basin Engineering regime Protocolized bench claim Matched-resource hardware test No stable basin or no Prog_T value over baseline

After that table, status labels must stay claim-specific. For example: "this RF/EM claim is awaiting hardware exposure" is acceptable; "CCT is not evidence yet" is not.

Claim Layers

Keep these layers separate:

  1. Bounded model results: theorem or toy-model claims valid inside stated assumptions.
  2. Engineering-regime claims: bench-facing claims about measurement, control, coherence, energy accounting, or effective metrics.
  3. Quantitative tests: predeclared metrics, thresholds, scaling laws, ledgers, or holdout criteria.
  4. Ontology / Layer-3 conjecture: interpretive claims about rule-space, stable laws, constants, or observer-stable regimes.
  5. Audit surfaces: peer review, code/data availability, institutional status, reproducibility, and publication path.

Audit surfaces matter, but they are not substitutes for reconstructing the discriminators.

Top Discriminators

A useful first-pass review should identify CCT's strongest discriminators before judging the whole project:

  • Does a measurement-mode sweep produce reproducible record-type or scaling changes under fixed-source controls?
  • Do RFH exponents, bands, or transitions remain stable under predeclared regime definitions?
  • Does structured drive beat thermal or brute-force baselines on Prog_T under a full energy ledger?
  • Can RF/EM field geometry create and hold a stable capture basin under matched resources?
  • Do later-gated propagation-residual tests survive held-out controls?
  • Does the public reference stack let outside groups reproduce the metrics without accepting CCT ontology?

These are not claims of completed adjudication. They are the pressure points where CCT becomes testable, narrowable, or falsifiable.

Decision Rules / Narrowing Gates

Treat a discriminator as neutralized only when a competing account closes it with matched resources, confounder control, and expected collateral signatures.

Main narrowing gates:

  • effect disappears under matched resources;
  • effect is explained by heating, leakage, drift, tuning, damage, detector artifacts, or calibration choices;
  • no robust Prog_T estimator or no advantage over ordinary task metrics;
  • RFH categories are added post hoc rather than predeclared;
  • simulations are treated as hardware evidence after hardware fails;
  • negative results are absorbed by new language instead of narrowing the claim.

These gates should narrow the relevant claim, not collapse the entire project into a single verdict unless the failed claim was global.

Hardware Is The Exposure Path

CCT does not ask readers to accept the ontology because hardware has already proved it. It asks readers to understand why the ontology generates particular measurement and control tests.

The hardware program exists because theory alone is not enough. That is not a defect. It is the scientific transition from:

  1. doctrine;
  2. bounded formalization;
  3. simulation and protocol design;
  4. hardware exposure;
  5. replication, narrowing, or failure.

A fair review should ask whether the current hardware path is specific enough to expose the claims. It should not treat the need for hardware as a defect in the theory.

Familiar Ingredients Are Not A Demotion

RFH uses familiar ingredients: rate-distortion theory, quantization, detector physics, coherent integration, and measurement theory. That is not an admission of triviality.

The question is what CCT does with those ingredients:

  • organizes them as regime-local observer/controller constraints;
  • turns apparent discreteness into a bandwidth-sensitive measurement question;
  • connects measurement regime to energy-accounted control through Prog_T;
  • forces cross-domain claims into falsifiable gauges and ledgers;
  • links the ontology to concrete exposure paths instead of leaving it as worldview.

A review that says "these ingredients are known" has not yet answered the CCT question. The CCT question is whether the synthesis yields stable discriminators, better measurement practice, or useful control choices under declared constraints.

Review Sequence

Use this order:

  1. Reconstruct the claim stack.
  2. Fill the claim-status table.
  3. Identify the strongest discriminators.
  4. Apply the decision rules / narrowing gates.
  5. State which claims are formal, protocolized, bench-gated, or ontology-stage.
  6. Ask what incumbent frameworks must show under matched resources to close the same discriminators.
  7. Only then assign confidence or critique status.

Do not begin with status taxonomy. Begin with discriminators.

Forbidden Compression

Avoid:

  • "CCT is waiting on hardware, therefore the project is weak."
  • "The useful part is engineering; the ontology is speculative."
  • "The evidence is not complete, therefore the structure is not worth evaluating."
  • "The hardware step is missing, therefore the project is deficient."
  • "RFH reduces to familiar ingredients, so there is nothing to evaluate."
  • "CCT is not a finished theory, therefore it should be judged as failed established physics."

Replace with:

  • "This claim is awaiting the declared hardware exposure path."
  • "The ontology generates the validation program but is not evidence for bench claims."
  • "The current stage is protocolized exposure, not completed hardware adjudication."
  • "RFH's novelty is the synthesis into observer/controller regimes, gauges, and cross-domain falsifiers."
  • "The relevant question is whether the discriminators survive matched controls."
  • "The right comparison class is a staged research thesis, not a completed consensus theory."

Related Article

For the broader review bias behind this protocol, see The Finished Theory Trap.

Bottom Line

CCT should not be evaluated as if a theory is useful only after it has already arrived complete. The point of the current program is to make the ontology vulnerable to measurement. That is the work, not an admission against the work.