Skip to content

[codex] Clean calibration eval warnings#162

Merged
DavidJBianco merged 1 commit into
devfrom
codex/calibration-cleanliness-fixes
May 15, 2026
Merged

[codex] Clean calibration eval warnings#162
DavidJBianco merged 1 commit into
devfrom
codex/calibration-cleanliness-fixes

Conversation

@DavidJBianco
Copy link
Copy Markdown
Collaborator

Summary

  • Fix actionable calibration cleanliness issues found while regenerating and evaluating scenarios/iteration-test from current dev.
  • Tighten evaluator matching so observation-aware/pivot-related expected gaps do not turn into false positives, while real contradictions still fail.
  • Fix OCSP optional field rendering, network observation-manifest accounting for sensor-filtered evidence, and visible Windows logon-before-process ordering.

Validation

  • uv run eforge validate-config
  • uv run eforge validate scenarios/iteration-test/scenario.yaml
  • uv run eforge generate scenarios/iteration-test/scenario.yaml --verbose --force
  • uv run eforge eval scenarios/iteration-test/data --scenario scenarios/iteration-test/scenario.yaml --format json --verbose -> overall 94.64, all hard gates passing
  • uv run ruff check .
  • uv run ruff format --check .
  • Focused regressions: 164 passed
  • uv run pytest -v -> 3075 passed, 15 skipped

Notes

Pivot-linkability misses that are realistic or explained by the observation profile are intentionally left as calibration signals rather than cleanup blockers.

@DavidJBianco DavidJBianco marked this pull request as ready for review May 15, 2026 18:06
@DavidJBianco DavidJBianco merged commit 87ac753 into dev May 15, 2026
4 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant