From 1579b58b0c6805b51565d7ec4440226417a87fdf Mon Sep 17 00:00:00 2001 From: "google-labs-jules[bot]" <161369871+google-labs-jules[bot]@users.noreply.github.com> Date: Tue, 31 Mar 2026 19:42:06 +0000 Subject: [PATCH] =?UTF-8?q?=F0=9F=A7=B9=20Rename=20auto-extracted=20method?= =?UTF-8?q?s=20in=20evaluate=5Fbaseline=5Fmetrics.py?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Renamed `_extracted_from_main_68` and `_extracted_from_main_25` to descriptive names based on their flags: `_scan_all_datasets` and `_run_dry_run_evaluation` respectively. Removed the accompanying TODO comments. Co-authored-by: daggerstuff <261005129+daggerstuff@users.noreply.github.com> --- training/scripts/evaluate_baseline_metrics.py | 10 ++++------ 1 file changed, 4 insertions(+), 6 deletions(-) diff --git a/training/scripts/evaluate_baseline_metrics.py b/training/scripts/evaluate_baseline_metrics.py index 41d44906..55f1acae 100644 --- a/training/scripts/evaluate_baseline_metrics.py +++ b/training/scripts/evaluate_baseline_metrics.py @@ -261,9 +261,9 @@ def main(): evaluator = BaselineMetricsEvaluator() if args.dry_run: - return _extracted_from_main_25(evaluator) + return _run_dry_run_evaluation(evaluator) if args.scan_all: - return _extracted_from_main_68(evaluator) + return _scan_all_datasets(evaluator) if args.scan_all_s3: return _scan_s3_datasets(evaluator, args.scan_all_s3) if args.input_file: @@ -410,8 +410,7 @@ def _scan_s3_datasets(evaluator, s3_prefix: str) -> int: return 0 -# TODO Rename this here and in `main` -def _extracted_from_main_68(evaluator): +def _scan_all_datasets(evaluator): logger.info("Scanning all dataset directories...") all_results: dict[str, Any] = {} @@ -441,8 +440,7 @@ def _extracted_from_main_68(evaluator): return 0 -# TODO Rename this here and in `main` -def _extracted_from_main_25(evaluator): +def _run_dry_run_evaluation(evaluator): logger.info("Running DRY RUN evaluation...") mock_data = [ {