Skip to content

Enhanced Evaluation Metrics with Macro/Micro Scores and Per-Label Analysis#304

Open
Xiaomin-HUANG wants to merge 1 commit intourchade:mainfrom
Xiaomin-HUANG:feature/evaluation
Open

Enhanced Evaluation Metrics with Macro/Micro Scores and Per-Label Analysis#304
Xiaomin-HUANG wants to merge 1 commit intourchade:mainfrom
Xiaomin-HUANG:feature/evaluation

Conversation

@Xiaomin-HUANG
Copy link

Added Macro F1 and per-label metrics to complement the existing Micro F1 score, enabling better visibility into model performance across different entity types.

What's New

  • Macro F1: Unweighted average across all labels (0.51)
  • Per-label breakdown: Precision, Recall, and F1 for each entity type
  • Formatted table: Sorted output for quick identification of best/worst performers
Enhenced output : {
        
        "per_class":{"tag1":{"precision":float, "recall":float,"f_score":float},
                    "tag2":{}...
                    },
        "micro":{"precision":float, "recall":float,"f_score":float},
        "macro":{"precision":float, "recall":float,"f_score":float},
        }

Formatted table eg :
image

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant