Skip to content
Permalink
Fetching contributors…
Cannot retrieve contributors at this time
65 lines (47 sloc) 1.99 KB
title taxonomy
BRATS 2012
category
docs

Ranking Table - Basics

  • Cases: Total number of datasets to be segmented
  • Segmentations: User submitted segmentations (can differ from cases)
  • Label: Regions in the segmentation.

Ranking calculations

  1. Metrics: We calculate metrics results for each label of submitted segmentations. These are displayed on the evaluation page of each user.
  2. Metric Average: Sum up all results per metric and devide by the number cases (not the number of segmentations submitted!)
  3. Metric Rank: We assign a rank for this average (comparing to other users)
  4. Label Rank: Average of the metric ranks.
  5. Final Rank: Average of the label ranks
  6. The table is sorted by the final rank

Example table for a single label

User Metric 1 Metric 2 Overall
User1 1.1 (1) 0.9 (2) 1 (1.5)
User2 3.1 (3) 0.92 (1) 2 (2)
User3 1.4 (2) 0.85 (3) 3 (2.5)
user Metric Average (Metric Rank) Metric Average (Metric Rank) Final Rank (Rank Average)

Updated Four-Class Labels - Challenge Metrics

  1. Dice,
  2. Positive Predictive Value (Specificity),
  3. Sensitivity,
  4. Kappa

Updated Four-Class Labels - Training Metrics

  1. Dice,
  2. Positive Predictive Value (Specificity),
  3. Sensitivity,
  4. Kappa

MICCAI 2012 on-site challenge (2 labels only) Challenge Metrics

  1. Dice,
  2. Positive Predictive Value (Specificity),
  3. Sensitivity,
  4. Average Distance,
  5. Hausdorff Distance
  6. Kappa

MICCAI 2012 on-site challenge (2 labels only) Training Metrics

  1. Dice,
  2. Positive Predictive Value (Specificity),
  3. Sensitivity,
  4. Average Distance,
  5. Hausdorff Distance
  6. Kappa
You can’t perform that action at this time.