-
Notifications
You must be signed in to change notification settings - Fork 1.5k
Evaluation subpackage and metric implementations #303
Copy link
Copy link
Closed
Labels
Design discussionsrelated to the generic API designsrelated to the generic API designsModule: metricsmetric for model quality assessmentsmetric for model quality assessmentsWG: EvaluationFor the evaluation working groupFor the evaluation working groupenhancementNew feature or requestNew feature or request
Milestone
Description
I'd like to use this issue to discuss the structure of the evaluation part of MONAI, essentially what is currently monai.metrics. I plan to do the following in a PR, but obviously would like to coordinate with you guys.
- Rename to
monai.evaluationto make it more generic, at first there will be modulesmetrics(Implementations of metrics)util(For now a confusion matrix helper to hold TP, FP, FN, TN and a wrapper to convert metrics into aignite.metrics.Metric)
- Implement the most commonly used metrics
- Confusion matrix based metrics (see https://en.wikipedia.org/wiki/Confusion_matrix for an overview)
- Hausdorff distance (95), medpy has an implementation
- Surface distance, medpy has an implementation
- Check medpy, NiftyNet, Clara Train for other metrics (see also Port metrics from niftynet and clara train #85)
I'm assuming you also have ideas what that package should look like and requirements that I'm not aware of. Would love to hear your thoughts! I'm also a bit biased towards segmentation, I'm guessing there are special metrics for other tasks that should be integrated.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
Design discussionsrelated to the generic API designsrelated to the generic API designsModule: metricsmetric for model quality assessmentsmetric for model quality assessmentsWG: EvaluationFor the evaluation working groupFor the evaluation working groupenhancementNew feature or requestNew feature or request