Coverage Score
{{ (coverage_score * 100)|round(2)|string + "%" }}
Average empirical coverage across alpha levels
{% if coverage_score >= 0.95 %} {% elif coverage_score >= 0.9 %} {% elif coverage_score >= 0.85 %} {% elif coverage_score >= 0.8 %} {% else %} {% endif %}
Calibration Error
{{ calibration_error|round(4) }}
Expected calibration error (lower is better)
{% if calibration_error < 0.02 %} {% elif calibration_error < 0.05 %} {% elif calibration_error < 0.1 %} {% elif calibration_error < 0.15 %} {% else %} {% endif %}
Sharpness
{{ sharpness|round(4) }}
Average precision of uncertainty estimates
{% if sharpness > 0.9 %} {% elif sharpness > 0.8 %} {% elif sharpness > 0.7 %} {% elif sharpness > 0.6 %} {% else %} {% endif %}
Consistency
{{ consistency|round(4) }}
Consistency of uncertainty across predictions
{% if consistency > 0.9 %} {% elif consistency > 0.8 %} {% elif consistency > 0.7 %} {% elif consistency > 0.6 %} {% else %} {% endif %}