CI - Confidence Index
\[\text{CI}(y, \hat{y}) = \text{R}(y, \hat{y}) * \text{WI}(y, \hat{y})\]
Latex equation code:
\text{CI}(y, \hat{y}) = \text{R}(y, \hat{y}) * \text{WI}(y, \hat{y})
Confidence Index [10] or Performance Index (CI/PI) is score that measures the performance of each estimation method, with a higher value indicating better performance. The range of the CI/PI is (-inf, 1], meaning it can take any value less than or equal to 1, but not including negative infinity.
Best possible score is 1.0, bigger value is better. Range = (-inf, 1], meaning of values:
> 0.85 Excellent Model 0.76-0.85 Very good 0.66-0.75 Good 0.61-0.65 Satisfactory 0.51-0.60 Poor 0.41-0.50 Bad < 0.40 Very bad
Example to use CI metric:
from numpy import array
from permetrics.regression import RegressionMetric
## For 1-D array
y_true = array([3, -0.5, 2, 7])
y_pred = array([2.5, 0.0, 2, 8])
evaluator = RegressionMetric(y_true, y_pred)
print(evaluator.confidence_index())
## For > 1-D array
y_true = array([[0.5, 1], [-1, 1], [7, -6]])
y_pred = array([[0, 2], [-1, 2], [8, -5]])
evaluator = RegressionMetric(y_true, y_pred)
print(evaluator.CI(multi_output="raw_values"))