Feature importance
PredictionValuesChange
The individual importance values for each of the input features (the default feature importances calculation method for nonranking metrics).
For each feature, PredictionValuesChange shows how much on average the prediction changes if the feature value changes. The bigger the value of the importance the bigger on average is the change to the prediction value, if this feature is changed.
See the Regular feature importance file format.
Calculation principles
Leaf pairs that are compared have different split values in the node on the path to these leaves. If the split condition is met (this condition depends on the feature F), the object goes to the left subtree; otherwise it goes to the right one.
$feature\_importance_{F} = \displaystyle\sum\limits_{trees, leafs_{F}} \left(v_{1}  avr \right)^{2} \cdot c_{1} + \left( v_{2}  avr \right)^{2} \cdot c_{2} { , }$
$avr = \displaystyle\frac{v_{1} \cdot c_{1} + v_{2} \cdot c_{2}}{c_{1} + c_{2}} { , where}$
 $c_1, c_2$ represent the total weight of objects in the left and right leaves respectively. This weight is equal to the number of objects in each leaf if weights are not specified for the dataset.
 $v_1, v_2$ represent the formula value in the left and right leaves respectively.
If the model uses a combination of some of the input features instead of using them individually, an average feature importance for these features is calculated and output. For example, the model uses a combination of features f54
, c56
and f77
. First, the feature importance is calculated for the combination of these features. Then the resulting value is divided by three and is assigned to each of the features.
If the model uses a feature both individually and in a combination with other features, the total importance value of this feature is defined using the following formula:
$feature\_total\_importance_{j} = feature\_importance + \sum\limits_{i=1}^{N}average\_feature\_importance_{i} { , where}$
 $feature\_importance_{j}$ is the individual feature importance of the jth feature.
 $average\_feature\_importance_{i}$ is the average feature importance of the jth feature in the ith combinational feature.
Complexity of computation
$O(trees\_count \cdot depth \cdot 2 ^ {depth} \cdot dimension)$
Specifics

Feature importance values are normalized so that the sum of importances of all features is equal to 100. This is possible because the values of these importances are always nonnegative.

Formula values inside different groups may vary significantly in ranking modes. This might lead to high importance values for some groupwise features, even though these features don't have a large impact on the resulting metric value.
LossFunctionChange
The individual importance values for each of the input features (the default feature importances calculation method for ranking metrics). This type of feature importance can be used for any model, but is particularly useful for ranking models, where other feature importance types might give misleading results.
For each feature the value represents the difference between the loss value of the model with this feature and without it. The model without this feature is equivalent to the one that would have been trained if this feature was excluded from the dataset. Since it is computationally expensive to retrain the model without one of the features, this model is built approximately using the original model with this feature removed from all the trees in the ensemble. The calculation of this feature importance requires a dataset and, therefore, the calculated value is datasetdependent.
See the Regular feature importance file format.
Calculation principles
The value of LossFunctionChange is defined so that the more important is the feature, the higher is its importance value.

Minimum best value objective metric:
$feature\_importance_{i} = metric (E_{i}v))  metric(v)$

Maximum best value objective metric:
$feature\_importance_{i} = metric(v)  metric(E_{i}v)$

Exact best value objective metric:
$feature\_importance_{i} = metric(E_{i}v)  best\_value  metric(v)  best\_value$
In general, the value of LossFunctionChange can be negative.
Variables description:

$E_{i}v$ is the mathematical expectation of the formula value without the $i$th feature. If the feature $i$ is on the path to a leaf, the new leaf value is set to the weighted average of values of leaves that have different paths by feature value. Weights represent the total weight of objects in the corresponding leaf. This weight is equal to the number of objects in each leaf, if weights are not specified in the dataset.
For feature combinations $F = (f_{1}, ..., f_{n})$, the average value in a leaf is calculated as follows:
$E_{f_i}v = \displaystyle\left(\frac{(n  1) v + E_{F}v}{n}\right)$ 
$v$ is the vector with formula values for the dataset. The training dataset are used, if both training and validation datasets are provided.

$metric$ is the loss function specified in the training parameters.
The size of the random subsample used for calculation is determined as follows:
$subsamples\_count = \min(samples_count, \max(2\cdot 10^5, \frac{2\cdot 10^9}{features\_count}))$
Complexity of computation
$O(trees\_count \cdot (2 ^ {depth} + subsamples\_count) \cdot depth +$
$+ Eval\_metric\_complexity(model, subsamples\_count) \cdot features\_count)$
This feature importance approximates the difference between metric values calculated on the following models:
 The model with the $i$th feature excluded.
 The original model with all features.
InternalFeatureImportance
The importance values both for each of the input features and for their combinations (if any).
See the InternalFeatureImportance file format.
Calculation principles
Leaf pairs that are compared have different split values in the node on the path to these leaves. If the split condition is met (this condition depends on the feature F), the object goes to the left subtree; otherwise it goes to the right one.
$feature\_importance_{F} = \displaystyle\sum\limits_{trees, leafs_{F}} \left(v_{1}  avr \right)^{2} \cdot c_{1} + \left( v_{2}  avr \right)^{2} \cdot c_{2} { , }$
$avr = \displaystyle\frac{v_{1} \cdot c_{1} + v_{2} \cdot c_{2}}{c_{1} + c_{2}} { , where}$
 $c_{1}, c_{2}$ represent the total weight of objects in the left and right leaves respectively. This weight is equal to the number of objects in each leaf if weights are not specified for the dataset.
 $v_{1}, v_{2}$ represent the formula value in the left and right leaves respectively.
If the model uses a combination of some of the input features instead of using them individually, an average feature importance for these features is calculated and output. For example, the model uses a combination of features f54
, c56
and f77
. First, the feature importance is calculated for the combination of these features. Then the resulting value is divided by three and is assigned to each of the features.
If the model uses a feature both individually and in a combination with other features, the total importance value of this feature is defined using the following formula:
$feature\_total\_importance_{j} = feature\_importance + \sum\limits_{i=1}^{N}average\_feature\_importance_{i} { , where}$
 $feature\_importance_{j}$ is the individual feature importance of the jth feature.
 $average\_feature\_importance_{i}$ is the average feature importance of the jth feature in the ith combinational feature.
Complexity of computation
$O(trees\_count \cdot depth \cdot 2 ^ {depth} \cdot dimension)$
PredictionDiff
The impact of a feature on the prediction results for a pair of objects. This type of feature importance is designed for analyzing the reasons for wrong ranking in a pair of documents, but it also can be used for any onedimensional model.
For each feature PredictionDiff reflects the maximum possible change in the predictions difference if the value of the feature is changed for both objects. The change is considered only if there is an improvement in the direction of changing the order of documents.
Alert
 Only models trained on datasets that do not contain categorical features are supported.
 Multiclassification modes are not supported.
Related information
Detailed information regarding usage specifics for different Catboost implementations.