Selecting the best model for you
When analyzing the results of your experiment, it is important to look for models with specific characteristics that are important for your use case. For example, in addition to consistently accurate predictions, you might also require models that can deliver predictions quickly. In the Models tab in your experiment, models are recommended to you based on several angles of analysis.
Analyzing the top models for an experiment

Based on your filters, recommended models are presented to help you consider several different quality perspectives. A single model might be considered a top model in more than one way. The top model types are:
Viewing the top models in the user interface
For information about finding and exploring the top models for your experiment, see Analyzing the model metrics table.
Best model
Based on your filters, the best model is automatically selected for analysis. The best model is highlighted with a icon.
In Qlik Predict, the best model is determined from a balanced calculation that takes accuracy metrics and prediction speed into account.
To determine the best model, the following process is automatically performed:
-
Select the model with the highest score for the predictive performance metric determined by the model type. The metrics used are:
-
Binary classification: F1
-
Multiclass classification: F1 Macro
-
Regression: R2
-
-
Using the performance scores from step 1, select all models that are within five percent of the score of the highest scoring model.
-
Out of all models selected, select the model with the fastest prediction speed (see Prediction speed). This model is the best model.
Most accurate
It is important that your model is able to generate predictions with high accuracy on a consistent basis. Although F1, F1 Macro, and R2 provide balanced scoring that comprehensively reflect model accuracy, you might also have an interest in the raw accuracy and precision metrics for your models.
The most accurate model is highlighted with a icon. To determine the most accurate model, the following process is automatically performed:
-
Select the model with the highest score for the predictive performance metric determined by the model type. The metrics used are:
-
Using the performance scores from step 1, select all models that are within ten percent of the score of the highest scoring model.
-
One of the following two pathways are used depending on the model type:
-
Binary classification:
-
If the training dataset is balanced, select the model with the highest accuracy score. This is the most accurate model. For information about the specific metric used, see Accuracy.
-
If the training dataset is imbalanced, select the model with the highest precision score. For information about the specific metric used, see Precision.
-
-
Multiclass classification or regression:
-
Fastest model
When choosing a model, you might want to place value on how fast the model can deliver predictions. The fastest model is highlighted with a icon.
Prediction speed determines which model is the fastest. However, the models' predictive accuracy is still considered. This is because a model might be able to generate predictions quickly, but it must also be able to predict with reasonable accuracy as well.
To determine the fastest model, the following process is automatically performed:
-
Select the model with the highest score for the predictive performance metric determined by the model type. The metrics used are:
-
One of the following two pathways are used depending on the model type:
-
Binary classification:
-
If the training dataset is balanced, select all models that have an accuracy score within ten percent of the accuracy score of the model selected in step 1. For information about the specific metric used, see Accuracy.
-
If the training dataset is imbalanced, select all models that are within ten percent of the score of the highest scoring model from step 1. The metrics from step 1 are used.
-
-
Multiclass classification or regression:
-
-
Out of all models selected, select the model with the fastest prediction speed (see Prediction speed). This model is the fastest model.
Prediction speed
Prediction speed is a model metric that applies to all model types: binary classification, multiclass classification, and regression. Prediction speed measures how fast a machine learning model is able to generate predictions.
In Qlik Predict, prediction speed is calculated using the combined feature computing time and test dataset prediction time. It is displayed in rows per second.
Prediction speed can be analyzed in the Model metrics table after running your experiment version. You can also view prediction speed data when analyzing models with embedded analytics. For more information, see:
Considerations
The measured prediction speed is based on the size of the training dataset rather than the data on which predictions are made. After deploying a model, you might notice differences between how fast predictions are created if training and prediction data differ greatly in size, or when creating real-time predictions on one or a handful of data rows.
Overfitting
Overfitting occurs when a model's predictive behavior is too closely mapped to the training dataset. When a model is overfitted, it has likely only memorized patterns in the training dataset, and will not be able to accurately predict future values.
Overfitting can have several causes, including issues related to training algorithms and overly short or complex training datasets.
In Qlik Predict, overfitting is automatically identified through an analysis of test-train results for all metrics used in the process for top model selection, except prediction speed:
If there is a greater than ten percent difference between any of these metrics when comparing the testing and training results, the model is suspected of being overfitted.
If a model is suspected of being overfitted, it is never presented as a recommended model, even if it scores well. The model is marked with a warning in the Model metrics table.
If all models shown in your filters are suspected to be overfitted, models recommendations are not provided.
Addressing overfitting
You can address overfitting by:
-
Not deploying models suspected of overfitting.
-
If you suspect an issue with your training dataset, see Getting your dataset ready for training to learn about how you can prepare your training data to avoid overfitting.