Training
This page covers everything you need to know about the Training Box and the nodes you can use inside it.
Now that you have initialized the main components of your experiment, it's time to define the training process for your experiment. Inside, can use two essential nodes:
Train Model: Define your pipeline's learning process.
Combine Models: Combine models from different pipelines.

Train Model: Model Training and Optimization Node
This node provides comprehensive control over model development through four key functions:
Base Model Training
Hyperparameter Tuning
Model Ensembling
Probability Calibration
The configuration options correspond to PyCaret's create_model()
function parameters (excluding the estimator
parameters, which are defined in the Model Node).

In the machine learning workflow, the Train Model node is used in the section shown below:

Configuration Options:
1. Hyperparameter Tuning (Optional): Enable this feature to optimize your model's performance. This functionality directly implements PyCaret's tune_model()
function. The function parameters can be set in the node.
Tuning Options:
Default PyCaret Tuning Grid: The system will automatically:
Test optimal parameter ranges
Apply cross-validation
Return the best performing configuration
Custom Tuning Grid: For advanced control:
Select parameters to tune from your model's options
Specify either:
Exact values to test (discrete)
Search ranges (continuous)

2. Model Ensembling (Optional): Activate to ensemble your trained model. This functionality directly implements PyCaret's ensemble_model()
function.
Configuration:
Select Ensemble Method (
method
):Bagging: Parallel training with bootstrap samples
Boosting: Sequential training with error correction
Select the number of estimators
n_estimators
: Number of models to ensemble (default: 10)
3. Probability Calibration (Optional)
Improve classification probability reliability. This functionality directly implements PyCaret's calibrate_model()
function.
To set up, choose Calibration Method:
Logistic Regression: Better for smaller datasets (<1,000 samples)
Isotonic Regression: More flexible for complex distributions

Combine Models: Combine trained models
This node enables model combination techniques to improve predictive performance. Connect trained models from Train Model nodes to create either stacked ensembles or blended predictions. It represents the combination section of the machine learning workflow, as shown below:

Combination Methods
Model Stacking: Implements PyCaret's
stack_models()
function to train a meta-model on the base models' outputs:Requires at least 2 models
Meta-model (default: logistic regression) learns optimal combination weights
Model Blending: Executes PyCaret's
blend_models()
function to average predictions:No meta-model training (faster execution)
Ideal for models with similar performance profiles
Note that the final combined model can be calibrated using PyCaret's calibrate_model()
function by simply turning on the Calibrate switch in the node.

This summarizes everything you need to know about the Training Box. Although it only uses one or two nodes, it is essential for your ML experiment. On the next page, you will learn about the Analysis Box as well as the Analysis Mode, which are essential to analyze your experiment's results.
Last updated