Experiment configuration¶
max_runtime_minutes
¶
Max. runtime in minutes before triggering the ‘Finish’ button. Approximately enforced. (0 = disabled) (Number)
Default value 1440
If the experiment is not done after this many minutes, stop feature engineering and model tuning as soon as possible and proceed with building the final modeling pipeline and deployment artifacts, independent of model score convergence or pre-determined number of iterations. Only active is not in reproducible mode. Depending on the data and experiment settings, overall experiment runtime can differ significantly from this setting.
max_runtime_minutes_until_abort
¶
Max. runtime in minutes before triggering the ‘Abort’ button.(0 = disabled) (Number)
Default value 10080
If the experiment is not done after this many minutes, push the abort button. Preserves experiment artifacts made so far for summary and log zip files, but further artifacts are made.
time_abort
¶
Time to trigger the ‘Abort’ button. (String)
Default value ''
If the experiment is not done by this time, push the abort button. Accepts time in format given by time_abort_format (defaults to %Y-%m-%d %H:%M:%S)assuming a time zone set by time_abort_timezone (defaults to UTC). One can also give integer seconds since 1970-01-01 00:00:00 UTC. Applies to time on a DAI worker that runs experiments. Preserves experiment artifacts made so far for summary and log zip files, but further artifacts are made.NOTE: If start new experiment with same parameters, restart, or refit, thisabsolute time will apply to such experiments or set of leaderboard experiments.
time_abort_format
¶
Time string format for time_abort. (String)
Default value '%Y-%m-%d %H:%M:%S'
Any format is allowed as accepted by datetime.strptime.
time_abort_timezone
¶
Time zone for time_abort. (String)
Default value 'UTC'
Any time zone in format accepted by datetime.strptime.
delete_model_dirs_and_files
¶
delete_model_dirs_and_files (Boolean)
Default value True
Whether to delete all directories and files matching experiment pattern when call do_delete_model (True), or whether to just delete directories (False). False can be used to preserve experiment logs that do not take up much space.
recipe
¶
Pipeline Building Recipe (String)
Default value 'auto'
# Recipe type ## Recipes override any GUI settings - ‘auto’: all models and features automatically determined by experiment settings, toml settings, and feature_engineering_effort
- ‘compliant’like ‘auto’ except:
interpretability=10 (to avoid complexity, overrides GUI or python client chose for interpretability)
enable_glm=’on’ (rest ‘off’, to avoid complexity and be compatible with algorithms supported by MLI)
fixed_ensemble_level=0: Don’t use any ensemble
*feature_brain_level=0*(: No feature brain used (to ensure every restart is identical)
max_feature_interaction_depth=1: interaction depth is set to 1 (no multi-feature interactions to avoid complexity)
target_transformer=’identity’: for regression (to avoid complexity)
check_distribution_shift_drop=’off’: Don’t use distribution shift between train, valid, and test to drop features (bit risky without fine-tuning)
- ‘monotonic_gbm’like ‘auto’ except:
monotonicity_constraints_interpretability_switch=1: enable monotonicity constraints
self.config.monotonicity_constraints_correlation_threshold = 0.01: see below
monotonicity_constraints_drop_low_correlation_features=true: drop features that aren’t correlated with target by at least 0.01 (specified by parameter above)
fixed_ensemble_level=0: Don’t use any ensemble (to avoid complexity)
included_models=[‘LightGBMModel’]
included_transformers=[‘OriginalTransformer’]: only original (numeric) features will be used
feature_brain_level=0: No feature brain used (to ensure every restart is identical)
monotonicity_constraints_log_level=’high’
autodoc_pd_max_runtime=-1: no timeout for PDP creation in AutoDoc
- ‘kaggle’like ‘auto’ except:
external validation set is concatenated with train set, with target marked as missing
test set is concatenated with train set, with target marked as missing
transformers that do not use the target are allowed to fit_transform across entire train + validation + test
several config toml expert options open-up limits (e.g. more numerics are treated as categoricals)
- Note: If plentiful memory, can:
choose kaggle mode and then change fixed_feature_interaction_depth to large negative number,
- otherwise default number of features given to transformer is limited to 50 by default
choose mutation_mode = “full”, so even more types are transformations are done at once per transformer
‘nlp_model’: Only enables NLP models that process pure text
‘nlp_transformer’: Only enables NLP transformers that process pure text, while any model type is allowed
‘image_model’: Only enables Image models that process pure images
‘image_transformer’: Only enables Image transformers that process pure images, while any model type is allowed
‘unsupervised’: Only enables unsupervised transformers, models and scorers
‘gpus_max’: Maximize use of GPUs (e.g. use XGBoost, rapids, Optuna hyperparameter search, etc.)
‘more_overfit_protection’: Potentially improve overfit, esp. for small data, by disabling target encoding and making GA behave like final model for tree counts and learning rate
Each pipeline building recipe mode can be chosen, and then fine-tuned using each expert settings. Changing the pipeline building recipe will reset all pipeline building recipe options back to default and then re-apply the specific rules for the new mode, which will undo any fine-tuning of expert options that are part of pipeline building recipe rules.
If choose to do new/continued/refitted/retrained experiment from parent experiment, the recipe rules are not re-applied and any fine-tuning is preserved. To reset recipe behavior, one can switch between ‘auto’ and the desired mode. This way the new child experiment will use the default settings for the chosen recipe.
enable_genetic_algorithm
¶
Enable genetic algorithm for selection and tuning of features and models (String)
Default value 'auto'
Whether to enable genetic algorithm for selection and hyper-parameter tuning of features and models. - If disabled (‘off’), will go directly to final pipeline training (using default feature engineering and feature selection). - ‘auto’ is same as ‘on’ unless pure NLP or Image experiment. - “Optuna”: Uses DAI genetic algorithm for feature engineering, but model hyperparameters are tuned with Optuna.
In the Optuna case, the scores shown in the iteration panel are the best score and trial scores.
Optuna mode currently only uses Optuna for XGBoost, LightGBM, and CatBoost (custom recipe).
If Pruner is enabled, as is default, Optuna mode disables mutations of eval_metric so pruning uses same metric across trials to compare properly.
Currently does not supported when pre_transformers or multi-layer pipeline used, which must go through at least one round of tuning or evolution.
make_python_scoring_pipeline
¶
Make Python scoring pipeline (String)
Default value 'auto'
Whether to create the Python scoring pipeline at the end of each experiment.
make_mojo_scoring_pipeline
¶
Make MOJO scoring pipeline (String)
Default value 'auto'
Whether to create the MOJO scoring pipeline at the end of each experiment. If set to “auto”, will attempt to create it if possible (without dropping capabilities). If set to “on”, might need to drop some models, transformers or custom recipes.
mojo_for_predictions_benchmark
¶
mojo_for_predictions_benchmark (Boolean)
Default value True
Perform timing and accuracy benchmarks for Injected MOJO scoring vs Python scoring. This is for full scoring data, and can be slow. This also requires hard asserts. Doesn’t force MOJO scoring by itself, so depends on mojo_for_predictions=’on’ if want full coverage.
mojo_for_predictions_benchmark_slower_than_python_threshold
¶
mojo_for_predictions_benchmark_slower_than_python_threshold (Number)
Default value 10
Fail hard if MOJO scoring is this many times slower than Python scoring.
mojo_for_predictions_benchmark_slower_than_python_min_rows
¶
mojo_for_predictions_benchmark_slower_than_python_min_rows (Number)
Default value 100
Fail hard if MOJO scoring is slower than Python scoring by a factor specified by mojo_for_predictions_benchmark_slower_than_python_threshold, but only if have at least this many rows. To reduce false positives.
mojo_for_predictions_benchmark_slower_than_python_min_seconds
¶
mojo_for_predictions_benchmark_slower_than_python_min_seconds (Float)
Default value 2.0
Fail hard if MOJO scoring is slower than Python scoring by a factor specified by mojo_for_predictions_benchmark_slower_than_python_threshold, but only if takes at least this many seconds. To reduce false positives.
inject_mojo_for_predictions
¶
inject_mojo_for_predictions (Boolean)
Default value True
Inject MOJO into fitted Python state if mini acceptance test passes, so can use C++ MOJO runtime when calling predict(enable_mojo=True, IS_SCORER=True, …). Prerequisite for mojo_for_predictions=’on’ or ‘auto’.
mojo_for_predictions
¶
Allow use of MOJO for making predictions (String)
Default value 'auto'
Use MOJO for making fast low-latency predictions after experiment has finished (when applicable, for AutoDoc/Diagnostics/Predictions/MLI and standalone Python scoring via scorer.zip). For ‘auto’, only use MOJO if number of rows is equal or below mojo_for_predictions_max_rows. For larger frames, it can be faster to use the Python backend since used libraries are more likely already vectorized.
mojo_for_predictions_max_rows
¶
Max number of rows for C++ MOJO predictions (Number)
Default value 10000
For smaller datasets, the single-threaded but low latency C++ MOJO runtime can lead to significantly faster scoring times than the regular in-Driverless AI Python scoring environment. If enable_mojo=True is passed to the predict API, and the MOJO exists and is applicable, then use the MOJO runtime for datasets that have fewer or equal number of rows than this threshold. MLI/AutoDoc set enable_mojo=True by default, so this setting applies. This setting is only used if mojo_for_predictions is ‘auto’.
mojo_for_predictions_batch_size
¶
Batch size for C++ MOJO predictions. (Number)
Default value 100
Batch size (in rows) for C++ MOJO predictions. Only when enable_mojo=True is passed to the predict API, and when the MOJO is applicable (e.g., fewer rows than mojo_for_predictions_max_rows). Larger values can lead to faster scoring, but use more memory.
mojo_acceptance_test_rtol
¶
Relative tolerance for mini MOJO acceptance test. (Float)
Default value 0.0
Relative tolerance for mini MOJO acceptance test. If Python/C++ MOJO differs more than this from Python, won’t use MOJO inside Python for later scoring. Only applicable if mojo_for_predictions=True. Disabled if <= 0.
mojo_acceptance_test_atol
¶
Absolute tolerance for mini MOJO acceptance test. (Float)
Default value 0.0
Absolute tolerance for mini MOJO acceptance test. If Python/C++ MOJO differs more than this from Python, won’t use MOJO inside Python for later scoring. Only applicable if mojo_for_predictions=True. Disabled if <= 0.
reduce_mojo_size
¶
Attempt to reduce the size of the MOJO (Boolean)
Default value False
Whether to attempt to reduce the size of the MOJO scoring pipeline. A smaller MOJO will also lead to less memory footprint during scoring. It is achieved by reducing some other settings like interaction depth, and hence can affect the predictive accuracy of the model.
make_pipeline_visualization
¶
Make pipeline visualization (String)
Default value 'auto'
Whether to create the pipeline visualization at the end of each experiment. Uses MOJO to show pipeline, input features, transformers, model, and outputs of model. MOJO-capable tree models show first tree.
make_python_pipeline_visualization
¶
Make python pipeline visualization (String)
Default value 'auto'
Whether to create the python pipeline visualization at the end of each experiment. Each feature and transformer includes a variable importance at end in brackets. Only done when forced on, and artifacts as png files will appear in summary zip. Each experiment has files per individual in final population: 1) preprune_False_0.0 : Before final pruning, without any additional variable importance threshold pruning 2) preprune_True_0.0 : Before final pruning, with additional variable importance <=0.0 pruning 3) postprune_False_0.0 : After final pruning, without any additional variable importance threshold pruning 4) postprune_True_0.0 : After final pruning, with additional variable importance <=0.0 pruning 5) posttournament_False_0.0 : After final pruning and tournament, without any additional variable importance threshold pruning 6) posttournament_True_0.0 : After final pruning and tournament, with additional variable importance <=0.0 pruning 1-5 are done with ‘on’ while ‘auto’ only does 6 corresponding to the final post-pruned individuals. Even post pruning, some features have zero importance, because only those genes that have value+variance in variable importance of value=0.0 get pruned. GA can have many folds with positive variance for a gene, and those are not removed in case they are useful features for final model. If small mojo option is chosen (reduce_mojo_size True), then the variance of feature gain is ignored for which genes and features are pruned as well as for what appears in the graph.
max_cols_make_autoreport_automatically
¶
Number of columns beyond which will not automatically build autoreport at end of experiment. (Number)
Default value 1000
max_cols_make_pipeline_visualization_automatically
¶
Number of columns beyond which will not automatically build pipeline visualization at end of experiment. (Number)
Default value 5000
pass_env_to_deprecated_python_scoring
¶
Pass environment variables to deprecated python scoring package (Boolean)
Default value False
- Pass environment variables from running Driverless AI instance to Python scoring pipeline for
- deprecated models, when they are used to make predictions. Use with caution.
If config.toml overrides are set by env vars, and they differ from what the experiment’s env
looked like when it was trained, then unexpected consequences can occur. Enable this only to ” override certain well-controlled settings like the port for H2O-3 custom recipe server.
transformer_description_line_length
¶
Line length for autoreport descriptions of transformers. -1 means use autodoc_keras_summary_line_length (Number)
Default value -1
benchmark_mojo_latency
¶
Measure MOJO scoring latency (String)
Default value 'auto'
Whether to measure the MOJO scoring latency at the time of MOJO creation.
benchmark_mojo_latency_auto_size_limit
¶
Max size of pipeline.mojo file (in MB) for when benchmark_mojo_latency is set to ‘auto’ (Number)
Default value 500
Max size of pipeline.mojo file (in MB) for automatic mode of MOJO scoring latency measurement
mojo_building_timeout
¶
Timeout in seconds to wait for MOJO creation at end of experiment. (Float)
Default value 1800.0
If MOJO creation times out at end of experiment, can still make MOJO from the GUI or from the R/Py clients (timeout doesn’t apply there).
mojo_building_parallelism
¶
Number of parallel workers to use during MOJO creation (-1 = all cores) (Number)
Default value -1
If MOJO creation is too slow, increase this value. Higher values can finish faster, but use more memory. If MOJO creation fails due to an out-of-memory error, reduce this value to 1. Set to -1 for all physical cores.
mojo_building_parallelism_base_model_size_limit
¶
Size of base models to allow mojo_building_parallelism (Number)
Default value 100000000
- Size in bytes that all pickled and compressed base models have to satisfy to use parallel MOJO building.
For large base models, parallel MOJO building can use too much memory. Only used if final_fitted_model_per_model_fold_files is true.
show_pipeline_sizes
¶
Whether to show model and pipeline sizes in logs (String)
Default value 'auto'
- Whether to show model and pipeline sizes in logs.
If ‘auto’, then not done if more than 10 base models+folds, because expect not concerned with size.
max_workers
¶
max_workers (Number)
Default value 1
Maximum number of workers for Driverless AI server pool (only 1 needed currently)
max_cores_dai
¶
max_cores_dai (Number)
Default value -1
Max number of CPU cores to use across all of DAI experiments and tasks. -1 is all available, with stall_subprocess_submission_dai_fork_threshold_count=0 means restricted to core count.
stall_subprocess_submission_dai_fork_threshold_count
¶
stall_subprocess_submission_dai_fork_threshold_count (Number)
Default value 0
Stall submission of tasks if total DAI fork count exceeds count (-1 to disable, 0 for automatic of max_cores_dai)
stall_subprocess_submission_mem_threshold_pct
¶
stall_subprocess_submission_mem_threshold_pct (Number)
Default value 2
Stall submission of tasks if system memory available is less than this threshold in percent (set to 0 to disable). Above this threshold, the number of workers in any pool of workers is linearly reduced down to 1 once hitting this threshold.
max_cores_by_physical
¶
max_cores_by_physical (Boolean)
Default value True
Whether to set automatic number of cores by physical (True) or logical (False) count. Using all logical cores can lead to poor performance due to cache thrashing.
max_cores_limit
¶
max_cores_limit (Number)
Default value 100
Absolute limit to core count
assumed_simultaneous_dt_forks_stats_openblas
¶
assumed_simultaneous_dt_forks_stats_openblas (Number)
Default value 1
Expected maximum number of forks by computing statistics during ingestion, used to ensure datatable doesn’t overload system
max_max_dt_threads_stats_openblas
¶
max_max_dt_threads_stats_openblas (Number)
Default value 8
Expected maximum of threads for datatable no matter if many more cores
max_dt_threads_do_timeseries_split_suggestion
¶
max_dt_threads_do_timeseries_split_suggestion (Number)
Default value 1
Maximum number of threads for datatable during TS properties preview panel computations).
kaggle_username
¶
Kaggle username (String)
Default value ''
Kaggle username for automatic submission and scoring of test set predictions. See https://github.com/Kaggle/kaggle-api#api-credentials for details on how to obtain Kaggle API credentials”,
kaggle_key
¶
Kaggle key (String)
Default value ''
Kaggle key for automatic submission and scoring of test set predictions. See https://github.com/Kaggle/kaggle-api#api-credentials for details on how to obtain Kaggle API credentials”,
kaggle_timeout
¶
Kaggle submission timeout in seconds (Number)
Default value 120
Max. number of seconds to wait for Kaggle API call to return scores for given predictions
kaggle_keep_submission
¶
Whether to keep Kaggle submission file in experiment directory (Boolean)
Default value False
kaggle_competitions
¶
Custom Kaggle competitions to make automatic test set submissions for. (String)
Default value ''
If provided, can extend the list to arbitrary and potentially future Kaggle competitions to make submissions for. Only used if kaggle_key and kaggle_username are provided. Provide a quoted comma-separated list of tuples (target column name, number of test rows, competition, metric) like this: kaggle_competitions=’(“target”, 200000, “santander-customer-transaction-prediction”, “AUC”), (“TARGET”, 75818, “santander-customer-satisfaction”, “AUC”)’
ping_period
¶
ping_period (Number)
Default value 60
Period (in seconds) of ping by Driverless AI server to each experiment (in order to get logger info like disk space and memory usage). 0 means don’t print anything.
ping_autodl
¶
Whether to enable ping of system status during DAI experiments. (Boolean)
Default value True
Whether to enable ping of system status during DAI experiments.
disk_limit_gb
¶
disk_limit_gb (Number)
Default value 5
Minimum amount of disk space in GB needed to run experiments. Experiments will fail if this limit is crossed. This limit exists because Driverless AI needs to generate data for model training feature engineering, documentation and other such processes.
stall_disk_limit_gb
¶
stall_disk_limit_gb (Number)
Default value 1
Minimum amount of disk space in GB needed to before stall forking of new processes during an experiment.
memory_limit_gb
¶
memory_limit_gb (Number)
Default value 5
Minimum amount of system memory in GB needed to start experiments. Similarly with disk space, a certain amount of system memory is needed to run some basic operations.
min_num_rows
¶
Min. number of rows needed to run experiment (Number)
Default value 100
Minimum number of rows needed to run experiments (values lower than 100 might not work). A minimum threshold is set to ensure there is enough data to create a statistically reliable model and avoid other small-data related failures.
min_rows_per_class
¶
min_rows_per_class (Number)
Default value 5
Minimum required number of rows (in the training data) for each class label for classification problems.
min_rows_per_split
¶
min_rows_per_split (Number)
Default value 5
Minimum required number of rows for each split when generating validation samples.
reproducibility_level
¶
Reproducibility Level (Number)
Default value 1
Level of reproducibility desired (for same data and same inputs). Only active if ‘reproducible’ mode is enabled (GUI button enabled or a seed is set from the client API). Supported levels are:
reproducibility_level = 1 for same experiment results as long as same O/S, same CPU(s) and same GPU(s) reproducibility_level = 2 for same experiment results as long as same O/S, same CPU architecture and same GPU architecture reproducibility_level = 3 for same experiment results as long as same O/S, same CPU architecture, not using GPUs reproducibility_level = 4 for same experiment results as long as same O/S, (best effort)
seed
¶
Random seed (Number)
Default value 1234
Seed for random number generator to make experiments reproducible, to a certain reproducibility level (see above). Only active if ‘reproducible’ mode is enabled (GUI button enabled or a seed is set from the client API).
missing_values
¶
missing_values (List)
Default value ['', '?', 'None', 'nan', 'NA', 'N/A', 'unknown', 'inf', '-inf', '1.7976931348623157e+308', '-1.7976931348623157e+308']
The list of values that should be interpreted as missing values during data import. This applies to both numeric and string columns. Note that the dataset must be reloaded after applying changes to this config via the expert settings. Also note that ‘nan’ is always interpreted as a missing value for numeric columns.
glm_nan_impute_training_data
¶
glm_nan_impute_training_data (Boolean)
Default value False
Whether to impute (to mean) for GLM on training data.
glm_nan_impute_validation_data
¶
glm_nan_impute_validation_data (Boolean)
Default value False
Whether to impute (to mean) for GLM on validation data.
glm_nan_impute_prediction_data
¶
glm_nan_impute_prediction_data (Boolean)
Default value True
Whether to impute (to mean) for GLM on prediction data (required for consistency with MOJO).
tf_nan_impute_value
¶
tf_nan_impute_value (Number)
Default value -5
For tensorflow, what numerical value to give to missing values, where numeric values are standardized. So 0 is center of distribution, and if Normal distribution then +-5 is 5 standard deviations away from the center. In many cases, an out of bounds value is a good way to represent missings, but in some cases the mean (0) may be better.
statistical_threshold_data_size_small
¶
statistical_threshold_data_size_small (Number)
Default value 100000
Internal threshold for number of rows x number of columns to trigger certain statistical techniques (small data recipe like including one hot encoding for all model types, and smaller learning rate) to increase model accuracy
statistical_threshold_data_size_large
¶
statistical_threshold_data_size_large (Number)
Default value 500000000
Internal threshold for number of rows x number of columns to trigger certain statistical techniques (fewer genes created, removal of high max_depth for tree models, etc.) that can speed up modeling. Also controls maximum rows used in training final model, by sampling statistical_threshold_data_size_large / columns number of rows
aux_threshold_data_size_large
¶
aux_threshold_data_size_large (Number)
Default value 10000000
Internal threshold for number of rows x number of columns to trigger sampling for auxiliary data uses, like imbalanced data set detection and bootstrap scoring sample size and iterations
set_method_sampling_row_limit
¶
set_method_sampling_row_limit (Number)
Default value 5000000
Internal threshold for set-based method for sampling without replacement. Can be 10x faster than np_random_choice internal optimized method, and up to 30x faster than np.random.choice to sample 250k rows from 1B rows etc.
performance_threshold_data_size_small
¶
performance_threshold_data_size_small (Number)
Default value 100000
Internal threshold for number of rows x number of columns to trigger certain changes in performance (fewer threads if beyond large value) to help avoid OOM or unnecessary slowdowns (fewer threads if lower than small value) to avoid excess forking of tasks
performance_threshold_data_size_large
¶
performance_threshold_data_size_large (Number)
Default value 100000000
Internal threshold for number of rows x number of columns to trigger certain changes in performance (fewer threads if beyond large value) to help avoid OOM or unnecessary slowdowns (fewer threads if lower than small value) to avoid excess forking of tasks
max_relative_cols_mismatch_allowed
¶
max_relative_cols_mismatch_allowed (Float)
Default value 0.5
Maximum fraction of mismatched columns to allow between train and either valid or test. Beyond this value the experiment will fail with invalid data error.
max_cols
¶
max_cols (Number)
Default value 10000000
Maximum number of columns to start an experiment. This threshold exists to constraint the # complexity and the length of the Driverless AI’s processes.
max_rows_col_stats
¶
max_rows_col_stats (Number)
Default value 1000000
Largest number of rows to use for column stats, otherwise sample randomly
max_rows_cv_in_cv_gini
¶
max_rows_cv_in_cv_gini (Number)
Default value 100000
Largest number of rows to use for cv in cv for target encoding when doing gini scoring test
max_rows_constant_model
¶
max_rows_constant_model (Number)
Default value 1000000
Largest number of rows to use for constant model fit, otherwise sample randomly
max_rows_final_ensemble_base_model_fold_scores
¶
max_rows_final_ensemble_base_model_fold_scores (Number)
Default value 1000000
Largest number of rows to use for final ensemble base model fold cores, otherwise sample randomly
max_rows_final_blender
¶
max_rows_final_blender (Number)
Default value 1000000
Largest number of rows to use for final ensemble blender, otherwise sample randomly
max_rows_final_train_score
¶
max_rows_final_train_score (Number)
Default value 5000000
Largest number of rows to use for final training score (no holdout), otherwise sample randomly
max_rows_final_roccmconf
¶
max_rows_final_roccmconf (Number)
Default value 1000000
Largest number of rows to use for final ROC, lift-gains, confusion matrix, residual, and actual vs. predicted. Otherwise sample randomly
max_rows_final_holdout_score
¶
max_rows_final_holdout_score (Number)
Default value 5000000
Largest number of rows to use for final holdout scores, otherwise sample randomly
max_rows_final_holdout_bootstrap_score
¶
max_rows_final_holdout_bootstrap_score (Number)
Default value 1000000
Largest number of rows to use for final holdout bootstrap scores, otherwise sample randomly
max_rows_leak
¶
Max. rows for leakage detection if wide rules used on wide data (Number)
Default value 100000
max_workers_fs
¶
Num. simultaneous predictions for feature selection (0 = auto) (Number)
Default value 0
How many workers to use for feature selection by permutation for predict phase. (0 = auto, > 0: min of DAI value and this value, < 0: exactly negative of this value)
max_workers_shift_leak
¶
Num. simultaneous fits for shift and leak checks if using LightGBM on CPU (0 = auto) (Number)
Default value 0
How many workers to use for shift and leakage checks if using LightGBM on CPU. (0 = auto, > 0: min of DAI value and this value, < 0: exactly negative of this value)
num_folds
¶
num_folds (Number)
Default value 3
Number of folds for models used during the feature engineering process. Increasing this will put a lower fraction of data into validation and more into training (e.g., num_folds=3 means 67%/33% training/validation splits). Actual value will vary for small or big data cases.
fold_balancing_repeats_times_rows
¶
fold_balancing_repeats_times_rows (Float)
Default value 100000000.0
max_fold_balancing_repeats
¶
max_fold_balancing_repeats (Number)
Default value 10
fixed_split_seed
¶
fixed_split_seed (Number)
Default value 0
show_fold_stats
¶
show_fold_stats (Boolean)
Default value True
allow_different_classes_across_fold_splits
¶
Allow different sets of classes across all train/validation fold splits (Boolean)
Default value True
For multiclass problems only. Whether to allow different sets of target classes across (cross-)validation fold splits. Especially important when passing a fold column that isn’t balanced w.r.t class distribution.
full_cv_accuracy_switch
¶
full_cv_accuracy_switch (Number)
Default value 9
Accuracy setting equal and above which enables full cross-validation (multiple folds) during feature evolution as opposed to only a single holdout split (e.g. 2/3 train and 1/3 validation holdout)
ensemble_accuracy_switch
¶
ensemble_accuracy_switch (Number)
Default value 5
Accuracy setting equal and above which enables stacked ensemble as final model. Stacking commences at the end of the feature evolution process.. It quite often leads to better model performance, but it does increase the complexity and execution time of the final model.
num_ensemble_folds
¶
num_ensemble_folds (Number)
Default value 4
Number of fold splits to use for ensemble_level >= 2. The ensemble modelling may require predictions to be made on out-of-fold samples hence the data needs to be split on different folds to generate these predictions. Less folds (like 2 or 3) normally create more stable models, but may be less accurate More folds can get to higher accuracy at the expense of more time, but the performance may be less stable when the training data is not enough (i.e. higher chance of overfitting). Actual value will vary for small or big data cases.
save_validation_splits
¶
Store internal validation split row indices (Boolean)
Default value False
Includes pickles of (train_idx, valid_idx) tuples (numpy row indices for original training data) for all internal validation folds in the experiment summary zip. For debugging.
fold_reps
¶
fold_reps (Number)
Default value 1
Number of repeats for each fold for all validation (modified slightly for small or big data cases)
max_num_classes_hard_limit
¶
max_num_classes_hard_limit (Number)
Default value 10000
max_num_classes
¶
Max. number of classes for classification problems (Number)
Default value 200
Maximum number of classes to allow for a classification problem. High number of classes may make certain processes of Driverless AI time-consuming. Memory requirements also increase with higher number of classes
max_num_classes_compute_roc
¶
Max. number of classes to compute ROC and confusion matrix for classification problems (Number)
Default value 200
Maximum number of classes to compute ROC and CM for, beyond which roc_reduce_type choice for reduction is applied. Too many classes can take much longer than model building time.
max_num_classes_client_and_gui
¶
Max. number of classes to show in GUI for confusion matrix (Number)
Default value 10
Maximum number of classes to show in GUI for confusion matrix, showing first max_num_classes_client_and_gui labels. Beyond 6 classes the diagnostics launched from GUI are visually truncated. This will only modify client-GUI launched diagnostics if changed in config.toml and server is restarted, while this value can be changed in expert settings to control experiment plots.
roc_reduce_type
¶
ROC/CM reduction technique for large class counts (String)
Default value 'rows'
If too many classes when computing roc, reduce by “rows” by randomly sampling rows, or reduce by truncating classes to no more than max_num_classes_compute_roc. If have sufficient rows for class count, can reduce by rows.
min_roc_sample_size
¶
min_roc_sample_size (Number)
Default value 1
max_rows_cm_ga
¶
Maximum number of rows to obtain confusion matrix related plots during feature evolution (Number)
Default value 500000
Maximum number of rows to obtain confusion matrix related plots during feature evolution. Does not limit final model calculation.
num_actuals_vs_predicted
¶
num_actuals_vs_predicted (Number)
Default value 100
Number of actuals vs. predicted data points to use in order to generate in the relevant plot/graph which is shown at the right part of the screen within an experiment.
use_feature_brain_new_experiments
¶
Whether to use Feature Brain for new experiments. (Boolean)
Default value False
- Whether to use feature_brain results even if running new experiments.
Feature brain can be risky with some types of changes to experiment setup. Even rescoring may be insufficient, so by default this is False. For example, one experiment may have training=external validation by accident, and get high score, and while feature_brain_reset_score=’on’ means we will rescore, it will have already seen during training the external validation and leak that data as part of what it learned from. If this is False, feature_brain_level just sets possible models to use and logs/notifies, but does not use these feature brain cached models.
resume_data_schema
¶
Whether to reuse dataset schema. (String)
Default value 'auto'
Whether reuse dataset schema, such as data types set in UI for each column, from parent experiment (‘on’) or to ignore original dataset schema and only use new schema (‘off’). resume_data_schema=True is a basic form of data lineage, but it may not be desirable if data colunn names changed to incompatible data types like int to string. ‘auto’: for restart, retrain final pipeline, or refit best models, default is to resume data schema, but new experiments would not by default reuse old schema. ‘on’: force reuse of data schema from parent experiment if possible ‘off’: don’t reuse data schema under any case. The reuse of the column schema can also be disabled by: in UI: selecting Parent Experiment as None in client: setting resume_experiment_id to None
resume_data_schema_old_logic
¶
resume_data_schema_old_logic (Boolean)
Default value False
feature_brain_level
¶
Model/Feature Brain Level (0..10) (Number)
Default value 2
Whether to show (or use) results from H2O.ai brain: the local caching and smart re-use of prior experiments, in order to generate more useful features and models for new experiments. See use_feature_brain_new_experiments for how new experiments by default do not use brain cache. It can also be used to control checkpointing for experiments that have been paused or interrupted. DAI will use H2O.ai brain cache if cache file has a) any matching column names and types for a similar experiment type b) exactly matches classes c) exactly matches class labels d) matches basic time series choices e) interpretability of cache is equal or lower f) main model (booster) is allowed by new experiment. Level of brain to use (for chosen level, where higher levels will also do all lower level operations automatically) -1 = Don’t use any brain cache and don’t write any cache 0 = Don’t use any brain cache but still write cache
Use case: Want to save model for later use, but want current model to be built without any brain models
- 1 = smart checkpoint from latest best individual model
Use case: Want to use latest matching model, but match can be loose, so needs caution
- 2 = smart checkpoint from H2O.ai brain cache of individual best models
Use case: DAI scans through H2O.ai brain cache for best models to restart from
- 3 = smart checkpoint like level #1, but for entire population. Tune only if brain population insufficient size
(will re-score entire population in single iteration, so appears to take longer to complete first iteration)
- 4 = smart checkpoint like level #2, but for entire population. Tune only if brain population insufficient size
(will re-score entire population in single iteration, so appears to take longer to complete first iteration)
- 5 = like #4, but will scan over entire brain cache of populations to get best scored individuals
(can be slower due to brain cache scanning if big cache)
- 1000 + feature_brain_level (above positive values) = use resumed_experiment_id and actual feature_brain_level,
to use other specific experiment as base for individuals or population, instead of sampling from any old experiments
GUI has 3 options and corresponding settings: 1) New Experiment: Uses feature brain level default of 2 2) New Experiment With Same Settings: Re-uses the same feature brain level as parent experiment 3) Restart From Last Checkpoint: Resets feature brain level to 1003 and sets experiment ID to resume from
(continued genetic algorithm iterations)
Retrain Final Pipeline: Like Restart but also time=0 so skips any tuning and heads straight to final model (assumes had at least one tuning iteration in parent experiment)
Other use cases: a) Restart on different data: Use same column names and fewer or more rows (applicable to 1 - 5) b) Re-fit only final pipeline: Like (a), but choose time=1 and feature_brain_level=3 - 5 c) Restart with more columns: Add columns, so model builds upon old model built from old column names (1 - 5) d) Restart with focus on model tuning: Restart, then select feature_engineering_effort = 3 in expert settings e) can retrain final model but ignore any original features except those in final pipeline (normal retrain but set brain_add_features_for_new_columns=false) Notes: 1) In all cases, we first check the resumed experiment id if given, and then the brain cache 2) For Restart cases, may want to set min_dai_iterations to non-zero to force delayed early stopping, else may not be enough iterations to find better model. 3) A “New experiment with Same Settings” of a Restart will use feature_brain_level=1003 for default Restart mode (revert to 2, or even 0 if want to start a fresh experiment otherwise)
feature_brain_reset_score
¶
Whether to re-score models from brain cache (String)
Default value 'auto'
- Whether to smartly keep score to avoid re-munging/re-training/re-scoring steps brain models (‘auto’), always
force all steps for all brain imports (‘on’), or never rescore (‘off’). ‘auto’ only re-scores if a difference in current and prior experiment warrants re-scoring, like column changes, metric changes, etc. ‘on’ is useful when smart similarity checking is not reliable enough. ‘off’ is uesful when know want to keep exact same features and model for final model refit, despite changes in seed or other behaviors in features that might change the outcome if re-scored before reaching final model. If set off, then no limits are applied to features during brain ingestion, while can set brain_add_features_for_new_columns to false if want to ignore any new columns in data. In addition, any unscored individuals loaded from parent experiment are not rescored when doing refit or retrain. Can also set refit_same_best_individual True if want exact same best individual (highest scored model+features) to be used regardless of any scoring changes.
enable_strict_confict_key_check_for_brain
¶
enable_strict_confict_key_check_for_brain (Boolean)
Default value True
allow_change_layer_count_brain
¶
For feature brain or restart/refit, whether to allow brain ingest to use different feature engineering layer count. (Boolean)
Default value False
brain_maximum_diff_score
¶
brain_maximum_diff_score (Float)
Default value 0.1
Relative number of columns that must match between current reference individual and brain individual. 0.0: perfect match 1.0: All columns are different, worst match e.g. 0.1 implies no more than 10% of columns mismatch between reference set of columns and brain individual.
max_num_brain_indivs
¶
max_num_brain_indivs (Number)
Default value 3
Maximum number of brain individuals pulled from H2O.ai brain cache for feature_brain_level=1, 2
feature_brain_save_every_iteration
¶
Feature Brain Save every which iteration (0 = disable) (Number)
Default value 0
Save feature brain iterations every iter_num % feature_brain_iterations_save_every_iteration == 0, to be able to restart/refit with which_iteration_brain >= 0 0 means disable
which_iteration_brain
¶
Feature Brain Restart from which iteration (-1 = auto) (Number)
Default value -1
When doing restart or re-fit type feature_brain_level with resumed_experiment_id, choose which iteration to start from, instead of only last best -1 means just use last best Usage: 1) Run one experiment with feature_brain_iterations_save_every_iteration=1 or some other number 2) Identify which iteration brain dump one wants to restart/refit from 3) Restart/Refit from original experiment, setting which_iteration_brain to that number in expert settings Note: If restart from a tuning iteration, this will pull in entire scored tuning population and use that for feature evolution
refit_same_best_individual
¶
Feature Brain refit uses same best individual (Boolean)
Default value False
When doing re-fit from feature brain, if change columns or features, population of individuals used to refit from may change order of which was best, leading to better result chosen (False case). But sometimes want to see exact same model/features with only one feature added, and then would need to set this to True case. E.g. if refit with just 1 extra column and have interpretability=1, then final model will be same features, with one more engineered feature applied to that new original feature.
restart_refit_redo_origfs_shift_leak
¶
For restart-refit, select which steps to do (List)
Default value []
When doing restart or re-fit of experiment from feature brain, sometimes user might change data significantly and then warrant redoing reduction of original features by feature selection, shift detection, and leakage detection. However, in other cases, if data and all options are nearly (or exactly) identical, then these steps might change the features slightly (e.g. due to random seed if not setting reproducible mode), leading to changes in features and model that is refitted. By default, restart and refit avoid these steps assuming data and experiment setup have no changed significantly. If check_distribution_shift is forced to on (instead of auto), then this option is ignored. In order to ensure exact same final pipeline is fitted, one should also set: 1) brain_add_features_for_new_columns false 2) refit_same_best_individual true 3) feature_brain_reset_score ‘off’ 4) force_model_restart_to_defaults false The score will still be reset if the experiment metric chosen changes, but changes to the scored model and features will be more frozen in place.
brain_rel_dir
¶
brain_rel_dir (String)
Default value 'H2O.ai_brain'
Directory, relative to data_directory, to store H2O.ai brain meta model files
brain_max_size_GB
¶
brain_max_size_GB (Number)
Default value 20
Maximum size in bytes the brain will store We reserve this memory to save data in order to ensure we can retrieve an experiment if for any reason it gets interrupted. -1: unlimited >=0 number of GB to limit brain to
brain_add_features_for_new_columns
¶
Feature Brain adds features with new columns even during retraining final model (Boolean)
Default value True
Whether to take any new columns and add additional features to pipeline, even if doing retrain final model. In some cases, one might have a new dataset but only want to keep same pipeline regardless of new columns, in which case one sets this to False. For example, new data might lead to new dropped features, due to shift or leak detection. To avoid change of feature set, one can disable all dropping of columns, but set this to False to avoid adding any columns as new features, so pipeline is perfectly preserved when changing data.
force_model_restart_to_defaults
¶
Restart-refit use default model settings if model switches (Boolean)
Default value True
If restart/refit and no longer have the original model class available, be conservative and go back to defaults for that model class. If False, then try to keep original hyperparameters, which can fail to work in general.
early_stopping
¶
early_stopping (Boolean)
Default value True
Whether to enable early stopping Early stopping refers to stopping the feature evolution/engineering process when there is no performance uplift after a certain number of iterations. After early stopping has been triggered, Driverless AI will initiate the ensemble process if selected.
early_stopping_per_individual
¶
early_stopping_per_individual (Boolean)
Default value True
Whether to enable early stopping per individual Each individual in the generic algorithm will stop early if no improvement, and it will no longer be mutated. Instead, the best individual will be additionally mutated.
min_dai_iterations
¶
Min. DAI iterations (Number)
Default value 0
Minimum number of Driverless AI iterations to stop the feature evolution/engineering process even if score is not improving. Driverless AI needs to run for at least that many iterations before deciding to stop. It can be seen a safeguard against suboptimal (early) convergence.
tensorflow_nlp_have_gpus_in_production
¶
tensorflow_nlp_have_gpus_in_production (Boolean)
Default value False
bert_migration_timeout_secs
¶
bert_migration_timeout_secs (Number)
Default value 600
enable_bert_transformer_acceptance_test
¶
enable_bert_transformer_acceptance_test (Boolean)
Default value False
enable_bert_model_acceptance_test
¶
enable_bert_model_acceptance_test (Boolean)
Default value False
string_col_as_text_min_relative_cardinality
¶
string_col_as_text_min_relative_cardinality (Float)
Default value 0.1
Mininum fraction of unique values for string columns to be considered as possible text (otherwise categorical)
string_col_as_text_min_absolute_cardinality
¶
string_col_as_text_min_absolute_cardinality (Number)
Default value 10000
Mininum number of uniques for string columns to be considered as possible text (if not already)
supported_image_types
¶
supported_image_types (List)
Default value ['jpg', 'jpeg', 'png', 'bmp', 'ppm', 'tif', 'tiff', 'JPG', 'JPEG', 'PNG', 'BMP', 'PPM', 'TIF', 'TIFF']
Supported image types. URIs with these endings will be considered as image paths (local or remote).
image_paths_absolute
¶
image_paths_absolute (Boolean)
Default value False
Whether to create absolute paths for images when importing datasets containing images. Can faciliate testing or re-use of frames for scoring.
text_dl_token_pad_percentile
¶
text_dl_token_pad_percentile (Number)
Default value 99
Percentile value cutoff of input text token lengths for nlp deep learning models
text_dl_token_pad_max
¶
text_dl_token_pad_max (Number)
Default value 512
Maximum token length of input text to be used in nlp deep learning models
tune_parameters_accuracy_switch
¶
tune_parameters_accuracy_switch (Number)
Default value 3
Accuracy setting equal and above which enables tuning of model parameters Only applicable if parameter_tuning_num_models=-1 (auto)
tune_target_transform_accuracy_switch
¶
tune_target_transform_accuracy_switch (Number)
Default value 5
Accuracy setting equal and above which enables tuning of target transform for regression. This is useful for time series when instead of predicting the actual target value, it might be better to predict a transformed target variable like sqrt(target) or log(target) as a means to control for outliers.
target_transformer
¶
Select target transformation of the target for regression problems (String)
Default value 'auto'
Select a target transformation for regression problems. Must be one of: [‘auto’, ‘identity’, ‘identity_noclip’, ‘center’, ‘standardize’, ‘unit_box’, ‘log’, ‘log_noclip’, ‘square’, ‘sqrt’, ‘double_sqrt’, ‘inverse’, ‘anscombe’, ‘logit’, ‘sigmoid’]. If set to ‘auto’, will automatically pick the best target transformer (if accuracy is set to tune_target_transform_accuracy_switch or larger, considering interpretability level of each target transformer), otherwise will fall back to ‘identity_no_clip’ (easiest to interpret, Shapley values are in original space, etc.). All transformers except for ‘center’, ‘standardize’, ‘identity_noclip’ and ‘log_noclip’ perform clipping to constrain the predictions to the domain of the target in the training data. Use ‘center’, ‘standardize’, ‘identity_noclip’ or ‘log_noclip’ to disable clipping and to allow predictions outside of the target domain observed in the training data (for parametric models or custom models that support extrapolation).
target_transformer_tuning_choices
¶
Select all allowed target transformations of the target for regression problems when doing target transformer tuning (List)
Default value ['identity', 'identity_noclip', 'center', 'standardize', 'unit_box', 'log', 'square', 'sqrt', 'double_sqrt', 'anscombe', 'logit', 'sigmoid']
Select list of target transformers to use for tuning. Only for target_transformer=’auto’ and accuracy >= tune_target_transform_accuracy_switch.
tournament_style
¶
Tournament model for genetic algorithm (String)
Default value 'auto'
Tournament style (method to decide which models are best at each iteration) ‘auto’ : Choose based upon accuracy and interpretability ‘uniform’ : all individuals in population compete to win as best (can lead to all, e.g. LightGBM models in final ensemble, which may not improve ensemble performance due to lack of diversity) ‘model’ : individuals with same model type compete (good if multiple models do well but some models that do not do as well still contribute to improving ensemble) ‘feature’ : individuals with similar feature types compete (good if target encoding, frequency encoding, and other feature sets lead to good results) ‘fullstack’ : Choose among optimal model and feature types ‘model’ and ‘feature’ styles preserve at least one winner for each type (and so 2 total indivs of each type after mutation) For each case, a round robin approach is used to choose best scores among type of models to choose from. If enable_genetic_algorithm==’Optuna’, then every individual is self-mutated without any tournament during the genetic algorithm. The tournament is only used to prune-down individuals for, e.g., tuning -> evolution and evolution -> final model.
tournament_uniform_style_interpretability_switch
¶
tournament_uniform_style_interpretability_switch (Number)
Default value 8
Interpretability above which will use ‘uniform’ tournament style
tournament_uniform_style_accuracy_switch
¶
tournament_uniform_style_accuracy_switch (Number)
Default value 6
Accuracy below which will use uniform style if tournament_style = ‘auto’ (regardless of other accuracy tournament style switch values)
tournament_model_style_accuracy_switch
¶
tournament_model_style_accuracy_switch (Number)
Default value 6
Accuracy equal and above which uses model style if tournament_style = ‘auto’
tournament_feature_style_accuracy_switch
¶
tournament_feature_style_accuracy_switch (Number)
Default value 13
Accuracy equal and above which uses feature style if tournament_style = ‘auto’
tournament_fullstack_style_accuracy_switch
¶
tournament_fullstack_style_accuracy_switch (Number)
Default value 13
Accuracy equal and above which uses fullstack style if tournament_style = ‘auto’
tournament_use_feature_penalized_score
¶
tournament_use_feature_penalized_score (Boolean)
Default value True
Whether to use penalized score for GA tournament or actual score
num_individuals
¶
num_individuals (Number)
Default value 2
Driverless AI uses a genetic algorithm (GA) to find the best features, best models and best hyper parameters for these models. The GA facilitates getting good results while not requiring torun/try every possible model/feature/parameter. This version of GA has reinforcement learning elements - it uses a form of exploration-exploitation to reach optimum solutions. This means it will capitalise on models/features/parameters that seem # to be working well and continue to exploit them even more, while allowing some room for trying new (and semi-random) models/features/parameters to avoid settling on a local minimum. These models/features/parameters tried are what-we-call individuals of a population. More # individuals connote more models/features/parameters to be tried and compete to find the best # ones.
fixed_fold_reps
¶
fixed_fold_reps (Number)
Default value 0
set fixed number of fold reps (if > 0) - useful for quick runs regardless of data
sanitize_natural_sort_limit
¶
sanitize_natural_sort_limit (Number)
Default value 1000
number of unique targets or folds counts after which switch to faster/simpler non-natural sorting and print outs
excluded_transformers
¶
Exclude specific transformers (List)
Default value []
- Auxiliary to included_transformers
e.g. to disable all Target Encoding: excluded_transformers = ‘[‘NumCatTETransformer’, ‘CVTargetEncodeF’, ‘NumToCatTETransformer’, ‘ClusterTETransformer’]’. Does not affect transformers used for preprocessing with included_pretransformers.
excluded_genes
¶
Exclude specific genes (List)
Default value []
Exclude list of genes (i.e. genes (built on top of transformers) to not use, independent of the interpretability setting) Some transformers are used by multiple genes, so this allows different control over feature engineering
for multi-class: ‘[‘InteractionsGene’, ‘WeightOfEvidenceGene’, ‘NumToCatTargetEncodeSingleGene’, ‘OriginalGene’, ‘TextGene’, ‘FrequentGene’, ‘NumToCatWeightOfEvidenceGene’, ‘NumToCatWeightOfEvidenceMonotonicGene’, ‘ CvTargetEncodeSingleGene’, ‘DateGene’, ‘NumToCatTargetEncodeMultiGene’, ‘ DateTimeGene’, ‘TextLinRegressorGene’, ‘ClusterIDTargetEncodeSingleGene’, ‘CvCatNumEncodeGene’, ‘TruncSvdNumGene’, ‘ClusterIDTargetEncodeMultiGene’, ‘NumCatTargetEncodeMultiGene’, ‘CvTargetEncodeMultiGene’, ‘TextLinClassifierGene’, ‘NumCatTargetEncodeSingleGene’, ‘ClusterDistGene’]’
for regression/binary: ‘[‘CvTargetEncodeSingleGene’, ‘NumToCatTargetEncodeSingleGene’, ‘CvCatNumEncodeGene’, ‘ClusterIDTargetEncodeSingleGene’, ‘TextLinRegressorGene’, ‘CvTargetEncodeMultiGene’, ‘ClusterDistGene’, ‘OriginalGene’, ‘DateGene’, ‘ClusterIDTargetEncodeMultiGene’, ‘NumToCatTargetEncodeMultiGene’, ‘NumCatTargetEncodeMultiGene’, ‘TextLinClassifierGene’, ‘WeightOfEvidenceGene’, ‘FrequentGene’, ‘TruncSvdNumGene’, ‘InteractionsGene’, ‘TextGene’, ‘DateTimeGene’, ‘NumToCatWeightOfEvidenceGene’, ‘NumToCatWeightOfEvidenceMonotonicGene’, ‘’NumCatTargetEncodeSingleGene’]’
This list appears in the experiment logs (search for ‘Genes used’) e.g. to disable interaction gene, use: excluded_genes = ‘[‘InteractionsGene’]’. Does not affect transformers used for preprocessing with included_pretransformers.
excluded_models
¶
Exclude specific models (List)
Default value []
Auxiliary to included_models
excluded_pretransformers
¶
Exclude specific pretransformers (List)
Default value []
Auxiliary to included_pretransformers
include_all_as_pretransformers_if_none_selected
¶
include_all_as_pretransformers_if_none_selected (Boolean)
Default value False
force_include_all_as_pretransformers_if_none_selected
¶
force_include_all_as_pretransformers_if_none_selected (Boolean)
Default value False
excluded_datas
¶
Exclude specific data recipes (List)
Default value []
Auxiliary to included_datas
excluded_individuals
¶
Exclude specific individual recipes (List)
Default value []
Auxiliary to included_individuals
excluded_scorers
¶
Exclude specific scorers (List)
Default value []
Auxiliary to included_scorers
enable_glm_rapids
¶
Enable RAPIDS-cudf extensions to GLM (Boolean)
Default value False
Whether to enable RAPIDS extensions to GLM models (not available until fixes are in xgboost 1.3.0)
use_dask_for_1_gpu
¶
use_dask_for_1_gpu (Boolean)
Default value False
Whether to use dask_cudf even for 1 GPU. If False, will use plain cudf.
dask_retrials_allreduce_empty_issue
¶
dask_retrials_allreduce_empty_issue (Number)
Default value 5
Number of retrials for dask fit to protect against known xgboost issues https://github.com/dmlc/xgboost/issues/6272 https://github.com/dmlc/xgboost/issues/6551
optuna_pruner_kwargs
¶
Set Optuna pruner constructor args. (Dict)
Default value {'n_startup_trials': 5, 'n_warmup_steps': 20, 'interval_steps': 20, 'percentile': 25.0, 'min_resource': 'auto', 'max_resource': 'auto', 'reduction_factor': 4, 'min_early_stopping_rate': 0, 'n_brackets': 4, 'min_early_stopping_rate_low': 0, 'upper': 1.0, 'lower': 0.0}
Set Optuna constructor arguments for particular applicable pruners. https://optuna.readthedocs.io/en/stable/reference/pruners.html
optuna_sampler_kwargs
¶
Set Optuna sampler constructor args. (Dict)
Default value {}
Set Optuna constructor arguments for particular applicable samplers. https://optuna.readthedocs.io/en/stable/reference/samplers.html
use_xgboost_xgbfi
¶
use_xgboost_xgbfi (Boolean)
Default value False
Whether to use (and expect exists) xgbfi feature interactions for xgboost.
drop_constant_model_final_ensemble
¶
drop_constant_model_final_ensemble (Boolean)
Default value True
xgboost_rf_exact_threshold_num_rows_x_cols
¶
xgboost_rf_exact_threshold_num_rows_x_cols (Number)
Default value 10000
lossguide_drop_factor
¶
Factor by which to drop max_leaves from effective max_depth value when doing loss_guide. E.g. if max_depth is normally 12, this makes leaves 2**11 not 2**12 (Float)
Default value 4.0
lossguide_max_depth_extend_factor
¶
Factor by which to extend max_depth mutations when doing loss_guide. E.g. if max_leaves ends up as x let max_depth be factor * x. (Float)
Default value 8.0
params_tune_grow_policy_simple_trees
¶
params_tune_grow_policy_simple_trees (Boolean)
Default value True
Whether to force max_leaves and max_depth to be 0 if grow_policy is depthwise and lossguide, respectively.
default_max_bin
¶
default_max_bin (Number)
Default value 256
Default max_bin for tree methods
default_lightgbm_max_bin
¶
default_lightgbm_max_bin (Number)
Default value 249
Default max_bin for LightGBM (64 recommended for GPU LightGBM for speed)
min_max_bin
¶
min_max_bin (Number)
Default value 32
Minimum max_bin for any tree
scale_mem_for_max_bin
¶
scale_mem_for_max_bin (Number)
Default value 10737418240
Amount of memory which can handle max_bin = 256 can handle 125 columns and max_bin = 32 for 1000 columns As available memory on system goes higher than this scale, can handle proportionally more columns at higher max_bin Currently set to 10GB
factor_rf
¶
factor_rf (Float)
Default value 1.25
Factor by which rf gets more depth than gbdt
tensorflow_use_all_cores
¶
tensorflow_use_all_cores (Boolean)
Default value True
Whether TensorFlow will use all CPU cores, or if it will split among all transformers
tensorflow_use_all_cores_even_if_reproducible_true
¶
tensorflow_use_all_cores_even_if_reproducible_true (Boolean)
Default value False
Whether TensorFlow will use all CPU cores if reproducible is set, or if it will split among all transformers
tensorflow_disable_memory_optimization
¶
tensorflow_disable_memory_optimization (Boolean)
Default value True
Whether to disable TensorFlow memory optimizations. Can help fix tensorflow.python.framework.errors_impl.AlreadyExistsError
tensorflow_cores
¶
tensorflow_cores (Number)
Default value 0
How many cores to use for each TensorFlow model, regardless if GPU or CPU based (0 = auto mode)
validate_meta_learner
¶
Enable basic logging and notifications for ensemble meta learner (Boolean)
Default value True
validate_meta_learner_extra
¶
Enable extra logging for ensemble meta learner: ensemble must be at least as good as each base model (Boolean)
Default value False
fixed_num_folds_evolution
¶
Number of cross-validation folds for feature evolution (-1 = auto) (Number)
Default value -1
Specify the fixed number of cross-validation folds (if >= 2) for feature evolution. (The actual number of splits allowed can be less and is determined at experiment run-time).
fixed_num_folds
¶
Number of cross-validation folds for final model (-1 = auto) (Number)
Default value -1
Specify the fixed number of cross-validation folds (if >= 2) for the final model. (The actual number of splits allowed can be less and is determined at experiment run-time).
fixed_only_first_fold_model
¶
Force only first fold for models (String)
Default value 'auto'
set “on” to force only first fold for models - useful for quick runs regardless of data
num_fold_ids_show
¶
Maximum number of fold IDs to show in logs (Number)
Default value 10
fold_scores_instability_warning_threshold
¶
Declare positive fold scores as unstable if stddev / mean is larger than this value (Float)
Default value 0.25
feature_evolution_data_size
¶
Max. num. of rows x num. of columns for feature evolution data splits (not for final pipeline) (Number)
Default value 300000000
Upper limit on the number of rows x number of columns for feature evolution (applies to both training and validation/holdout splits) feature evolution is the process that determines which features will be derived. Depending on accuracy settings, a fraction of this value will be used
final_pipeline_data_size
¶
Max. num. of rows x num. of columns for reducing training data set (for final pipeline) (Number)
Default value 1000000000
Upper limit on the number of rows x number of columns for training final pipeline.
max_validation_to_training_size_ratio_for_final_ensemble
¶
Max. size of validation data relative to training data (for final pipeline), otherwise will sample (Float)
Default value 2.0
Smaller values can speed up final pipeline model training, as validation data is only used for early stopping. Note that final model predictions and scores will always be provided on the full dataset provided.
force_stratified_splits_for_imbalanced_threshold_binary
¶
Perform stratified sampling for binary classification if the target is more imbalanced than this. (Float)
Default value 0.01
Ratio of minority to majority class of the target column beyond which stratified sampling is done for binary classification. Otherwise perform random sampling. Set to 0 to always do random sampling. Set to 1 to always do stratified sampling.
force_stratified_splits_for_binary_max_rows
¶
Perform stratified sampling for binary classification if the dataset has fewer rows than this. (Number)
Default value 1000000
stratify_for_regression
¶
Perform stratified sampling for regression problems (using binning). (Boolean)
Default value True
Specify whether to do stratified sampling for validation fold creation for iid regression problems. Otherwise perform random sampling.
imbalance_ratio_multiclass_threshold
¶
Ratio of most frequent to least frequent class for imbalanced multiclass classification problems equal and above which to trigger special handling due to class imbalance (Number)
Default value 5
Special handling can include special models, special scorers, special feature engineering.
heavy_imbalance_ratio_multiclass_threshold
¶
Ratio of most frequent to least frequent class for imbalanced multiclass classification problems equal and above which to trigger special handling due to heavy class imbalance (Number)
Default value 25
Special handling can include special models, special scorers, special feature engineering.
imbalance_sampling_rank_averaging
¶
Whether to do rank averaging bagged models inside of imbalanced models, instead of probability averaging (String)
Default value 'auto'
- Rank averaging can be helpful when ensembling diverse models when ranking metrics like AUC/Gini
metrics are optimized. No MOJO support yet.
imbalance_ratio_notification_threshold
¶
imbalance_ratio_notification_threshold (Float)
Default value 2.0
For binary classification: ratio of majority to minority class equal and above which to notify
of imbalance in GUI to say slightly imbalanced.
More than imbalance_ratio_sampling_threshold
will say problem is imbalanced.
nbins_ftrl_list
¶
nbins_ftrl_list (List)
Default value [1000000, 10000000, 100000000]
List of possible bins for FTRL (largest is default best value)
te_bin_list
¶
te_bin_list (List)
Default value [25, 10, 100, 250]
List of possible bins for target encoding (first is default value)
woe_bin_list
¶
woe_bin_list (List)
Default value [25, 10, 100, 250]
List of possible bins for weight of evidence encoding (first is default value) If only want one value: woe_bin_list = [2]
ohe_bin_list
¶
ohe_bin_list (List)
Default value [10, 25, 50, 75, 100]
List of possible bins for ohe hot encoding (first is default value)
cols_to_drop_sanitized
¶
cols_to_drop_sanitized (List)
Default value []
cols_to_group_by_sanitized
¶
cols_to_group_by_sanitized (List)
Default value []
leaderboard_mode
¶
Control the automatic leaderboard mode (String)
Default value 'baseline'
‘baseline’: Explore explemplar set of models with baselines as reference. ‘random’: Explore 10 random seeds for same setup. Useful since nature of genetic algorithm is noisy and repeats might get better results, or one can ensemble the custom individuals from such repeats. ‘line’: Explore good model with all features and original features with all models. Useful as first exploration. ‘line_all’: Like ‘line’, but enable all models and transformers possible instead of only what base experiment setup would have inferred. ‘product’: Explore one-by-one Cartesian product of each model and transformer. Useful for exhaustive exploration.
default_knob_offset_accuracy
¶
Offset for default accuracy knob (Number)
Default value 0
- Allows control over default accuracy knob setting.
If default models are too complex, set to -1 or -2, etc. If default models are not accurate enough, set to 1 or 2, etc.
default_knob_offset_time
¶
Offset for default time knob (Number)
Default value 0
- Allows control over default time knob setting.
If default experiments are too slow, set to -1 or -2, etc. If default experiments finish too fast, set to 1 or 2, etc.
default_knob_offset_interpretability
¶
Offset for default interpretability knob (Number)
Default value 0
- Allows control over default interpretability knob setting.
If default models are too simple, set to -1 or -2, etc. If default models are too complex, set to 1 or 2, etc.
shift_check_text
¶
shift_check_text (Boolean)
Default value False
Whether to enable checking text for shift, currently only via label encoding.
use_rf_for_shift_if_have_lgbm
¶
use_rf_for_shift_if_have_lgbm (Boolean)
Default value True
Whether to use LightGBM random forest mode without early stopping for shift detection.
shift_key_features_varimp
¶
shift_key_features_varimp (Float)
Default value 0.01
Normalized training variable importance above which to check the feature for shift Useful to avoid checking likely unimportant features
shift_check_reduced_features
¶
shift_check_reduced_features (Boolean)
Default value True
Whether to only check certain features based upon the value of shift_key_features_varimp
shift_trees
¶
shift_trees (Number)
Default value 100
Number of trees to use to train model to check shift in distribution No larger than max_nestimators
shift_max_bin
¶
shift_max_bin (Number)
Default value 256
The value of max_bin to use for trees to use to train model to check shift in distribution
shift_min_max_depth
¶
shift_min_max_depth (Number)
Default value 4
The min. value of max_depth to use for trees to use to train model to check shift in distribution
shift_max_max_depth
¶
shift_max_max_depth (Number)
Default value 8
The max. value of max_depth to use for trees to use to train model to check shift in distribution
detect_features_distribution_shift_threshold_auc
¶
detect_features_distribution_shift_threshold_auc (Float)
Default value 0.55
If distribution shift detection is enabled, show features for which shift AUC is above this value (AUC of a binary classifier that predicts whether given feature value belongs to train or test data)
drop_features_distribution_shift_min_features
¶
drop_features_distribution_shift_min_features (Number)
Default value 1
Minimum number of features to keep, keeping least shifted feature at least if 1
shift_high_notification_level
¶
shift_high_notification_level (Float)
Default value 0.8
Shift beyond which shows HIGH notification, else MEDIUM
leakage_check_text
¶
leakage_check_text (Boolean)
Default value True
Whether to enable checking text for leakage, currently only via label encoding.
leakage_key_features_varimp
¶
leakage_key_features_varimp (Float)
Default value 0.001
Normalized training variable importance (per 1 minus AUC/R2 to control for leaky varimp dominance) above which to check the feature for leakage Useful to avoid checking likely unimportant features
leakage_key_features_varimp_if_no_early_stopping
¶
leakage_key_features_varimp_if_no_early_stopping (Float)
Default value 0.05
Like leakage_key_features_varimp, but applies if early stopping disabled when can trust multiple leaks to get uniform varimp.
leakage_check_reduced_features
¶
leakage_check_reduced_features (Boolean)
Default value True
Whether to only check certain features based upon the value of leakage_key_features_varimp. If any feature has AUC near 1, will consume all variable importance, even if another feature is also leaky. So False is safest option, but True generally good if many columns.
use_rf_for_leakage_if_have_lgbm
¶
use_rf_for_leakage_if_have_lgbm (Boolean)
Default value True
Whether to use LightGBM random forest mode without early stopping for leakage detection.
leakage_trees
¶
leakage_trees (Number)
Default value 100
Number of trees to use to train model to check for leakage No larger than max_nestimators
leakage_max_bin
¶
leakage_max_bin (Number)
Default value 256
The value of max_bin to use for trees to use to train model to check for leakage
leakage_min_max_depth
¶
leakage_min_max_depth (Number)
Default value 6
The value of max_depth to use for trees to use to train model to check for leakage
leakage_max_max_depth
¶
leakage_max_max_depth (Number)
Default value 8
The value of max_depth to use for trees to use to train model to check for leakage
drop_features_leakage_min_features
¶
drop_features_leakage_min_features (Number)
Default value 1
Minimum number of features to keep, keeping least leakage feature at least if 1
leakage_train_test_split
¶
leakage_train_test_split (Float)
Default value 0.25
Ratio of train to validation holdout when testing for leakage
check_system
¶
Whether to check system installation on server startup (Boolean)
Default value True
abs_tol_for_perfect_score
¶
abs_tol_for_perfect_score (Float)
Default value 0.0001
How close to the optimal value (usually 1 or 0) does the validation score need to be to be considered perfect (to stop the experiment)?
data_ingest_timeout
¶
data_ingest_timeout (Float)
Default value 86400.0
Timeout in seconds to wait for data ingestion.
gpu_locking_trust_pool_submission
¶
gpu_locking_trust_pool_submission (Boolean)
Default value True
- Whether to trust GPU locking for submission of GPU jobs to limit memory usage.
If False, then wait for as GPU submissions to be less than number of GPUs, even if later jobs could be purely CPU jobs that did not need to wait. Only applicable if not restricting number of GPUs via num_gpus_per_experiment, else have to use resources instead of relying upon locking.
gpu_locking_free_dead
¶
gpu_locking_free_dead (Boolean)
Default value True
Whether to steal GPU locks when process is neither on GPU PID list nor using CPU resources at all (e.g. sleeping). Only steal from multi-GPU locks that are incomplete. Prevents deadlocks in case multi-GPU model hangs.
tensorflow_allow_cpu_only
¶
tensorflow_allow_cpu_only (Boolean)
Default value False
check_pred_contribs_sum
¶
check_pred_contribs_sum (Boolean)
Default value False
debug_daimodel_level
¶
debug_daimodel_level (Number)
Default value 0
debug_debug_xgboost_splits
¶
debug_debug_xgboost_splits (Boolean)
Default value False
log_predict_info
¶
Whether to show detailed predict information in logs. (Boolean)
Default value True
log_fit_info
¶
Whether to show detailed fit information in logs. (Boolean)
Default value True
stalled_time_kill_ref
¶
stalled_time_kill_ref (Float)
Default value 440.0
Amount of time to stall (in seconds) before killing the job (assumes it hung). Reference time is scaled by train data shape of rows * cols to get used stalled_time_kill
long_time_psdump
¶
long_time_psdump (Number)
Default value 1800
Amount of time between checks for some process taking long time, every cycle full process list will be dumped to console or experiment logs if possible.
do_psdump
¶
do_psdump (Boolean)
Default value False
Whether to dump ps every long_time_psdump
livelock_signal
¶
livelock_signal (Boolean)
Default value False
Whether to check every long_time_psdump seconds and SIGUSR1 to all children to see where maybe stuck or taking long time.
num_cpu_sockets_override
¶
num_cpu_sockets_override (Number)
Default value 0
Value to override number of sockets, in case DAIs determination is wrong, for non-trivial systems. 0 means auto.
num_gpus_override
¶
num_gpus_override (Number)
Default value -1
Value to override number of GPUs, in case DAIs determination is wrong, for non-trivial systems. -1 means auto.Can also set min_num_cores_per_gpu=-1 to allowany number of GPUs for each experiment regardlessof number of cores.
show_gpu_usage_only_if_locked
¶
show_gpu_usage_only_if_locked (String)
Default value 'auto'
Whether to show GPU usage only when locking. ‘auto’ means ‘on’ if num_gpus_override is different than actual total visible GPUs, else it means ‘off’
show_inapplicable_models_preview
¶
show_inapplicable_models_preview (Boolean)
Default value False
Show inapplicable models in preview, to be sure not missing models one could have used
show_inapplicable_transformers_preview
¶
show_inapplicable_transformers_preview (Boolean)
Default value False
Show inapplicable transformers in preview, to be sure not missing transformers one could have used
show_warnings_preview
¶
show_warnings_preview (Boolean)
Default value False
Show warnings for models (image auto, Dask multinode/multi-GPU) if conditions are met to use but not chosen to avoid missing models that could benefit accuracy/performance
show_warnings_preview_unused_map_features
¶
show_warnings_preview_unused_map_features (Boolean)
Default value True
Show warnings for models that have no transformers for certain features.
max_cols_show_unused_features
¶
max_cols_show_unused_features (Number)
Default value 1000
Up to how many input features to determine, during GUI/client preview, unused features. Too many slows preview down.
max_cols_show_feature_transformer_mapping
¶
max_cols_show_feature_transformer_mapping (Number)
Default value 1000
Up to how many input features to show transformers used for each input feature.
warning_unused_feature_show_max
¶
warning_unused_feature_show_max (Number)
Default value 3
Up to how many input features to show, in preview, that are unused features.
interaction_finder_max_rows_x_cols
¶
interaction_finder_max_rows_x_cols (Float)
Default value 200000.0
interaction_finder_corr_threshold
¶
interaction_finder_corr_threshold (Float)
Default value 0.95
min_bootstrap_samples
¶
Minimum number of bootstrap samples (Number)
Default value 1
Minimum number of bootstrap samples to use for estimating score and its standard deviation Actual number of bootstrap samples will vary between the min and max, depending upon row count (more rows, fewer samples) and accuracy settings (higher accuracy, more samples)
max_bootstrap_samples
¶
Maximum number of bootstrap samples (Number)
Default value 100
Maximum number of bootstrap samples to use for estimating score and its standard deviation Actual number of bootstrap samples will vary between the min and max, depending upon row count (more rows, fewer samples) and accuracy settings (higher accuracy, more samples)
min_bootstrap_sample_size_factor
¶
Minimum fraction of rows to use for bootstrap samples (Float)
Default value 1.0
Minimum fraction of row size to take as sample size for bootstrap estimator Actual sample size used for bootstrap estimate will vary between the min and max, depending upon row count (more rows, smaller sample size) and accuracy settings (higher accuracy, larger sample size)
max_bootstrap_sample_size_factor
¶
Maximum fraction of rows to use for bootstrap samples (Float)
Default value 10.0
Maximum fraction of row size to take as sample size for bootstrap estimator Actual sample size used for bootstrap estimate will vary between the min and max, depending upon row count (more rows, smaller sample size) and accuracy settings (higher accuracy, larger sample size)
bootstrap_final_seed
¶
Seed to use for final model bootstrap sampling (Number)
Default value -1
Seed to use for final model bootstrap sampling, -1 means use experiment-derived seed. E.g. one can retrain final model with different seed to get different final model error bars for scores.
benford_mad_threshold_int
¶
benford_mad_threshold_int (Float)
Default value 0.03
Benford’s law: mean absolute deviance threshold equal and above which integer valued columns are treated as categoricals too
benford_mad_threshold_real
¶
benford_mad_threshold_real (Float)
Default value 0.1
Benford’s law: mean absolute deviance threshold equal and above which real valued columns are treated as categoricals too
stabilize_features
¶
Use tuning-evolution search result for final model transformer. (Boolean)
Default value True
- Whether final pipeline uses fixed features for some transformers that would normally
perform search, such as InteractionsTransformer. Use what learned from tuning and evolution (True) or to freshly search for new features (False). This can give a more stable pipeline, especially for small data or when using interaction transformer as pretransformer in multi-layer pipeline.
fraction_std_bootstrap_ladder_factor
¶
Factor of standard deviation of bootstrap scores by which to accept new model in genetic algorithm. Too small a fraction will lead to accepting new models easily even if no significant improvement in score, while too large a fraction will reject too many good models. Non-zero value is a bit risky when no folds are used in GA, because bootstrap score is only rough estimate of error. (Float)
Default value 0.01
bootstrap_ladder_samples_limit
¶
Minimum number of bootstrap samples that are required to limit accepting new model. If less than this, then new model is always accepted. (Number)
Default value 10
rdelta_percent_score_penalty_per_feature_by_interpretability
¶
rdelta_percent_score_penalty_per_feature_by_interpretability (String)
Default value '{1: 0.0, 2: 0.1, 3: 1.0, 4: 2.0, 5: 5.0, 6: 10.0, 7: 20.0, 8: 30.0, 9: 50.0, 10: 100.0, 11: 100.0, 12: 100.0, 13: 100.0}'
drop_low_meta_weights
¶
drop_low_meta_weights (Boolean)
Default value True
meta_weight_allowed_by_interpretability
¶
meta_weight_allowed_by_interpretability (String)
Default value '{1: 1E-7, 2: 1E-5, 3: 1E-4, 4: 1E-3, 5: 1E-2, 6: 0.03, 7: 0.05, 8: 0.08, 9: 0.10, 10: 0.15, 11: 0.15, 12: 0.15, 13: 0.15}'
meta_weight_allowed_for_reference
¶
Min. weight of meta learner for reference models during ensembling. If 1.0, then reference model must be the clear winner to be kept. Set to 0.0 to never drop reference models (Float)
Default value 1.0
show_full_pipeline_details
¶
Whether to show full pipeline details (Boolean)
Default value False
num_transformed_features_per_pipeline_show
¶
Number of features to show when logging size of fitted transformers (Number)
Default value 10
fs_data_vary_for_interpretability
¶
fs_data_vary_for_interpretability (Number)
Default value 7
fs_data_frac
¶
fs_data_frac (Float)
Default value 0.5
many_columns_count
¶
Number of columns beyond which reduce expensive tasks at cost of some accuracy. (Number)
Default value 400
columns_count_interpretable
¶
Number of columns beyond which do not set default knobs to high interpretability even if bigger data. (Number)
Default value 200
round_up_indivs_for_busy_gpus
¶
Whether to round-up individuals to ensure all GPUs used. Not always best if (say) have 16 GPUs, better to have multiple experiments if in multi-user environment on single node. (Boolean)
Default value True
check_timeout_per_gpu
¶
check_timeout_per_gpu (Number)
Default value 20
How long to wait per GPU for tensorflow/torch to run during system checks.
gpu_exit_if_fails
¶
gpu_exit_if_fails (Boolean)
Default value True
Whether to fail start-up if cannot successfully run GPU checks
require_graphviz
¶
Whether to require Graphviz package at startup (Boolean)
Default value True
Graphviz is an optional requirement for native installations (RPM/DEP/Tar-SH, outside of Docker)to convert .dot files into .png files for pipeline visualizations as part of experiment artifacts
fast_approx_max_num_trees_ever
¶
fast_approx_max_num_trees_ever (Number)
Default value -1
Max. number of trees to use for all tree model predictions. For testing, when predictions don’t matter. -1 means disabled.
fast_approx_num_trees
¶
fast_approx_num_trees (Number)
Default value 250
Max. number of trees to use for fast_approx=True (e.g., for AutoDoc/MLI).
fast_approx_do_one_fold
¶
fast_approx_do_one_fold (Boolean)
Default value True
Whether to speed up fast_approx=True further, by using only one fold out of all cross-validation folds (e.g., for AutoDoc/MLI).
fast_approx_do_one_model
¶
fast_approx_do_one_model (Boolean)
Default value False
Whether to speed up fast_approx=True further, by using only one model out of all ensemble models (e.g., for AutoDoc/MLI).
fast_approx_contribs_num_trees
¶
fast_approx_contribs_num_trees (Number)
Default value 50
Max. number of trees to use for fast_approx_contribs=True (e.g., for ‘Fast Approximation’ in GUI when making Shapley predictions, and for AutoDoc/MLI).
fast_approx_contribs_do_one_fold
¶
fast_approx_contribs_do_one_fold (Boolean)
Default value True
Whether to speed up fast_approx_contribs=True further, by using only one fold out of all cross-validation folds (e.g., for ‘Fast Approximation’ in GUI when making Shapley predictions, and for AutoDoc/MLI).
fast_approx_contribs_do_one_model
¶
fast_approx_contribs_do_one_model (Boolean)
Default value True
Whether to speed up fast_approx_contribs=True further, by using only one model out of all ensemble models (e.g., for ‘Fast Approximation’ in GUI when making Shapley predictions, and for AutoDoc/MLI).
use_187_prob_logic
¶
use_187_prob_logic (Boolean)
Default value True
Whether to use exploit-explore logic like DAI 1.8.x. False will explore more.
enable_ohe_linear
¶
enable_ohe_linear (Boolean)
Default value False
Whether to enable cross-validated OneHotEncoding+LinearModel transformer
max_absolute_feature_expansion
¶
max_absolute_feature_expansion (Number)
Default value 1000
booster_for_fs_permute
¶
booster_for_fs_permute (String)
Default value 'auto'
model_class_name_for_fs_permute
¶
model_class_name_for_fs_permute (String)
Default value 'auto'
switch_from_tree_to_lgbm_if_can
¶
switch_from_tree_to_lgbm_if_can (Boolean)
Default value True
default_model_class_name
¶
default_model_class_name (String)
Default value 'LightGBMModel'
textlin_num_classes_switch
¶
Class count above which do not use TextLin Transformer (Number)
Default value 5
Class count above which do not use TextLin Transformer.
text_gene_dim_reduction_choices
¶
text_gene_dim_reduction_choices (List)
Default value [50]
text_gene_max_ngram
¶
text_gene_max_ngram (List)
Default value [1, 2, 3]
number_of_texts_to_cache_in_bert_transformer
¶
number_of_texts_to_cache_in_bert_transformer (Number)
Default value -1
Enables caching of BERT embeddings by temporally saving the embedding vectors to the experiment directory. Set to -1 to cache all text, set to 0 to disable caching.
gbm_early_stopping_rounds_min
¶
gbm_early_stopping_rounds_min (Number)
Default value 1
gbm_early_stopping_rounds_max
¶
gbm_early_stopping_rounds_max (Number)
Default value 10000000000
max_varimp_to_save
¶
max_varimp_to_save (Number)
Default value 100
Max. number of top variable importances to save per iteration (GUI can only display a max. of 14)
max_num_varimp_to_log
¶
max_num_varimp_to_log (Number)
Default value 10
Max. number of top variable importances to show in logs during feature evolution
max_num_varimp_shift_to_log
¶
max_num_varimp_shift_to_log (Number)
Default value 10
Max. number of top variable importance shifts to show in logs and GUI after final model built
can_skip_final_upper_layer_failures
¶
can_skip_final_upper_layer_failures (Boolean)
Default value True
Whether can skip final model transformer failures for layer > first layer for multi-layer pipeline.
config_overrides
¶
Add to config.toml via toml string (String)
Default value ''
Instructions for ‘Add to config.toml via toml string’ in GUI expert page Self-referential toml parameter, for setting any other toml parameters as string of tomls separated by
(spaces around are ok).
Useful when toml parameter is not in expert mode but want per-experiment control. Setting this will override all other choices. In expert page, each time expert options saved, the new state is set without memory of any prior settings. The entered item is a fully compliant toml string that would be processed directly by toml.load(). One should include 2 double quotes around the entire setting, or double quotes need to be escaped. One enters into the expert page text as follows: e.g. ``enable_glm=”off”
enable_xgboost_gbm=”off” enable_lightgbm=”on”``
- e.g. ``””enable_glm=”off”
enable_xgboost_gbm=”off” enable_lightgbm=”off” enable_tensorflow=”on”””``
e.g. fixed_num_individuals=4
e.g. params_lightgbm="{'objective':'poisson'}"
e.g. ""params_lightgbm="{'objective':'poisson'}"""
e.g. ``max_cores=10
data_precision=”float32” max_rows_feature_evolution=50000000000 ensemble_accuracy_switch=11 feature_engineering_effort=1 target_transformer=”identity” tournament_feature_style_accuracy_switch=5 params_tensorflow=”{‘layers’: (100, 100, 100, 100, 100, 100)}”``
- e.g. “”max_cores=10
data_precision=”float32” max_rows_feature_evolution=50000000000 ensemble_accuracy_switch=11 feature_engineering_effort=1 target_transformer=”identity” tournament_feature_style_accuracy_switch=5 params_tensorflow=”{‘layers’: (100, 100, 100, 100, 100, 100)}”””
If you see: “toml.TomlDecodeError” then ensure toml is set correctly. When set in the expert page of an experiment, these changes only affect experiments and not the server Usually should keep this as empty string in this toml file.
dump_modelparams_every_scored_indiv_feature_count
¶
dump_modelparams_every_scored_indiv_feature_count (Number)
Default value 3
Number of features to show in model dump every scored individual
dump_modelparams_every_scored_indiv_mutation_count
¶
dump_modelparams_every_scored_indiv_mutation_count (Number)
Default value 3
Number of past mutations to show in model dump every scored individual
dump_modelparams_separate_files
¶
dump_modelparams_separate_files (Boolean)
Default value False
Whether to append (false) or have separate files, files like: individual_scored_id%d.iter%d*params*, (true) for modelparams every scored indiv
delete_preview_trans_timings
¶
delete_preview_trans_timings (Boolean)
Default value True
whether to delete preview timings if wrote transformer timings
use_random_text_file
¶
use_random_text_file (Boolean)
Default value False
runtime_estimation_train_frame
¶
runtime_estimation_train_frame (String)
Default value ''
enable_bad_scorer
¶
enable_bad_scorer (Boolean)
Default value False
debug_col_dict_prefix
¶
debug_col_dict_prefix (String)
Default value ''
return_early_debug_col_dict_prefix
¶
return_early_debug_col_dict_prefix (Boolean)
Default value False
return_early_debug_preview
¶
return_early_debug_preview (Boolean)
Default value False
autoviz_enable_recommendations
¶
Autoviz Use Recommended Transformations (Boolean)
Default value True
When enabled, experiment will try to use feature transformations recommended by Autoviz
autoviz_recommended_transformation
¶
Autoviz Recommended Transformations (Dict)
Default value {}
Key-value pairs of column names, and transformations that Autoviz recommended
last_recipe
¶
last_recipe (String)
Default value ''
Internal helper to allow memory of if changed recipe
make_mojo_scoring_pipeline_for_features_only
¶
Create MOJO for feature engineering pipeline only (no predictions) (Boolean)
Default value False
Create MOJO for feature engineering pipeline only (no predictions)
mojo_replace_target_encoding_with_grouped_input_cols
¶
Replaces target encoding features with concatenated input features. (Boolean)
Default value False
Replaces target encoding features by their input columns. Instead of CVTE_Age:Income:Zip, this will create Age:Income:Zip. Only when make_mojo_scoring_pipeline_for_features_only is enabled.
time_series_causal_split_recipe
¶
Whether causal recipe is used for non-lag-based recipe (Boolean)
Default value False
Whether causal splits are used when time_series_recipe is false orwhether to use same train-gap-test splits when lag transformers are disabled (default behavior).For train-test gap, period, etc. to be used when lag-based recipe is disabled, this must be false.
use_lags_if_causal_recipe
¶
Use lag transformers when using causal time-series recipe (Boolean)
Default value False
- Whether to use lag transformers when using causal-split for validation
(as occurs when not using time-based lag recipe). If no time groups columns, lag transformers will still use time-column as sole time group column.
min_ymd_timestamp
¶
min_ymd_timestamp (Number)
Default value 19000101
Earliest allowed datetime (in %Y%m%d format) for which to allow automatic conversion of integers to a time column during parsing. For example, 2010 or 201004 or 20100402 or 201004022312 can be converted to a valid date/datetime, but 1000 or 100004 or 10000402 or 10004022313 can not, and neither can 201000 or 20100500 etc.
max_ymd_timestamp
¶
max_ymd_timestamp (Number)
Default value 21000101
Latest allowed datetime (in %Y%m%d format) for which to allow automatic conversion of integers to a time column during parsing. For example, 2010 or 201004 or 20100402 can be converted to a valid date/datetime, but 3000 or 300004 or 30000402 or 30004022313 can not, and neither can 201000 or 20100500 etc.
max_rows_datetime_format_detection
¶
max_rows_datetime_format_detection (Number)
Default value 100000
maximum number of data samples (randomly selected rows) for date/datetime format detection
disallowed_datetime_formats
¶
List of disallowed datetime formats. (List)
Default value ['%y']
Manually disables certain datetime formats during data ingest and experiments. For example, [‘%y’] will avoid parsing columns that contain ‘00’, ‘01’, ‘02’ string values as a date column.
use_datetime_cache
¶
use_datetime_cache (Boolean)
Default value True
Whether to use datetime cache
datetime_cache_min_rows
¶
datetime_cache_min_rows (Number)
Default value 10000
Minimum amount of rows required to utilize datetime cache
holiday_country
¶
holiday_country (String)
Default value ''
max_time_series_properties_sample_size
¶
max_time_series_properties_sample_size (Number)
Default value 250000
Max. sample size for automatic determination of time series train/valid split properties, only if time column is selected
max_lag_sizes
¶
max_lag_sizes (Number)
Default value 30
Maximum number of lag sizes to use for lags-based time-series experiments. are sampled from if sample_lag_sizes==True, else all are taken (-1 == automatic)
min_lag_autocorrelation
¶
min_lag_autocorrelation (Float)
Default value 0.1
Minimum required autocorrelation threshold for a lag to be considered for feature engineering
max_signal_lag_sizes
¶
max_signal_lag_sizes (Number)
Default value 100
How many samples of lag sizes to use for a single time group (single time series signal)
single_model_vs_cv_score_reldiff
¶
single_model_vs_cv_score_reldiff (Float)
Default value 0.05
single_model_vs_cv_score_reldiff2
¶
single_model_vs_cv_score_reldiff2 (Float)
Default value 0.0
blend_in_link_space
¶
Whether to blend ensembles in link space (applies to classification only) (Boolean)
Default value True
Whether to blend ensembles in link space, so that can apply inverse link function to get predictions after blending. This allows to get Shapley values to sum up to final predictions, after applying inverse link function: preds = inverse_link( (blend(base learner predictions in link space ))) = inverse_link(sum(blend(base learner shapley values in link space))) = inverse_link(sum( ensemble shapley values in link space ))For binary classification, this is only supported if inverse_link = logistic = 1/(1+exp(-x))For multiclass classification, this is only supported if inverse_link = softmax = exp(x)/sum(exp(x))For regression, this behavior happens naturally if all base learners use the identity link function, otherwise not possible
tgc_via_ui_max_ncols
¶
tgc_via_ui_max_ncols (Number)
Default value 10
Maximum amount of columns send from UI to backend in order to auto-detect TGC
tgc_dup_tolerance
¶
tgc_dup_tolerance (Float)
Default value 0.01
Maximum frequency of duplicated timestamps for TGC detection