How It Works
Here is the algorithm followed by Lightwood starting from the input data setup, through model building and training, up to getting predictions.BaseMixer
class.
To learn more about Lightwood philosophy, follow this link.
Accuracy Metrics
Lightwood provides ways to score the accuracy of the model using one of the accuracy functions. The accuracy functions includemean_absolute_error
, mean_squared_error
, precision_score
, recall_score
, and f1_score
.
USING
clause of the CREATE MODEL
statement.
Here are the accuracy functions used by default:
- the
r2_score
value for regression predictions. - the
balanced_accuracy_score
value for classification predictions. - the
complementary_smape_array_accuracy
value for time series predictions.
DESCRIBE
statement.Tuning the Lightwood ML Engine Features
Description
In MindsDB, the underlying AutoML models are based on the Lightwood engine by default. This library generates models automatically based on the data and declarative problem definition. But the default configuration can be overridden using theUSING
statement that provides an option to configure specific parameters of the training process.
In the upcoming version of MindsDB, it will be possible to choose from more ML frameworks. Please note that the Lightwood engine is used by default.
Syntax
Here is the syntax:encoders
Key
It grants access to configure how each column is encoded. By default, the AutoML engine tries to get the best match for the data.
encoders
and their options, visit the Lightwood documentation page on encoders.
model
Key
It allows you to specify the type of machine learning algorithm to learn from the encoder data.
Model | Description | |
---|---|---|
BaseMixer | It is a base class for all mixers. | |
LightGBM | This mixer configures and uses LightGBM for regression or classification tasks depending on the problem definition. | |
LightGBMArray | This mixer consists of several LightGBM mixers in regression mode aimed at time series forecasting tasks. | |
NHitsMixer | This mixer is a wrapper around an MQN-HITS deep learning model. | |
Neural | This mixer trains a fully connected dense network from concatenated encoded outputs of each feature in the dataset to predict the encoded output. | |
NeuralTs | This mixer inherits from Neural mixer and should be used for time series forecasts. | |
ProphetMixer | This mixer is a wrapper around the popular time series library Prophet. | |
RandomForest | This mixer supports both regression and classification tasks. It inherits from sklearn.ensemble.RandomForestRegressor and sklearn.ensemble.RandomForestClassifier. | |
Regression | This mixer inherits from scikit-learn’s Ridge class. | |
SkTime | This mixer is a wrapper around the popular time series library sktime. | |
Unit | This is a special mixer that passes along whatever prediction is made by the target encoder without modifications. It is used for single-column predictive scenarios that may involve complex and/or expensive encoders (e.g. free-form text classification with transformers). | |
XGBoostMixer | This mixer is a good all-rounder, due to the generally great performance of tree-based ML algorithms for supervised learning tasks with tabular data. |
Please note that not all mixers are available in our cloud environment. In particular, LightGBM, LightGBMArray, NHITS, and Prophet.
model
options, visit the Lightwood documentation page on mixers.
Other Keys Supported by Lightwood in JsonAI
The most common use cases of configuring predictors useencoders
and model
keys explained above. To see all the available keys, check out the Lightwood documentation page on JsonAI.
Example
Here we use thehome_rentals
dataset and specify particular encoders
for some columns and a LightGBM model
.
You can visit the comprehensive Lightwood docs here.
Check out the Lightwood tutorials here.