Learner Technique
|
Description
|
Model Complexity
|
Linear Regression
|
Predicts a continuous goal and tries to assess the weight of each input feature in reducing the difference between actual and predicted values. This technique is good for predicting values, quantities, counts, and other continuous variable outcomes. It’s not well suited for Boolean or categorical outcomes.
Example: Predict the number of microns a drill will be off-target based on features such as hours of operation, pump pressure, temperature, drill bit changes.
|
Simple
|
Logistic Regression
|
Mainly a classification algorithm that predicts the probability that a goal belongs to a specific category. Tries to assess the weight of each input feature in reducing the number of misclassifications. This technique is best suited to predicting categorical and Boolean outcomes.
Example: Predict a pump failure based on features such as hours of operation, pump pressure, pump type.
|
Simple
|
Decision Tree
|
Builds a tree model that predicts a class or value for the goal. Tries to find the best way to partition the data, according to the input features, to improve classification accuracy. This technique is easy to interpret, can handle both numerical and categorical data, automatically learns feature interactions, and can train quickly even on large datasets. Tree models are most appropriate when a model needs to explain rules such as “If this AND that, THEN outcome.”
|
Simple
|
Support Vector Machines (SVM)
|
Can be applied only to a Boolean goal. Builds a model that predicts the classification of a goal into one of two categories. Tries to classify data in a way that maps the largest separation between the categories. Currently this technique performs linear classifications only.
An SVM learner can only be used with a Soloist or a Majority Vote ensemble technique.
|
Simple
|
Neural Network
|
Uses a set of interconnected nodes and layers to predicts a class or value for a goal. Tries to create a network that learns a set of adaptive weights to predict outcomes accurately based on the input features. This technique can handle both numerical and categorical data. It can be used to represent various linear or non-linear functions and it can train quickly even on large datasets.
|
Moderate
|
Random Forest
|
Generates multiple, random decision trees and outputs predictions by computing the average from all the trees. This technique provides an unbiased estimate of model accuracy and can handle both large datasets and large numbers of features.
|
High
|
Gradient Boost
|
Generates multiple, sequential decisions trees, where each tree is constructed using information from the previous tree. The output is a final boosted model which is the sequential culmination of the iterative process. This technique sometimes performs better than other decision tree techniques but multiple trials may be necessary to decide on the best number of trees to include. Training can be slow and noisy data can lead to overfitting.
|
High
|
|
When using a machine learning ensemble, model complexity can increase.
|