A Brief Guide to Amazon SageMaker Algorithms (Part 2/4): Tabular Data

Technical Deep Dive
February 4, 2025
by
Aaron West
Brief Guide to Amazon SageMaker Algorithms: Tabular Data

Brief Guide to Amazon SageMaker Algorithms: Tabular Data

Image produced with Amazon Bedrock

This article is the second in a four-part series exploring Amazon SageMaker's machine learning algorithms, focusing on those designed for tabular data processing. It delves into key concepts, functionalities, and practical applications of algorithms such as XGBoost, AutoGluon, CatBoost, Factorization Machines, K-Nearest Neighbors, Linear Learner, LightGBM, and TabTransformer. This overview highlights aspects essential for the AWS Certified Machine Learning Specialty exam, making it a valuable resource for both certification candidates and professionals seeking to enhance their understanding of SageMaker's structured data capabilities. Be sure to check out Part 1 on Natural Language Processing and Text if you haven't yet - Part 3 in the series focuses on unsupervised learning models and provides some code examples.

XGBoost

XGBoost is an open-source, efficient, and popular implementation of the gradient boosted trees algorithm. It is known for its performance in machine learning competitions due to its ability to handle various data types and relationships, and its wide range of hyperparameters for fine-tuning. Amazon SageMaker offers a managed XGBoost environment, providing an estimator to execute training scripts.

  • 🤖 Supervised Learning: XGBoost is a supervised learning algorithm that combines estimates from multiple simpler models to make accurate predictions.
  • 💪 Versatility: It excels in handling different data types, relationships, and distributions, making it robust and versatile.
  • 💻 Portable Deployment: XGBoost can be statically compiled, allowing you to create a lightweight build that can run on resource-constrained environments, such as a Raspberry Pi. This flexibility makes it possible to deploy models even in edge computing scenarios or devices with limited hardware resources.

Memory Efficiency and Performance

  1. In-Memory Processing: XGBoost can load the entire dataset into memory when resources allow, enabling extremely fast processing and training times.
  2. Out-of-Core Computing: For datasets too large to fit in memory, XGBoost employs out-of-core computing techniques, allowing it to handle large datasets efficiently even with limited RAM.
  3. Cache-Aware Access: It uses cache-aware access patterns to optimize data processing, further enhancing performance.

Practical Applications

XGBoost is widely used in various domains due to its versatility and performance:

  1. Financial Services: Predicting credit risk and fraud detection
  2. Retail: Customer segmentation and product recommendation
  3. Healthcare: Disease prediction and patient outcome analysis
  4. Marketing: Click-through rate prediction and customer churn analysis
  5. Manufacturing: Predictive maintenance and quality control

Hyperparameters

XGBoost in SageMaker requires just two mandatory hyperparameters: num_class and num_round. num_class is only required if the objective is set to multi:softmax or multi:softprob and specifies the number of classes or categories, while num_round dictates the number of boosting rounds for training. These parameters are fundamental to configuring XGBoost for classification tasks or setting up the overall training process.

Like other algorithms in this series, XGBoost offers a variety of optional hyperparameters that allow for further tuning. Here are a few valuable ones:

  • alpha: L1 regularization term on weights. A higher value makes the model more conservative.
  • base_score: Initial prediction score of all instances. This can be useful for setting the global bias.
  • booster: Determines the boosting algorithm to use, with gbtree, gblinear, and dart as available options.
  • colsample_bytree: Specifies the subsample ratio of columns when constructing each tree, controlling feature selection.

AutoGluon

AutoGluon-Tabular is an open-source automated machine learning (AutoML) framework designed for generating high-performing machine learning models from raw tabular datasets. It distinguishes itself from other AutoML frameworks by employing a unique approach that emphasizes model ensembling and stacking, rather than solely focusing on model and hyperparameter selection.

  • 💻 Open-Source Framework: AutoGluon-Tabular is an open-source AutoML framework for training machine learning models on tabular datasets.
  • 🤖 Automation: It automates the process of feature engineering, model selection, hyperparameter tuning, and model ensembling.
  • 🏆 Ensembling: It excels in ensembling multiple models and stacking them in multiple layers to achieve high accuracy.

Practical Applications

AutoGluon-Tabular is particularly useful in scenarios where quick, automated model development is needed:

  1. Rapid Prototyping: Quickly generating baseline models for new projects
  2. Kaggle Competitions: Achieving competitive results with minimal manual tuning
  3. Business Intelligence: Automating predictive analytics for various business metrics
  4. Research: Comparing automated approaches with manual model development
  5. Educational Purposes: Teaching students about AutoML concepts and practices

Hyperparameters

One of the strengths of AutoGluon-Tabular is its ability to automatically tune hyperparameters, allowing it to adapt to different types of tasks without requiring the user to manually set specific values. This flexibility makes it ideal for automated machine learning (AutoML) workflows, where manual tuning can often be a bottleneck. However, AutoGluon-Tabular offers several optional hyperparameters for users who wish to exert more control over the process. Here are a few key examples:

  • eval_metric: Determines the evaluation metric used for validation, with options such as root_mean_squared_error for regression, roc_auc for binary classification, and accuracy for multi-class classification. By default, it's set to auto, letting the algorithm select the metric based on the task.
  • presets: Provides a set of predefined configurations for various training needs. For instance, best_quality achieves high predictive accuracy with slower inference, while optimize_for_deployment reduces model size for faster inference and deployment.
  • auto_stack: Automatically enables bagging and multi-layer stacking, which can significantly boost predictive accuracy at the cost of longer training times. This is a useful feature for those looking to maximize performance.
  • num_bag_folds: Configures the number of folds used for bagging models. Bagging can help to reduce model variance, and adjusting the number of folds allows for balancing between bias and variance.
  • num_stack_levels: Specifies the number of stacking levels in the ensemble model. Higher stacking levels can improve model accuracy but may increase training time, with a typical recommended range of 1 to 3 levels.

CatBoost

CatBoost is a powerful open-source machine learning algorithm known as Gradient Boosting Decision Tree (GBDT), which combines multiple simple models to make accurate predictions. It is specifically designed to address prediction shift issues commonly found in GBDT implementations.

  • 🌳 Categorical Boosting: CatBoost is a gradient boosting algorithm known for its performance and ability to handle categorical features effectively.
  • ⚙️ Innovative Improvements: It introduces two major improvements: ordered boosting, which mitigates target leakage, and an innovative algorithm for handling categorical features.

Practical Applications

CatBoost is particularly useful in scenarios involving categorical data and where prediction accuracy is crucial:

  1. E-commerce: Product categorization and recommendation systems
  2. Finance: Credit scoring and risk assessment
  3. Healthcare: Patient diagnosis and treatment outcome prediction
  4. Marketing: Customer segmentation and targeted advertising
  5. Natural Language Processing: Text classification and sentiment analysis

Hyperparameters

CatBoost in SageMaker is designed with default hyperparameters that allow for quick and easy setup. While there are no strictly required hyperparameters, CatBoost offers several customizable options to improve model performance depending on your specific needs. Here are a few key hyperparameters you can adjust:

  • iterations: This hyperparameter controls the maximum number of trees that will be built during training. By default, it's set to 500, but increasing or decreasing this value can help tune the model for longer or shorter training times depending on your dataset.
  • early_stopping_rounds: If the model's performance stops improving, training will halt after a set number of rounds. The default value is 5, and this parameter helps prevent overfitting while speeding up the training process.
  • eval_metric: The evaluation metric used for validation during training. CatBoost automatically selects the appropriate metric based on the type of problem (RMSE for regression, AUC for binary classification, or MultiClass for multi-class classification). However, you can customize this based on your specific needs by setting values such as logloss or accuracy.
  • learning_rate: Controls how quickly the model adapts during training. Smaller values lead to slower, more precise training, while larger values speed up training but risk overshooting the optimal solution. The default value is 0.009.
  • depth: Refers to the depth of the trees used in the model. Deeper trees can capture more complex patterns in the data but may also lead to overfitting. The default value is 6, and it can be adjusted between 1 and 16.
  • l2_leaf_reg: A regularization parameter for the L2 regularization of leaf scores, which helps avoid overfitting. The default value is 3.

Factorization Machines

Amazon SageMaker Factorization Machines is a supervised learning algorithm used for classification and regression tasks. It combines the advantages of Support Vector Machines with factorization models, effectively capturing interactions between features in high-dimensional, sparse datasets.

  • 🤖 Supervised Learning: Used for both classification and regression tasks.
  • 📊 Feature Interaction: Leverages the strengths of Support Vector Machines and factorization models to capture complex feature interactions.
  • 📈 High-Dimensional Data: Designed for high-dimensional sparse datasets, making it effective for problems like recommender systems.

Practical Applications

Factorization Machines are particularly useful in scenarios involving sparse data and complex feature interactions:

  1. Recommender Systems: Personalized product or content recommendations
  2. Click-Through Rate Prediction: Improving online advertising effectiveness
  3. User Behavior Analysis: Understanding and predicting user interactions
  4. Feature Extraction: Generating meaningful features from sparse data
  5. Social Network Analysis: Predicting connections or interactions between users

Hyperparameters

Factorization Machines in SageMaker require three key hyperparameters that must be set by the user: feature_dim, num_factors, and predictor_type. These are crucial for estimating model parameters from data, especially for sparse inputs.

  • feature_dim: This defines the dimensionality of the input feature space, which can be very high with sparse input data. Valid values are positive integers within a suggested range of [10000,10000000].
  • num_factors: Determines the number of factors used in factorization. A range of [2,1000] is suggested, with 64 being a good starting point.
  • predictor_type: Defines whether the task is binary classification or regression. Valid options are binary_classifier for classification tasks and regressor for regression.

In addition to these required parameters, there are several optional hyperparameters that can be tuned for specific needs:

  • bias_init_method: Controls how the bias term is initialized, with options including normal (default), uniform, or constant.
  • bias_init_scale: Specifies the scale range for initializing bias when bias_init_method is set to uniform.
  • bias_init_sigma: Defines the standard deviation for initializing bias when using the normal method.
  • bias_init_value: Sets a constant value for the bias term when the initialization method is set to constant.

K-Nearest Neighbors (KNN)

The K-Nearest Neighbors (k-NN) algorithm is an index-based, non-parametric method used for classification and regression tasks. It predicts based on the majority vote or average of the k-nearest data points.

  • 🧮 Versatile Algorithm: KNN is used for both classification and regression tasks.
  • 📊 Non-Parametric Method: It makes no assumptions about the underlying data distribution.
  • 👨 ‍🏫 Classification Approach: In classification, it assigns the most frequent class label among the k-nearest neighbors.
  • 📈 Regression Approach: In regression, it calculates the average of the k-nearest neighbors' target values.

Practical Applications

KNN is widely used in various domains due to its simplicity and effectiveness:

  1. Image Recognition: Classifying images based on similar known images
  2. Recommendation Systems: Suggesting items based on user similarities
  3. Finance: Credit scoring and fraud detection
  4. Healthcare: Disease diagnosis based on patient similarities
  5. Anomaly Detection: Identifying outliers in datasets

Hyperparameters

There are a several several core parameters must be defined to ensure the KNN algorithm is appropriately configured:

  • feature_dim: This defines the number of features in the input data. It is crucial for setting up the dimensionality of the data points.
  • k: Represents the number of nearest neighbors the algorithm will consider when making predictions.
  • predictor_type: Specifies whether the task is classification or regression, using the values classifier for classification tasks or regressor for regression tasks.
  • sample_size: Determines how many data points will be sampled from the training dataset for the k-NN algorithm to process.
  • dimension_reduction_target: Required when dimension reduction is applied, this parameter sets the target dimension for the reduction.

There are also several optional hyperparameters that can be adjusted to optimize performance, especially when dealing with large datasets or specific use cases:

  • dimension_reduction_type: Controls the type of dimension reduction, such as sign for random projection or fjlt for the Johnson-Lindenstrauss transform.
  • faiss_index_ivf_nlists: For the FAISS library, this specifies the number of centroids to use when building the index, which is essential for large-scale nearest neighbor search.
  • faiss_index_pq_m: This defines the number of sub-vector components in FAISS when applying Product Quantization (PQ) indexing, helping to optimize memory usage during nearest neighbor search.

Linear Learner

Amazon SageMaker Linear Learner is a versatile supervised machine learning algorithm used for both classification and regression problems. It allows users to explore different training objectives and choose the best solution based on a validation set.

  • 🤖 Dual Functionality: Solves both classification (binary and multiclass) and regression problems.
  • 📊 Flexible Input: Accepts labeled examples (x, y) as input, where x is a high-dimensional vector and y is a numeric label.
  • 📈 Linear Mapping: Learns a linear function or threshold function to map input vectors to approximate labels.
  • 🎯 Objective Exploration: Allows exploring different training objectives and choosing the best solution using a validation set.

Practical Applications

Linear Learner is useful in a wide range of scenarios where linear relationships between features and targets are assumed:

  1. Financial Forecasting: Predicting stock prices or economic indicators
  2. Marketing Analytics: Customer lifetime value prediction
  3. Healthcare: Patient length of stay prediction
  4. Manufacturing: Quality control and defect prediction
  5. Environmental Science: Climate variable prediction

Hyperparameters

For the Linear Learner algorithm, the following are the key required hyperparameters:

  • num_classes: The number of classes for the response variable. This parameter is required if you're performing multi-class classification and determines how many classes the algorithm will predict (e.g., if you have three classes, this would be set to 3).
  • predictor_type: Specifies the type of problem, whether it's binary classification, multi-class classification, or regression.

There are several optional ones, allowing for greater flexibility and optimization based on the specific characteristics of your dataset and learning task:

  • accuracy_top_k: Defines the top-K accuracy metric, which evaluates how well the model ranks the true label among the top K predictions. It is useful for multi-class classification tasks.
  • optimizer: Determines which optimization algorithm will be used. The valid options include auto, which is the default and selects adam as the optimizer; sgd, which stands for stochastic gradient descent; adam, an adaptive momentum estimation technique; and rmsprop, a method that uses a moving average of squared gradients to normalize the gradient. The default value is auto.
  • beta_1 and beta_2: Control the decay rates for first and second-moment estimates in optimization. These are specific to the Adam optimizer.
  • bias_lr_mult and bias_wd_mult: Allow for different learning rates and regularization for the bias term, which can help fine-tune the model's performance.

LightGBM

LightGBM is an open-source implementation of the Gradient Boosting Decision Tree (GBDT) algorithm, known for its efficiency and use in supervised learning. It excels at predicting target variables by combining estimates from simpler models.

  • 💻 Efficient Implementation: LightGBM is an open-source, efficient implementation of the Gradient Boosting Decision Tree (GBDT) algorithm.
  • 📈 Ensemble Learning: GBDT combines estimates from a set of simpler, weaker models to predict a target variable accurately.
  • 🚀 Enhanced Performance: LightGBM improves the efficiency and scalability of traditional GBDT algorithms.

Practical Applications

LightGBM is widely used in various domains due to its efficiency and performance:

  1. Online Advertising: Click-through rate prediction
  2. Retail: Sales forecasting and inventory optimization
  3. Finance: Risk assessment and fraud detection
  4. Healthcare: Disease progression prediction
  5. Energy Sector: Energy demand forecasting

Hyperparameters

For LightGBM, there are no strictly required hyperparameters, as the algorithm can automate much of the tuning process. However, several optional hyperparameters can significantly impact the performance and behavior of the model:

  • num_leaves: Controls the maximum number of leaves in one tree. A larger number can improve accuracy but also increases the risk of overfitting. Default value is 31.
  • learning_rate: Adjusts how quickly the model learns. A smaller value may lead to better generalization but requires more training iterations. Default value is 0.1.
  • max_depth: Limits the depth of a tree to prevent overfitting, especially with smaller datasets. Setting it to -1 removes any limit. Default value is -1.

TabTransformer

TabTransformer is a novel deep tabular data modeling architecture for supervised learning, leveraging self-attention-based Transformers to transform categorical feature embeddings into robust contextual embeddings for improved prediction accuracy.

  • 🤖 Transformer-Based: Built on self-attention-based Transformers, a powerful architecture from natural language processing.
  • 📊 Tabular Data Focus: Specifically designed for deep tabular data modeling.
  • 🎯 Improved Accuracy: Aims for higher prediction accuracy in supervised learning tasks compared to traditional methods.
  • 💪 Robust Embeddings: Creates robust contextual embeddings, improving prediction accuracy.
  • 🛡 ️ Data Handling: Effectively handles missing and noisy data, a common issue in real-world datasets.

Practical Applications

TabTransformer is particularly useful in scenarios involving complex tabular data:

  1. Financial Services: Credit scoring and fraud detection with improved accuracy
  2. Healthcare: Patient outcome prediction considering complex interactions between features
  3. Marketing: Customer behavior prediction for targeted campaigns
  4. Human Resources: Employee performance prediction and talent management
  5. IoT and Sensor Data: Analyzing and predicting patterns in sensor data with temporal aspects

Hyperparameters

Like AutoGluon, CatBoost and LightGBM, TabTransformer doesn't have strictly required hyperparameters, and like every other algorithm, it offers various optional settings that enable users to fine-tune training and performance based on their data. Here are three key optional hyperparameters to consider when configuring the model:

  • n_blocks: Specifies the number of transformer blocks used in the architecture, which determines the depth of the model. Valid values range from 1 to 12, and the default is 4.
  • learning_rate: Controls the rate at which the model weights are updated during training. Lower values lead to more gradual updates and can help in achieving better convergence. Valid values range from 0.0 to 1.0, with a default of 0.001.
  • mlp_dropout: Applies dropout to the feedforward network between layers, helping to prevent overfitting by randomly setting a fraction of input units to 0 during training. This parameter has valid values from 0.0 to 1.0, with a default of 0.1.

Choosing the Right Algorithm and Complementary Use Cases

The selection of a machine learning algorithm depends on the data structure, the problem you're addressing, and the specific requirements for your output. Although various algorithms can handle a range of tasks, some are more suitable for certain scenarios than others. Below is a guide on when to choose one algorithm over another and how they can be combined for better outcomes.

XGBoost vs. CatBoost for Tabular Data

  • When to Choose XGBoost: XGBoost is a highly efficient algorithm suited for structured tabular data. It excels in handling missing values, non-linear relationships, and unbalanced datasets. If your goal is to perform classification or regression tasks with high control over hyperparameter tuning and fine-grained adjustments, XGBoost provides robust performance and flexibility.
  • When to Choose CatBoost: CatBoost is an ideal choice when dealing with categorical features, as it handles categorical variables natively without requiring extensive preprocessing. It also shines in tasks requiring interpretable models, such as finance or healthcare, where transparency and interpretability are crucial. CatBoost often reduces the need for hyperparameter tuning and automatically optimizes many parameters, making it easier to set up and use effectively.

LightGBM vs. Linear Learner for High-Speed Processing

  • LightGBM: If you are working with extremely large datasets and need a model that can train quickly while consuming minimal memory, LightGBM is an excellent choice. It's optimized for both speed and performance, making it ideal for high-dimensional data where computational efficiency is essential.
  • Linear Learner: On the other hand, if you are dealing with linear classification or regression tasks where the model's simplicity and interpretability are key, Linear Learner is an effective and straightforward option. It's particularly useful for smaller datasets where more complex models may not be necessary or when interpretability is a priority.

AutoGluon vs. TabTransformer for Automating Model Tuning

  • When to Choose AutoGluon: AutoGluon automates much of the machine learning pipeline, from model selection to hyperparameter optimization, making it ideal for those who need rapid prototyping with minimal manual intervention. If you're looking for an approach that quickly delivers high-quality results without requiring expertise in model tuning, AutoGluon is your best bet.
  • When to Choose TabTransformer: TabTransformer, on the other hand, is ideal when your tabular data contains both categorical and continuous features that can benefit from deep learning-based feature representations. It's a strong choice for complex classification tasks where capturing interactions between features is critical for model performance.

Combining Algorithms for Enhanced Tabular Data Workflows

  • XGBoost for Classification + TabTransformer for Feature Representation: You can use TabTransformer to learn richer feature representations from tabular data, followed by XGBoost for classification or regression tasks. This combination is effective for complex problems where feature interaction plays a significant role in model performance.
  • AutoGluon for Rapid Prototyping + LightGBM for Optimization: AutoGluon can be employed for rapid prototyping and model selection, while LightGBM can be used to optimize performance on large datasets for more demanding production tasks.

Implementation in SageMaker

The tabular data algorithms we've explored in this article all benefit from Amazon SageMaker's unified implementation framework. Each algorithm, from XGBoost to TabTransformer, is integrated into SageMaker as an Estimator class, providing a consistent interface for training and deployment. Data preprocessing for these tabular algorithms often involves converting your dataset into the SageMaker-compatible format, typically using the record_set() utility or similar data handling methods specific to each algorithm.

Training these models in SageMaker involves configuring algorithm-specific hyperparameters, which allow you to fine-tune performance for your particular use case. Whether you're dealing with gradient boosting algorithms like XGBoost, CatBoost, and LightGBM, or more specialized algorithms like Factorization Machines or TabTransformer, SageMaker provides a standardized approach to model training and evaluation.

Once trained, these tabular data models can be deployed to SageMaker Endpoints for real-time inference, or used for batch predictions on large datasets. The deployment process is streamlined across all algorithms, allowing for easy integration into your existing data pipelines and applications.

Further Resources

For those looking to dive deeper into these algorithms or seeking the most up-to-date information, I recommend visiting the official Amazon SageMaker SDK documentation. The Algorithms section here also provides comprehensive details on all the algorithms we've discussed here, as well as others available in SageMaker. This resource offers in-depth explanations of algorithm parameters, input/output specifications, and implementation details that are crucial for fine-tuning your models. While this article serves as an overview and study aid, the official AWS documentation should be your go-to source for the most current and detailed information as you work with these machine learning tools in your projects.

Image Generated by Amazon Bedrock
Got Questions? Contact us

Your data is trying to tell you something

Contact us

... are you listening?