Convect AI Flow Platform Logo

Forecasting Models

Understanding the AI algorithms behind demand forecasting

Forecasting Models

The forecast application provides a comprehensive suite of forecasting algorithms, from traditional statistical methods to advanced machine learning and neural network approaches. Understanding each model's strengths helps you choose the best algorithms for your specific use case.

Model Categories

Traditional Statistical Methods

Exponential Smoothing (ETS)

Best for: Data with trend and seasonality patterns

Characteristics:

  • Strengths: Handles trend and seasonal components well, robust to outliers
  • Weaknesses: May struggle with complex patterns or regime changes
  • Use Cases: Retail sales, inventory management, seasonal products
  • Parameters: Automatically optimized for your data

ARIMA (Autoregressive Integrated Moving Average)

Best for: Stationary time series with clear patterns

Characteristics:

  • Strengths: Excellent for stationary data, well-established methodology
  • Weaknesses: Requires stationary data, may miss complex relationships
  • Use Cases: Financial data, economic indicators, stable demand patterns
  • Parameters: Automatically optimized for your data

Holt-Winters

Best for: Data with both trend and seasonality

Characteristics:

  • Strengths: Specifically designed for seasonal data, handles trends
  • Weaknesses: Assumes additive seasonality, may not capture complex patterns
  • Use Cases: Seasonal retail, tourism, agricultural products
  • Parameters: Automatically optimized for your data

Machine Learning Methods

Linear Regression

Best for: Simple linear relationships with external factors

Characteristics:

  • Strengths: Simple, interpretable, fast computation
  • Weaknesses: Assumes linear relationships, limited to simple patterns
  • Use Cases: Basic trend analysis, simple demand patterns
  • Features: Can incorporate external variables

LightGBM (Gradient Boosting Machine)

Best for: Complex patterns with multiple features

Characteristics:

  • Strengths: Handles complex relationships, feature importance, robust
  • Weaknesses: May overfit with small datasets, less interpretable
  • Use Cases: Multi-factor demand, complex business patterns
  • Features: Can use external and categorical data effectively

Deep Learning Methods

LSTM (Long Short-Term Memory)

Best for: Complex temporal patterns and long sequences

Characteristics:

  • Strengths: Captures complex temporal dependencies, handles long sequences
  • Weaknesses: Requires large datasets, computationally intensive
  • Use Cases: Complex demand patterns, multi-step forecasting
  • Architecture: Recurrent neural network with memory cells

N-BEATS (Neural Basis Expansion Analysis)

Best for: Univariate time series with interpretable components

Characteristics:

  • Strengths: Interpretable, handles trend and seasonality, no preprocessing
  • Weaknesses: Univariate only, may not capture external factors
  • Use Cases: Clean time series data, interpretable forecasts
  • Architecture: Deep neural network with backward residual links

N-HITS (Neural Hierarchical Interpolation for Time Series)

Best for: Hierarchical time series and multi-scale patterns

Characteristics:

  • Strengths: Handles hierarchical data, multi-scale forecasting
  • Weaknesses: Complex architecture, requires hierarchical structure
  • Use Cases: Product hierarchies, geographic hierarchies
  • Architecture: Hierarchical interpolation with neural networks

PatchTST (Patch-based Time Series Transformer)

Best for: Long sequence forecasting with attention mechanisms

Characteristics:

  • Strengths: Handles long sequences, attention-based, flexible
  • Weaknesses: Computationally intensive, requires sufficient data
  • Use Cases: Long-term forecasting, complex temporal patterns
  • Architecture: Transformer with patch-based input processing

Autoformer

Best for: Long sequence forecasting with auto-correlation

Characteristics:

  • Strengths: Efficient for long sequences, auto-correlation mechanism
  • Weaknesses: May not capture all complex patterns
  • Use Cases: Long-horizon forecasting, energy consumption
  • Architecture: Transformer with auto-correlation attention

Ensemble Methods

Mean Ensemble

Best for: Robust, stable predictions across different scenarios

Characteristics:

  • Strengths: Reduces overfitting, stable predictions, simple
  • Weaknesses: May not capture model-specific strengths
  • Use Cases: General forecasting, when unsure about best model
  • Method: Averages predictions from all selected models

Model Selection Guidelines

Data Characteristics

Data Volume

  • Small Datasets (< 100 items): ETS, ARIMA, Linear Regression
  • Medium Datasets (100-1000 items): LightGBM, N-BEATS, Mean Ensemble
  • Large Datasets (> 1000 items): LSTM, PatchTST, Autoformer

Data Quality

  • Clean Data: Any model, focus on performance
  • Noisy Data: ETS, LightGBM, ensemble methods
  • Missing Values: Models with robust preprocessing

Pattern Complexity

  • Simple Trends: Linear Regression, ETS
  • Seasonal Patterns: Holt-Winters, N-BEATS
  • Complex Patterns: LSTM, PatchTST, Autoformer
  • Multiple Factors: LightGBM, ensemble methods

Business Context

Forecast Horizon

  • Short-term (1-4 periods): Any model, focus on accuracy
  • Medium-term (5-12 periods): LSTM, N-BEATS, ensemble methods
  • Long-term (13+ periods): N-HITS, PatchTST, Autoformer

Update Frequency

  • Real-time: LightGBM, simple models
  • Batch processing: Deep learning models, ensemble methods
  • Periodic updates: Any model, consider computational cost

Business Requirements

  • Interpretability: ETS, ARIMA, Linear Regression
  • Accuracy: Ensemble methods, deep learning
  • Speed: LightGBM, simple statistical models
  • Robustness: Ensemble methods, ETS

Model Performance Metrics

Accuracy Measures

WMAPE (Weighted Mean Absolute Percentage Error)

  • Definition: Weighted percentage error
  • Interpretation: Accounts for item importance or volume
  • Use: When some items are more important than others

Best Practices

Model Selection Strategy

Start Simple

  1. Test Individual Models: Understand strengths and weaknesses
  2. Use Backtests: Validate performance on historical data
  3. Iterate and Improve: Refine based on results

Consider Data Characteristics

  1. Data Volume: Match model complexity to data size
  2. Pattern Type: Choose models suited to your data patterns
  3. Update Frequency: Consider computational requirements
  4. Business Context: Align with business needs and constraints

Performance Optimization

Data Preparation

  1. Clean Data: Remove outliers and handle missing values
  2. Feature Engineering: Create relevant features for ML models
  3. Scale Appropriately: Normalize data for neural networks
  4. Validate Relationships: Ensure external data relevance

Model Configuration

  1. Parameter Tuning: Let the system optimize automatically
  2. Ensemble Diversity: Include different model types
  3. Validation Strategy: Use backtests for model selection
  4. Monitoring: Track performance over time

Continuous Improvement

Regular Evaluation

  1. Periodic Backtesting: Test models as data patterns change
  2. Performance Monitoring: Track forecast accuracy
  3. Model Updates: Incorporate new algorithms
  4. Business Feedback: Align with changing requirements

Model Maintenance

  1. Data Quality: Maintain high-quality input data
  2. Model Selection: Update models based on performance
  3. Parameter Optimization: Retune as needed
  4. Documentation: Keep track of model choices and rationale

Choosing the right forecasting models is crucial for accurate predictions. Start with ensemble methods and refine based on your specific data characteristics and business requirements.