AIPrimer.AI
  • 🚦AI Primer In Transportation
  • CHAPTER 1 - INTRODUCTION TO MACHINE LEARNING
    • Machine Learning in Transportation
    • What is Machine Learning?
    • Types of Machine Learning
      • Supervised Learning
      • Unsupervised Learning
      • Semi-supervised Learning
      • Reinforced Learning
    • Fundamental concepts of machine learning
      • Model Training and Testing
      • Evaluating the Model’s Prediction Accuracy
      • The Underfitting and Overfitting Problems
      • Bias-Variance Tradeoff in Overfitting
      • Model Validation Techniques
      • Hyperparameter Tuning
      • Model Regularization
      • The Curse of Ddimensionality
    • Machine Learning versus Statistics
  • CHAPTER 2 - SUPERVISED METHODS
    • Supervised Learning_Complete Draft
    • K-Nearest Neighbor (KNN) Algorithm
    • Tree-Based Methods
    • Boosting
    • Support Vector Machines (SVMs)
  • CHAPTER 3 - UNSUPERVISED LEARNING
    • Principal Component Analysis
      • How Does It Work?
      • Interpretation of PCA result
      • Applications in Transportation
    • CLUSTERING
      • K-MEANS
      • SPECTRAL CLUSTERING
      • Hierarchical Clustering
    • REFERENCE
  • CHAPTER 4 - NEURAL NETWORK
    • The Basic Paradigm: Multilayer Perceptron
    • Regression and Classification Problems with Neural Networks
    • Advanced Topologies
      • Modular Network
      • Coactive Neuro–Fuzzy Inference System
      • Recurrent Neural Networks
      • Jordan-Elman Network
      • Time-Lagged Feed-Forward Network
      • Deep Neural Networks
  • CHAPTER 5 - DEEP LEARNING
    • Convolutional Neural Networks
      • Introduction
      • Convolution Operation
      • Typical Layer Structure
      • Parameters and Hyperparameters
      • Summary of Key Features
      • Training of CNN
      • Transfer Learning
    • Recurrent Neural Networks
      • Introduction
      • Long Short-Term Memory Neural Network
      • Application in transportation
    • Recent Development
      • AlexNet, ZFNet, VggNet, and GoogLeNet
      • ResNet
      • U-Net: Full Convolutional Network
      • R-CNN, Fast R-CNN, and Faster R-CNN
      • Mask R-CNN
      • SSD and YOLO
      • RetinaNet
      • MobileNets
      • Deformable Convolution Networks
      • CenterNet
      • Exemplar Applications in Transportation
    • Reference
  • CHAPTER 6 - REINFORCEMENT LEARNING
    • Introduction
    • Reinforcement Learning Algorithms
    • Model-free v.s. Model-based Reinforcement Learning
    • Applications of Reinforcement Learning to Transportation and Traffic Engineering
    • REFERENCE
  • CHAPTER 7 - IMPLEMENTING ML AND COMPUTATIONAL REQUIREMENTS
    • Data Pipeline for Machine Learning
      • Introduction
      • Problem Definition
      • Data Ingestion
      • Data Preparation
      • Data Segregation
      • Model Training
      • Model Deployment
      • Performance Monitoring
    • Implementation Tools: The Machine Learning Ecosystem
      • Machine Learning Framework
      • Data Ingestion tools
      • Databases
      • Programming Languages
      • Visualization Tools
    • Cloud Computing
      • Types and Services
    • High-Performance Computing
      • Deployment on-premise vs on-cloud
      • Case Study: Data-driven approach for the implementation of Variable Speed Limit
      • Conclusion
  • CHAPTER 8 - RESOURCES
    • Mathematics and Statistics
    • Programming, languages, and software
    • Machine learning environments
    • Tools of the Trade
    • Online Learning Sites
    • Key Math Concepts
  • REFERENCES
  • IMPROVEMENT BACKLOG
Powered by GitBook
On this page
  1. CHAPTER 1 - INTRODUCTION TO MACHINE LEARNING

Machine Learning versus Statistics

Data analysis and modeling is probably the most important research tool in the field of transportation. There are two different schools of thought in the world of transportation data analysis: statistical models and machine learning. Statistical models have been traditionally used to analyze the transportation data, while machine learning techniques have recently gained a lot of attention from transportation researchers and practitioners.

Statistical techniques are the mathematics of collecting, cleaning, organizing, modeling and interpreting the data or a sample of the data. Machine learning techniques are the algorithms that combine learning, adaption, evaluation, and in general structuring an unstructured and complex data. This section describes some of the differences and similarities between machine learning and traditional statistical models.

Modeling Framework

Machine learning techniques are known for their ability to work with multidimensional, complex and non-linear data with a massive amount of observations and variables (i.e., big data). Statistical models, though, have solid mathematical frameworks that make them capable to discover the relationships between the variables, fit a project-specific probability model, compute quantitative measures of confidence, and verify certain statistical assumptions.

Statistical models are generally defined in terms of their mathematical formulation and their statistical assumptions, whereas machine learning techniques are mostly defined in terms of their architecture and learning algorithms.

Objectives

Statistical models largely emphasize on inference by providing a model, predictor or classifier, and offering insights on the data and its structure. Machine learning techniques find generalizable predictive patterns in the data and mostly emphasize on implementation.

Statistical models are self-explanatory and will be helpful to interpret the data by providing statistical measures, such as marginal effects, elasticities and coefficient’s signs, magnitude and significance. On the other hand, machine learning techniques provide an efficient, accurate and flexible representation of the underlying and hidden relationships between the variables in the data. Machine learning techniques are well-known for their high prediction capability.

Flexibility

Statistical models follow rigid assumptions and mostly result in one final model / conclusion, while machine learning techniques mostly result in more than one model / conclusion, as their learning curve may have various local minima. In fact, machine learning techniques are more flexible, since their functional form is approximated through a learning process, compared to the statistical models with multiple priori statistical assumptions. Overall, machine learning techniques have a higher learning, better generalization ability, and a higher adaptability.

Self-Exploratory

Since the machine learning techniques mostly have a hidden mechanism, some researchers call them black box. In contrast, the statistical models are self-exploratory, and one can interpret them by looking at the estimated coefficients, their signs and other statistical tests, such as t-statistics or p-value.

Assumptions & Limitations

Statistical models mostly have multiple specified, priori and strict hypotheses and place some restrictions on the data, distributions and error terms. Multicollinearity, nonlinearity, autocorrelation, and non-normality are some of the situations where statistical models will face problematic issues. On the other hand, machine learning techniques are extremely adaptable, with no prior model or error distribution specifications.

Data Requirements

Machine learning techniques require more data, compared to statistical models. A higher number of observations in a machine learning model (such as neural networks), will lead to a better learning fit and a higher prediction accuracy. It has been a common practice between machine learning modelers and transportation practitioners to put aside 10-20 and 20-30 percent of observations for cross-validation and model evaluation, respectively. Thus, most of the machine learning models are being developed by using 60-70 percent of the data, which requires even a greater number of observations.

Predicative Ability & Accuracy

One of the major differences between machine learning and statistical models is their prediction ability. Machine learning techniques are designed to make the most accurate predictions possible, while statistical models emphasize on the inference about the relationships between variables. The prediction process in machine learning techniques happens by using general-purpose learning algorithms to find patterns in large databases, with minimal assumptions about the data distribution. Although, the lack of an explicit model output can make machine learning techniques very difficult to interpret.

Modeling Effort

Modeling effort can be divided into four stages:

A) Data preparation: since machine learning techniques have less assumptions and limitations, the data preparation stage for statistical models is more difficult and more time-consuming.

B) Model calibration (running the model): machine learning models can handle larger data with more complexities, which requires more computations. Thus, the model calibration (model run) stage for machine learning techniques is more time-consuming compared to the statistical models.

C) Model validation (model improvement): since machine learning techniques have a higher predictive ability, compared to the statistical models, one can improve the model prediction easier. Less statistical and mathematical assumptions are associated with a more straightforward model improvement for machine learning techniques.

D) Model Interpretation statistical models provide multiple interpretative measures, such as coefficients, signs, t-statistics, p-values, confidence intervals, marginal effects, elasticities, etc.

Category
Property
Statistical Models
Machine Learning

General Comparison

Flexibility

βœ“

Self-exploratory

βœ“

Statistical framework

βœ“

Probability maximization

βœ“

Error minimization

βœ“

Rigid assumptions

βœ“

Rigid assumptions

βœ“

Data Availability

Small sample data

βœ“

Dimension reduction

βœ“

Complex data

βœ“

Big data

βœ“

Real-time data

βœ“

Dealing with missing data

βœ“

βœ“

Dealing with outliers

βœ“

βœ“

Modeling Purpose

Hypothesis check

βœ“

βœ“

Interpretation

βœ“

Comparison

βœ“

Prediction

βœ“

Modeling Type

Regression

βœ“

βœ“

Classification

βœ“

βœ“

Count outcomes

βœ“

βœ“

Discrete outcomes

βœ“

βœ“

Rule-based

βœ“

Pattern recognition

βœ“

Modeling Effort

Data preparation

βœ“

Running time

βœ“

Model validation / improvement

βœ“

Post Modeling

Sensitivity analysis

βœ“

βœ“

Elasticity / Marginal effect

βœ“

Ranking variables’ importance

βœ“

Probability (significance) tests

βœ“

Confidence intervals

βœ“

Model learning improvement

βœ“

PreviousThe Curse of DdimensionalityNextCHAPTER 2 - SUPERVISED METHODS

Last updated 1 year ago