Deep neural networks for choice analysis: A statistical learning theory perspective

•Used statistical learning theory to evaluate DNNs in choice analysis.•Operationalized DNN interpretability by using the choice probability functions.•Provided a tight upper bound on the estimation error of DNNs.•Conducted experiments to identify when DNNs outperform classical models.•DNNs can be mo...

Full description

Saved in:
Bibliographic Details
Published in:Transportation research. Part B: methodological Vol. 148; pp. 60 - 81
Main Authors: Wang, Shenhao, Wang, Qingyi, Bailey, Nate, Zhao, Jinhua
Format: Journal Article
Language:English
Published: Oxford Elsevier Ltd 01.06.2021
Elsevier Science Ltd
Subjects:
ISSN:0191-2615, 1879-2367
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:•Used statistical learning theory to evaluate DNNs in choice analysis.•Operationalized DNN interpretability by using the choice probability functions.•Provided a tight upper bound on the estimation error of DNNs.•Conducted experiments to identify when DNNs outperform classical models.•DNNs can be more predictive and interpretable than BNL and MNL models. Although researchers increasingly use deep neural networks (DNN) to analyze individual choices, overfitting and interpretability issues remain obstacles in theory and practice. This study presents a statistical learning theoretical framework to examine the tradeoff between estimation and approximation errors, and between the quality of prediction and of interpretation. It provides an upper bound on the estimation error of the prediction quality in DNN, measured by zero-one and log losses, shedding light on why DNN models do not overfit. It proposes a metric for interpretation quality by formulating a function approximation loss that measures the difference between true and estimated choice probability functions. It argues that the binary logit (BNL) and multinomial logit (MNL) models are the specific cases of DNNs, since the latter always has smaller approximation errors. We explore the relative performance of DNN and classical choice models through three simulation scenarios comparing DNN, BNL, and binary mixed logit models (BXL), as well as one experiment comparing DNN to BNL, BXL, MNL, and mixed logit (MXL) in analyzing the choice of trip purposes based on the National Household Travel Survey 2017. The results indicate that DNN can be used for choice analysis beyond the current practice of demand forecasting because it has the inherent utility interpretation and the power of automatically learning utility specification. Our results suggest DNN outperforms BNL, BXL, MNL, and MXL models in both prediction and interpretation when the sample size is large (≥O(104)), the input dimension is high, or the true data generating process is complex, while performing worse when the opposite is true. DNN outperforms BNL and BXL in zero-one, log, and approximation losses for most of the experiments, and the larger sample size leads to greater incremental value of using DNN over classical discrete choice models. Overall, this study introduces the statistical learning theory as a new foundation for high-dimensional data, complex statistical models, and non-asymptotic data regimes in choice analysis, and the experiments show the effective prediction and interpretation of DNN for its applications to policy and behavioral analysis.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:0191-2615
1879-2367
DOI:10.1016/j.trb.2021.03.011