Skip to content

Commit

Permalink
FIX: added DOIs
Browse files Browse the repository at this point in the history
  • Loading branch information
VincentAuriau committed Jun 13, 2024
1 parent cc20f32 commit 0ef1aed
Showing 1 changed file with 16 additions and 4 deletions.
20 changes: 16 additions & 4 deletions docs/paper/paper.bib
Original file line number Diff line number Diff line change
Expand Up @@ -15,6 +15,7 @@ @article{Brathwaite:2018
journal={Journal of Choice Modelling},
publisher={Elsevier BV},
author={Brathwaite, Timothy and Walker, Joan L.},
doi = {https://doi.org/10.1016/j.jocm.2018.01.002},
year={2018},
month=dec,
pages={78–112}
Expand All @@ -25,12 +26,14 @@ @article{Du:2023
author={Tianyu Du and Ayush Kanodia and Susan Athey},
year={2023},
journal={arXiv preprint arXiv:{2304.01906}},
doi = {https://doi.org/10.48550/arXiv.2304.01906},
}

@article{Aouad:2023,
title={Representing random utility choice models with neural networks},
author={Aouad, Ali and D{\'e}sir, Antoine},
journal={arXiv preprint arXiv:2207.12877},
doi = {https://doi.org/10.48550/arXiv.2207.12877},
year={2022}
}

Expand All @@ -43,13 +46,15 @@ @article{Han:2022
issn = {0191-2615},
author = {Yafei Han and Francisco Camara Pereira and Moshe Ben-Akiva and Christopher Zegras},
keywords = {Discrete choice models, Neural networks, Taste heterogeneity, Interpretability, Utility specification, Machine learning, Deep learning},
doi = {https://doi.org/10.1016/j.trb.2022.07.001},
abstract = {Discrete choice models (DCMs) require a priori knowledge of the utility functions, especially how tastes vary across individuals. Utility misspecification may lead to biased estimates, inaccurate interpretations and limited predictability. In this paper, we utilize a neural network to learn taste representation. Our formulation consists of two modules: a neural network (TasteNet) that learns taste parameters (e.g., time coefficient) as flexible functions of individual characteristics; and a multinomial logit (MNL) model with utility functions defined with expert knowledge. Taste parameters learned by the neural network are fed into the choice model and link the two modules. Our approach extends the L-MNL model (Sifringer et al., 2020) by allowing the neural network to learn the interactions between individual characteristics and alternative attributes. Moreover, we formalize and strengthen the interpretability condition — requiring realistic estimates of behavior indicators (e.g., value-of-time, elasticity) at the disaggregated level, which is crucial for a model to be suitable for scenario analysis and policy decisions. Through a unique network architecture and parameter transformation, we incorporate prior knowledge and guide the neural network to output realistic behavior indicators at the disaggregated level. We show that TasteNet-MNL reaches the ground-truth model’s predictability and recovers the nonlinear taste functions on synthetic data. Its estimated value-of-time and choice elasticities at the individual level are close to the ground truth. In contrast, exemplary logit models with misspecified systematic utility lead to biased parameter estimates and lower prediction accuracy. On a publicly available Swissmetro dataset, TasteNet-MNL outperforms benchmarking MNLs and Mixed Logit model’s predictability. It learns a broader spectrum of taste variations within the population and suggests a higher average value-of-time. Our source code is available for research and application.}
}

@article{Salvadé:2024,
title={RUMBoost: Gradient Boosted Random Utility Models},
author={Salvad{\'e}, Nicolas and Hillel, Tim},
journal={arXiv preprint arXiv:2401.11954},
doi={https://doi.org/10.48550/arXiv.2401.11954},
year={2024}
}

Expand All @@ -63,7 +68,8 @@ @article{Train:1987
publisher = {[RAND Corporation, Wiley]},
title = {The Demand for Local Telephone Service: A Fully Discrete Model of Residential Calling Patterns and Service Choices},
volume = {18},
year = {1987}
year = {1987},
doi = {https://doi.org/10.2307/2555538},
}

@Article{Harris:2020,
Expand All @@ -83,6 +89,7 @@ @Article{Harris:2020
volume = {585},
number = {7825},
pages = {357--362},
doi = {10.1038/s41586-020-2649-2},
publisher = {Springer Science and Business Media {LLC}},
}

Expand All @@ -91,7 +98,8 @@ @software{Abadi:2015
license = {Apache-2.0},
month = nov,
title = {{TensorFlow, Large-scale machine learning on heterogeneous systems}},
year = {2015}
year = {2015},
doi = {10.5281/zenodo.4724125}
}

@Inbook{Nocedal:2006,
Expand All @@ -103,6 +111,7 @@ @Inbook{Nocedal:2006
address="New York, NY",
pages="164--192",
isbn="978-0-387-40065-5",
doi = {https://doi.org/10.1007/0-306-48332-7_250}
}

@article{Kingma:2017,
Expand All @@ -112,6 +121,7 @@ @article{Kingma:2017
archivePrefix={arXiv},
primaryClass={cs.LG},
journal={arXiv preprint arXiv:{1412.6980}},
doi = {https://doi.org/10.48550/arXiv.1412.6980}
}

@article{Tieleman:2012,
Expand All @@ -138,7 +148,8 @@ @article{AouadMarket:2023
number={2},
pages={648--667},
year={2023},
publisher={INFORMS}
publisher={INFORMS},
doi={https://doi.org/10.1287/msom.2023.1195},
}

@article{MendezDiaz:2014,
Expand All @@ -151,7 +162,7 @@ @article{MendezDiaz:2014
issn = {0166-218X},
author = {Isabel Méndez-Díaz and Juan José Miranda-Bront and Gustavo Vulcano and Paula Zabala},
keywords = {Retail operations, Revenue management, Choice behavior, Multinomial logit, Integer programming, Fractional programming},
abstract = {We study the product assortment problem of a retail operation that faces a stream of customers who are heterogeneous with respect to preferences. Each customer belongs to a market segment characterized by a consideration set that includes the alternatives viewed as options, and by the preference weights that the segment assigns to each of those alternatives. Upon arrival, he checks the offer set displayed by the firm, and either chooses one of those products or quits without purchasing according to a multinomial-logit (MNL) criterion. The firm’s goal is to maximize the expected revenue extracted during a fixed time horizon. This problem also arises in the growing area of choice-based, network revenue management, where computational speed is a critical factor for the practical viability of a solution approach. This so-called latent-class, logit assortment problem is known to be NP-Hard. In this paper, we analyze unconstrained and constrained (i.e., with a limited number of products to display) versions of it, and propose a branch-and-cut algorithm that is computationally fast and leads to (nearly) optimal solutions.}
doi = {https://doi.org/10.1016/j.dam.2012.03.003},
}

@software{pandas:2020,
Expand All @@ -160,6 +171,7 @@ @software{pandas:2020
month = feb,
year = 2020,
publisher = {Zenodo},
doi = {10.5281/zenodo.3509134},
}

@inproceedings{Bierlaire:2001,
Expand Down

0 comments on commit 0ef1aed

Please sign in to comment.