Fair models for impartial policies: controlling algorithmic bias in transport behavioural modelling
The increasing use of new data sources and machine learning models in transport modelling raises concerns with regards to potentially unfair model-based decisions that rely on gender, age, income or other socio-economic and demographic data. We demonstrate the impact of such algorithmic bias and explore the best practices to address it using three different supervised learning models of varying levels of complexity. We also analyse how the different kinds of data (survey data vs. big data) could be associated with different levels of bias. The methodology we propose detects the model’s bias and implements measures to mitigate it. Specifically, three bias mitigation algorithms are implemented, one at each stage of the model development pipeline—before the classifier is trained (pre-processing), when training the classifier (in-processing) and after the classification (post-processing). As these debiasing techniques have an inevitable impact on the accuracy of predicting the behaviour of individuals, the comparison of different types of models and algorithms allows us to determine which techniques provide the best balance between bias mitigation and accuracy loss for each case. This approach improves model transparency and provides an objective assessment of model fairness. The results reveal that mode choice models are affected by algorithmic bias, and it is proven that the implementation of off-the-shelf mitigation techniques allows to achieve fairer classification models.
VEGA GONZALO Maria;
CHRISTIDIS Panayotis;
2022-07-12
MDPI
JRC128441
2071-1050 (online),
https://www.mdpi.com/2071-1050/14/14/8416,
https://publications.jrc.ec.europa.eu/repository/handle/JRC128441,
10.3390/su14148416 (online),
Additional supporting files
File name | Description | File type | |