Document Type
Conference Proceeding
Publication Title
Proceedings of Machine Learning Research
Abstract
Given an algorithmic predictor that is accurate on some source population consisting of strategic human decision subjects, will it remain accurate if the population respond to it? In our setting, an agent or a user corresponds to a sample (X, Y ) drawn from a distribution D and will face a model h and its classification result h(X). Agents can modify X to adapt to h, which will incur a distribution shift on (X, Y ). Our formulation is motivated by applications where the deployed machine learning models are subjected to human agents, and will ultimately face responsive and interactive data distributions. We formalize the discussions of the transferability of a model by studying how the performance of the model trained on the available source distribution (data) would translate to the performance on its induced domain. We provide both upper bounds for the performance gap due to the induced domain shift, as well as lower bounds for the trade-offs that a classifier has to suffer on either the source training distribution or the induced target distribution. We provide further instantiated analysis for two popular domain adaptation settings, including covariate shift and target shift.
First Page
4514
Last Page
4528
Publication Date
7-2023
Keywords
Algorithmics, Classification results, Data distribution, Decision subject, Human agent, Human decisions, Machine learning models, Model transferabilities, Performance, Source population
Recommended Citation
Y. Chen, Z. Tang, K. Zhang and Y. Liu, "Model Transferability with Responsive Decision Subjects," Proceedings of Machine Learning Research, vol. 202, pp. 4514 - 4528, Jul 2023.
Comments
Open Access version from PMLR
Uploaded on June 13, 2024