Open Access System for Information Sharing

Login Library

 

Conference
Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads
Full metadata record
Files in This Item:
There are no files associated with this item.
DC FieldValueLanguage
dc.contributor.authorJang, Sooyong-
dc.contributor.authorPark, Sangdon-
dc.contributor.authorLee, Insup-
dc.contributor.authorBastani, Osbert-
dc.date.accessioned2024-02-01T01:55:08Z-
dc.date.available2024-02-01T01:55:08Z-
dc.date.created2024-01-30-
dc.date.issued2022-07-20-
dc.identifier.urihttps://oasis.postech.ac.kr/handle/2014.oak/120010-
dc.description.abstractA standard assumption in supervised learning is that the training data and test data are from the same distribution. However, this assumption often fails to hold in practice, which can cause the learned model to perform poorly. We consider the problem of detecting covariate shift, where the covariate distribution shifts but the conditional distribution of labels given covariates remains the same. This problem can naturally be solved using a two-sample test-i.e., test whether the current test distribution of covariates equals the training distribution of covariates. Our algorithm builds on classifier tests, which train a discriminator to distinguish train and test covariates, and then use the accuracy of this discriminator as a test statistic. A key challenge is that classifier tests assume given a fixed set of test covariates. In practice, test covariates often arrive sequentially over time-e.g., a self-driving car observes a stream of images while driving. Furthermore, covariate shift can occur multiple times-i.e., shift and then shift back later or gradually shift over time. To address these challenges, our algorithm trains the discriminator online. Additionally, it evaluates test accuracy using each new covariate before taking a gradient step; this strategy avoids constructing a held-out test set, which can improve sample efficiency. We prove that this optimization preserves the correctness-i.e., our algorithm achieves a desired bound on the false positive rate. In our experiments, we show that our algorithm efficiently detects covariate shifts on multiple datasets-ImageNet, IWildCam, and Py150.-
dc.languageEnglish-
dc.publisherML Research Press-
dc.relation.isPartOf39th International Conference on Machine Learning, ICML 2022-
dc.relation.isPartOfProceedings of Machine Learning Research-
dc.titleSequential Covariate Shift Detection Using Classifier Two-Sample Tests-
dc.typeConference-
dc.type.rimsCONF-
dc.identifier.bibliographicCitation39th International Conference on Machine Learning, ICML 2022, pp.9845 - 9880-
dc.citation.conferenceDate2022-07-17-
dc.citation.conferencePlaceUS-
dc.citation.endPage9880-
dc.citation.startPage9845-
dc.citation.title39th International Conference on Machine Learning, ICML 2022-
dc.contributor.affiliatedAuthorPark, Sangdon-
dc.description.journalClass1-
dc.description.journalClass1-

qr_code

  • mendeley

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher

박상돈PARK, SANGDON
Grad. School of AI
Read more

Views & Downloads

Browse