Skip to content

priyanshujiiii/Bayesian_Adaptation_for_Covariate_Shift

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

15 Commits
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

Training on Test Data with Bayesian Adaptation for Covariate Shift

Reimplementation of NeurIPS 2022 paper by Aurick Zhou & Sergey Levine
UC Berkeley, Department of Electrical Engineering and Computer Sciences


๐Ÿ“Œ Project Summary

This project reimplements the key ideas from the paper:

Training on Test Data with Bayesian Adaptation for Covariate Shift
NeurIPS 2021 โ€” Aurick Zhou, Sergey Levine

The goal is to improve model performance under distribution shifts at test time, using a Bayesian adaptation framework. Instead of training a model to handle all possible shifts, this approach adapts to the specific test-time data using unlabeled inputs via entropy minimization.


๐Ÿง  Core Idea

  • At test time, the model sees inputs from a different distribution.
  • The method uses Bayesian reasoning to adapt the model to the test data without needing labels.
  • This is achieved through a regularized entropy minimization process.
  • Leads to better accuracy and uncertainty estimation on shifted data.

๐Ÿ“Œ Project Summary

This project reimplements the key ideas from the paper:

Training on Test Data with Bayesian Adaptation for Covariate Shift
NeurIPS 2021 โ€” Aurick Zhou, Sergey Levine

The goal is to improve model performance under distribution shifts at test time, using a Bayesian adaptation framework. Instead of training a model to handle all possible shifts, this approach adapts to the specific test-time data using unlabeled inputs via entropy minimization.


๐Ÿง  Core Idea

  • At test time, the model sees inputs from a different distribution.
  • The method uses Bayesian reasoning to adapt the model to the test data without needing labels.
  • This is achieved through a regularized entropy minimization process.
  • Leads to better accuracy and uncertainty estimation on shifted data.

๐Ÿš€ Update & Experimental Notes

  • This project successfully replicates the core idea of the BACS (Bayesian Adaptation with Calibration and Smoothing) algorithm.
  • We observe that increasing the number of BACS updates significantly improves test-time performance, especially under heavy corruption.
  • All experiments were run on Google Colab due to computational limitations. As a result, we did not perform a full-scale replication of the original results.
  • Despite this, the core mechanism works well, and our results demonstrate strong performance gains under covariate shift scenarios.

๐Ÿ™ Acknowledgment

We sincerely thank the authors for introducing this elegant and practical method.
Their work inspired this reimplementation and enabled our exploration of robust test-time adaptation.


About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

โšก