Idempotent Test-Time Training (
For a quick tryout of the method, check the Colab notebooks below!
Please also refer to the Zigzag paper, which served as the foundation for our work.
This paper introduces Idempotent Test-Time Training (
We train the networks on the MNIST training set and evaluate both the vanilla model and our Idempotent Test-Time Training (IT$^3$) approach on the test set with added Gaussian noise. As expected, the vanilla model shows a significant drop in performance compared to its results on the clean test data. In contrast, our IT$^3$ approach demonstrates better results with smaller accuracy degradation, showcasing its effectiveness in handling noisy inputs.
We train the networks on the CIFAR training set and evaluate both the vanilla model and our Idempotent Test-Time Training (IT$^3$) approach on a Gaussian-noised test set, similar to our experiments with MNIST. As expected, the vanilla model experiences a substantial drop in performance compared to its results on the clean test data. In contrast, our IT$^3$ approach achieves better results with less accuracy degradation, highlighting its effectiveness in handling noisy inputs.
More coming soon...
If you find this code useful, please consider citing our paper:
Durasov, Nikita, et al. "IT$^3$: Idempotent Test-Time Training." arXiv 2024.
@article{durasov2024ittt,
title = {IT $\^{} 3$: Idempotent Test-Time Training},
author = {Durasov, Nikita and Shocher, Assaf and Oner, Doruk and Chechik, Gal and Efros, Alexei A and Fua, Pascal},
journal = {arXiv preprint arXiv:2410.04201},
year = {2024}
}