Project page
| Paper
| BibTeX
This repo contains the reference source code of the CROW algorithm in PyTorch. CROW is a transductive learning algorithm for the cross-domain open-world discovery setting, where the goal is to recognize previously seen classes and discover novel classes simultaneously under the domain shift between labeled and unlabeled samples. CROW achieves state-of-the-art performance on the four universal domain adaptation benchmark datasets. For more details, please check our paper Cross-domain Open-world Discovery (ICML '24).
The code is built with the following libraries
Before running the code, let's prepare the data. We use standard benchmark domain adaptation datasets:
We provide the txt file for loading the data in the folder 'Dataset'. Please put the data into the same folder 'Dataset'. For example, put the data of OfficeHome domain Art into path './Dataset/OfficeHome/Art'.
In our paper, we consider 75 different experiment settings. Here, we use OfficeHome, 50% seen/unseen splitting, as an example.
- Get the CLIP pretrained model First, let's download the checkpoint of the CLIP pretrained model from here. In our experiment, we use ViT-L-14-336px. Please put it in the path './ckpt/clip/ViT-L-14-336px.pt'
After getting the CLIP pretrained model, we begin to run CROW. Steps 2 and 3 are used to get the prototypes by clustering the target samples and supervised learning to the source samples in the CLIP feature space. We provide the precomputed prototypes of both source and target samples. If you use our prototypes, please put them in the path './ckpt/prototype_source' and './ckpt/prototype_target'. Then, you can skip steps 2 and 3.
Please download the prototypes here.
Update: By using the provided prototypes, you should be able to get the results in the paper. If you run steps 2 and 3 to get the prototypes, you can get better results for Office and OfficeHome datasets than in the paper since we have further optimized some steps. In short, do not be surprised if you get better results compared to the performance in our paper :)
- Precompute and save the representation features
python get_features.py --dataset OfficeHome
The dataset should be one of the following: Office, OfficeHome, VisDA, or DomainNet. The precomputed features will be saved in the folder './representations'.
- Clustering step: find the prototypes for the source and target samples using the precomputed representation features
python get_prototype_source.py --dataset OfficeHome
python get_prototype_target.py --dataset OfficeHome
- Matching and Finetuning step:
python matching.py --dataset OfficeHome --seen_class 0.5
Seen_class is the ratio of seen classes to the total classes. For example, the OfficeHome dataset has 65 total classes, and 0.5 means we consider the setting of 33 seen classes and 32 unseen classes.
While developing CROW we greatly benefited from the open-source repositories:
If you find our code useful, please consider citing:
@inproceedings{shuo2024crow,
title={Cross-domain Open-world Discovery},
author={Wen, Shuo and Brbi’c, Maria},
booktitle={International Conference on Machine Learning},
year={2024},
}