-
Notifications
You must be signed in to change notification settings - Fork 3.6k
public datasets for evaluation #45
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Same q here. Would be great and appreciate it if you could share which sources you get those public datasets for evaluation. Thanks. For example, here is the table of stats about FER-2013 dataset in the the other paper which is consistent with Kaggle page, but different than the stats reported from CLIP paper In the paper cited by CLIP paper about FER-2013, it says "The resulting |
Hi, thanks for pointing out some of the details we were cursory or missing; upon investigating, we found that:
root = f"{DATA_ROOT}/eurosat/2750"
seed = 42
random.seed(seed)
train_paths, valid_paths, test_paths = [], [], []
for folder in [os.path.basename(folder) for folder in sorted(glob.glob(os.path.join(root, "*")))]:
keep_paths = random.sample(glob.glob(os.path.join(root, folder, "*")), 1500)
keep_paths = [os.path.relpath(path, root) for path in keep_paths]
train_paths.extend(keep_paths[:500])
valid_paths.extend(keep_paths[500:1000])
test_paths.extend(keep_paths[1000:]) We could’ve used a better setup such as mean-per-class using all available data and would rather encourage future studies to do so, while we note that the comparisons in the paper used this same subset across all models, so their relative scores can still be considered “fair”.
root = f"{DATA_ROOT}/resisc45"
seed = 42
paths = sorted(glob.glob(os.path.join(root, "*.jpg")))
random.seed(seed)
random.shuffle(paths)
if split == 'train':
paths = paths[:len(paths) // 10]
elif split == 'valid':
paths = paths[len(paths) // 10:(len(paths) // 10) * 2]
elif split == 'test':
paths = paths[(len(paths) // 10) * 2:]
else:
raise NotImplementedError
|
Hi Jong, thanks so much for all the information! It is super helpful. I have some questions about two more datasets and would really appreciate it if you could help. Thanks in advance. Birdsnap: the official site only provides the image urls. When using the associated script from the official dataset to download the images, I ended up with ""“NEW_OK:40318, ALREADY_OK:0, DOWNLOAD_FAILED:5030, SAVE_FAILED:0, MD5_FAILED:4481, MYSTERY_FAILED:0.”. Have you folks experienced similar problems? CLEVR(Counts): the CLIP paper says "2,500 random samples of the CLEVR Thanks! |
@jongwook Thank you so much for sharing these details. I have two more detailed questions: |
Hi there,
I'm trying to set up public datasets for evaluation listed in Table 9, but got different train/test size for some datasets:
Dataset I found on Kaggle has train dataset 28,709, Val(public test) 3,589, (Train+Val 32,298 in total) and Test (private test) 3,589.
Tensorflow stl10 has training dataset with 5,000 images and testing dataset with 8,000.
Tensorflow eurosat only has training dataset with 27,000 images.
The site Tensorflow refers to only have training dataset, which is 31,500 images.
This archive I found has 2 training datasets (GTSRB_Final_Training_Images.zip and GTSRB-Training_fixed.zip), but both have size different from Table 9.
This is what Table 9 shows:
It would be greatly appreciated if you could point me to the source of data split shown in Table 9.
The text was updated successfully, but these errors were encountered: