Supplementary for Trustworthy model evaluation on a budget
[Paper]
Published at ICLR 2023 Workshop on Trustworthy and Reliable Large-Scale Machine Learning Models
@inproceedings{fostiropoulos2023trustworthy,
title={Trustworthy model evaluation on a budget},
author={Fostiropoulos, Iordanis and Brown, Bowman Noah and Itti, Laurent},
booktitle={ICLR 2023 Workshop on Trustworthy and Reliable Large-Scale Machine Learning Models}
}
Instructions based on Ubuntu 18.04 with python 3.10+
-
git clone https://github.com/fostiropoulos/trust_ml
-
cd trust_ml
-
pip install -e .
The raw data used for our experiments:
Dong, Xuanyi, et al. "Nats-bench: Benchmarking nas algorithms for architecture topology and size." IEEE transactions on pattern analysis and machine intelligence 44.7 (2021): 3634-3646.
NOTE We provide the preprocessed dataset in dataset.pickle
https://github.com/D-X-Y/NATS-Bench
https://drive.google.com/file/d/1vzyK0UVH2D3fTpa1_dSWnp1gvGpAxRul/view?usp=sharing
You can download from command-line:
pip install gdown
gdown https://drive.google.com/uc?id=1vzyK0UVH2D3fTpa1_dSWnp1gvGpAxRul
python -m trustml.exp2.evaluate
To re-make the NATS-Bench dataset you can run
python -m trustml.exp1.data
The preprocessed dataset is saved at data/dataset.pickle
To re-run the ablation of the sampler you can delete data/results.pickle
python -m trustml.exp1.main
python -m trustml.exp2.train