This is the code for our paper "PINAT: A Permutation INvariance Augmented Transformer for NAS Predictor".
As AAAI does not publish the supplementary material, we provide this file in the folder of assets.
This repository contains the code of PINAT. To run the codes, please see the prerequisites below:
- Download the datasets of NAS-Bench-101 and NAS-Bench-201, and pre-trained checkpoints from the Google Drive.
- Install the necessary packages below via pip or conda. We used Anaconda Python 3.8 for our experiments.
dgl==0.6.1
h5py==3.7.0
matplotlib==3.6.2
networkx==2.8.7
opencv-python==4.6.0.66
pandas==1.5.1
Pillow==9.2.0
prettytable==3.4.1
pynvml==11.4.1
PyYAML==6.0
schema==0.7.5
scikit-learn==1.1.2
scipy==1.9.3
seaborn==0.12.1
tensorboard==2.10.1
tensorflow-gpu==2.10.0
torch==1.9.0
torch-cluster==1.6.0
torch-geometric==2.1.0.post1
torch-scatter==2.0.9
torch-sparse==0.6.12
torchvision==0.10.0
tqdm==4.64.1- We also summarize our environments into the
requirements.txt. To install a same environment, simply run:
pip install -r requirements.txtGo to the folder of nasbench and please refer to nasbench/README.md
Go to the folder of darts and please refer to darts/README.md
If you find this work helpful in your research, please consider citing our paper:
@inproceedings{lu2023pinat,
title = {PINAT: A Permutation INvariance Augmented Transformer for NAS Predictor},
author = {Lu, Shun and Hu, Yu and Wang, Peihao and Han, Yan and Tan, Jianchao and Li, Jixiang and Yang, Sen and Liu, Ji},
booktitle = {Proceedings of the AAAI Conference on Artificial Intelligence (AAAI)},
year = {2023}
}
