Supplementary Material: Neural Predictor for Neural Architecture Search

semanticscholar(2020)

引用 0|浏览38
暂无评分
摘要
Figure 1 includes ablation study of different architectures for the Neural Predictor on NASBench-101. We compared Graph Convolutional Networks (GCN) and Multi-layer Perceptrons (MLP) in the figure. To generate inputs for a MLP, we simply concatenate the one-hot codes of node operations with the upper triangle of the adjacency matrix. From the figure, we can see such a simple MLP can outperform state-of-the-art Regularized Evolution; more importantly, the GCN that we selected achieves the best. We also tried Convolutional Neural Networks (CNN) but completely failed with a performance near to random search. During our development, we also proposed a data augmentation to improve the performance of MLP and CNN. In this augmentation, we randomly permute the order of nodes to generate new inputs online. However, we needed to perform the permutation during validation; otherwise, the validation data distribution is different from training data distribution. More importantly, GCN encodes the inductive bias that the prediction should be permutation invariant. Therefore, GCN is our final decision.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要