Adapted from HGB benchmark.
For the results of original HGB, run
python run_new.py --dataset DBLP --seeds 1 2 3 4 5
python run_new.py --dataset ACM --feats-type 2 --lr 0.001 --seeds 1 2 3 4 5
To eliminate the influence of neighbor attention, ${\rm HGB}^*$ adds additional functions after attention calculation to average the weights of neighbors with the same node type. Similar methods are used to eliminate the influence of semantic attention, and the results are represented as ${\rm {HGB}^\dagger}$.
For the results of our MeanAggregator (i.e. ${\rm HGB}^*$) which removes neighbor attention, run
python run_new.py --dataset DBLP --average-attention-values --seeds 1 2 3 4 5
python run_new.py --dataset ACM --feats-type 2 --average-attention-values --lr 0.001 --seeds 1 2 3 4 5
For the results ${\rm HAN}^\dagger$ which removes semantic attention, run
python run_new.py --dataset DBLP --average-semantic-values --seeds 1 2 3 4 5
python run_new.py --dataset ACM --feats-type 2 --average-semantic-values --lr 0.001 --seeds 1 2 3 4 5