[go: nahoru, domu]

Skip to content

Commit

Permalink
reorg Mamba folder
Browse files Browse the repository at this point in the history
  • Loading branch information
ChloeXWang committed Feb 1, 2024
1 parent 7579c47 commit f6aa31e
Show file tree
Hide file tree
Showing 37 changed files with 1,256 additions and 154 deletions.
2 changes: 1 addition & 1 deletion configs/Exphormer/cifar10.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@ metric_best: accuracy
wandb:
use: True
project: GNN-Benchmark-RandomSeed-Exphormer
entity: tf-map
entity: project
dataset:
format: PyG-GNNBenchmarkDataset
name: CIFAR10
Expand Down
2 changes: 1 addition & 1 deletion configs/Exphormer/cluster.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@ metric_best: accuracy-SBM
wandb:
use: True
project: GNN-Benchmark-RandomSeed-Exphormer #flop_experiment_re #graph-mamba
entity: tf-map
entity: project
dataset:
format: PyG-GNNBenchmarkDataset
name: CLUSTER
Expand Down
2 changes: 1 addition & 1 deletion configs/Exphormer/malnettiny.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@ metric_best: accuracy
wandb:
use: True
project: GNN-Benchmark-RandomSeed-Exphormer
entity: tf-map
entity: project
dataset:
format: PyG-MalNetTiny
name: LocalDegreeProfile
Expand Down
2 changes: 1 addition & 1 deletion configs/Exphormer/mnist.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@ metric_best: accuracy
wandb:
use: True
project: GNN-Benchmark-RandomSeed-Exphormer #graph-mamba
entity: tf-map
entity: project
dataset:
format: PyG-GNNBenchmarkDataset
name: MNIST
Expand Down
2 changes: 1 addition & 1 deletion configs/Exphormer/pattern.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@ metric_best: accuracy-SBM
wandb:
use: True
project: GNN-Benchmark-RandomSeed-Exphormer #graph-mamba
entity: tf-map
entity: project
dataset:
format: PyG-GNNBenchmarkDataset
name: PATTERN
Expand Down
4 changes: 2 additions & 2 deletions configs/Exphormer_LRGB/cocosuperpixels-EX1.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -2,8 +2,8 @@ out_dir: results
metric_best: f1
wandb:
use: True
project: LRGB-Benchmark-RandomSeed-Exphormer #graph-mamba #flop_experiment_re
entity: tf-map
project: LRGB-Benchmark-RandomSeed-Exphormer
entity: project
dataset:
format: PyG-COCOSuperpixels
name: edge_wt_region_boundary
Expand Down
4 changes: 2 additions & 2 deletions configs/Exphormer_LRGB/pcqm-contact-EX1.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,8 +3,8 @@ metric_best: mrr
num_workers: 4
wandb:
use: True
project: Graph-Mamba-Benchmark-GNN #LRGB-Benchmark-RandomSeed-Exphormer #flop_experiment_re #pcqm4m-contact
entity: tf-map
project: Graph-Mamba-Benchmark-GNN
entity: project
dataset:
format: OGB
name: PCQM4Mv2Contact-shuffle
Expand Down
4 changes: 2 additions & 2 deletions configs/Exphormer_LRGB/peptides-func-EX1.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -2,8 +2,8 @@ out_dir: results
metric_best: ap
wandb:
use: True
project: LRGB-Benchmark-RandomSeed-Exphormer #flop_experiment_re #graph-mamba
entity: tf-map
project: LRGB-Benchmark-RandomSeed-Exphormer
entity: project
dataset:
format: OGB
name: peptides-functional
Expand Down
4 changes: 2 additions & 2 deletions configs/Exphormer_LRGB/peptides-struct-EX.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,8 +3,8 @@ metric_best: mae
metric_agg: argmin
wandb:
use: True
project: LRGB-Benchmark-RandomSeed-Exphormer #graph-mamba #flop_experiment_re
entity: tf-map
project: LRGB-Benchmark-RandomSeed-Exphormer
entity: project
dataset:
format: OGB
name: peptides-structural
Expand Down
4 changes: 2 additions & 2 deletions configs/Exphormer_LRGB/vocsuperpixels-EX1.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -2,8 +2,8 @@ out_dir: results
metric_best: f1
wandb:
use: True
project: LRGB-Benchmark-RandomSeed-Exphormer #graph-mamba #flop_experiment_re
entity: tf-map
project: LRGB-Benchmark-RandomSeed-Exphormer
entity: project
dataset:
format: PyG-VOCSuperpixels
name: edge_wt_region_boundary
Expand Down
15 changes: 3 additions & 12 deletions configs/GPS/cifar10-GPS.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ model:
edge_decoding: dot
graph_pooling: mean
gt: # Hyperparameters optimized for ~100k budget.
layer_type: CustomGatedGCN+Mamba_Hybrid_Degree_Noise
layer_type: CustomGatedGCN+Transformer
layers: 3
n_heads: 4
dim_hidden: 52 # `gt.dim_hidden` must match `gnn.dim_inner`
Expand All @@ -58,17 +58,8 @@ gnn:
optim:
clip_grad_norm: True
optimizer: adamW
weight_decay: 0.01
base_lr: 0.005
weight_decay: 1e-5
base_lr: 0.001
max_epoch: 100
scheduler: cosine_with_warmup
num_warmup_epochs: 5
seed: 0
# optim:
# clip_grad_norm: True
# optimizer: adamW
# weight_decay: 1e-5
# base_lr: 0.001
# max_epoch: 100
# scheduler: cosine_with_warmup
# num_warmup_epochs: 5
15 changes: 3 additions & 12 deletions configs/GPS/cluster-GPS.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ model:
loss_fun: weighted_cross_entropy
edge_decoding: dot
gt: # Hyperparameters optimized for ~500k budget.
layer_type: CustomGatedGCN+Mamba_Hybrid_Degree_Noise
layer_type: CustomGatedGCN+Transformer
layers: 16
n_heads: 8
dim_hidden: 48 # `gt.dim_hidden` must match `gnn.dim_inner`
Expand All @@ -59,17 +59,8 @@ gnn:
optim:
clip_grad_norm: True
optimizer: adamW
weight_decay: 0.0001
base_lr: 0.001
weight_decay: 1e-5
base_lr: 0.0005
max_epoch: 100
scheduler: cosine_with_warmup
num_warmup_epochs: 5
seed: 0
# optim:
# clip_grad_norm: True
# optimizer: adamW
# weight_decay: 1e-5
# base_lr: 0.0005
# max_epoch: 100
# scheduler: cosine_with_warmup
# num_warmup_epochs: 5
27 changes: 9 additions & 18 deletions configs/GPS/cocosuperpixels-GPS.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@ out_dir: results
metric_best: f1
wandb:
use: True
project: LRGB-Benchmark-RandomSeed #flop_experiment_re
project: LRGB-Benchmark-RandomSeed
entity: project
dataset:
format: PyG-COCOSuperpixels
Expand Down Expand Up @@ -36,7 +36,7 @@ model:
type: GPSModel
loss_fun: weighted_cross_entropy
gt:
layer_type: CustomGatedGCN+Mamba_Hybrid_Degree_Noise_Bucket
layer_type: CustomGatedGCN+Transformer
layers: 4
n_heads: 8
dim_hidden: 96 # `gt.dim_hidden` must match `gnn.dim_inner`
Expand All @@ -54,20 +54,11 @@ gnn:
dropout: 0.0
agg: mean
normalize_adj: False
# optim:
# clip_grad_norm: True
# optimizer: adamW
# weight_decay: 0.0
# base_lr: 0.0005
# max_epoch: 300
# scheduler: cosine_with_warmup
# num_warmup_epochs: 10
optim:
clip_grad_norm: True
optimizer: adamW
weight_decay: 0.001
base_lr: 0.0015
max_epoch: 300
scheduler: cosine_with_warmup
num_warmup_epochs: 10
seed: 0
clip_grad_norm: True
optimizer: adamW
weight_decay: 0.0
base_lr: 0.0005
max_epoch: 300
scheduler: cosine_with_warmup
num_warmup_epochs: 10
7 changes: 3 additions & 4 deletions configs/GPS/malnettiny-GPS.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@ out_dir: results
metric_best: accuracy
wandb:
use: True
project: Graph-Mamba-Benchmark-GNN
project: GNN-Benchmark-RandomSeed
entity: project
dataset:
format: PyG-MalNetTiny
Expand All @@ -26,7 +26,7 @@ model:
edge_decoding: dot
graph_pooling: max
gt:
layer_type: CustomGatedGCN+Mamba_Hybrid_Degree_Noise_Bucket
layer_type: CustomGatedGCN+Transformer
layers: 5
n_heads: 4
dim_hidden: 64 # `gt.dim_hidden` must match `gnn.dim_inner`
Expand All @@ -48,11 +48,10 @@ optim:
clip_grad_norm: True
optimizer: adamW
weight_decay: 1e-5
base_lr: 0.0015
base_lr: 0.0005
max_epoch: 150
scheduler: cosine_with_warmup
num_warmup_epochs: 10
seed: 0
# optim:
# clip_grad_norm: True
# optimizer: adamW
Expand Down
17 changes: 4 additions & 13 deletions configs/GPS/mnist-GPS.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ model:
edge_decoding: dot
graph_pooling: mean
gt: # Hyperparameters optimized for ~100k budget.
layer_type: CustomGatedGCN+Mamba_Hybrid_Degree_Noise
layer_type: CustomGatedGCN+Transformer
layers: 3
n_heads: 4
dim_hidden: 52 # `gt.dim_hidden` must match `gnn.dim_inner`
Expand All @@ -58,17 +58,8 @@ gnn:
optim:
clip_grad_norm: True
optimizer: adamW
weight_decay: 0.01
base_lr: 0.005
weight_decay: 1e-5
base_lr: 0.001
max_epoch: 100
scheduler: cosine_with_warmup
num_warmup_epochs: 5
seed: 0
# optim:
# clip_grad_norm: True
# optimizer: adamW
# weight_decay: 1e-5
# base_lr: 0.001
# max_epoch: 100
# scheduler: cosine_with_warmup
# num_warmup_epochs: 5
num_warmup_epochs: 5
15 changes: 3 additions & 12 deletions configs/GPS/pattern-GPS.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ model:
loss_fun: weighted_cross_entropy
edge_decoding: dot
gt: # Hyperparameters optimized for up to ~500k budget.
layer_type: CustomGatedGCN+Mamba_Hybrid_Degree_Noise
layer_type: CustomGatedGCN+Transformer
layers: 6
n_heads: 4
dim_hidden: 64 # `gt.dim_hidden` must match `gnn.dim_inner`
Expand All @@ -59,17 +59,8 @@ gnn:
optim:
clip_grad_norm: True
optimizer: adamW
weight_decay: 0.0001
base_lr: 0.001
weight_decay: 1e-5
base_lr: 0.0005
max_epoch: 100
scheduler: cosine_with_warmup
num_warmup_epochs: 5
seed: 0
# optim:
# clip_grad_norm: True
# optimizer: adamW
# weight_decay: 1e-5
# base_lr: 0.0005
# max_epoch: 100
# scheduler: cosine_with_warmup
# num_warmup_epochs: 5
25 changes: 3 additions & 22 deletions configs/GPS/pcqm-contact-GPS.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ model:
loss_fun: cross_entropy
edge_decoding: dot
gt:
layer_type: CustomGatedGCN+Mamba_Hybrid_Degree_Noise #Mamba_Hybrid_Degree_Noise #Mamba
layer_type: CustomGatedGCN+Transformer #Mamba_Hybrid_Degree_Noise #Mamba
layers: 4
n_heads: 4
dim_hidden: 96 # `gt.dim_hidden` must match `gnn.dim_inner`
Expand All @@ -57,27 +57,8 @@ gnn:
optim:
clip_grad_norm: True
optimizer: adamW
weight_decay: 0.01
base_lr: 0.002
weight_decay: 0.0
base_lr: 0.0003
max_epoch: 100
scheduler: cosine_with_warmup
num_warmup_epochs: 10
seed: 0
# optim:
# clip_grad_norm: True
# optimizer: adamW
# weight_decay: 0.0
# base_lr: 0.0003
# max_epoch: 200
# scheduler: cosine_with_warmup
# num_warmup_epochs: 10
#optim:
# batch_accumulation: 1
# optimizer: adamW
# weight_decay: 0.0
# base_lr: 0.0003
# max_epoch: 200
# scheduler: reduce_on_plateau
# reduce_factor: 0.5
# schedule_patience: 10
# min_lr: 1e-5
18 changes: 4 additions & 14 deletions configs/GPS/peptides-func-GPS.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ model:
loss_fun: cross_entropy
graph_pooling: mean
gt:
layer_type: CustomGatedGCN+Mamba_Hybrid_Degree_Noise
layer_type: CustomGatedGCN+Transformer
n_heads: 4
dim_hidden: 96 # `gt.dim_hidden` must match `gnn.dim_inner`
dropout: 0.0
Expand All @@ -55,18 +55,8 @@ gnn:
optim:
clip_grad_norm: True
optimizer: adamW
weight_decay: 0.01
base_lr: 0.001
weight_decay: 0.0
base_lr: 0.0003
max_epoch: 200
scheduler: cosine_with_warmup
num_warmup_epochs: 10
seed: 0
#optim:
# optimizer: adamW
# weight_decay: 0.0
# base_lr: 0.0003
# max_epoch: 500
# scheduler: reduce_on_plateau
# reduce_factor: 0.5
# schedule_patience: 20
# min_lr: 1e-5
num_warmup_epochs: 10
18 changes: 4 additions & 14 deletions configs/GPS/peptides-struct-GPS.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ model:
loss_fun: l1
graph_pooling: mean
gt:
layer_type: CustomGatedGCN+Mamba_Hybrid_Degree_Noise
layer_type: CustomGatedGCN+Transformer
layers: 4
n_heads: 4
dim_hidden: 96 # `gt.dim_hidden` must match `gnn.dim_inner`
Expand All @@ -56,18 +56,8 @@ gnn:
optim:
clip_grad_norm: True
optimizer: adamW
weight_decay: 0.01
base_lr: 0.001
weight_decay: 0.0
base_lr: 0.0003
max_epoch: 200
scheduler: cosine_with_warmup
num_warmup_epochs: 10
seed: 0
#optim:
# optimizer: adamW
# weight_decay: 0.0
# base_lr: 0.0003
# max_epoch: 500
# scheduler: reduce_on_plateau
# reduce_factor: 0.5
# schedule_patience: 20
# min_lr: 1e-5
num_warmup_epochs: 10
Loading

0 comments on commit f6aa31e

Please sign in to comment.