diff --git a/LRA/code/lra_config.py b/LRA/code/lra_config.py index 787df57..700abef 100755 --- a/LRA/code/lra_config.py +++ b/LRA/code/lra_config.py @@ -40,7 +40,6 @@ "nystrom-256":32, "linformer-256":32, "reformer-2":32, - "longformer-128":32, "performer-256":32, "linear":32, }, @@ -52,7 +51,6 @@ "nystrom-256":{"attention_grad_checkpointing":False, "num_landmarks":256, "conv_kernel_size":35}, "linformer-256":{"attention_grad_checkpointing":False, "linformer_k":256}, "reformer-2":{"attention_grad_checkpointing":False, "num_hash":2}, - "longformer-128":{"attention_grad_checkpointing":False, "window_size":128, "first_token_view":True}, "performer-256":{"attention_grad_checkpointing":False, "rp_dim":256, "kernel_type":"relu"}, "linear":{"attention_grad_checkpointing":False}, } @@ -97,7 +95,6 @@ "nystrom-256":128, "linformer-256":128, "reformer-2":128, - "longformer-128":128, "performer-256":128, "linear":128, }, @@ -109,7 +106,6 @@ "nystrom-256":{"attention_grad_checkpointing":False, "num_landmarks":256, "conv_kernel_size":35}, "linformer-256":{"attention_grad_checkpointing":False, "linformer_k":256}, "reformer-2":{"attention_grad_checkpointing":False, "num_hash":2}, - "longformer-128":{"attention_grad_checkpointing":False, "window_size":128, "first_token_view":True}, "performer-256":{"attention_grad_checkpointing":False, "rp_dim":256, "kernel_type":"relu"}, "linear":{"attention_grad_checkpointing":False}, } @@ -149,7 +145,6 @@ "nystrom-256":128, "linformer-256":128, "reformer-2":128, - "longformer-128":128, "performer-256":128, "linear":128, }, @@ -161,7 +156,6 @@ "nystrom-256":{"attention_grad_checkpointing":False, "num_landmarks":256, "conv_kernel_size":35}, "linformer-256":{"attention_grad_checkpointing":False, "linformer_k":256}, "reformer-2":{"attention_grad_checkpointing":False, "num_hash":2}, - "longformer-128":{"attention_grad_checkpointing":False, "window_size":128, "first_token_view":True}, "performer-256":{"attention_grad_checkpointing":False, "rp_dim":256, "kernel_type":"relu"}, "linear":{"attention_grad_checkpointing":False}, } @@ -206,7 +200,6 @@ "nystrom-256":32, "linformer-256":32, "reformer-2":32, - "longformer-128":32, "performer-256":32, "linear":32, }, @@ -218,7 +211,6 @@ "nystrom-256":{"attention_grad_checkpointing":False, "num_landmarks":256, "conv_kernel_size":35}, "linformer-256":{"attention_grad_checkpointing":False, "linformer_k":256}, "reformer-2":{"attention_grad_checkpointing":False, "num_hash":2}, - "longformer-128":{"attention_grad_checkpointing":False, "window_size":128, "first_token_view":True}, "performer-256":{"attention_grad_checkpointing":False, "rp_dim":256, "kernel_type":"relu"}, "linear":{"attention_grad_checkpointing":False}, } @@ -263,7 +255,6 @@ "nystrom-256":32, "linformer-256":32, "reformer-2":32, - "longformer-128":32, "performer-256":32, "linear":32, }, @@ -275,7 +266,6 @@ "nystrom-256":{"attention_grad_checkpointing":False, "num_landmarks":256, "conv_kernel_size":35}, "linformer-256":{"attention_grad_checkpointing":False, "linformer_k":256}, "reformer-2":{"attention_grad_checkpointing":False, "num_hash":2}, - "longformer-128":{"attention_grad_checkpointing":False, "window_size":128, "first_token_view":True}, "performer-256":{"attention_grad_checkpointing":False, "rp_dim":256, "kernel_type":"relu"}, "linear":{"attention_grad_checkpointing":False}, }