diff --git a/configs/sst2/16-100-sst2-conti-demon-prompting.json b/configs/sst2/16-100-sst2-conti-demon-prompting.json new file mode 100644 index 0000000..170eb58 --- /dev/null +++ b/configs/sst2/16-100-sst2-conti-demon-prompting.json @@ -0,0 +1,26 @@ +{ + "train_batch_size": 8, + "eval_batch_size": 8, + "task_name": "sst-2", + "random_seed": 13, + "k_shot": 16, + "dataset_dir": "./rsc/eng_few_shot/k-shot/SST-2/16-100", + "learning_rate": 3e-05, + "max_grad_norm": 1.0, + "lm_model": "roberta-large", + "manual_template": "It was .", + "max_seq_length": 128, + "gradient_accumulation_steps": 1, + "num_train_epochs": 50, + "log_step_count_steps": 20, + "lambda_rate": 0.3, + "eval_every_step": null, + "demonstration_tokens": [ + ["well worth the effort", "a total waste of my time", "a real treat to watch", "so sad", "a cultural revolution"], + ["an instant hit", "a gift", "entertaining on an inferior level", "an unforgettable experience", "a thriller with an edge"] + ], + "soft_token_length": 10, + "is_cuda": true, + "warmup_proportion": 0.6, + "tensorboard_output_dir": "./logs/sst2/16-100-soft-10-k5-lambda0.3" +} \ No newline at end of file diff --git a/configs/sst2/16-13-sst2-conti-demon-prompting.json b/configs/sst2/16-13-sst2-conti-demon-prompting.json new file mode 100644 index 0000000..f620958 --- /dev/null +++ b/configs/sst2/16-13-sst2-conti-demon-prompting.json @@ -0,0 +1,26 @@ +{ + "train_batch_size": 8, + "eval_batch_size": 8, + "task_name": "sst-2", + "random_seed": 13, + "k_shot": 16, + "dataset_dir": "./rsc/eng_few_shot/k-shot/SST-2/16-13", + "learning_rate": 3e-05, + "max_grad_norm": 1.0, + "lm_model": "roberta-large", + "manual_template": "It was .", + "max_seq_length": 128, + "gradient_accumulation_steps": 1, + "num_train_epochs": 50, + "log_step_count_steps": 20, + "lambda_rate": 0.3, + "eval_every_step": null, + "demonstration_tokens": [ + ["well worth the effort", "a total waste of my time", "a real treat to watch", "so sad", "a cultural revolution"], + ["an instant hit", "a gift", "entertaining on an inferior level", "an unforgettable experience", "a thriller with an edge"] + ], + "soft_token_length": 10, + "is_cuda": true, + "warmup_proportion": 0.6, + "tensorboard_output_dir": "./logs/sst2/16-13-soft-10-k5" +} \ No newline at end of file diff --git a/configs/sst2/16-21-sst2-conti-demon-prompting.json b/configs/sst2/16-21-sst2-conti-demon-prompting.json new file mode 100644 index 0000000..c91c2d0 --- /dev/null +++ b/configs/sst2/16-21-sst2-conti-demon-prompting.json @@ -0,0 +1,26 @@ +{ + "train_batch_size": 8, + "eval_batch_size": 8, + "task_name": "sst-2", + "random_seed": 13, + "k_shot": 16, + "dataset_dir": "./rsc/eng_few_shot/k-shot/SST-2/16-21", + "learning_rate": 3e-05, + "max_grad_norm": 1.0, + "lm_model": "roberta-large", + "manual_template": "It was .", + "max_seq_length": 128, + "gradient_accumulation_steps": 1, + "num_train_epochs": 50, + "log_step_count_steps": 20, + "lambda_rate": 0.3, + "eval_every_step": null, + "demonstration_tokens": [ + ["well worth the effort", "a total waste of my time", "a real treat to watch", "so sad", "a cultural revolution"], + ["an instant hit", "a gift", "entertaining on an inferior level", "an unforgettable experience", "a thriller with an edge"] + ], + "soft_token_length": 10, + "is_cuda": true, + "warmup_proportion": 0.6, + "tensorboard_output_dir": "./logs/sst2/16-21-soft-10-k5" +} \ No newline at end of file diff --git a/configs/sst2/16-42-sst2-conti-demon-prompting.json b/configs/sst2/16-42-sst2-conti-demon-prompting.json new file mode 100644 index 0000000..0c86a0f --- /dev/null +++ b/configs/sst2/16-42-sst2-conti-demon-prompting.json @@ -0,0 +1,26 @@ +{ + "train_batch_size": 8, + "eval_batch_size": 8, + "task_name": "sst-2", + "random_seed": 13, + "k_shot": 16, + "dataset_dir": "./rsc/eng_few_shot/k-shot/SST-2/16-42", + "learning_rate": 3e-05, + "max_grad_norm": 1.0, + "lm_model": "roberta-large", + "manual_template": "It was .", + "max_seq_length": 128, + "gradient_accumulation_steps": 1, + "num_train_epochs": 50, + "log_step_count_steps": 20, + "lambda_rate": 0.3, + "eval_every_step": null, + "demonstration_tokens": [ + ["well worth the effort", "a total waste of my time", "a real treat to watch", "so sad", "a cultural revolution"], + ["an instant hit", "a gift", "entertaining on an inferior level", "an unforgettable experience", "a thriller with an edge"] + ], + "soft_token_length": 10, + "is_cuda": true, + "warmup_proportion": 0.6, + "tensorboard_output_dir": "./logs/sst2/16-42-soft-10-k5" +} \ No newline at end of file diff --git a/configs/sst2/16-87-sst2-conti-demon-prompting.json b/configs/sst2/16-87-sst2-conti-demon-prompting.json new file mode 100644 index 0000000..76a32b9 --- /dev/null +++ b/configs/sst2/16-87-sst2-conti-demon-prompting.json @@ -0,0 +1,26 @@ +{ + "train_batch_size": 8, + "eval_batch_size": 8, + "task_name": "sst-2", + "random_seed": 13, + "k_shot": 16, + "dataset_dir": "./rsc/eng_few_shot/k-shot/SST-2/16-87", + "learning_rate": 3e-05, + "max_grad_norm": 1.0, + "lm_model": "roberta-large", + "manual_template": "It was .", + "max_seq_length": 128, + "gradient_accumulation_steps": 1, + "num_train_epochs": 50, + "log_step_count_steps": 20, + "lambda_rate": 0.3, + "eval_every_step": null, + "demonstration_tokens": [ + ["well worth the effort", "a total waste of my time", "a real treat to watch", "so sad", "a cultural revolution"], + ["an instant hit", "a gift", "entertaining on an inferior level", "an unforgettable experience", "a thriller with an edge"] + ], + "soft_token_length": 10, + "is_cuda": true, + "warmup_proportion": 0.6, + "tensorboard_output_dir": "./logs/sst2/16-87-soft-10-k5-lambda0.3" +} \ No newline at end of file