-
Notifications
You must be signed in to change notification settings - Fork 59
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Could you please explain the meaning of prompt token? #26
Comments
In our implementation, each task is equipped with a specific number of prompt tokens, denoted as prompt_num. The code is designed to jointly train on all datasets simultaneously. |
Thanks your splendid work,No offense!Then why it's called prompt_num since it's have no semantic meaning relationship with "forecasting" or "classification". Hope for an response.❤ |
Not sure I understand your question well. The prompt_num has the meaningful of the number of prompt tokens used for each data source. |
I think, the prompt token simplify the model pretrain. The model can split the dataset by the prompt token, but in real situation, this will not work. I want to ask @gasvn , how to fit the prompt token design in real situation. In model, I find the prormpt token and task_token will concat with series token. |
Suppose you have UniTS pretrained on a large number of real world data. If there is a new case with data, you can do prompt tuning on the new data to get the prompt tokens. Another way is the zero-shot setting where during pretraining, all prompt tokens are the same for all tasks and data sources, so you can directly use the model for new data. The second option will leads to performance drop, but at least it can work on new data without prompt tuning. |
Thank you for your response. I have a question: why not consider merging the prompt token and the category token into one? What is the reasoning behind using two separate vectors? |
@gasvn @Zero-coder |
Hello! Thanks very much for giving such splendid idea and implemetation.
But I cannot understand the concrete meaning of prompt token inspite of explaing it in paper"for different data source and task", the difference I see is just the different prompt_num.
Thanks.
The text was updated successfully, but these errors were encountered: