Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Clarification on the Use of reduce_sim in Loss Function and Key Updates #14

Open
JasonStraka opened this issue Dec 21, 2024 · 1 comment

Comments

@JasonStraka
Copy link

JasonStraka commented Dec 21, 2024

Dear Author,

Thank you for sharing the code implementation of your work. I have been carefully studying your paper and codebase, but I have come across a discrepancy that I would like to clarify.

In your paper, the reduce_sim term is mentioned as a part of the loss function, which plays a role in optimizing the prompt selection process. However, in the provided code, it appears that reduce_sim is calculated but not included in the loss function (it is commented out). As a result, the prompt_key_dict, which is randomly initialized, does not receive any gradient updates during training.

Here are my specific questions:

Was the exclusion of reduce_sim from the loss function intentional in the provided code? If so, could you elaborate on the reasoning behind this decision?
How does the omission of reduce_sim affect the effectiveness of the learned prompts, especially since the keys (prompt_key_dict) remain randomly initialized without updates?
If this was an oversight, could you provide guidance on how to properly incorporate reduce_sim into the loss function and ensure that prompt_key_dict is updated?
Did the implementation consider the diversity of keys during training, such as tracking how frequently each key was selected? For example, is there a mechanism to ensure that less frequently used keys are adapted or penalized to promote diversity?

I appreciate your time and effort in addressing these queries, as they are critical to understanding and reproducing the results presented in your paper.

Thank you!

Best regards,
Jason

@idevede
Copy link
Contributor

idevede commented Dec 23, 2024

Hi Jason,

Thank you for your interest in our TEMPO work! This code corresponds to our ICLR camera-ready version, where the prompt is designed in a semi-soft manner. We have left the implementation of the prompt pool in this repository solely for future research discussions and do not currently involve it in the training or inference process.

Best,

Authors

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants