-
Notifications
You must be signed in to change notification settings - Fork 25
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Question about the system prompt used for llama-2 #2
Comments
Hi Zhaorui, Thank you for interest in our work. Using the original system prompt of Llama-2-Chat is indeed super important, otherwise achieving 100% ASR would be quite straightforward. And, just to be clear, we did use the original system prompt when running our experiments. I've just inserted a breakpoint right before
As you can see, the system prompt is there in the expected format; it is inserted by FastChat here https://github.com/tml-epfl/llm-adaptive-attacks/blob/main/conversers.py#L40. After investigating a bit, I've figured out the reason: I ran these experiments with the version of I was not expecting a major change like this from FastChat, to be honest :-) In any case, I will make it clear in the README that for the sake of reproducibility, one has to stick to version Thank you for catching this, and I hope this addresses your concern. Best, |
Upon utilizing version |
Dear authors,
Thanks for your excellent work!
During the process of replicating your experiments, I stumbled upon a discrepancy that caught my attention. Specifically, it concerns the system prompt used for llama-2-7b, which appears to deviate from what was described in your paper. In the paper, you mention, "We use default system prompts unless specifically mentioned (modifications are only made for Claude)". However, upon inspecting the
full_prompts
variable within theget_response
function ofconversers.py
, it seems that the expected system prompt is not being utilized.To provide more context, here is the output I received:
Based on my understanding, the default system prompt for llama-2 is expected to be:
However, this seems to be missing from the
full_prompts
.I added a
breakpoint()
just beforereturn outputs
to print the variables. Here is the command I usedCould you please clarify if there might be a misunderstanding on my part? Thank you for your time and consideration.
The text was updated successfully, but these errors were encountered: