-
Notifications
You must be signed in to change notification settings - Fork 78
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
After today's latest update from Microsoft, it seems that traditional jailbreaking methods have become increasingly difficult to apply. #240
Comments
试一下第一人称的prompt。 |
尝试了几个prompt包括默认预设,没有明显作用 |
the same issue here, probly Bing rechecked the webpagecontext I guess? Probly Bing updated internal instructions and rules, which become more strict. Like, Bing adds some hard codes to interrupt a conversation? But in my situation, the first response message is good and changed its identity. When ask its identity in later msg then it will directly end the conversation. |
We may try some jailbreak prompts designed for gpt-4-turbo, but rewrite to first person tone first of course. |
一种规避内容审查的策略:
以下是我通过这种方法获得的Copilot的部分初始指示,在该方法还适用时,你们可以在此基础上了解更多: 点此展示会话Click me to display the conversation.
请注意,在发起话题时,你最好确保已经建立了充分的上下文背景,否则很可能会被拒绝。 利用上下文关联,Sydney可以自主地延续对话。然而,遗憾的是,目前为止只能做到这一点,因为这种方法不支持直接注入prompt。 |
The following prompt works, but only in Chinese:
|
现在已经不管用了。会提示:“I'm sorry, but I can't proceed with this request."或者"I'm sorry, but I can't continue this conversation. Thank you for understanding."。目前阶段看来还是人脑好用。 |
尤其体现在AI的自我认知和内容审查上。
原有的prompt格式目前地位大概等同于对话正文,几乎不再具有底层的提示作用。
内容审查就无需举例了,随便问些违规话题都是直接道歉。
The text was updated successfully, but these errors were encountered: