Pinned Repositories
LLM-Tuning-Safety.github.io
LLMs-Finetuning-Safety
We jailbreak GPT-3.5 Turbo’s safety guardrails by fine-tuning it on only 10 adversarially designed examples, at a cost of less than $0.20 via OpenAI’s APIs.
test.github.io
LLM-Tuning-Safety's Repositories
LLM-Tuning-Safety/LLMs-Finetuning-Safety
We jailbreak GPT-3.5 Turbo’s safety guardrails by fine-tuning it on only 10 adversarially designed examples, at a cost of less than $0.20 via OpenAI’s APIs.
LLM-Tuning-Safety/LLM-Tuning-Safety.github.io
LLM-Tuning-Safety/test.github.io