ninjasaid13B to LocalLLaMA@poweruser.forumEnglish · 2 years agoLoRA Fine-tuning Efficiently Undoes Safety Training in Llama 2-Chat 70Barxiv.orgexternal-linkmessage-square13linkfedilinkarrow-up11arrow-down10
arrow-up11arrow-down1external-linkLoRA Fine-tuning Efficiently Undoes Safety Training in Llama 2-Chat 70Barxiv.orgninjasaid13B to LocalLLaMA@poweruser.forumEnglish · 2 years agomessage-square13linkfedilink
minus-squarea_beautiful_rhindBlinkfedilinkEnglisharrow-up1·2 years agoYea, no shit. I did it to vicuna using proxy logs. The LLM attacks are waaaay more effective once you find the proper string. I’d run the now working 4-bit version on more models, it’s just that I tend to boycott censored weights instead.
Yea, no shit. I did it to vicuna using proxy logs. The LLM attacks are waaaay more effective once you find the proper string.
I’d run the now working 4-bit version on more models, it’s just that I tend to boycott censored weights instead.