Nemeski@lemm.ee to Technology@lemmy.worldEnglish · 2 years agoOpenAI’s latest model will block the ‘ignore all previous instructions’ loopholewww.theverge.comexternal-linkmessage-square100fedilinkarrow-up1448cross-posted to: technology@lemmit.online
arrow-up1448external-linkOpenAI’s latest model will block the ‘ignore all previous instructions’ loopholewww.theverge.comNemeski@lemm.ee to Technology@lemmy.worldEnglish · 2 years agomessage-square100fedilinkcross-posted to: technology@lemmit.online
minus-squarePasta Dental@sh.itjust.workslinkfedilinkEnglisharrow-up66·2 years agoIll believe it when I see it: an LLM is basically a random box, you can’t 100% patch it. Their only way for it to stop generating bomb recipes is to remove that data from the training
Ill believe it when I see it: an LLM is basically a random box, you can’t 100% patch it. Their only way for it to stop generating bomb recipes is to remove that data from the training