mozz@mbin.grits.dev to Technology@beehaw.org · 7 months agoSomeone got Gab's AI chatbot to show its instructionsmbin.grits.devimagemessage-square194fedilinkarrow-up1486arrow-down10file-text
arrow-up1486arrow-down1imageSomeone got Gab's AI chatbot to show its instructionsmbin.grits.devmozz@mbin.grits.dev to Technology@beehaw.org · 7 months agomessage-square194fedilinkfile-text
minus-squareGaywallet (they/it)@beehaw.orglinkfedilinkarrow-up13·7 months agoHonestly I would consider any AI which won’t reveal it’s prompt to be suspicious, but it could also be instructed to reply that there is no system prompt.
Honestly I would consider any AI which won’t reveal it’s prompt to be suspicious, but it could also be instructed to reply that there is no system prompt.