
cstanhope@social.coop ("Your friendly 'net denizen") wrote:
I love this type of LLM "instruction" that you find when prompts get leaked: "Do not hallucinate". Like, *that's* been the problem. We just keep forgetting to tell them not to hallucinate. So silly!