Mark Pesce on Nostr: Now it can be told: While doing some AI engineering work for a client, I developed a ...
Now it can be told:
While doing some AI engineering work for a client, I developed a prompt - completely inadvertently - that reduced every AI chatbot to gibberish (except Anthropic's Claude 3). I then spent a week trying to alert the LLM vendors to this issue - and largely failed. There is _no_ mechanism to report flaws in these models that are already deployed to billions of users. Read the whole story in The Register (npub10cz…4lpy)
https://www.theregister.com/2024/05/23/ai_untested_unstable/
While doing some AI engineering work for a client, I developed a prompt - completely inadvertently - that reduced every AI chatbot to gibberish (except Anthropic's Claude 3). I then spent a week trying to alert the LLM vendors to this issue - and largely failed. There is _no_ mechanism to report flaws in these models that are already deployed to billions of users. Read the whole story in The Register (npub10cz…4lpy)
https://www.theregister.com/2024/05/23/ai_untested_unstable/