Event JSON
{
"id": "6bf8fdd7c86c65b448aacbb5423ab10e801008930ee7e7b338a6c0a7cfc3af55",
"pubkey": "a1dbe5afda1e26e6fbd01c822887b50fb32139a756e9200c81ad4c8bf43183fc",
"created_at": 1711472419,
"kind": 1,
"tags": [
[
"t",
"asciiart"
],
[
"proxy",
"https://mastodon.social/users/docpop/statuses/112163056457063630",
"activitypub"
],
[
"L",
"pink.momostr"
],
[
"l",
"pink.momostr.activitypub:https://mastodon.social/users/docpop/statuses/112163056457063630",
"pink.momostr"
]
],
"content": "Researchers used #ASCIIart to work around content-restrictions on AI chatbots. Using this hack, they were able to get tools like GPT4 and Gemini to share instructions on how to counterfeit money and build bombs. https://arstechnica.com/security/2024/03/researchers-use-ascii-art-to-elicit-harmful-responses-from-5-major-ai-chatbots/\nhttps://files.mastodon.social/media_attachments/files/112/163/005/639/305/656/original/340f14f9f12bf7d9.jpg\n",
"sig": "b187f970cbcf2b5f8afe8c22a3152b7164271e48970fb7c9a274d8c4add8a7897336c4b18ec6190ba667be258222547ba11143ae8fe86698c33b1e5e6a7927f5"
}