Alex Gantman on Nostr: The fundamental conclusion emerging from AI security research seems to be that the ...
The fundamental conclusion emerging from AI security research seems to be that the system's output should be considered as sensitive as the most sensitive data in the training set (everything leaks) and as untrustworthy as the least trustworthy data in the training set (everything contaminates). This has echoes of the Bell-LaPadula and Biba models. Of course, these models failed because even if they were theoretically sound, they were unworkable in practice. The workaround was to insert humans in the loop to make authorization decisions. But the whole point of AI is to take the human out of the loop. Interesting dilemma.
Published at
2024-03-22 17:16:44Event JSON
{
"id": "30011cdcdb18d51880673927f047ff2bf3f087fc4e3a7a1a16af91e8475a85e7",
"pubkey": "48208ce6ed49a196d88d90a06e5fce17907d057e4d8966b0fb62a34b8f033298",
"created_at": 1711127804,
"kind": 1,
"tags": [
[
"proxy",
"https://ioc.exchange/users/againsthimself/statuses/112140471778437282",
"activitypub"
]
],
"content": "The fundamental conclusion emerging from AI security research seems to be that the system's output should be considered as sensitive as the most sensitive data in the training set (everything leaks) and as untrustworthy as the least trustworthy data in the training set (everything contaminates). This has echoes of the Bell-LaPadula and Biba models. Of course, these models failed because even if they were theoretically sound, they were unworkable in practice. The workaround was to insert humans in the loop to make authorization decisions. But the whole point of AI is to take the human out of the loop. Interesting dilemma.",
"sig": "fcbae2785fd7379d6ee07ebb96f3b1737a06b4ecfe346643bdb2190682314e51f3d50c7317ea89aa561eb7d0d29e303e731ee50f0759ae31abe439d64cc9ea14"
}