Event JSON
{
"id": "087a292b5abb0fb1601e19764b7988fd2a29b9e046283c2dfec9c0ce9aa59794",
"pubkey": "6e5bb3f4df765d024066fb8b1b0ab196b68726b8092782a387c9698fd05d8c35",
"created_at": 1711206221,
"kind": 1,
"tags": [
[
"p",
"721c9237286f6863c8a390e5ef819fd0cf347850344b21ca1bb6f29ee1e69a2d",
"wss://relay.mostr.pub"
],
[
"p",
"97b7416af6549a9a115317ef79eb8be793e1f8a2a08d5dc00eedf490a4751b31",
"wss://relay.mostr.pub"
],
[
"e",
"31b11812c575bd1cb9f1b8989bda5f76f8861bef21c6499085cd8bea9c1a939e",
"wss://relay.mostr.pub",
"reply"
],
[
"proxy",
"https://tldr.nettime.org/users/tante/statuses/112145610963097728",
"activitypub"
]
],
"content": "nostr:npub1wgwfydegda5x8j9rjrj7lqvl6r8ng7zsx39jrjsmkmefac0xngkskdesx5 I mean it depends a bit on how you define \"bias\" but in general no. Unless you define a certain level of bias (or bias that corresponds with your belief) as \"bias free\" every dataset (and that's what an LLM is: a compressed dataset) has biases, some more obvious than others.",
"sig": "70d6925cb457a1b1bae57d53400b48cc35af89828c1f624c0eb60ea3b756da637448c56ec59604adb8f268e0aacbb94526c213b1eb3f6c7065e803a51c76b5bf"
}