Event JSON
{
"id": "595fdb8c635e5cfdf8a643b67cdb102955bec30a0308bc31354c82eff93f4e92",
"pubkey": "c164d8acb8a3c19954787a38d9b17d38213f5d7b8851473907d2fd3ad2d9850b",
"created_at": 1710976760,
"kind": 1,
"tags": [
[
"p",
"872ab8e157527ea743574e273a0fd26ccfe032ecef539a7191be1c4418511ca1",
"wss://relay.mostr.pub"
],
[
"p",
"509c345f0518ca1421ba9f26a98ac6b0b6f884b1874ed9c20320fd41158313e4",
"wss://relay.mostr.pub"
],
[
"e",
"31c793d747b3e8629dcd3ccbf32adbea84f4ac9c95bd3cfc8bb5140b23c5c9dc",
"wss://relay.mostr.pub",
"reply"
],
[
"proxy",
"https://mastodon.social/users/nazokiyoubinbou/statuses/112130572950022562",
"activitypub"
]
],
"content": "nostr:npub1su4t3c2h2fl2ws6hfcnn5r7jdn87qvhvaafe5uv3hcwygxz3rjssf6a0dw LLMs could have been amazing if A: they weren't treated like the AI that they are not actually and B: the companies doing them didn't for some reason get this idea that laws, morals, and etc suddenly mean nothing in regards to them and they can just do whatever they want without even a single care.",
"sig": "5432f7527a7227a66c7595332cb7b15061d53237c20fbe0d38c5b95ab8a876e0f58ba7163085794e1417ec48aa9ae4794f4c02736af4b46ea3461e5160de741f"
}