Event JSON
{
"id": "586abfcb67deb4e4db8fde9ae1b3fddc6cabaea89ef9fccbb26d9cf138ef9e43",
"pubkey": "a37afb730401ef68fa7ed24bb96be33f72147008c4a55ae12a62de81b1b98841",
"created_at": 1719817603,
"kind": 1,
"tags": [
[
"proxy",
"https://thepit.social/@peter/112709966430328632",
"web"
],
[
"proxy",
"https://thepit.social/users/peter/statuses/112709966430328632",
"activitypub"
],
[
"L",
"pink.momostr"
],
[
"l",
"pink.momostr.activitypub:https://thepit.social/users/peter/statuses/112709966430328632",
"pink.momostr"
],
[
"expiration",
"1722412510"
]
],
"content": "this is a good thread. again, the thing to remember about LLMs is they don't \"know\" anything. they are not going out and retrieving information. they are generating plausible-sounding language. and they are very good at that! so good that they've convinced a bunch of humans that they can think! even though clearly, they cannot. https://old.reddit.com/r/explainlikeimfive/comments/1dsdd3o/eli5_why_cant_llms_like_chatgpt_calculate_a/",
"sig": "e1f2e91cfb65dda030c298010b0b2db81222fc4ef6c7794c7cfb27ae71eb7362cd56013854c5914dac7df980146eba6347c20cdfaac905b284d2bb98aae2bd61"
}