Event JSON
{
"id": "af6a8d745176191b7a4f313ed5aed386ba426c59a4acbdf718cfa19b271a9b9c",
"pubkey": "90108340405aa825f165c1e9224cc0b8e791202c3295642d6995683bc5a8e060",
"created_at": 1729312125,
"kind": 1,
"tags": [
[
"p",
"6af7103d13aa5ac90b78b7b1b3a6df0473603ae04fa731e88e6c52583371744c",
"wss://relay.mostr.pub"
],
[
"p",
"edb81957549b17e4e65f8fee3e0d5ac96b8b97d4a52d58d2b6a132ae5d64dd08",
"wss://relay.mostr.pub"
],
[
"t",
"meenakim"
],
[
"t",
"llm"
],
[
"t",
"chatgpt"
],
[
"proxy",
"https://astrodon.social/users/SkipHuffman/statuses/113332199465798534",
"activitypub"
]
],
"content": "Listening to #MeenaKim on nostr:npub1dtm3q0gn4fdvjzmck7cm8fklq3ekqwhqf7nnr6ywd3f9svm3w3xqfcple5 just now discussing #LLM and #chatGPT. The guest commented that the systems \"sometimes\" hallucinate. That isn't exactly correct. These systems ALWAYS hallucinate because they have no theory of mind. Their hallucinations are often congruent with objective reality, but that is just happenstance. To the systems all utterances are equivalent.",
"sig": "34605e621113744539cbd10571ea851ed74064681ffce7712a49d36275582a7946de8fc39898361d55e163a57f1ca3e8f6db1bbff84c6a6e6a29e5a640985ebb"
}