Event JSON
{
"id": "a42b099220bbf9b798fd8c8c17de4e0e2014e6e6b44fab6ff100eab4f5b1ab95",
"pubkey": "a4dd031459b989e5f3896a6407160b014517568eb96a9be0e9e6ca5594855c45",
"created_at": 1725723295,
"kind": 1,
"tags": [
[
"p",
"dee34601686a23b7558564186e77a135e1d08eac09a54270b59650631e9df9b3",
"wss://relay.mostr.pub"
],
[
"p",
"358b313c445eccc8485ec27702158b45fa6b768ba7a1bcb99c829ebdc979f27f",
"wss://relay.mostr.pub"
],
[
"e",
"e271ecfc82cd1ded2a0113b4fd359583559ba1dd30852b6014d25d0434a13897",
"wss://relay.mostr.pub",
"reply"
],
[
"proxy",
"https://thepit.social/users/Steve/statuses/113097001901038340",
"activitypub"
]
],
"content": "nostr:npub1mm35vqtgdg3mw4v9vsvxuaapxhsapr4vpxj5yu94jegxx85alxesmwft6j ultimately there's probably a case here where using LLMs that are actually helpful require tasks specific LLMs trained on known material... I wonder if that's where this all shakes out.",
"sig": "716125341df16f71edeb50c7396a5b1d6e07e7da45e6b8651cc1deb30cc9873a13897b8a0f492714855da161cca16c48a6ece951b9b5dc9b79b9a80211b2af90"
}