Event JSON
{
"id": "65fd8aa948c7cc5770d166d55737c2ce11c79a4f830d66544df6c1a52afffa3e",
"pubkey": "8b0be93ed69c30e9a68159fd384fd8308ce4bbf16c39e840e0803dcb6c08720e",
"created_at": 1725721500,
"kind": 1,
"tags": [
[
"p",
"a1451f62ed88aa362fbbd84dbf8980745a10b34117e380c157c387b6edf93c0a",
"wss://relay.mostr.pub"
],
[
"p",
"fb7d6f616fc903fac631580a9eb2ab26623d611e1a78c5cc2086a05d85bb960e",
"wss://relay.mostr.pub"
],
[
"e",
"a2f4af4c152a5e116e8852a302801e705b79f2b857dfc5c53f7fbfe78bc91b96",
"wss://relay.mostr.pub",
"reply"
],
[
"proxy",
"https://fedi.simonwillison.net/users/simon/statuses/113096884238279405",
"activitypub"
]
],
"content": "nostr:npub159z37chd3z4rvtammpxmlzvqw3dppv6pzl3cps2hcwrmdm0e8s9qqdddac oh fantastic! I’ve played with https://github.com/mlc-ai/web-llm but I didn’t know about the llama.cpp port, that’s awesome",
"sig": "ac94b5102f3875848a3f37468240274a286d021aa5086fe10bdd9b4bf42a8dced2be3c296f7a2953306032c8928d9d92066d4e3fda49b0317e6edf54cba085d7"
}