Event JSON
{
"id": "d327dbf6e1eac3c35bacf8a88cf0a5659183bd0dc4bab3b2eaedfbd76d3b2385",
"pubkey": "6fe41d47b96c3a39ab992b4fb91dffcaeea328ba3b32e7487fc878e6ad728f66",
"created_at": 1700464452,
"kind": 1,
"tags": [
[
"p",
"ad951bfdf8f75f9c42aa74da8322bff38accbe325b7e8a9eec916226ac0ac24a",
"wss://relay.mostr.pub"
],
[
"p",
"dfd7b128bbb1222873eca016c1411fa7037a2d3b71dc0bf736480db54d3a61c3",
"wss://relay.mostr.pub"
],
[
"e",
"f261138d6dfb954be2edbebbdac98241b77b8be496803dafc2d8282bd5b36f50",
"wss://relay.mostr.pub",
"reply"
],
[
"proxy",
"https://infosec.exchange/users/luigirenna/statuses/111441638339609218",
"activitypub"
]
],
"content": "nostr:npub14k23hl0c7a0ecs42wndgxg4l7w9ve03jtdlg48hvj93zdtq2cf9q3ted68 LLaMA2 is quite interesting. The 7B model can be run locally (slowly), and it's well suited for experiments and tinkering. Right now I am writing my thesis on adversarial attacks against LLM, and after an initial attempt against GPT, LLaMA2 has become my go-to test model",
"sig": "b91fd4ccd4af5bf1efdbda10935233a5a2e1cb792728dfe416ace5d9cee4334e7f191ed87fe512535bf51f8d46b908411e2776b7acf7f5ddb19f822ab105e46a"
}