Event JSON
{
"id": "8ab6061fd3c1a4b16c3e8acf03239d2e2390e34ae56e7011c0caf2353318db96",
"pubkey": "2936462bda8612e290f17231fddca9a658b472680cb661b537b1121d5b3d683b",
"created_at": 1740059787,
"kind": 1,
"tags": [
[
"e",
"9ca31705212eb454d564fc8b5e7c4b383cf61cd70beab5ae771c32a8c81c1a8f",
"",
"root"
],
[
"p",
"2936462bda8612e290f17231fddca9a658b472680cb661b537b1121d5b3d683b"
],
[
"r",
"wss://140.f7z.io/"
],
[
"r",
"wss://at.nostrworks.com/"
],
[
"r",
"wss://adre.su/"
],
[
"r",
"wss://bitcoiner.social/"
],
[
"r",
"wss://bitcoinmaximalists.online/"
],
[
"r",
"wss://bevo.nostr1.com/"
],
[
"r",
"wss://bostr.bitcointxoko.com/"
]
],
"content": "手写了一个Transformer架构计算注意力机制的简单例子,希望明天具备一定理科能力的文科生,能看懂。另外个别理科生可以更好的理解注意力机制。\n\n别看这个例子,不熟悉excel,搞了2h才弄好,太麻烦了。 \n\nhttps://postimg.cc/fkZHLM9q\n\nhttps://i.postimg.cc/ryR7fmFT/593af2a7gy1hyq4ecmw5oj21cj0qjwpv.jpg",
"sig": "7d197f05af3169b8fbc958c2b7106f77bdf3645cb189c7521bb5644d25b6f929a92de8d41d3947d63653dc9132bb9bb92a68172c18012e9f32db333b67870e33"
}