shawn on Nostr: It’s a long shot, but has anyone successfully compiled ollama, llama.cpp, or ...
It’s a long shot, but has anyone successfully compiled ollama, llama.cpp, or similar using AMD’s rocm library with a Ryzen 760M iGPU? I’m stumped. #asknostr
Published at
2025-01-30 15:37:02Event JSON
{
"id": "56d1d5bd9f076b0c2dc70f245a30e60031ddce7b1816dc5498bad681cf541473",
"pubkey": "c7eda660a6bc8270530e82b4a7712acdea2e31dc0a56f8dc955ac009efd97c86",
"created_at": 1738251422,
"kind": 1,
"tags": [
[
"t",
"asknostr"
]
],
"content": "It’s a long shot, but has anyone successfully compiled ollama, llama.cpp, or similar using AMD’s rocm library with a Ryzen 760M iGPU? I’m stumped. #asknostr",
"sig": "ca96ebd2ba452357757eac83ba04d1059b6318f4debe3add60e20f904828c734df61c3657a3113cd9af3e7cfc92e63eb99d3ade7985c9e41c63d77be4a0b5b22"
}