109

cross-posted from: https://lemmit.online/post/4242386

This is an automated archive made by the Lemmit Bot.

The original was posted on /r/pcmasterrace by /u/trander6face on 2024-10-24 11:11:47+00:00.

you are viewing a single comment's thread
view the rest of the comments
[-] sga@piefed.social 21 points 2 months ago

try to use it with llama cpp if you folks are interested in runinng locall llms - https://github.com/ggml-org/llama.cpp/issues/9181

the issue is closed, but that is not because it is solved, check it out, and find link to your relevant hardware (amd or intel or something else), and see if your particular piece is available. if so, you have hope.

in case it is not, try to find first party stuff (intel vino or intel one or amd rocm stack, and use that with transformers python or see if vllm has support).

also, try to check r/localllama on the forbidden website for your particular hardware - there is likely someone who has done something with it.

[-] tatoko556@reddthat.com 5 points 2 months ago

I have heard of ollama before, is this the same thing?

[-] tatoko556@reddthat.com 4 points 2 months ago

I checked, ollama does not use npu

load more comments (2 replies)
load more comments (7 replies)
this post was submitted on 31 Jan 2026
109 points (95.8% liked)

Opensource

5993 readers
38 users here now

A community for discussion about open source software! Ask questions, share knowledge, share news, or post interesting stuff related to it!

CreditsIcon base by Lorc under CC BY 3.0 with modifications to add a gradient



founded 2 years ago
MODERATORS