I did some research on SBCs for running LLMs. My target is to run LLM (with tool calling support) and make a Voice AI.
At first, searching SBCs, I came up with Raspberry Pi 5. But I watched my youtube reviews and came to know that, RPi 5 doesn’t have any GPU for LLMs to run. The 3b models lags terrible in it. As I have seen in some videos, LLAMA 3.2 3b only generates 1.78 tokens per second. That is not expected.
Then I came up with the name Radxa. Radxa SBCs doesn’t have much videos on youtube. But, I watched a few and saw that Radxa X4 has 4GB GPU according to https://youtu.be/F2atAHDOaIA video at 7:45
That would be enough for running a 3B model, wouldn’t it?
Again, according to the docs of Radxa, Rk3588 Chip SBC has another advantage. The Radxa Rock 5B plus can generate 6 tokens per second on Phi3 3.8B model. (Docs: https://docs.radxa.com/en/rock5/rock5c/app-development/rkllm_usage)
Main Question
To sum up, Raspberry Pi 5 is terrible for running 3B LLMs, Radxa X4 has 4 GB GPU which should be enough for running 3B Models, and Radxa Rock 5B plus has The RK3588 chip advantage. My question is, If I run Qwen 3B Model, which one should I buy? Radxa X4? or Rock 5B plus?
N.B: I don’t have any CPU/GPU knowledge, so I don’t know which is better. A community support would be great🙂