Replies: 1 comment 1 reply
-
If you get bored and need something to do, I would love to see a side by side comparison between the RPI5 and the RK1. It would be interesting to see how the RK3588 stacks up against the A76. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Moved from this issue. Author: @segabor
I promised to share results of Turing RK1 module. It arrived yesterday so I took the chance to run distributed llama on it.
Capability: 8 cores, 32 GB RAM. Storage: 1 TB NVMe SSD
OS: custom Ubuntu Server
Model: llama-2-7b
Command
sudo nice -n -20 ./main inference \ --model /mnt/bigdata/llama-2-7b/dllama_llama-2-7b_q40.bin \ --tokenizer ./tokenizer.bin \ --weights-float-type q40 \ --buffer-float-type q80 \ --prompt "Hello world" \ --steps 16 \ --nthreads 4
Result
Beta Was this translation helpful? Give feedback.
All reactions