Running same prompt with different hardware(CPU & Customer hardware) lead to different prompt response even with temp:0.0 value #12167
Closed
akapoor3518
started this conversation in
General
Replies: 1 comment 8 replies
-
@akapoor3518 Does it happen when you set the same integer random seed value? (-s or --seed parameter). |
Beta Was this translation helpful? Give feedback.
8 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
I am running below prompt (with --temp 0.0) on CPU and my customer hardware and i expected same response. But i am getting different response. I will looking through Llama.cpp code and see why this is difference. It will help if you guys can also give your suggestion on this.
Below what i had run:
./build/bin/llama-cli -p "my cat name" -m ./models/tinyllama-vo-5m-para.gguf --device none -c 12288 --temp 0.0 --n-predict 4 --repeat-penalty 1.5 -b 1024 --top-k 50 --top-p 0.9 --repeat-last-n 5 --no-warmup
Prompt response
###########
my cat name was a little girl
With Customer Hardware only following operation: GGML_OP_NONE, GGML_OP_ADD, GGML_OP_SUB, GGML_OP_DIV, GGML_OP_MUL were offloaded to customer hardware, rest going to CPU
./build/bin/llama-cli -p "my cat name" -m ./models/tinyllama-vo-5m-para.gguf --device customer-hardware -c 12288 --temp 0.0 --n-predict 4 --repeat-penalty 1.5 -b 1024 --top-k 50 --top-p 0.9 --repeat-last-n 5 --no-warmup
Prompt response
###########
my cat name the sun was shining
Beta Was this translation helpful? Give feedback.
All reactions