ollama-multirun: this_is_the_ollama_bash_lib_bash_script_a: 20250725-211539

models: deepseek-r1:14b deepseek-r1:8b dolphin3:8b gemma3n:e4b mistral:7b qwen2.5vl:7b qwen3:8b

Prompt: (raw) (yaml) words:845 bytes:5895

Model Response
words
Response
bytes
Total
duration
Load
duration
Prompt eval
count
Prompt eval
duration
Prompt eval
rate
Eval
count
Eval
duration
Eval
rate
Model
params
Model
size
Model
context
Ollama
context
Ollama
proc
deepseek-r1:14b 597 4243 6m15.6504715s 2.7879425s 1783 token(s) 4.0784418s 437.18 tokens/s 2349 token(s) 6m8.7834033s 6.37 tokens/s 14.8B 14 GB 131072 16384 48%/52% CPU/GPU
deepseek-r1:8b 647 4699 14m21.2021266s 3.2627999s 1782 token(s) 1.7477969s 1019.57 tokens/s 9534 token(s) 14m16.1910182s 11.14 tokens/s 8.2B 9.6 GB 131072 16384 23%/77% CPU/GPU
dolphin3:8b 432 2887 30.6933286s 3.0549306s 1797 token(s) 1.4746932s 1218.56 tokens/s 549 token(s) 26.1626261s 20.98 tokens/s 8.0B 8.7 GB 131072 16384 15%/85% CPU/GPU
gemma3n:e4b 981 7030 1m51.3876623s 4.2704795s 2235 token(s) 8.1724663s 273.48 tokens/s 1870 token(s) 1m38.9431165s 18.90 tokens/s 6.9B 8.5 GB 32768 16384 12%/88% CPU/GPU
mistral:7b 486 3153 33.0599898s 2.6213209s 2065 token(s) 1.8082413s 1141.99 tokens/s 703 token(s) 28.629913s 24.55 tokens/s 7.2B 8.3 GB 32768 16384 11%/89% CPU/GPU
qwen2.5vl:7b 1534 10222 1m50.4283671s 2.7010685s 1800 token(s) 1.1966003s 1504.26 tokens/s 2780 token(s) 1m46.5294661s 26.10 tokens/s 8.3B 10 GB 128000 16384 34%/66% CPU/GPU
qwen3:8b 841 6411 4m21.8895935s 3.2772767s 1790 token(s) 1.7939954s 997.77 tokens/s 3154 token(s) 4m16.8178007s 12.28 tokens/s 8.2B 9.6 GB 40960 16384 23%/77% CPU/GPU


System
Ollama proc23%/77% CPU/GPU
Ollama context16384
Ollama version0.10.0-rc0
Multirun timeout99999 seconds
Sys archx86_64
Sys processorunknown
sys memory14G + 16G
Sys OSCYGWIN_NT-10.0-22631 3.6.4-1.x86_64