<<< Back to Index


LM Studio Mini Head-to-Head (No-Thinking Grading) — 2026-02-22


Run dir: /home/slime/.openclaw/workspace-base/.run/lmstudio-mini-benchmark/20260222-151912-no-thinking


Scoring uses normalized final answer (thinking text stripped).


RankModelScoreAvg sec/prompt
1qwen/qwen3-vl-8b5/51.23
2openai/gpt-oss-20b5/52.51
3qwen3-coder-next5/59.78
4openai/gpt-oss-120b5/522.61
5google/gemma-3-4b4/50.9
6google/gemma-3-12b3/51.58
7zai-org/glm-4.6v-flash1/52.52
8deepseek/deepseek-r1-0528-qwen3-8b0/52.33
9mistralai/ministral-3-14b-reasoning0/53.14
10zai-org/glm-4.7-flash0/55.27

Artifacts


<<< Back to Index