| DeepSeek-R1-Distill-Qwen-1.5B | 128 | W8A16 | 1075.04 | 3.57 | Module LLM Kit / LLM630 Compute Kit |
| DeepSeek-R1-Distill-Qwen-1.5B | 256 | W8A16 | 3056.86 | 3.57 | Module LLM Kit / LLM630 Compute Kit |
| DeepSeek-R1-Distill-Qwen-1.5B | 256 | W4A16 | - | 13.29 | LLM8850 |
| Llama-3.2-1B-Instruct | 128 | W8A16 | 891 | 4.48 | Module LLM Kit / LLM630 Compute Kit |
| Llama-3.2-1B-Instruct | 256 | W8A16 | 2601.11 | 4.49 | Module LLM Kit / LLM630 Compute Kit |
| MiniCPM4-0.5B | 512 | W8A16 | 212.91 | 21.05 | LLM8850 |
| openbuddy-llama3.2-1b-v23.1-131k | 128 | W8A16 | 891.02 | 4.52 | Module LLM Kit / LLM630 Compute Kit |
| Qwen2.5-0.5B-Instruct | 128 | W8A16 | 359.8 | 10.32 | Module LLM Kit / LLM630 Compute Kit |
| Qwen2.5-0.5B-Instruct | 256 | W8A16 | 1126.19 | 10.3 | Module LLM Kit / LLM630 Compute Kit |
| Qwen2.5-0.5B-Instruct | 128 | W4A16 | 442.95 | 12.52 | Module LLM Kit / LLM630 Compute Kit |
| Qwen2.5-0.5B-Instruct | 128 | W4A16 | 140.17 | 37.11 | AI Pyramid |
| Qwen2.5-0.5B-Instruct | 128 | W4A16 | - | 27.05 | LLM8850 |
| Qwen2.5-1.5B-Instruct | 128 | W8A16 | 3056.54 | 3.57 | Module LLM Kit / LLM630 Compute Kit |
| Qwen2.5-1.5B-Instruct | 128 | W4A16 | 1219.54 | 4.63 | Module LLM Kit / LLM630 Compute Kit |
| Qwen2.5-1.5B-Instruct | 128 | W4A16 | 289.06 | 16.77 | AI Pyramid |
| Qwen2.5-1.5B-Instruct | 128 | W4A16 | - | 15.06 | LLM8850 |
| Qwen2.5-3B-Instruct | 128 | W4A16 | 550.3 | 9.46 | AI Pyramid |
| Qwen2.5-0.5B-Instruct | 1024 | W8A16 | 533.19 | 9.76 | Module LLM Kit / LLM630 Compute Kit |
| Qwen2.5-0.5B-Instruct | 1024 | W8A16 | 143.02 | 25.5 | AI Pyramid |
| Qwen2.5-0.5B-Instruct | - | - | 8210 | 1.54 | RaspberryPi5 CPU (ollama) |
| Qwen3-0.6B | 128 | W8A16 | 361.81 | 10.28 | Module LLM Kit / LLM630 Compute Kit |
| Qwen3-0.6B | 2048 | W8A16 | 670.51 | 12.88 | LLM8850 |
| Qwen3-1.7B | 2048 | W8A16 | 796.38 | 7.38 | LLM8850 |