| Model | Inference Library | Backend | Tool Calling Support | Key Issues |
|---|---|---|---|---|
| Qwen3-0.6B | LiteRT-LM | ❌ CPU (slow) | ❌ Not supported | Multiturn conversation broken due to crashes with thinking models; unable to deactivate thinking mode; google-ai-edge/LiteRT-LM#1027 |
| Hammer2.1-0.5B | LiteRT-LM | ❌ CPU (slow) | ❌ Not supported | Multiturn conversation broken due internal API fail |
| Hammer2.1-0.5B | MediaPipe | ❌ CPU (slow) | ❌ Not supported | Internal failure when loading model; solution aparently abandoned by Google |
| Gemma3-1B | LiteRT-LM | ✅ GPU (fast) | ❌ Not supported | No tool calling training; unable to identify tools or maintain input/output schema |
Last active
December 9, 2025 12:07
-
-
Save monday8am/adfb33e3dec374d1d6f55d4550f1baec to your computer and use it in GitHub Desktop.
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment