Skip to content

Instantly share code, notes, and snippets.

@monday8am
Last active December 9, 2025 12:07
Show Gist options
  • Select an option

  • Save monday8am/adfb33e3dec374d1d6f55d4550f1baec to your computer and use it in GitHub Desktop.

Select an option

Save monday8am/adfb33e3dec374d1d6f55d4550f1baec to your computer and use it in GitHub Desktop.
Model Inference Library Backend Tool Calling Support Key Issues
Qwen3-0.6B LiteRT-LM ❌ CPU (slow) ❌ Not supported Multiturn conversation broken due to crashes with thinking models; unable to deactivate thinking mode; google-ai-edge/LiteRT-LM#1027
Hammer2.1-0.5B LiteRT-LM ❌ CPU (slow) ❌ Not supported Multiturn conversation broken due internal API fail
Hammer2.1-0.5B MediaPipe ❌ CPU (slow) ❌ Not supported Internal failure when loading model; solution aparently abandoned by Google
Gemma3-1B LiteRT-LM ✅ GPU (fast) ❌ Not supported No tool calling training; unable to identify tools or maintain input/output schema
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment