on-device-llm-bench

Three backends, same prompts, same harness: gemma-tjs (Gemma 4, Transformers.js + WebGPU) · phi4-edge (Phi-4-mini, Edge Prompt API, native) · phi4-tjs (Phi-4-mini, Transformers.js + WebGPU). phi4-edge vs phi4-tjs isolates the runtime; gemma-tjs vs phi4-tjs isolates the model. Source on GitHub.

Results

Loading…

How to read this
Contribute a run from your device

Run all three backends per the README, then add a folder under results/<YYYYMMDD>-<os-arch>/ with the three JSONs and a report.md. Append an entry to docs/manifest.json and open a PR.