Post
38
Qwen3.5 on-device benchmarks on the Nvidia Jetson lineup are now live 🚀
We've added the latest Qwen3.5 models (0
8B - 9B) to our on-device inference benchmarks (Nvidia Jetson Orin Nano Super, AGX Orin, AGX Thor).
👉 Explore TPS, TTFT, E2E latency, and TPOT. Measured on real hardware: embedl/Edge-Inference-Benchmarks
🌟 Stay tuned for additional benchmarks and Embedl-optimized models: Enabling models run faster and on less expensive hardware.
If you're working on edge LLM deployment, we'd love to discuss your use case.
We've added the latest Qwen3.5 models (0
8B - 9B) to our on-device inference benchmarks (Nvidia Jetson Orin Nano Super, AGX Orin, AGX Thor).
👉 Explore TPS, TTFT, E2E latency, and TPOT. Measured on real hardware: embedl/Edge-Inference-Benchmarks
🌟 Stay tuned for additional benchmarks and Embedl-optimized models: Enabling models run faster and on less expensive hardware.
If you're working on edge LLM deployment, we'd love to discuss your use case.