Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
JonnaMatΒ 
posted an update about 16 hours ago
Post
122
Qwen3.5 on-device benchmarks on the Nvidia Jetson lineup are now live πŸš€

We've added the latest Qwen3.5 models (0
8B - 9B) to our on-device inference benchmarks (Nvidia Jetson Orin Nano Super, AGX Orin, AGX Thor).

πŸ‘‰ Explore TPS, TTFT, E2E latency, and TPOT. Measured on real hardware: embedl/Edge-Inference-Benchmarks

🌟 Stay tuned for additional benchmarks and Embedl-optimized models: Enabling models run faster and on less expensive hardware.

If you're working on edge LLM deployment, we'd love to discuss your use case.

Additional Jetson Orin Nano Super benchmarks will be added continuously.

In this post