Datalab’s Post

Launch Week — Day 2: We made Chandra faster (Again) 🚀 Today we’re introducing Chandra Small, our new latency-optimized OCR model available now via the Datalab API. Chandra Small is 2–3x faster than the standard Chandra model with minimal performance degradation. We trained it with quantization-aware training (QAT), making it quantization-friendly and enabling even lower latency in production. A few highlights: ⭐ 2–3x faster inference ⭐ ~30% latency reduction from reduced token usage ⭐ 2–4 pages/sec on an H100 ⭐Maintains strong performance on benchmarks like olmOCR You can try Chandra Small today by using Fast mode in the API. Stay tuned for tomorrow's launch and in the meantime, check out the links below!

  • No alternative text description for this image
  • No alternative text description for this image

To view or add a comment, sign in

Explore content categories