Launch Week — Day 2: We made Chandra faster (Again) 🚀 Today we’re introducing Chandra Small, our new latency-optimized OCR model available now via the Datalab API. Chandra Small is 2–3x faster than the standard Chandra model with minimal performance degradation. We trained it with quantization-aware training (QAT), making it quantization-friendly and enabling even lower latency in production. A few highlights: ⭐ 2–3x faster inference ⭐ ~30% latency reduction from reduced token usage ⭐ 2–4 pages/sec on an H100 ⭐Maintains strong performance on benchmarks like olmOCR You can try Chandra Small today by using Fast mode in the API. Stay tuned for tomorrow's launch and in the meantime, check out the links below!
Datalab’s Post
Explore content categories
- Career
- Productivity
- Finance
- Soft Skills & Emotional Intelligence
- Project Management
- Education
- Technology
- Leadership
- Ecommerce
- User Experience
- Recruitment & HR
- Customer Experience
- Real Estate
- Marketing
- Sales
- Retail & Merchandising
- Science
- Supply Chain Management
- Future Of Work
- Consulting
- Writing
- Economics
- Artificial Intelligence
- Employee Experience
- Workplace Trends
- Fundraising
- Networking
- Corporate Social Responsibility
- Negotiation
- Communication
- Engineering
- Hospitality & Tourism
- Business Strategy
- Change Management
- Organizational Culture
- Design
- Innovation
- Event Planning
- Training & Development