Batch-GPT Documentation
A jump-server to convert OpenAI chat completion API requests to batched requests
Overview
Batch-GPT enables seamless integration with OpenAI’s Batch API by acting as a drop-in replacement for standard OpenAI API endpoints. It intelligently collects and batches requests, providing significant cost savings while maintaining compatibility with existing OpenAI clients.
Quick Integration
It really is as simple as:
from openai import OpenAI
- client = OpenAI(api_key="sk-...")
+ client = OpenAI(api_key="dummy_openai_api_key", base_url="http://batch-gpt")
Highlights
- Cost Savings: Up to 50% reduction using OpenAI’s Batch API
- Zero Config: Drop-in replacement for OpenAI API clients
- Reliability: Automatic recovery of interrupted batches
- Monitoring: Real-time terminal-based batch status tracking
- Flexibility: Sync/Async/Cache operation modes