Basic Chat Completion
Streaming Responses
For real-time applications, use streaming to receive responses as they’re generated:Usage Metadata
Include"usage": {"include": true} to get token counts and cost:
Advanced Parameters
Prime Inference supports all standard OpenAI API parameters:Next Steps
Team Accounts
Using inference with team accounts
API Reference
Complete API documentation