Create Infinite Context Response
Respond to an input prompt with Captain's infinite context window management.
Under the hood, Captain splits an input into 30-40% context chunks (for good accuracy).
LLMs process these chunks in parallel, and the responses are generatively merged
for a comprehensive output.
Headers:
- Authorization: Bearer {api_key} - Captain API key for authentication
- X-Organization-ID: Organization UUID
Parameters:
- input: Natural language text content (can be very large, 100M+ tokens)
- query: Optional specific question to answer from the input (if not provided, generates overview)
- stream: Enable real-time streaming via S2 (default: False)
Returns:
- output: Natural language response to the input
Request
This endpoint expects an object.
input
query
stream
tools
tool_choice
Response
Successful Response