Headers
Body
Request model for completion generation
Natural-language query used to retrieve relevant chunks or documents.
1Metadata filters supporting logical operators ($and/$or/$not/$nor) and field predicates ($eq/$ne/$gt/$gte/$lt/$lte/$in/$nin/$exists/$type/$regex/$contains).
Maximum number of chunks or documents to return.
x > 0Minimum similarity score a result must meet before it is returned.
When provided, overrides the workspace reranking configuration for this request.
When provided, uses Morphik's finetuned ColPali style embeddings (recommended to be True for high quality retrieval).
How to return image chunks: base64 data URI (default) or a presigned URL
base64, url Number of additional chunks/pages to retrieve before and after matched chunks (ColPali only)
x >= 0Name of the graph to use for knowledge graph-enhanced retrieval
Number of relationship hops to traverse in the graph
1 <= x <= 3Whether to include relationship paths in the response
Optional folder scope for the operation. Accepts a single folder name or a list of folder names.
Optional end-user scope for the operation
Maximum number of tokens allowed in the generated completion.
Sampling temperature passed to the completion model (None uses provider default).
Optional customizations for entity extraction, resolution, and query prompts Container for query-related prompt overrides.
Use this class when customizing prompts for query operations, which may include customizations for entity extraction, entity resolution, and the query/response generation itself.
This is the most feature-complete override class, supporting all customization types.
Available customizations:
- entity_extraction: Customize how entities are identified in text
- entity_resolution: Customize how entity variants are grouped
- query: Customize response generation style, format, and tone
Each type has its own required placeholders. See the specific class documentation for details and examples.
Schema for structured output, can be a Pydantic model or JSON schema dict
Optional chat session ID for persisting conversation history
Whether to stream the response back in chunks
LiteLLM-compatible model configuration (e.g., model name, API key, base URL)
Whether to include inline citations with filename and page number in the response

