Expand description
Inference request management for AI model providers.
Provides a minimal blocking interface for sending prompts via configured
Providers and returning a unified InferenceResponse. Supports
OpenAI-compatible and Ollama chat APIs using the Prompt abstraction.
Structs§
- Http
Response - Internal, provider-agnostic HTTP response used by the inference client abstraction.
- Inference
Response - Unified response for a single provider request, including text and token usage.
- Ollama
Chat Response - Response structure from Ollama chat API
- Ollama
Message - Ollama
Message Response - Ollama
Options - Open
AiChoice - Open
AiMessage - Open
AiRequest - Open
AiResponse - Open
AiUsage - Reqwest
Inference Client - Default blocking HTTP client implementation backed by
reqwest.
Traits§
- Inference
Http Client - Minimal HTTP client abstraction for inference requests.
Functions§
- send_
request - Public entry point: send a request using the default blocking HTTP client.
- send_
request_ with_ client - Internal helper: core logic parameterized over an
InferenceHttpClient.