Core Types
PromptMessageExtended
Extended message format used across Fast Agent for representing conversation messages with additional metadata.RequestParams
Parameters for configuring LLM generation requests.Maximum number of tokens to generate
Model identifier to use for generation. Overrides modelPreferences when specified.
Whether to maintain conversation history
Maximum number of tool calls allowed in a conversation turn
Whether to allow simultaneous tool execution
Override response format for structured output. Prefer using Pydantic models.
Template variables for dynamic templates (TensorZero backend)
Metadata to pass through to MCP tool calls via _meta field
Emit monotonic progress updates for the internal tool loop
Skip post-tool LLM synthesis and return tool results directly
Maximum time in seconds to wait for streaming completion. Set to None to disable.
Sampling temperature (provider support varies)
Nucleus sampling parameter (provider support varies)
Top-k sampling parameter (provider support varies)
Minimum probability threshold for sampling (provider support varies)
Presence penalty (provider support varies)
Frequency penalty (provider support varies)
Repetition penalty (provider support varies)
Responses-family service tier override (fast/priority or flex)
ResourceLink
Reference to an MCP resource, re-exported from the MCP SDK.Enumerations
LlmStopReason
Enumeration of reasons why LLM message generation stopped."endTurn" - Normal completion of a conversation turn"stopSequence" - Stopped due to a stop sequence match"maxTokens" - Reached maximum token limit"toolUse" - Stopped to execute tool calls"pause" - Generation paused"error" - Error occurred during generation"cancelled" - Generation cancelled by user"timeout" - Generation timed out"safety" - Safety or content warning triggeredAssistantMessagePhase
Phase metadata for assistant messages, aligned with OpenAI Responses SDK."commentary"- Assistant is providing commentary or thinking"final_answer"- Assistant is providing the final answer
COMMENTARY_PHASE: Type-safe constant for"commentary"FINAL_ANSWER_PHASE: Type-safe constant for"final_answer"
Analysis Types
ConversationSummary
Analyzes conversation history and provides computed statistics.List of messages to analyze
Total number of messages in the conversation
Number of messages from the user
Number of messages from the assistant
Total number of tool calls made
Number of tool calls that resulted in errors
Number of successful tool calls
Proportion of tool calls that failed (0.0 to 1.0)
Mapping of tool names to call counts. Example:
{"fetch_weather": 3}Mapping of tool names to error counts
Whether any tool calls were made
Whether any tool errors occurred
Messages split into logical conversation turns
Number of conversation turns
Total LLM generation time in milliseconds
Average response time for assistant messages
Wall-clock time from first LLM call to last, including tool execution
Unix timestamp when first LLM call started
Unix timestamp when last LLM call ended
Tool Timing Types
ToolTimingInfo
Timing metadata for individual tool calls.Tool execution time in milliseconds
Transport channel identifier for the tool
