package provider import ( "context" "errors" "fmt" "io" openai "github.com/sashabaranov/go-openai" ) // OpenAIProvider implements AIProvider for OpenAI-compatible APIs // (OpenAI, Qwen, Zhipu, DeepSeek, etc.) type OpenAIProvider struct { client *openai.Client name string } // NewOpenAIProvider creates a new OpenAI-compatible provider. // baseUrl can be empty to use the default OpenAI endpoint. func NewOpenAIProvider(baseUrl, apiKey string) *OpenAIProvider { config := openai.DefaultConfig(apiKey) if baseUrl != "" { config.BaseURL = baseUrl } return &OpenAIProvider{ client: openai.NewClientWithConfig(config), name: "openai_compat", } } // Name returns the provider name. func (p *OpenAIProvider) Name() string { return p.name } // Chat sends a synchronous chat completion request. func (p *OpenAIProvider) Chat(ctx context.Context, req *ChatRequest) (*ChatResponse, error) { messages := convertToOpenAIMessages(req.Messages) openaiReq := openai.ChatCompletionRequest{ Model: req.Model, Messages: messages, MaxTokens: req.MaxTokens, Temperature: float32(req.Temperature), } resp, err := p.client.CreateChatCompletion(ctx, openaiReq) if err != nil { return nil, fmt.Errorf("openai chat completion failed: %w", err) } if len(resp.Choices) == 0 { return nil, fmt.Errorf("openai chat completion returned no choices") } return &ChatResponse{ Content: resp.Choices[0].Message.Content, Model: resp.Model, InputTokens: resp.Usage.PromptTokens, OutputTokens: resp.Usage.CompletionTokens, FinishReason: string(resp.Choices[0].FinishReason), }, nil } // ChatStream sends a streaming chat completion request. It returns a channel // that delivers StreamChunk values. The channel is closed when the stream // ends or an error occurs. The final chunk has Done=true and may include // token usage if the API provides it. func (p *OpenAIProvider) ChatStream(ctx context.Context, req *ChatRequest) (<-chan *StreamChunk, error) { messages := convertToOpenAIMessages(req.Messages) openaiReq := openai.ChatCompletionRequest{ Model: req.Model, Messages: messages, MaxTokens: req.MaxTokens, Temperature: float32(req.Temperature), Stream: true, StreamOptions: &openai.StreamOptions{ IncludeUsage: true, }, } stream, err := p.client.CreateChatCompletionStream(ctx, openaiReq) if err != nil { return nil, fmt.Errorf("openai stream creation failed: %w", err) } ch := make(chan *StreamChunk, 64) go func() { defer close(ch) defer stream.Close() for { response, err := stream.Recv() if errors.Is(err, io.EOF) { // Stream finished normally select { case ch <- &StreamChunk{Done: true}: case <-ctx.Done(): } return } if err != nil { // Send error indication as the final chunk select { case ch <- &StreamChunk{ Content: fmt.Sprintf("[stream error: %v]", err), Done: true, FinishReason: "error", }: case <-ctx.Done(): } return } chunk := &StreamChunk{} // Extract content delta from choices if len(response.Choices) > 0 { chunk.Content = response.Choices[0].Delta.Content if response.Choices[0].FinishReason != "" { chunk.FinishReason = string(response.Choices[0].FinishReason) } } // Extract usage from the final usage chunk (when StreamOptions.IncludeUsage is true) if response.Usage != nil { chunk.InputTokens = response.Usage.PromptTokens chunk.OutputTokens = response.Usage.CompletionTokens } select { case ch <- chunk: case <-ctx.Done(): return } } }() return ch, nil } // convertToOpenAIMessages converts our unified ChatMessage slice to OpenAI format. func convertToOpenAIMessages(messages []ChatMessage) []openai.ChatCompletionMessage { result := make([]openai.ChatCompletionMessage, 0, len(messages)) for _, msg := range messages { result = append(result, openai.ChatCompletionMessage{ Role: msg.Role, Content: msg.Content, }) } return result }