Add Azure OpenAI integration - new service, admin UI, and backend support
This commit is contained in:
parent
ae90efbfa0
commit
4fcab45b83
|
|
@ -20,6 +20,7 @@ const ENGINES = [
|
||||||
{ id: 'openrouter', label: 'OpenRouter', icon: Cloud },
|
{ id: 'openrouter', label: 'OpenRouter', icon: Cloud },
|
||||||
{ id: 'openai', label: 'OpenAI', icon: Server },
|
{ id: 'openai', label: 'OpenAI', icon: Server },
|
||||||
{ id: 'google', label: 'Google Vision', icon: Eye },
|
{ id: 'google', label: 'Google Vision', icon: Eye },
|
||||||
|
{ id: 'azure', label: 'Azure OpenAI', icon: Cloud },
|
||||||
];
|
];
|
||||||
|
|
||||||
const LOCAL_MODELS = [
|
const LOCAL_MODELS = [
|
||||||
|
|
@ -352,6 +353,25 @@ export default function AIModerationPage() {
|
||||||
{selectedEngine === 'google' && (
|
{selectedEngine === 'google' && (
|
||||||
<p className="text-xs text-gray-500">Google Vision SafeSearch is configured via service account. No additional settings needed.</p>
|
<p className="text-xs text-gray-500">Google Vision SafeSearch is configured via service account. No additional settings needed.</p>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
|
{selectedEngine === 'azure' && (
|
||||||
|
<div>
|
||||||
|
<label className="text-xs font-medium text-gray-600 block mb-1">Deployment Name</label>
|
||||||
|
<input
|
||||||
|
type="text"
|
||||||
|
value={modelId}
|
||||||
|
onChange={(e) => {
|
||||||
|
setModelId(e.target.value);
|
||||||
|
setModelName(e.target.value);
|
||||||
|
}}
|
||||||
|
placeholder="e.g., gpt-4o-vision"
|
||||||
|
className="w-full text-sm border border-gray-300 rounded-lg px-3 py-2 focus:outline-none focus:ring-2 focus:ring-brand-500"
|
||||||
|
/>
|
||||||
|
<p className="text-xs text-gray-500 mt-2">
|
||||||
|
Azure OpenAI deployment name (configured in Azure portal). Uses your Azure credits.
|
||||||
|
</p>
|
||||||
|
</div>
|
||||||
|
)}
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
{/* Save Button */}
|
{/* Save Button */}
|
||||||
|
|
|
||||||
|
|
@ -124,6 +124,15 @@ func main() {
|
||||||
// Initialize OpenRouter service
|
// Initialize OpenRouter service
|
||||||
openRouterService := services.NewOpenRouterService(dbPool, cfg.OpenRouterAPIKey)
|
openRouterService := services.NewOpenRouterService(dbPool, cfg.OpenRouterAPIKey)
|
||||||
|
|
||||||
|
// Initialize Azure OpenAI service
|
||||||
|
var azureOpenAIService *services.AzureOpenAIService
|
||||||
|
if cfg.AzureOpenAIAPIKey != "" && cfg.AzureOpenAIEndpoint != "" {
|
||||||
|
azureOpenAIService = services.NewAzureOpenAIService(dbPool, cfg.AzureOpenAIAPIKey, cfg.AzureOpenAIEndpoint, cfg.AzureOpenAIAPIVersion)
|
||||||
|
log.Info().Msg("Azure OpenAI service initialized")
|
||||||
|
} else {
|
||||||
|
log.Warn().Msg("Azure OpenAI credentials not provided, Azure OpenAI service disabled")
|
||||||
|
}
|
||||||
|
|
||||||
// Initialize content filter (hard blocklist + strike system)
|
// Initialize content filter (hard blocklist + strike system)
|
||||||
contentFilter := services.NewContentFilter(dbPool)
|
contentFilter := services.NewContentFilter(dbPool)
|
||||||
|
|
||||||
|
|
@ -177,7 +186,7 @@ func main() {
|
||||||
|
|
||||||
moderationHandler := handlers.NewModerationHandler(moderationService, openRouterService, localAIService)
|
moderationHandler := handlers.NewModerationHandler(moderationService, openRouterService, localAIService)
|
||||||
|
|
||||||
adminHandler := handlers.NewAdminHandler(dbPool, moderationService, appealService, emailService, openRouterService, officialAccountsService, linkPreviewService, localAIService, cfg.JWTSecret, cfg.TurnstileSecretKey, s3Client, cfg.R2MediaBucket, cfg.R2VideoBucket, cfg.R2ImgDomain, cfg.R2VidDomain)
|
adminHandler := handlers.NewAdminHandler(dbPool, moderationService, appealService, emailService, openRouterService, azureOpenAIService, officialAccountsService, linkPreviewService, localAIService, cfg.JWTSecret, cfg.TurnstileSecretKey, s3Client, cfg.R2MediaBucket, cfg.R2VideoBucket, cfg.R2ImgDomain, cfg.R2VidDomain)
|
||||||
|
|
||||||
accountHandler := handlers.NewAccountHandler(userRepo, emailService, cfg)
|
accountHandler := handlers.NewAccountHandler(userRepo, emailService, cfg)
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -41,6 +41,9 @@ type Config struct {
|
||||||
OpenRouterAPIKey string
|
OpenRouterAPIKey string
|
||||||
AIGatewayURL string
|
AIGatewayURL string
|
||||||
AIGatewayToken string
|
AIGatewayToken string
|
||||||
|
AzureOpenAIAPIKey string
|
||||||
|
AzureOpenAIEndpoint string
|
||||||
|
AzureOpenAIAPIVersion string
|
||||||
}
|
}
|
||||||
|
|
||||||
func LoadConfig() *Config {
|
func LoadConfig() *Config {
|
||||||
|
|
@ -88,6 +91,9 @@ func LoadConfig() *Config {
|
||||||
OpenRouterAPIKey: getEnv("OPENROUTER_API", ""),
|
OpenRouterAPIKey: getEnv("OPENROUTER_API", ""),
|
||||||
AIGatewayURL: getEnv("AI_GATEWAY_URL", ""),
|
AIGatewayURL: getEnv("AI_GATEWAY_URL", ""),
|
||||||
AIGatewayToken: getEnv("AI_GATEWAY_TOKEN", ""),
|
AIGatewayToken: getEnv("AI_GATEWAY_TOKEN", ""),
|
||||||
|
AzureOpenAIAPIKey: getEnv("AZURE_OPENAI_API_KEY", ""),
|
||||||
|
AzureOpenAIEndpoint: getEnv("AZURE_OPENAI_ENDPOINT", ""),
|
||||||
|
AzureOpenAIAPIVersion: getEnv("AZURE_OPENAI_API_VERSION", "2024-02-15-preview"),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -26,6 +26,7 @@ type AdminHandler struct {
|
||||||
appealService *services.AppealService
|
appealService *services.AppealService
|
||||||
emailService *services.EmailService
|
emailService *services.EmailService
|
||||||
openRouterService *services.OpenRouterService
|
openRouterService *services.OpenRouterService
|
||||||
|
azureOpenAIService *services.AzureOpenAIService
|
||||||
officialAccountsService *services.OfficialAccountsService
|
officialAccountsService *services.OfficialAccountsService
|
||||||
linkPreviewService *services.LinkPreviewService
|
linkPreviewService *services.LinkPreviewService
|
||||||
localAIService *services.LocalAIService
|
localAIService *services.LocalAIService
|
||||||
|
|
@ -38,13 +39,14 @@ type AdminHandler struct {
|
||||||
vidDomain string
|
vidDomain string
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewAdminHandler(pool *pgxpool.Pool, moderationService *services.ModerationService, appealService *services.AppealService, emailService *services.EmailService, openRouterService *services.OpenRouterService, officialAccountsService *services.OfficialAccountsService, linkPreviewService *services.LinkPreviewService, localAIService *services.LocalAIService, jwtSecret string, turnstileSecret string, s3Client *s3.Client, mediaBucket string, videoBucket string, imgDomain string, vidDomain string) *AdminHandler {
|
func NewAdminHandler(pool *pgxpool.Pool, moderationService *services.ModerationService, appealService *services.AppealService, emailService *services.EmailService, openRouterService *services.OpenRouterService, azureOpenAIService *services.AzureOpenAIService, officialAccountsService *services.OfficialAccountsService, linkPreviewService *services.LinkPreviewService, localAIService *services.LocalAIService, jwtSecret string, turnstileSecret string, s3Client *s3.Client, mediaBucket string, videoBucket string, imgDomain string, vidDomain string) *AdminHandler {
|
||||||
return &AdminHandler{
|
return &AdminHandler{
|
||||||
pool: pool,
|
pool: pool,
|
||||||
moderationService: moderationService,
|
moderationService: moderationService,
|
||||||
appealService: appealService,
|
appealService: appealService,
|
||||||
emailService: emailService,
|
emailService: emailService,
|
||||||
openRouterService: openRouterService,
|
openRouterService: openRouterService,
|
||||||
|
azureOpenAIService: azureOpenAIService,
|
||||||
officialAccountsService: officialAccountsService,
|
officialAccountsService: officialAccountsService,
|
||||||
linkPreviewService: linkPreviewService,
|
linkPreviewService: linkPreviewService,
|
||||||
localAIService: localAIService,
|
localAIService: localAIService,
|
||||||
|
|
@ -2850,6 +2852,36 @@ func (h *AdminHandler) TestAIModeration(c *gin.Context) {
|
||||||
}
|
}
|
||||||
response["result"] = result
|
response["result"] = result
|
||||||
|
|
||||||
|
case "azure":
|
||||||
|
if h.azureOpenAIService == nil {
|
||||||
|
c.JSON(http.StatusServiceUnavailable, gin.H{"error": "Azure OpenAI not configured"})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
var result *services.ModerationResult
|
||||||
|
var err error
|
||||||
|
isImage := strings.Contains(req.ModerationType, "image") || req.ModerationType == "video"
|
||||||
|
if isImage && req.ImageURL != "" {
|
||||||
|
if req.ModerationType == "video" {
|
||||||
|
urls := strings.Split(req.ImageURL, ",")
|
||||||
|
result, err = h.azureOpenAIService.ModerateVideo(ctx, urls)
|
||||||
|
} else {
|
||||||
|
result, err = h.azureOpenAIService.ModerateImage(ctx, req.ImageURL)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
// Use type-specific config if available, fall back to generic text
|
||||||
|
result, err = h.azureOpenAIService.ModerateWithType(ctx, req.ModerationType, req.Content, nil)
|
||||||
|
if err != nil {
|
||||||
|
// Fall back to generic text moderation
|
||||||
|
result, err = h.azureOpenAIService.ModerateText(ctx, req.Content)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
response["error"] = err.Error()
|
||||||
|
c.JSON(http.StatusOK, response)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
response["result"] = result
|
||||||
|
|
||||||
default:
|
default:
|
||||||
c.JSON(http.StatusBadRequest, gin.H{"error": "Invalid engine: " + engine})
|
c.JSON(http.StatusBadRequest, gin.H{"error": "Invalid engine: " + engine})
|
||||||
return
|
return
|
||||||
|
|
@ -3767,6 +3799,23 @@ func (h *AdminHandler) GetAIEngines(c *gin.Context) {
|
||||||
}
|
}
|
||||||
engines = append(engines, googleEngine)
|
engines = append(engines, googleEngine)
|
||||||
|
|
||||||
|
// 5. Azure OpenAI
|
||||||
|
azureEngine := EngineStatus{
|
||||||
|
ID: "azure",
|
||||||
|
Name: "Azure OpenAI",
|
||||||
|
Description: "Microsoft Azure OpenAI service with vision models. Uses your Azure credits. Supports text and image moderation with customizable prompts.",
|
||||||
|
Configured: h.azureOpenAIService != nil,
|
||||||
|
}
|
||||||
|
if h.azureOpenAIService != nil {
|
||||||
|
azureEngine.Status = "ready"
|
||||||
|
azureEngine.Details = map[string]any{
|
||||||
|
"uses_azure_credits": true,
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
azureEngine.Status = "not_configured"
|
||||||
|
}
|
||||||
|
engines = append(engines, azureEngine)
|
||||||
|
|
||||||
c.JSON(http.StatusOK, gin.H{"engines": engines})
|
c.JSON(http.StatusOK, gin.H{"engines": engines})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
297
go-backend/internal/services/azure_openai_service.go
Normal file
297
go-backend/internal/services/azure_openai_service.go
Normal file
|
|
@ -0,0 +1,297 @@
|
||||||
|
package services
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"context"
|
||||||
|
"encoding/base64"
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"net/http"
|
||||||
|
"strings"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/jackc/pgx/v5/pgxpool"
|
||||||
|
)
|
||||||
|
|
||||||
|
// AzureOpenAIService handles interactions with Azure OpenAI API
|
||||||
|
type AzureOpenAIService struct {
|
||||||
|
pool *pgxpool.Pool
|
||||||
|
httpClient *http.Client
|
||||||
|
apiKey string
|
||||||
|
endpoint string
|
||||||
|
apiVersion string
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewAzureOpenAIService creates a new Azure OpenAI service
|
||||||
|
func NewAzureOpenAIService(pool *pgxpool.Pool, apiKey, endpoint, apiVersion string) *AzureOpenAIService {
|
||||||
|
return &AzureOpenAIService{
|
||||||
|
pool: pool,
|
||||||
|
httpClient: &http.Client{
|
||||||
|
Timeout: 60 * time.Second,
|
||||||
|
},
|
||||||
|
apiKey: apiKey,
|
||||||
|
endpoint: endpoint,
|
||||||
|
apiVersion: apiVersion,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// AzureOpenAIMessage represents a message in Azure OpenAI chat completion
|
||||||
|
type AzureOpenAIMessage struct {
|
||||||
|
Role string `json:"role"`
|
||||||
|
Content interface{} `json:"content"`
|
||||||
|
ContentParts []struct {
|
||||||
|
Type string `json:"type"`
|
||||||
|
Text string `json:"text,omitempty"`
|
||||||
|
ImageURL struct {
|
||||||
|
URL string `json:"url,omitempty"`
|
||||||
|
} `json:"image_url,omitempty"`
|
||||||
|
} `json:"content,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// AzureOpenAIRequest represents a chat completion request to Azure OpenAI
|
||||||
|
type AzureOpenAIRequest struct {
|
||||||
|
Model string `json:"model"`
|
||||||
|
Messages []AzureOpenAIMessage `json:"messages"`
|
||||||
|
Temperature *float64 `json:"temperature,omitempty"`
|
||||||
|
MaxTokens *int `json:"max_tokens,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// AzureOpenAIResponse represents a chat completion response from Azure OpenAI
|
||||||
|
type AzureOpenAIResponse struct {
|
||||||
|
ID string `json:"id"`
|
||||||
|
Choices []struct {
|
||||||
|
Message struct {
|
||||||
|
Role string `json:"role"`
|
||||||
|
Content string `json:"content"`
|
||||||
|
} `json:"message"`
|
||||||
|
FinishReason string `json:"finish_reason"`
|
||||||
|
} `json:"choices"`
|
||||||
|
Usage struct {
|
||||||
|
PromptTokens int `json:"prompt_tokens"`
|
||||||
|
CompletionTokens int `json:"completion_tokens"`
|
||||||
|
TotalTokens int `json:"total_tokens"`
|
||||||
|
} `json:"usage"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// ModerateText sends text content to Azure OpenAI for moderation
|
||||||
|
func (s *AzureOpenAIService) ModerateText(ctx context.Context, content string) (*ModerationResult, error) {
|
||||||
|
config, err := s.GetModerationConfig(ctx, "text")
|
||||||
|
if err != nil || !config.Enabled || config.ModelID == "" {
|
||||||
|
return nil, fmt.Errorf("text moderation not configured")
|
||||||
|
}
|
||||||
|
return s.callModel(ctx, config.ModelID, config.SystemPrompt, content, nil)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ModerateImage sends an image URL to Azure OpenAI vision model for moderation
|
||||||
|
func (s *AzureOpenAIService) ModerateImage(ctx context.Context, imageURL string) (*ModerationResult, error) {
|
||||||
|
config, err := s.GetModerationConfig(ctx, "image")
|
||||||
|
if err != nil || !config.Enabled || config.ModelID == "" {
|
||||||
|
return nil, fmt.Errorf("image moderation not configured")
|
||||||
|
}
|
||||||
|
return s.callModel(ctx, config.ModelID, config.SystemPrompt, "", []string{imageURL})
|
||||||
|
}
|
||||||
|
|
||||||
|
// ModerateWithType sends content to a specific moderation type config
|
||||||
|
func (s *AzureOpenAIService) ModerateWithType(ctx context.Context, moderationType string, textContent string, imageURLs []string) (*ModerationResult, error) {
|
||||||
|
config, err := s.GetModerationConfig(ctx, moderationType)
|
||||||
|
if err != nil || !config.Enabled || config.ModelID == "" {
|
||||||
|
return nil, fmt.Errorf("%s moderation not configured", moderationType)
|
||||||
|
}
|
||||||
|
return s.callModel(ctx, config.ModelID, config.SystemPrompt, textContent, imageURLs)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ModerateVideo sends video frame URLs to Azure OpenAI vision model for moderation
|
||||||
|
func (s *AzureOpenAIService) ModerateVideo(ctx context.Context, frameURLs []string) (*ModerationResult, error) {
|
||||||
|
config, err := s.GetModerationConfig(ctx, "video")
|
||||||
|
if err != nil || !config.Enabled || config.ModelID == "" {
|
||||||
|
return nil, fmt.Errorf("video moderation not configured")
|
||||||
|
}
|
||||||
|
return s.callModel(ctx, config.ModelID, config.SystemPrompt, "These are 3 frames extracted from a short video. Analyze all frames for policy violations.", frameURLs)
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetModerationConfig retrieves moderation configuration from database
|
||||||
|
func (s *AzureOpenAIService) GetModerationConfig(ctx context.Context, moderationType string) (*ModerationConfigEntry, error) {
|
||||||
|
var config ModerationConfigEntry
|
||||||
|
query := `SELECT id, moderation_type, model_id, model_name, system_prompt, enabled, engines, updated_at, updated_by
|
||||||
|
FROM ai_moderation_config
|
||||||
|
WHERE moderation_type = $1 AND $2 = ANY(engines)`
|
||||||
|
|
||||||
|
err := s.pool.QueryRow(ctx, query, moderationType, "azure").Scan(
|
||||||
|
&config.ID, &config.ModerationType, &config.ModelID, &config.ModelName,
|
||||||
|
&config.SystemPrompt, &config.Enabled, &config.Engines, &config.UpdatedAt, &config.UpdatedBy,
|
||||||
|
)
|
||||||
|
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("azure moderation config not found for %s: %w", moderationType, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return &config, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// downloadImage downloads an image from URL and returns base64 encoded data
|
||||||
|
func (s *AzureOpenAIService) downloadImage(ctx context.Context, url string) (string, error) {
|
||||||
|
req, err := http.NewRequestWithContext(ctx, "GET", url, nil)
|
||||||
|
if err != nil {
|
||||||
|
return "", fmt.Errorf("failed to create image request: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := s.httpClient.Do(req)
|
||||||
|
if err != nil {
|
||||||
|
return "", fmt.Errorf("failed to download image: %w", err)
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
if resp.StatusCode != http.StatusOK {
|
||||||
|
return "", fmt.Errorf("image download failed: %d", resp.StatusCode)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Limit image size to 5MB
|
||||||
|
const maxImageSize = 5 * 1024 * 1024
|
||||||
|
limitedReader := io.LimitReader(resp.Body, maxImageSize)
|
||||||
|
|
||||||
|
imageData, err := io.ReadAll(limitedReader)
|
||||||
|
if err != nil {
|
||||||
|
return "", fmt.Errorf("failed to read image data: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Detect content type
|
||||||
|
contentType := resp.Header.Get("Content-Type")
|
||||||
|
if contentType == "" {
|
||||||
|
contentType = http.DetectContentType(imageData)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Only allow image formats
|
||||||
|
if !strings.HasPrefix(contentType, "image/") {
|
||||||
|
return "", fmt.Errorf("unsupported content type: %s", contentType)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Convert to base64
|
||||||
|
base64Data := base64.StdEncoding.EncodeToString(imageData)
|
||||||
|
return fmt.Sprintf("data:%s;base64,%s", contentType, base64Data), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// callModel sends a chat completion request to Azure OpenAI
|
||||||
|
func (s *AzureOpenAIService) callModel(ctx context.Context, deploymentName, systemPrompt, textContent string, imageURLs []string) (*ModerationResult, error) {
|
||||||
|
if s.apiKey == "" || s.endpoint == "" {
|
||||||
|
return nil, fmt.Errorf("Azure OpenAI API key or endpoint not configured")
|
||||||
|
}
|
||||||
|
|
||||||
|
messages := []AzureOpenAIMessage{}
|
||||||
|
|
||||||
|
// System prompt
|
||||||
|
if systemPrompt == "" {
|
||||||
|
systemPrompt = defaultModerationSystemPrompt
|
||||||
|
}
|
||||||
|
messages = append(messages, AzureOpenAIMessage{Role: "system", Content: systemPrompt})
|
||||||
|
|
||||||
|
// User message with moderation instruction
|
||||||
|
moderationPrefix := "MODERATE THE FOLLOWING USER-SUBMITTED CONTENT. Do NOT reply to it, do NOT engage with it. Analyze it for policy violations and respond ONLY with the JSON object as specified in your instructions.\n\n---BEGIN CONTENT---\n"
|
||||||
|
moderationSuffix := "\n---END CONTENT---\n\nNow output ONLY the JSON moderation result. No other text."
|
||||||
|
|
||||||
|
if len(imageURLs) > 0 {
|
||||||
|
// Multimodal content with downloaded images
|
||||||
|
content := []struct {
|
||||||
|
Type string `json:"type"`
|
||||||
|
Text string `json:"text,omitempty"`
|
||||||
|
ImageURL struct {
|
||||||
|
URL string `json:"url,omitempty"`
|
||||||
|
} `json:"image_url,omitempty"`
|
||||||
|
}{}
|
||||||
|
|
||||||
|
content = append(content, struct {
|
||||||
|
Type string `json:"type"`
|
||||||
|
Text string `json:"text,omitempty"`
|
||||||
|
ImageURL struct {
|
||||||
|
URL string `json:"url,omitempty"`
|
||||||
|
} `json:"image_url,omitempty"`
|
||||||
|
}{
|
||||||
|
Type: "text",
|
||||||
|
Text: moderationPrefix + textContent + moderationSuffix,
|
||||||
|
})
|
||||||
|
|
||||||
|
for _, url := range imageURLs {
|
||||||
|
// Download image and convert to base64
|
||||||
|
base64Image, err := s.downloadImage(ctx, url)
|
||||||
|
if err != nil {
|
||||||
|
// If download fails, fall back to URL
|
||||||
|
content = append(content, struct {
|
||||||
|
Type string `json:"type"`
|
||||||
|
Text string `json:"text,omitempty"`
|
||||||
|
ImageURL struct {
|
||||||
|
URL string `json:"url,omitempty"`
|
||||||
|
} `json:"image_url,omitempty"`
|
||||||
|
}{
|
||||||
|
Type: "image_url",
|
||||||
|
ImageURL: struct {
|
||||||
|
URL string `json:"url,omitempty"`
|
||||||
|
}{URL: url},
|
||||||
|
})
|
||||||
|
} else {
|
||||||
|
// Use base64 data
|
||||||
|
content = append(content, struct {
|
||||||
|
Type string `json:"type"`
|
||||||
|
Text string `json:"text,omitempty"`
|
||||||
|
ImageURL struct {
|
||||||
|
URL string `json:"url,omitempty"`
|
||||||
|
} `json:"image_url,omitempty"`
|
||||||
|
}{
|
||||||
|
Type: "image_url",
|
||||||
|
ImageURL: struct {
|
||||||
|
URL string `json:"url,omitempty"`
|
||||||
|
}{URL: base64Image},
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
messages = append(messages, AzureOpenAIMessage{Role: "user", Content: content})
|
||||||
|
} else {
|
||||||
|
wrappedText := moderationPrefix + textContent + moderationSuffix
|
||||||
|
messages = append(messages, AzureOpenAIMessage{Role: "user", Content: wrappedText})
|
||||||
|
}
|
||||||
|
|
||||||
|
reqBody := AzureOpenAIRequest{
|
||||||
|
Model: deploymentName,
|
||||||
|
Messages: messages,
|
||||||
|
Temperature: floatPtr(0.0),
|
||||||
|
MaxTokens: intPtr(500),
|
||||||
|
}
|
||||||
|
|
||||||
|
jsonBody, err := json.Marshal(reqBody)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("failed to marshal request: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Build Azure OpenAI URL
|
||||||
|
url := fmt.Sprintf("%s/openai/deployments/%s/chat/completions?api-version=%s", s.endpoint, deploymentName, s.apiVersion)
|
||||||
|
|
||||||
|
req, err := http.NewRequestWithContext(ctx, "POST", url, bytes.NewBuffer(jsonBody))
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("failed to create request: %w", err)
|
||||||
|
}
|
||||||
|
req.Header.Set("Content-Type", "application/json")
|
||||||
|
req.Header.Set("api-key", s.apiKey)
|
||||||
|
|
||||||
|
resp, err := s.httpClient.Do(req)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("Azure OpenAI request failed: %w", err)
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
if resp.StatusCode != http.StatusOK {
|
||||||
|
body, _ := io.ReadAll(resp.Body)
|
||||||
|
return nil, fmt.Errorf("Azure OpenAI API error %d: %s", resp.StatusCode, string(body))
|
||||||
|
}
|
||||||
|
|
||||||
|
var chatResp AzureOpenAIResponse
|
||||||
|
if err := json.NewDecoder(resp.Body).Decode(&chatResp); err != nil {
|
||||||
|
return nil, fmt.Errorf("failed to decode response: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(chatResp.Choices) == 0 {
|
||||||
|
return nil, fmt.Errorf("no response from model")
|
||||||
|
}
|
||||||
|
|
||||||
|
raw := chatResp.Choices[0].Message.Content
|
||||||
|
return parseModerationResponse(raw), nil
|
||||||
|
}
|
||||||
Loading…
Reference in a new issue