package llm import ( "bytes" "encoding/json" "errors" "fmt" "io" "net/http" "strings" "time" ) func (s *llmService) SendGeminiRequest(systemPrompt string, userPrompt string, model string) (string, error) { apiKey := getEnvConfig("GEMINI_API_KEY") if apiKey == "" { return "", errors.New("GEMINI_API_KEY not found in .env or environment") } apiBase := "https://generativelanguage.googleapis.com/v1beta" url := fmt.Sprintf("%s/models/%s:generateContent?key=%s", strings.TrimRight(apiBase, "/"), model, apiKey) reqBody := map[string]interface{}{} if systemPrompt != "" { reqBody["system_instruction"] = map[string]interface{}{ "parts": []map[string]string{ {"text": systemPrompt}, }, } } reqBody["contents"] = []map[string]interface{}{ { "role": "user", "parts": []map[string]string{ {"text": userPrompt}, }, }, } jsonBody, err := json.Marshal(reqBody) if err != nil { return "", err } req, err := http.NewRequest("POST", url, bytes.NewBuffer(jsonBody)) if err != nil { return "", err } req.Header.Set("Content-Type", "application/json") client := &http.Client{Timeout: 120 * time.Second} resp, err := client.Do(req) if err != nil { return "", err } defer resp.Body.Close() bodyBytes, err := io.ReadAll(resp.Body) if err != nil { return "", err } if resp.StatusCode != http.StatusOK { return "", fmt.Errorf("Gemini API error status %d: %s", resp.StatusCode, string(bodyBytes)) } var result struct { Candidates []struct { Content struct { Parts []struct { Text string `json:"text"` } `json:"parts"` } `json:"content"` } `json:"candidates"` } if err := json.Unmarshal(bodyBytes, &result); err != nil { return "", err } if len(result.Candidates) > 0 && len(result.Candidates[0].Content.Parts) > 0 { return result.Candidates[0].Content.Parts[0].Text, nil } return "", errors.New("empty response from Gemini API") }