Transition google-generativeai pour google-genai
This commit is contained in:
parent
a9b79304f8
commit
a4d1bae9f0
1 changed files with 31 additions and 46 deletions
|
@ -3,8 +3,9 @@ import logging
|
||||||
import sys
|
import sys
|
||||||
from typing import Optional, Dict, Any
|
from typing import Optional, Dict, Any
|
||||||
|
|
||||||
# MODIFIÉ ICI: Supprime 'from google.genai import types' car types.GenerateContentConfig n'est plus utilisé de cette manière
|
# MODIFIÉ ICI: Importations pour google-genai
|
||||||
import google.generativeai as genai
|
from google import genai
|
||||||
|
from google.genai import types # Nécessaire pour configurer les types comme GenerateContentConfig
|
||||||
import mistralai
|
import mistralai
|
||||||
from mistralai.client import MistralClient
|
from mistralai.client import MistralClient
|
||||||
|
|
||||||
|
@ -28,53 +29,38 @@ try:
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Error during mistralai debug info collection: {e}")
|
logger.error(f"Error during mistralai debug info collection: {e}")
|
||||||
|
|
||||||
# --- Configuration globale du client Gemini (Ce bloc est maintenant supprimé car la configuration est faite via GenerativeModel) ---
|
|
||||||
# Vous pouvez retirer ce bloc si vous l'aviez :
|
|
||||||
# if settings.LLM_PROVIDER == "gemini" and settings.GEMINI_API_KEY:
|
|
||||||
# try:
|
|
||||||
# genai.configure(
|
|
||||||
# api_key=settings.GEMINI_API_KEY,
|
|
||||||
# client_options={"api_endpoint": "generativelanguage.googleapis.com"}
|
|
||||||
# )
|
|
||||||
# logger.info("GenAI client globally configured with API endpoint.")
|
|
||||||
# except Exception as e:
|
|
||||||
# logger.error(f"Erreur lors de la configuration globale de GenAI: {e}")
|
|
||||||
|
|
||||||
|
|
||||||
class AIService:
|
class AIService:
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
self.provider = settings.LLM_PROVIDER
|
self.provider = settings.LLM_PROVIDER
|
||||||
self.model_name = settings.GEMINI_MODEL_NAME if self.provider == "gemini" else settings.MISTRAL_MODEL_NAME
|
self.model_name = settings.GEMINI_MODEL_NAME if self.provider == "gemini" else settings.MISTRAL_MODEL_NAME
|
||||||
|
|
||||||
self.raw_safety_settings = [
|
|
||||||
{"category": "HARM_CATEGORY_HARASSMENT", "threshold": "BLOCK_NONE"},
|
|
||||||
{"category": "HARM_CATEGORY_HATE_SPEECH", "threshold": "BLOCK_NONE"},
|
|
||||||
{"category": "HARM_CATEGORY_SEXUALLY_EXPLICIT", "threshold": "BLOCK_NONE"},
|
|
||||||
{"category": "HARM_CATEGORY_DANGEROUS_CONTENT", "threshold": "BLOCK_NONE"},
|
|
||||||
]
|
|
||||||
|
|
||||||
self.raw_generation_config = {
|
|
||||||
"temperature": 0.7,
|
|
||||||
"top_p": 1,
|
|
||||||
"top_k": 1,
|
|
||||||
}
|
|
||||||
|
|
||||||
if self.provider == "gemini":
|
if self.provider == "gemini":
|
||||||
try:
|
try:
|
||||||
# MODIFICATION CRUCIALE ICI : On initialise directement le GenerativeModel
|
# Initialisation du client genai.Client()
|
||||||
# genai.Client() et types.GenerateContentConfig ne sont plus utilisés directement ici
|
self.client = genai.Client(
|
||||||
self.model = genai.GenerativeModel(
|
api_key=settings.GEMINI_API_KEY
|
||||||
model_name=self.model_name,
|
|
||||||
safety_settings=self.raw_safety_settings, # Passez safety_settings ici
|
|
||||||
generation_config=self.raw_generation_config # Passez generation_config ici
|
|
||||||
# La clé API est lue automatiquement depuis GEMINI_API_KEY si elle est configurée.
|
|
||||||
# Ou vous pouvez la passer explicitement: api_key=settings.GEMINI_API_KEY
|
|
||||||
)
|
)
|
||||||
logger.info(f"Modèle Gemini GenerativeModel initialisé avec modèle : {self.model_name}")
|
logger.info(f"Client Gemini genai.Client() initialisé.")
|
||||||
|
|
||||||
|
# Configuration de la génération avec types.GenerateContentConfig
|
||||||
|
self.gemini_config = types.GenerateContentConfig(
|
||||||
|
temperature=0.7,
|
||||||
|
top_p=1.0,
|
||||||
|
top_k=1,
|
||||||
|
safety_settings=[
|
||||||
|
types.SafetySetting(category="HARM_CATEGORY_HARASSMENT", threshold="BLOCK_NONE"),
|
||||||
|
types.SafetySetting(category="HARM_CATEGORY_HATE_SPEECH", threshold="BLOCK_NONE"),
|
||||||
|
types.SafetySetting(category="HARM_CATEGORY_SEXUALLY_EXPLICIT", threshold="BLOCK_NONE"),
|
||||||
|
types.SafetySetting(category="HARM_CATEGORY_DANGEROUS_CONTENT", threshold="BLOCK_NONE"),
|
||||||
|
]
|
||||||
|
)
|
||||||
|
logger.info(f"Configuration Gemini types.GenerateContentConfig créée.")
|
||||||
|
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Erreur d'initialisation du modèle Gemini: {e}")
|
logger.error(f"Erreur d'initialisation du client Gemini: {e}")
|
||||||
raise ValueError(f"Impossible d'initialiser le modèle Gemini. Vérifiez votre GEMINI_API_KEY et le nom du modèle. Erreur: {e}")
|
raise ValueError(f"Impossible d'initialiser le client Gemini. Vérifiez votre GEMINI_API_KEY. Erreur: {e}")
|
||||||
|
|
||||||
elif self.provider == "mistral":
|
elif self.provider == "mistral":
|
||||||
if not settings.MISTRAL_API_KEY:
|
if not settings.MISTRAL_API_KEY:
|
||||||
|
@ -117,15 +103,14 @@ class AIService:
|
||||||
response_content = ""
|
response_content = ""
|
||||||
if self.provider == "gemini":
|
if self.provider == "gemini":
|
||||||
try:
|
try:
|
||||||
contents = [
|
# MODIFIÉ ICI: Utilisation d'une lambda pour envelopper l'appel à generate_content
|
||||||
{"role": "user", "parts": [{"text": prompt}]}
|
# avec tous ses arguments, pour que run_sync reçoive une fonction sans arguments supplémentaires
|
||||||
]
|
|
||||||
|
|
||||||
# MODIFIÉ ICI : 'contents' est maintenant passé comme argument positionnel direct à generate_content
|
|
||||||
response = await anyio.to_thread.run_sync(
|
response = await anyio.to_thread.run_sync(
|
||||||
self.model.generate_content,
|
lambda: self.client.models.generate_content(
|
||||||
contents, # <-- Correction pour l'erreur "unexpected keyword argument 'contents'"
|
model=self.model_name,
|
||||||
# Les configurations (température, safety_settings) sont déjà définies lors de l'initialisation de self.model
|
contents=[{"role": "user", "parts": [{"text": prompt}]}],
|
||||||
|
config=self.gemini_config,
|
||||||
|
)
|
||||||
)
|
)
|
||||||
response_content = response.text
|
response_content = response.text
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue