diff --git a/main.py b/main.py index 52e6ad9..4b73f52 100644 --- a/main.py +++ b/main.py @@ -123,6 +123,22 @@ async def get_ai_response(user_input, system_prompt): messages=messages ) ai_msg = response.choices[0].message.content + + elif AI_PROVIDER == "nvidia": + if not NVIDIA_API_KEY: + return "Fehler: NVIDIA_API_KEY fehlt!" + + # NVIDIA nutzt ebenfalls das OpenAI Interface + client = openai.OpenAI( + base_url="https://integrate.api.nvidia.com/v1", + api_key=NVIDIA_API_KEY + ) + messages = [{"role": "system", "content": system_prompt}] + chat_history + response = client.chat.completions.create( + model=NVIDIA_MODEL, + messages=messages + ) + ai_msg = response.choices[0].message.content elif AI_PROVIDER == "google": # Für Google Gemini @@ -430,14 +446,12 @@ async def get_models(provider: str, url: str = None): elif provider == "nvidia": if not NVIDIA_API_KEY or "hier" in NVIDIA_API_KEY: return {"models": []} - # NVIDIA nutzt das OpenAI SDK! import openai client = openai.AsyncOpenAI( api_key=NVIDIA_API_KEY, base_url="https://integrate.api.nvidia.com/v1" ) - try: response = await client.models.list() # Bei NVIDIA heißen die Modelle oft nach dem Schema "nvidia/llama-3.1-405b"