Zum Hauptinhalt springen

OpenRouter

LiteLLM unterstützt alle Text-/Chat-/Vision-Modelle von OpenRouter

Open In Colab

Verwendung

import os
from litellm import completion
os.environ["OPENROUTER_API_KEY"] = ""
os.environ["OPENROUTER_API_BASE"] = "" # [OPTIONAL] defaults to https://openrouter.ai/api/v1


os.environ["OR_SITE_URL"] = "" # [OPTIONAL]
os.environ["OR_APP_NAME"] = "" # [OPTIONAL]

response = completion(
model="openrouter/google/palm-2-chat-bison",
messages=messages,
)

OpenRouter Completion Models

🚨 LiteLLM unterstützt ALLE OpenRouter-Modelle. Senden Sie model=openrouter/<your-openrouter-model>, um es an Open Router zu senden. Alle OpenRouter-Modelle finden Sie hier

ModellnameFunktionsaufruf
openrouter/openai/gpt-3.5-turbocompletion('openrouter/openai/gpt-3.5-turbo', messages)
openrouter/openai/gpt-3.5-turbo-16kcompletion('openrouter/openai/gpt-3.5-turbo-16k', messages)
openrouter/openai/gpt-4completion('openrouter/openai/gpt-4', messages)
openrouter/openai/gpt-4-32kcompletion('openrouter/openai/gpt-4-32k', messages)
openrouter/anthropic/claude-2completion('openrouter/anthropic/claude-2', messages)
openrouter/anthropic/claude-instant-v1completion('openrouter/anthropic/claude-instant-v1', messages)
openrouter/google/palm-2-chat-bisoncompletion('openrouter/google/palm-2-chat-bison', messages)
openrouter/google/palm-2-codechat-bisoncompletion('openrouter/google/palm-2-codechat-bison', messages)
openrouter/meta-llama/llama-2-13b-chatcompletion('openrouter/meta-llama/llama-2-13b-chat', messages)
openrouter/meta-llama/llama-2-70b-chatcompletion('openrouter/meta-llama/llama-2-70b-chat', messages)

OpenRouter-Parameter übergeben - Transforms, Models, Route

Übergeben Sie transforms, models, route als Argumente an litellm.completion()

import os
from litellm import completion
os.environ["OPENROUTER_API_KEY"] = ""

response = completion(
model="openrouter/google/palm-2-chat-bison",
messages=messages,
transforms = [""],
route= ""
)