Using Parameters

Configure temperature, max_tokens, logit_bias, and more with LangDB AI Gateway. Test easily via API, UI, or Playground.

LangDB AI Gateway supports every LLM parameter like temperature, max_tokens, stop sequences, logit_bias, and more.

API Usage:

from openai import OpenAI

response = client.chat.completions.create(
    model="gpt-4o", # Change Model
    messages=[
        {"role": "user", "content": "What are the earnings of Apple in 2022?"},
    ],
    temperature=0.7,               # temperature parameter
    max_tokens=150,                # max_tokens parameter
    stream=True                   # stream parameter
)

UI

You can also use the UI to test various parameters and getting code snippet

Playground

Use the Playground to tweak parameters in real time via the Virtual Model config and send test requests instantly.

Trying different Parameters for chat completions through LangDB Playground

Samples

Explore ready-made code snippets complete with preconfigured parameters—copy, paste, and customize to fit your needs.

Trying different Parameters for chat completions through LangDB Samples

Last updated

Was this helpful?