Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.llmstudio.ai/llms.txt

Use this file to discover all available pages before exploring further.

Make chat calls using your LLM.

Parameters

The llm.chat method can have the following parameters.
ParameterTypeDescription
input strThe input message to send to the chat model.
is_streamboolThe temperature parameter for the model.
**kwargsdictAdditional parameters to pass to the chat model.
Refer to your provider-specific documentation for additional kwargs you can use.

Returns

OutputTypeDescription
ChatCompletionobjectA chat completion object in the OpenAI format + metrics computed by LLMstudio.

Usage

Here’s how to use .chat() to make calls to your LLM.
1
Start by importing LLM.
from llmstudio import LLM
2
Set up an LLM from your desired provider.
llm = LLM('openai/gpt-4o')
3
Create your message. Your message can be a simple string or a message in the OpenAI format.
message = "Hello, how are you today?"
4
Get your response.
response = llm.chat(message)
Vizualize your response.
print(response)
You are done chating with your LLMstudio LLM!