Multi-model chat overview
LLM Chat fetches the list of active chat models registered in your project from the organization’s model registry. Every model the LLM proxy has marked asactive and mode=chat appears in the model selector.
Responses are returned as plain text and rendered with whitespace preserved, so you can inspect raw model output without Markdown processing.
Selecting a model
Open LLM Chat
Navigate to LLM Chat from the sidebar (or via the dashboard breadcrumb at Dashboard → LLM Chat).
Choose a model
Click the Select Model dropdown. The list shows every active chat model your project has access to, identified by name. The first available model is selected by default.
Direct API integration with model providers
All model calls in LLM Chat route through the TrayLinx LLM proxy (REACT_APP_OPENAI_API_BASE). The proxy:
- Authenticates requests using your project’s secret key as the OpenAI-compatible API key
- Forwards the
openai-organizationheader to identify the tenant - Returns responses in the OpenAI chat completion format
Conversation history
LLM Chat maintains a single conversation per session. The interface stores the exchange between your prompts and model responses in-memory for the duration of the page session. TheLLMService that powers LLM Chat supports passing conversation history to the model for multi-turn conversations. Messages in the current session are appended as HumanMessage objects in a LangChain chain, giving the model context from earlier in the conversation.
Conversation history in LLM Chat is not persisted across page reloads. For persistent, searchable conversation history, use the Personal Assistant.
Using prompt templates
LLM Chat integrates with TrayLinx’s prompt template system. When you have prompt template assets in your project (created in Studio Tools asprompt_template subtype assets), you can reference them to pre-populate the prompt field with a known template.
Templates let you standardize inputs across repeated tasks — for example, a fixed evaluation rubric or a structured data extraction prompt — so you can change only the variable parts between runs.
Response analysis
After the model responds, the response text appears in the Response card below the prompt input. The response is rendered withwhiteSpace: pre-wrap so line breaks and indentation in the model output are preserved exactly.
To analyze or compare responses:
- Copy the response text manually for comparison across models.
- Switch the model in the selector and re-submit the same prompt to compare outputs side by side.
- Use the prompt field as a scratchpad — modify the prompt and re-generate to iterate on wording.