How to Select a Model
Choose a model from the dropdown in the chat header.
Default Model
When you set a frequently used model as default, it’s selected automatically every time you start a new chat.The default model is per-user. If multiple models are selected when you save the default, the multi-model selection is saved as default.
Multi-model Conversations
Get answers from multiple models for the same question. Including the first one, you can select up to 4 models simultaneously.
Pick additional models
Choose models to compare from the new dropdown. Up to 4 total (including the first).

Multi-model use cases
Multi-model use cases
- Quality comparison: Compare model performance with the same question
- Cross-validation: Get multiple opinions before important decisions
- Best-answer selection: Combine each model’s strengths into the best answer
- Cost vs. performance: Compare results from low-cost and high-performance models
The multi-model feature can be restricted for regular users by admin settings. If you don’t have permission, the ”+” button isn’t shown.
Tuning Model Parameters
The Chat Controls panel lets you tune model parameters per conversation.Opening the Panel
| Environment | Path |
|---|---|
| Desktop | Top-right More (⋯) menu in the chat → select Overview or Artifacts |
| Mobile | Tap the slider icon on the right of the header, or open the More menu → Controls |
The desktop More (⋯) menu only appears after sending the first message.
In a fresh chat, send a prompt once to start the conversation, then tune parameters.
(Adjusted values apply to subsequent messages.)

System Prompt
Set a system prompt that applies to the entire conversation.Key Parameters
Temperature
Temperature
Controls response creativity (randomness).
Range:
| Value | Behavior | Use Case |
|---|---|---|
| 0.0 ~ 0.3 | Deterministic, consistent answers | Code generation, fact-checking, data extraction |
| 0.4 ~ 0.7 | Balanced answers | General chat, summarization, translation |
| 0.8 ~ 2.0 | Creative, varied answers | Brainstorming, creative writing, idea generation |
0.0 ~ 2.0 (UI slider initial value: 0.8 — the value shown when switching to custom; this may differ from the model’s actual default)Top P (Nucleus Sampling)
Top P (Nucleus Sampling)
Limits the probability range the model considers when picking the next token.
- 0.1: Considers only the top 10% probability tokens — very focused
- 0.9: Considers the top 90% — diverse expression
- Generally used together with Temperature; we recommend tuning only one at a time.
Max Tokens
Max Tokens
Limits the maximum length of the AI response in tokens. If unset, the model’s default max applies.
Frequency Penalty
Frequency Penalty
Discourages reuse of tokens that already appeared. Higher values reduce repetition.Range:
-2.0 ~ 2.0 (default: 0)Seed
Seed
Setting a fixed seed produces (nearly) identical responses for the same prompt.
Useful when you need reproducible results.
Reasoning Effort
Reasoning Effort
Applies only to reasoning models. Controls how much effort goes into reasoning.
- Use values like
low,medium,high. - Not supported by all models.
Stop Sequence
Stop Sequence
When the specified string appears, the model stops generating.
Use it to control specific output formats.
Other Advanced Parameters
| Parameter | Description |
|---|---|
| Stream Chat Response | Real-time streaming response on/off |
| Function Calling | Tool calling mode (Default / Native) |
| Top K | Considers only the top K tokens for sampling |
| Min P | Considers only tokens with at least the minimum probability |
| Presence Penalty | Applies a flat penalty to already-seen tokens to encourage new topics (range: -2.0 ~ 2.0) |
| Repeat Penalty | Penalty on repeated tokens (Ollama-only, range: 0.0 ~ 2.0) |
| Repeat Last N | Number of recent tokens checked for repetition |
| Mirostat | Algorithm that auto-regulates response perplexity |
| Context Length (Ollama) | Context window size (in tokens) |
| Logit Bias | Direct adjustment of the appearance probability of specific tokens (-100 ~ 100) |
The Chat Controls panel itself is visible to all users, but the System Prompt and Advanced Params sections are only shown to admins or users with the chat-controls permission (
permissions.chat.controls).Model Capability Display
Agent models manage each capability (web search, image generation, code execution) individually.| State | Description |
|---|---|
| on | Enabled by default |
| user | User can enable manually (default off) |
| off | Disabled (not shown in the input menu) |
base_model_id), all capabilities enabled by the admin in system settings are shown.