System (Instructions): Specify how you would like the LLM to respond (e.g., the style). It is common to specify here how the LLM should utilize the data received in Prompt.
Type: Text
Prompt: Provide the data the LLM should consider. Type {{ to open the variable builder.
Max tokens: The maximum number of output tokens for each LLM run.
Temperature: The diversity of the LLM generation. To have more diverse or creative generations, increase the temperature. To have more deterministic response, decrease the temperature.
Top P: The Top P parameter constrains how many tokens the LLM considers for generation at each step. For more diverse responses increase top p towards a maximum value of 1.0. This setting is found in the gear on the LLM node.
Stream Response: Check to have responses from the LLM stream. Ensure to change the Type on the output node to “Streamed Text”.
JSON Output: Check to to have the model return a structured JSON output rather than pure text.
Show Sources: Display sources of documents used from the knowledge base.
Show Confidence: Show the confidence level of the LLM’s answer.
Toxic Input Filtration: Filter out toxic content; if the LLM receives a toxic message, the LLM will respond with a respectful one.
Detect PII: Detect and remove PII from being sent to the LLM.