LLM model Node
Overview
The LLM Model Node integrates Large Language Models into your flow, enabling natural language processing, generation, and understanding. It supports multiple model providers, temperature control, and a fallback mechanism to ensure reliability.
Usage cost: 1 credit
Configuration
Settings
Model Selection
Primary Model*: Select the main LLM model
Fallback Model: Optional backup model if primary fails
Temperature (0-1): Controls response randomness and creativity
Lower values (closer to 0): More focused, deterministic responses
Higher values (closer to 1): More creative, varied responses
Prompts
System Prompt: Instructions/context for the model's behavior
User Prompt: The main input to be processed
Past Message History: Optional chat history for context
Output format
Output as JSON (Toggle):
Off (Default): The node outputs the model's response as a plain string.
On: Instructs the model to format its response as JSON and attempts to parse the output string. If parsing succeeds, the
response
output port will contain a JSON object/array. If it fails,response
will contain the original string.
Note: JSON structure and parsing success depend heavily on the model's ability to follow instructions. Clearly prompting for JSON format is recommended when this is enabled.
Output Ports
response
(string): The model's generated response
Best Practices
Model Selection
Choose models based on your specific needs (cost, speed, capabilities)
Always configure a fallback model for critical flows
Temperature Settings
Use lower temperatures (0.1-0.3) for:
Factual responses
Structured output
Consistent results
Use higher temperatures (0.6-0.9) for:
Creative writing
Brainstorming
Conversational responses
Prompt Engineering
Keep system prompts clear and specific
Use variables in prompts to make them dynamic
Include relevant context in the prompt
Structure prompts with clear input/output expectations
Message History
Consider memory limitations of the model
Clean or truncate long conversation histories
Using JSON Output
Enable 'Output as JSON' when structured data is needed for downstream nodes or the final workflow output.
Instruct the model clearly in the prompt (User or System) to output only valid JSON. Specifying the exact desired keys and structure enhances reliability (e.g., "Respond ONLY with a valid JSON object containing 'name' (string) and 'items' (array of strings). Do not include any other text.").
Lower temperatures often improve the reliability of JSON generation.
Common Issues
High temperature settings may lead to inconsistent outputs
Missing or poorly formatted system prompts can result in unexpected responses
Token limits may be exceeded with long prompts or chat histories
Rate limiting may affect response times
JSON Output Failures: Even when requested, the model might produce invalid JSON, include explanatory text around the JSON, or fail to follow formatting instructions.
Last updated