Hub
Pricing About
NodeNode / Predictor

LLM Prompter

KNIME LabsAIModels
Drag & drop
Like

For each row in the input table, this node sends one prompt to the LLM and receives a corresponding response. Rows and the corresponding prompts are treated in isolation, i.e. the LLM cannot remember the contents of the previous rows or how it responded to them.

Note : If you use the Credentials Configuration node and do not select the "Save password in configuration (weakly encrypted)" option for passing the API key for the LLM connector node, the Credentials Configuration node will need to be reconfigured upon reopening the workflow, as the credentials flow variable was not saved and will therefore not be available to downstream nodes.

Node details

Input ports
  1. Type: org.knime.python3.nodes.PythonBinaryBlobFileStorePortObject
    LLM or Chat Model

    A large language model or chat model.

  2. Type: Table
    Prompt Table

    A table containing a string column with prompts.

Output ports
  1. Type: Table
    Result Table

    A table containing prompts and their respective answer.

Extension

The LLM Prompter node is part of this extension:

  1. Go to item

Related workflows & nodes

  1. Go to item
  2. Go to item
  3. Go to item

KNIME
Open for Innovation

KNIME AG
Talacker 50
8001 Zurich, Switzerland
  • Software
  • Getting started
  • Documentation
  • Courses + Certification
  • Solutions
  • KNIME Hub
  • KNIME Forum
  • Blog
  • Events
  • Partner
  • Developers
  • KNIME Home
  • Careers
  • Contact us
Download KNIME Analytics Platform Read more about KNIME Business Hub
© 2025 KNIME AG. All rights reserved.
  • Trademarks
  • Imprint
  • Privacy
  • Terms & Conditions
  • Data Processing Agreement
  • Credits