Hub
Pricing About
WorkflowWorkflow

Generate Text Using a Many-To-One LSTM Network (Training)

Deep learningKerasText generationRNNLSTM
+6
alinebessa profile image
Draft Latest edits on 
Nov 23, 2018 1:28 PM
Drag & drop
Like
Download workflow
Workflow preview
The workflow builds, trains, and saves an RNN with an LSTM layer to generate new fictive fairy tales. The brown nodes define the network structure. The "Pre-Processing and Encoding" part of the workflow reads the fairy tales, index-encodes them, and creates semi-overlapping sequences. The Keras Network Learner node trains the network using the index-encoded fairy tales. Finally, the trained network is converted into a TensorFlow model, and saved to a file.

External resources

  • “Once Upon A Time … “ by LSTM Network Blogpost
Loading deploymentsLoading ad hoc jobs

Used extensions & nodes

Created with KNIME Analytics Platform version 4.6.1
  • Go to item
    KNIME Base nodesTrusted extension

    KNIME AG, Zurich, Switzerland

    Version 4.6.1

    knime
  • Go to item
    KNIME Data GenerationTrusted extension

    KNIME AG, Zurich, Switzerland

    Version 4.6.0

    knime
  • Go to item
    KNIME Deep Learning - Keras IntegrationTrusted extension

    KNIME AG, Zurich, Switzerland

    Version 4.6.0

    knime
  • Go to item
    KNIME Deep Learning - TensorFlow IntegrationTrusted extension

    KNIME AG, Zurich, Switzerland

    Version 4.6.0

    knime
  • Go to item
    KNIME JavasnippetTrusted extension

    KNIME AG, Zurich, Switzerland

    Version 4.6.0

    knime

Legal

By using or downloading the workflow, you agree to our terms and conditions.

KNIME
Open for Innovation

KNIME AG
Talacker 50
8001 Zurich, Switzerland
  • Software
  • Getting started
  • Documentation
  • Courses + Certification
  • Solutions
  • KNIME Hub
  • KNIME Forum
  • Blog
  • Events
  • Partner
  • Developers
  • KNIME Home
  • Careers
  • Contact us
Download KNIME Analytics Platform Read more about KNIME Business Hub
© 2025 KNIME AG. All rights reserved.
  • Trademarks
  • Imprint
  • Privacy
  • Terms & Conditions
  • Data Processing Agreement
  • Credits