Hub
Pricing About
NodeNode / Learner

Random Forest Learner (deprecated)

AnalyticsMiningDecision Tree EnsembleClassification

This node has been deprecated and its use is not recommended. Please search for updated nodes instead.

Like

Learns a random forest*, which consists of a chosen number of decision trees. Each of the decision tree models is learned on a different set of rows (records) and a different set of columns (describing attributes), whereby the latter can also be a bit-vector or byte-vector descriptor (e.g. molecular fingerprint). The row sets for each decision tree are created by bootstrapping and have the same size as the original input table. For each node of a decision tree a new set of attributes is determined by taking a random sample of size sqrt(m) where m is the total number of attributes. The output model describes a random forest and is applied in the corresponding predictor node using a simply majority vote.

This node provides a subset of the functionality of the Tree Ensemble Learner corresponding to a random forest. If you need additional functionality please check out the Tree Ensemble Learner .

Experiments have shown the results on different data sets are very similar to the random Forest implementation available in R . Known differences are in the missing value handling (currently not available in this node) and the split creation for nominal attributes (the original random forest classifier uses binary nominal splits; this implementation creates child nodes for each possible split attribute).

The decision tree construction takes place in main memory (all data and all models are kept in memory).


(*) RANDOM FORESTS is a registered trademark of Minitab, LLC and is used with Minitab’s permission.

Node details

Input ports
  1. Type: Table
    Input Data
    The data to learn from. It must contain at least one nominal target column and either a fingerprint (bit-vector/byte-vector) column or another numeric or nominal column.
Output ports
  1. Type: Table
    Out-of-bag error estimates
    The input data with the out-of-bag error estimates, i.e. for each input row the majority vote of all models that did not use the row in the training. If the entire data was used to train the individual models then this output will contain the input data with missing prediction and confidence values. The appended columns are equivalent to the columns appended by the corresponding predictor node. There is one additional column model count , which contains the number of models used for the voting (number of models not using the row throughout the learning.)
  2. Type: Table
    Attribute Statistics
    A statistics table on the attributes used in the different tree learners. Each row represents one training attribute with these statistics: #splits (level x) as the number of models, which use the attribute as split on level x (with level 0 as root split); #candidates (level x) is the number of times an attribute was in the attribute sample for level x (in a random forest setup these samples differ from node to node). If no attribute sampling is used #candidates is the number of models. Note, these numbers are uncorrected, i.e. if an attribute is selected on level 0 but is also in the candidate set of level 1 (but will not be split on level 1 because it has been split one level up), the #candidate number will still count the attribute as candidate.
  3. Type: Tree Ensembles
    Random Forest Model
    The trained model.

Extension

The Random Forest Learner (deprecated) node is part of this extension:

  1. Go to item

Related workflows & nodes

  1. Go to item
  2. Go to item
  3. Go to item

KNIME
Open for Innovation

KNIME AG
Talacker 50
8001 Zurich, Switzerland
  • Software
  • Getting started
  • Documentation
  • Courses + Certification
  • Solutions
  • KNIME Hub
  • KNIME Forum
  • Blog
  • Events
  • Partner
  • Developers
  • KNIME Home
  • Careers
  • Contact us
Download KNIME Analytics Platform Read more about KNIME Business Hub
© 2025 KNIME AG. All rights reserved.
  • Trademarks
  • Imprint
  • Privacy
  • Terms & Conditions
  • Data Processing Agreement
  • Credits