Creates a new Spark context via Apache Livy.
This node requires access to a remote file system such as HDFS/webHDFs/httpFS or S3/Blob Store/Cloud Store in order to exchange temporary files between KNIME and the Spark context (running on the cluster).
Note: Executing this node always creates a new Spark context. Resetting the node or closing the KNIME workflow will destroy the Spark context. Spark contexts created by this node cannot be shared between KNIME workflows.