Build Cube with Flink
Kylin v3.1 introduces the Flink cube engine, it uses Apache Flink to replace MapReduce in the build cube step; You can check KYLIN-3758. The current document uses the sample cube to demo how to try the new engine.
Preparation
To finish this tutorial, you need a Hadoop environment which has Kylin v3.1.0 or above installed. Here we will use Cloudera CDH 5.7 environment, the Hadoop components as well as Hive/HBase has already been started.
Install Kylin v3.1.0 or above
Download the Kylin binary for CDH 5.7+ from Kylin's download page, and then uncompress the tar ball into /usr/local/ folder:
wget http://www-us.apache.org/dist/kylin/apache-kylin-3.1.0/apache-kylin-3.1.0-bin-cdh57.tar.gz -P /tmp
tar -zxvf /tmp/apache-kylin-3.1.0-bin-cdh57.tar.gz -C /usr/local/
export KYLIN_HOME=/usr/local/apache-kylin-3.1.0-bin-cdh57
Prepare "kylin.env.hadoop-conf-dir"
To run Flink on Yarn, need specify HADOOP_CONF_DIR environment variable, which is the directory that contains the (client side) configuration files for Hadoop. In many Hadoop distributions the directory is "/etc/hadoop/conf"; Kylin can automatically detect this folder from Hadoop configuration, so by default you don't need to set this property. If your configuration files are not in default folder, please set this property explicitly.
Check Flink configuration
Point FLINK_HOME to your flink installation path:
export FLINK_HOME=/path/to/flink