Apache Kylin : Analytical Data Warehouse for Big Data

Page tree

Versions Compared


  • This line was added.
  • This line was removed.
  • Formatting was changed.


 With a read/write separation deployment, we can completely isolate both build and query workloads.



  1. Make sure the hadoop version(HDP or CDH) is supported by Kylin.
  2. Check commands like hdfs and hive are all working properly and can access cluster resources.
  3. If the two clusters have enabled the HDFS NameNode HA, please check and make sure their HDFS nameservice names are different. If they are the same, please change one of them to avoid conflict.
  4. Please make sure the network latency between the two clusters is low enough, as there will be a large number of data moved back and forth during model build process.


  1. Install Kylin 4.0 by the following guide on Kylin server.
  2. Prepare the hadoop configuration files of the two cluster and put them into Kylin server.
    1. Open $KYLIN_HOME/conf/kylin.properties
    2. Set  kylin.env.hadoop-conf-dir with the path of the directories of query cluster hadoop configuration files.
    3. Set kylin.engine.submit-hadoop-conf-dir with the path of the directories of build cluster hadoop configuration files.
  3. Put the hive-site.xml of the build cluster into the directory of query cluster hadoop configuration files.

Now read/write separation deployment is configured.


  • $KYLIN_HOME/bin/check-env.sh and $KYLIN_HOME/bin/sample.sh are not available in this deployment mode.

  • In this mode, kylin.engine.spark-conf.spark.yarn.queue in kylin.properties should be configured as the queue of the build cluster.