这是我参与11月更文挑战的第26天,活动详情查看:2021最后一次更文挑战
一、依赖环境
需要软件如下:
| 软件 | 版本 |
|---|---|
Hive | 2.3.7 |
Hadoop | 2.9.2 |
HBase | 1.3.1 |
Zookeeper | 3.4.14 |
Kafka | 1.0.2 |
Spark | 2.4.5 |
集群规划:
注意: 要求 hbase 的 hbase.zookeeper.quorum 值必须只能是 host1、host2、...。不允许出现 host:2181。
hbase-site.xml 文件,如下:
<configuration>
<!-- 指定hbase在HDFS上存储的路径 -->
<property>
<name>hbase.rootdir</name>
<value>hdfs://linux121:9000/hbase</value>
</property>
<!-- 指定hbase是分布式的 -->
<property>
<name>hbase.cluster.distributed</name>
<value>true</value>
</property>
<!-- 指定zk的地址,多个用“,”分割 -->
<property>
<name>hbase.zookeeper.quorum</name>
<value>linux121,linux122,linux123</value>
</property>
</configuration>
二、Kylin 安装配置
- 解压缩软件 (
apache-kylin-3.1.1-bin-hbase1x.tar.gz), 并移动到/opt/lagou/servers目录下
# linux122 机子下
cd /opt/lagou/software
tar zxvf apache-kylin-3.1.1-bin-hbase1x.tar.gz
mv apache-kylin-3.1.1-bin-hbase1x/ ../servers/kylin-3.1.1
cd ../servers/kylin-3.1.1
- 添加环境变量, 并使之生效
vi /etc/profile
# 增加以下内容
export KYLIN_HOME=/opt/lagou/servers/kylin-3.1.1
export PATH=$PATH:$KYLIN_HOME/bin
# 使生效
source /etc/profile
- 增加
kylin依赖组件的配置
cd $KYLIN_HOME/conf
ln -s $HADOOP_HOME/etc/hadoop/hdfs-site.xml hdfs-site.xml
ln -s $HADOOP_HOME/etc/hadoop/core-site.xml core-site.xml
ln -s $HBASE_HOME/conf/hbase-site.xml hbase-site.xml
ln -s $HIVE_HOME/conf/hive-site.xml hive-site.xml
ln -s $SPARK_HOME/conf/spark-defaults.conf spark-defaults.conf
- 修改
kylin.sh
cd $KYLIN_HOME/bin
vim kylin.sh
# 在 kylin.sh 文件头部添加
# 在 set verbose=true to print more logs during start up 后面添加即可
export HADOOP_HOME=/opt/lagou/servers/hadoop-2.9.2
export HIVE_HOME=/opt/lagou/servers/hive-2.3.7
export HBASE_HOME=/opt/lagou/servers/hbase-1.3.1
export SPARK_HOME=/opt/lagou/servers/spark-2.4.5
- 检查依赖
$KYLIN_HOME/bin/check-env.sh
# 实操如下:
[root@linux122 bin]# ./check-*
Retrieving hadoop conf dir...
...................................................[PASS]
KYLIN_HOME is set to /opt/lagou/servers/kylin-3.1.1
Checking HBase
...................................................[PASS]
Checking hive
...................................................[PASS]
Checking hadoop shell
...................................................[PASS]
Checking hdfs working dir
mkdir: End of File Exception between local host is: "linux122/172.16.64.122"; destination host is: "linux121":9000; : java.io.EOFException; For more details see: http://wiki.apache.org/hadoop/EOFException
...................................................[FAIL]
Failed to create /kylin. Please make sure the user has right to access /kylin
# 最后一个 FAIL 是因为 hadoop 端口 9000 与 clickhouse 相冲突了。需要把 clickhouse 停止
# 对应停止 ck 的命令如下:systemctl stop clickhouse-server
# kylin 中执行的:
# 对应执行的命令:hadoop ${hadoop_conf_param} fs -mkdir -p $WORKING_DIR
# 对应执行的命令:hadoop --config /opt/lagou/servers/hadoop-2.9.2/etc/hadoop fs -mkdir -p /kylin
错误处理:
/opt/lagou/servers/kylin-3.1.1/bin/check-port-availability.sh: line 30: netstat: command not found
# 解决:
yum install net-tools
三、启动集群
- 启动
zookeeper(linux121)
[root@linux121 shells]# pwd
/root/shells
# 启动
[root@linux121 shells]# zk.sh start
start zookeeper server...
ZooKeeper JMX enabled by default
Using config: /opt/lagou/servers/zookeeper-3.4.14/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED
ZooKeeper JMX enabled by default
Using config: /opt/lagou/servers/zookeeper-3.4.14/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED
ZooKeeper JMX enabled by default
Using config: /opt/lagou/servers/zookeeper-3.4.14/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED
# 关闭
[root@linux121 shells]# zk.sh stop
- 启动
HDFS(linux121)
[root@linux121 ~]# start-dfs.sh
Starting namenodes on [linux121]
linux121: starting namenode, logging to /opt/lagou/servers/hadoop-2.9.2/logs/hadoop-root-namenode-linux121.out
linux121: starting datanode, logging to /opt/lagou/servers/hadoop-2.9.2/logs/hadoop-root-datanode-linux121.out
linux122: starting datanode, logging to /opt/lagou/servers/hadoop-2.9.2/logs/hadoop-root-datanode-linux122.out
linux123: starting datanode, logging to /opt/lagou/servers/hadoop-2.9.2/logs/hadoop-root-datanode-linux123.out
Starting secondary namenodes [linux123]
linux123: starting secondarynamenode, logging to /opt/lagou/servers/hadoop-2.9.2/logs/hadoop-root-secondarynamenode-linux123.out
- 启动
YARN(linux123)
[root@linux123 ~]# start-yarn.sh
starting yarn daemons
starting resourcemanager, logging to /opt/lagou/servers/hadoop-2.9.2/logs/yarn-root-resourcemanager-linux123.out
linux122: starting nodemanager, logging to /opt/lagou/servers/hadoop-2.9.2/logs/yarn-root-nodemanager-linux122.out
linux121: starting nodemanager, logging to /opt/lagou/servers/hadoop-2.9.2/logs/yarn-root-nodemanager-linux121.out
linux123: starting nodemanager, logging to /opt/lagou/servers/hadoop-2.9.2/logs/yarn-root-nodemanager-linux123.out
- 启动
HBase(linux121)
[root@linux121 ~]# start-hbase.sh
starting master, logging to /opt/lagou/servers/hbase-1.3.1/logs/hbase-root-master-linux121.out
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option PermSize=128m; support was removed in 8.0
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=128m; support was removed in 8.0
linux121: starting regionserver, logging to /opt/lagou/servers/hbase-1.3.1/bin/../logs/hbase-root-regionserver-linux121.out
linux122: starting regionserver, logging to /opt/lagou/servers/hbase-1.3.1/bin/../logs/hbase-root-regionserver-linux122.out
linux123: starting regionserver, logging to /opt/lagou/servers/hbase-1.3.1/bin/../logs/hbase-root-regionserver-linux123.out
linux121: Java HotSpot(TM) 64-Bit Server VM warning: ignoring option PermSize=128m; support was removed in 8.0
linux121: Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=128m; support was removed in 8.0
linux122: Java HotSpot(TM) 64-Bit Server VM warning: ignoring option PermSize=128m; support was removed in 8.0
linux122: Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=128m; support was removed in 8.0
linux123: Java HotSpot(TM) 64-Bit Server VM warning: ignoring option PermSize=128m; support was removed in 8.0
linux123: Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=128m; support was removed in 8.0
linux122: starting master, logging to /opt/lagou/servers/hbase-1.3.1/bin/../logs/hbase-root-master-linux122.out
linux122: Java HotSpot(TM) 64-Bit Server VM warning: ignoring option PermSize=128m; support was removed in 8.0
linux122: Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=128m; support was removed in 8.0
- 启动
metastore(linux123)
$ nohup hive --service metastore &
# 实操如下:
[root@linux123 ~]# nohup hive --service metastore &
[1] 2722
[root@linux123 ~]# nohup: ignoring input and appending output to ‘nohup.out’
- 启动
Yarn history server(linux121)
$ mr-jobhistory-daemon.sh start historyserver
# 实操如下:
[root@linux121 ~]# mr-jobhistory-daemon.sh start historyserver
starting historyserver, logging to /opt/lagou/servers/hadoop-2.9.2/logs/mapred-root-historyserver-linux121.out
# 查看是否存在:
$ jps
- 启动
kylin
$ kylin.sh start
# 实操如下:
[root@linux122 bin]# pwd
/opt/lagou/servers/kylin-3.1.1/bin
[root@linux122 bin]# kylin.sh start
Retrieving hadoop conf dir...
...................................................[PASS]
KYLIN_HOME is set to /opt/lagou/servers/kylin-3.1.1
Checking HBase
...................................................[PASS]
Checking hive
...................................................[PASS]
Checking hadoop shell
...................................................[PASS]
Checking hdfs working dir --config /opt/lagou/servers/hadoop-2.9.2/etc/hadoop , /kylin
...................................................[PASS]
Retrieving Spark dependency...
Optional dependency spark not found, if you need this; set SPARK_HOME, or run bin/download-spark.sh
...................................................[PASS]
Retrieving Flink dependency...
Optional dependency flink not found, if you need this; set FLINK_HOME, or run bin/download-flink.sh
...................................................[PASS]
Retrieving kafka dependency...
Couldn't find kafka home. If you want to enable streaming processing, Please set KAFKA_HOME to the path which contains kafka dependencies.
...................................................[PASS]
/opt/lagou/servers/kylin-3.1.1/bin/check-port-availability.sh: line 30: netstat: command not found
Checking environment finished successfully. To check again, run 'bin/check-env.sh' manually.
Using cached dependency...
skip spark_dependency
- 登录
Kylin Web界面
| URL | http://linux122:7070/kylin |
|---|---|
| 默认用户名 | ADMIN |
| 默认密码 | KYLIN |
备注: 用户名和密码都必须是大写
如下图所示: