环境说明:
主机名:cmcc01为例
操作系统:centos7
安装部署软件 | 版本 | 部署方式 |
centos | 7 | |
zookeeper | zookeeper-3.4.10 | 伪分布式 |
hadoop | hadoop-3.1.3 | 伪分布式 |
hive | hive-3.1.3-bin | 伪分布式 |
clickhouse | 21.11.10.1-2 | 单节点多实例 |
dolphinscheduler | 3.0.0 | 单节点 |
kettle | pdi-ce-9.3.0.0 | 单节点 |
sqoop | sqoop-1.4.7 | 单节点 |
seatunnel | seatunnel-incubating-2.1.2 | 单节点 |
spark | spark-2.4.8 | 单节点 |
1.下载Hive:
官网:https://hive.apache.org/downloads.html
解压:
# 解压
tar -zxvf /opt/package/apache-hive-3.1.3-bin.tar.gz -C /opt/software/
# 修改目录名称
mv /opt/software/apache-hive-3.1.3-bin /opt/software/hive-3.1.3-bin
2.配置环境变量
vim ~/.bash_profile
# hive
export HIVE_HOME=/opt/software/hive-3.1.3-bin
export HIVE_CONF_DIR=${HIVE_HOME}/conf
export HADOOP_CLASSPATH=$HADOOP_CLASSPATH:$HIVE_HOME/lib/*
export PATH=$PATH:${HIVE_HOME}/bin
# 使用环境变量生效
source ~/.bash_profile
3.修改配置信息
cd /opt/software/hive-3.1.3-bin/conf
1.修改 hive-env.sh
# 复制配置文件
cp hive-env.sh.template hive-env.sh
# 修改文件
vim hive-env.sh
# 添加以下内容
HADOOP_HOME=/opt/software/hadoop-3.1.3
export HIVE_CONF_DIR=/opt/software/hive-3.1.3-bin/conf
2.修改 hive-site.xml,将mysql作为hive的元数据库
# 复制配置文件
cp hive-default.xml.template hive-site.xml
vim hive-site.xml
将<configuration>标签中所有内容替换为以下内容,并修改自己配置
注意:替换配置信息为自己的配置信息
<configuration>
<!-- WARNING!!! This file is auto generated for documentation purposes ONLY! -->
<!-- WARNING!!! Any changes you make to this file will be ignored by Hive. -->
<!-- WARNING!!! You must make your changes in hive-site.xml instead. -->
<!-- Hive Execution Parameters -->
<property>
<name>javax.jdo.option.ConnectionURL</name>
<!-- cmcc01 修改为你自己的Mysql ip地址或主机名 -->
<value>jdbc:mysql://cmcc01/hive?createDatabaseIfNotExist=true</value>
</property>
<property>
<name>javax.jdo.option.ConnectionDriverName</name>
<value>com.mysql.cj.jdbc.Driver</value>
</property>
<property>
<name>javax.jdo.option.ConnectionUserName</name>
<!-- root 修改为你自己的Mysql 用户名 -->
<value>root</value>
</property>
<property>
<name>javax.jdo.option.ConnectionPassword</name>
<!-- 123qwe 修改为你自己的Mysql 密码 -->
<value>123qwe</value>
</property>
<property>
<name>hive.cli.print.header</name>
<value>true</value>
<description>Whether to print the names of the columns in query output.</description>
</property>
<property>
<name>hive.cli.print.current.db</name>
<value>true</value>
<description>Whether to include the current database in the Hive prompt.</description>
</property>
<property>
<name>hive.server2.thrift.bind.host</name>
<!-- cmcc01 修改为你自己的主机名 -->
<value>cmcc01</value>
</property>
<property>
<name>hive.server2.thrift.port</name>
<value>10000</value>
</property>
</configuration>
3.修改 beeline-log4j2.properties hive-log4j2.properties
cd /opt/software/hive-3.1.3-bin/conf
# 复制日志文件
cp hive-log4j2.properties.template hive-log4j2.properties
cp beeline-log4j2.properties.template beeline-log4j2.properties
# 修改hive-log4j2.properties,将日志级别调整到ERROR
vim hive-log4j2.properties
property.hive.log.level = ERROR
rootLogger.level=ERROR
# 修改beeline-log4j2.properties,将日志级别调整到ERROR
vim beeline-log4j2.properties
logger.HiveConnection.level = ERROR
rootLogger.level = WARN
3.拷贝jar包
# 复制mysql jar包到hive lib
cp /opt/package/mysql-connector-java-8.0.20.jar /opt/software/hive-3.1.3-bin/lib
# 删除guava jar包
rm -rf /opt/software/hive-3.1.3-bin/lib/guava-*.jar
# 复制hadoop最新guava包到hive lib
cp /opt/software/hadoop-3.1.3/share/hadoop/common/lib/guava-27.0-jre.jar /opt/software/hive-3.1.3-bin/lib
缺少 guava jar包,在初始化时将报错如下:
4.初始化元数据库
schematool -initSchema -dbType mysql
5.启动hiveserver2
1.创建hiveserver2启动脚本
vim /opt/software/start_hiveserver2.sh
添加以下内容
#!bin/bash
nohup ${HIVE_HOME}/bin/hiveserver2 > ${HIVE_HOME}/logs/hiveserver2.log 2>&1 &
2.创建hiveserver2停止脚本
vim /opt/software/stop_hiveserver2.sh
添加以下内容
# !bin/bash
ps -ef| grep org.apache.hive.service.server.HiveServer2 | grep -v grep | awk '{print $2}' | xargs kill
3.启动hiveserver2
sh /opt/software/start_hiveserver2.sh
4.登录hive
# 正常方式
hive
# 使用hiveserver2(在启动服务后需要一两分钟后才能登录)
beeline -u jdbc:hive2://cmcc01:10000/default -n root
本文含有隐藏内容,请 开通VIP 后查看