简体   繁体   English

无法将 Spark 制作为 Hive 执行引擎

[英]Not able to make Spark as Hive execution engine

I am using Hive 3.1.2 and Spark 2.4.5(Hadoop 3 on Ubuntu 18.04) Getting below error while running queries我正在使用 Hive 3.1.2 和 Spark 2.4.5(Ubuntu 18.04 上的 Hadoop 3)运行查询时出现以下错误

0: jdbc:hive2://localhost:10000> select count(1) from retail_db.orders; 0: jdbc:hive2://localhost:10000> select count(1) 来自retail_db.orders; Error: Error while processing statement: FAILED: Execution Error, return code 30041 from org.apache.hadoop.hive.ql.exec.spark.SparkTask.错误:处理语句时出错:FAILED:执行错误,从org.apache.hadoop.hive.park.ql.execs.execs.execs.返回代码30041。 Failed to create Spark client for Spark session 4aa97085-8a23-4ad8-a8c6-b701d41d7110 (state=42000,code=30041)无法为 Spark session 4aa97085-8a23-4ad8-a8c6-b701d41d7110 创建 Spark 客户端(状态=42000,代码=30041)

Log: 2020-06-19 12:18:28,704 INFO [HiveServer2-Background-Pool: Thread-72] ql.Driver: Concurrency mode is disabled, not creating a lock manager 2020-06-19 12:18:28,731 ERROR [HiveServer2-Background-Pool: Thread-72] operation.Operation: Error running hive query: org.apache.hive.service.cli.HiveSQLException: Error while processing statement: FAILED: Execution Error, return code 30041 from org.apache.hadoop.hive.ql.exec.spark.SparkTask.日志:2020-06-19 12:18:28,704 INFO [HiveServer2-Background-Pool: Thread-72] ql.Driver:并发模式已禁用,未创建锁管理器 2020-06-19 12:18:28,731 错误 [ HiveServer2-Background-Pool: Thread-72] operation.Operation: Error running hive query: org.apache.hive.service.cli.HiveSQLException: Error while processing statement: FAILED: Execution Error, return code 30041 from org.apache.hadoop .hive.ql.exec.spark.SparkTask。 Failed to create Spark client for Spark session 4aa97085-8a23-4ad8-a8c6-b701d41d7110 at org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:335) at org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:226) at org.apache.hive.service.cli.operation.SQLOperation.access$700(SQLOperation.java:87) at org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork$1.run(SQLOperation.java:316) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at Failed to create Spark client for Spark session 4aa97085-8a23-4ad8-a8c6-b701d41d7110 at org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:335) at org.apache.hive.service.cli .operation.SQLOperation.runQuery(SQLOperation.java:226) at org.apache.hive.service.cli.operation.SQLOperation.access$700(SQLOperation.java:87) at org.apache.hive.service.cli.operation. SQLOperation$BackgroundWork$1.run(SQLOperation.java:316) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729) at org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork.run(SQLOperation.java:329) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Cau org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729) at org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork.run(SQLOperation.java:329) at java.util.concurrent. Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util .concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 在 java.lang.Thread.run(Thread.Z93F725A07423C31C889FZ48B3) sed by: org.apache.hadoop.hive.ql.metadata.HiveException: Failed to create Spark client for Spark session 4aa97085-8a23-4ad8-a8c6-b701d41d7110 at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.getHiveException(SparkSessionImpl.java:221) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:92) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionManagerImpl.getSession(SparkSessionManagerImpl.Z93F725A07423FE1C889F448B33D21F46 sed by: org.apache.hadoop.hive.ql.metadata.HiveException: Failed to create Spark client for Spark session 4aa97085-8a23-4ad8-a8c6-b701d41d7110 at org.apache.hadoop.hive.ql.exec.spark.session .SparkSessionImpl.getHiveException(SparkSessionImpl.java:221) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:92) at org.apache.hadoop.hive.ql.exec .spark.session.SparkSessionManagerImpl.getSession(SparkSessionManagerImpl.Z93F725A07423FE1C889F448B33D21F46 Z:115) at org.apache.hadoop.hive.ql.exec.spark.SparkUtilities.getSparkSession(SparkUtilities.java:136) at org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:115) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664) at org.ZB6EFD606D118D0F62066E Z:115) at org.apache.hadoop.hive.ql.exec.spark.SparkUtilities.getSparkSession(SparkUtilities.java:136) at org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask. java:115) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664) at org.ZB6EFD606D118D0F62066E 31419FF04CCZ.hadoop.hive.ql.Driver.execute(Driver.java:2335) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703) at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157) at org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOpera 31419FF04CCZ.hadoop.hive.ql.Driver.execute(Driver.java:2335) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011) at org.apache.hadoop.hive.ql. Driver.run(Driver.java:1709) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703) at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver. java:157) 在 org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation) tion.java:224)... 11 more Caused by: java.lang.NoClassDefFoundError: org/apache/spark/SparkConf at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.generateSparkConf(HiveSparkClientFactory.java:263) at org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient.(RemoteHiveSparkClient.java:98) at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.createHiveSparkClient(HiveSparkClientFactory.java:76) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessi tion.java:224)... 11 more Caused by: java.lang.NoClassDefFoundError: org/apache/spark/SparkConf at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.generateSparkConf(HiveSparkClientFactory.java: 263) at org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient.(RemoteHiveSparkClient.java:98) at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.createHiveSparkClient(HiveSparkClientFactory.java:76 ) 在 org.apache.hadoop.hive.ql.exec.spark.Z21D6F4Z40CFB5119AEE5SE4 onImpl.open(SparkSessionImpl.java:87) onImpl.open(SparkSessionImpl.java:87)

According to this error:根据这个错误:
Caused by: java.lang.NoClassDefFoundError: org/apache/spark/SparkConf

Hive cannot find Spark classes. Hive 找不到 Spark 类。
If you're using Cloudera, you can add them with this command:如果您使用的是 Cloudera,则可以使用以下命令添加它们:

cd /opt/cloudera/parcels/CDH/lib;
cp  spark/jars/*scala-lib* hive/lib/;
cp  spark/jars/*spark-core* hive/lib/;
cp  spark/jars/*spark-netw*  hive/lib/;
cp  spark/jars/chill-java*.jar  hive/lib/;
cp  spark/jars/jackson-module-*  hive/lib/;
cp  spark/jars/jersey-server-*.jar  hive/lib/;
cp  spark/jars/json4s-ast_*.jar  hive/lib/;
cp  spark/jars/kryo-shaded-*.jar  hive/lib/;
cp  spark/jars/minlog-*.jar  hive/lib/;
cp  spark/jars/scala-xml_*.jar  hive/lib/;
cp  spark/jars/spark-launcher_*.jar  hive/lib/;
cp  spark/jars/spark-network-shuffle_*.jar  hive/lib/;
cp  spark/jars/spark-launcher_*.jar  hive/lib/;
cp  spark/jars/spark-unsafe_*.jar  hive/lib/;
cp  spark/jars/xbean-asm7-shaded-*.jar  hive/lib/;

You can find more info here:您可以在此处找到更多信息:
https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started
in Configuring Hive section.配置 Hive部分。

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM