繁体   English   中英

如何在DSE中使用Java API访问Hive

[英]How to access Hive using Java API in DSE

我正在使用DSE 4.0.3上的具有2个cassandra,2个solr和1个hadoop节点的5节点集群,我正在尝试通过Java api连接Hive。 以下是我要执行的程序

import java.sql.SQLException;
import java.sql.Connection;
import java.sql.ResultSet;
import java.sql.Statement;
import java.sql.DriverManager;

public class HiveJdbcClient {
    private static String driverName = "org.apache.hadoop.hive.jdbc.HiveDriver";

 /**
 * @param args
 * @throws SQLException
  */
 public static void main(String[] args) throws SQLException {
    try {
      Class.forName(driverName);
    } catch (ClassNotFoundException e) {
    // TODO Auto-generated catch block
    e.printStackTrace();
    System.exit(1);
  }
  Connection con = DriverManager.getConnection("jdbc:hive://localhost:10000/default",  "", "");
  Statement stmt = con.createStatement();
  String tableName = "testHiveDriverTable";
  stmt.executeQuery("drop table " + tableName);
  ResultSet res = stmt.executeQuery("create table " + tableName + " (key int, value string)");
 // show tables
  String sql = "show tables '" + tableName + "'";
  System.out.println("Running: " + sql);
  res = stmt.executeQuery(sql);
  if (res.next()) {
    System.out.println(res.getString(1));
  }
  // describe table
  sql = "describe " + tableName;
  System.out.println("Running: " + sql);
  res = stmt.executeQuery(sql);
  while (res.next()) {
    System.out.println(res.getString(1) + "\t" + res.getString(2));
  }

  // load data into table
  // NOTE: filepath has to be local to the hive server
  // NOTE: /tmp/a.txt is a ctrl-A separated file with two fields per line
  String filepath = "/tmp/a.txt";
  sql = "load data local inpath '" + filepath + "' into table " + tableName;
  System.out.println("Running: " + sql);
  res = stmt.executeQuery(sql);

  // select * query
  sql = "select * from " + tableName;
  System.out.println("Running: " + sql);
  res = stmt.executeQuery(sql);
  while (res.next()) {
    System.out.println(String.valueOf(res.getInt(1)) + "\t" + res.getString(2));
  }

  // regular hive query
  sql = "select count(1) from " + tableName;
  System.out.println("Running: " + sql);
  res = stmt.executeQuery(sql);
  while (res.next()) {
    System.out.println(res.getString(1));
   }
  }
}

但是当我遇到错误时

Connection con = DriverManager.getConnection("jdbc:hive://localhost:10000/default", "", "");


Exception in thread "main" java.lang.NoSuchMethodError: org.apache.thrift.protocol.TProtocol.getScheme()Ljava/lang/Class;
at org.apache.hadoop.hive.service.ThriftHive$execute_args.write(ThriftHive.java:1076)
at org.apache.thrift.TServiceClient.sendBase(TServiceClient.java:63)
at org.apache.hadoop.hive.service.ThriftHive$Client.send_execute(ThriftHive.java:110)
at org.apache.hadoop.hive.service.ThriftHive$Client.execute(ThriftHive.java:102)
at org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:192)
at org.apache.hadoop.hive.jdbc.HiveStatement.execute(HiveStatement.java:132)
at org.apache.hadoop.hive.jdbc.HiveConnection.configureConnection(HiveConnection.java:132)
at org.apache.hadoop.hive.jdbc.HiveConnection.<init>(HiveConnection.java:122)
at org.apache.hadoop.hive.jdbc.HiveDriver.connect(HiveDriver.java:106)
at java.sql.DriverManager.getConnection(Unknown Source)
at java.sql.DriverManager.getConnection(Unknown Source)
at example.create.HiveTable.main(HiveTable.java:22)

注意:我已经在运行程序$ dse hive --service hiveserver时启动了Thrift服务器

而且我正在得到

Connection con = DriverManager.getConnection("jdbc:hive2://localhost:10000/default", "", "");

Exception in thread "main" java.sql.SQLException: Invalid URL: jdbc:hive2://54.243.203.229:10000/default
at org.apache.hadoop.hive.jdbc.HiveConnection.<init>(HiveConnection.java:86)
at org.apache.hadoop.hive.jdbc.HiveDriver.connect(HiveDriver.java:106)
at java.sql.DriverManager.getConnection(Unknown Source)
at java.sql.DriverManager.getConnection(Unknown Source)
at example.create.HiveTable.main(HiveTable.java:22)

注意:我已经在运行程序$ dse hive --service hiveserver2时启动了Thrift服务器。

Hite-site.xml是:

<configuration>
<!--Hive Execution Parameters -->
  <property>
          <name>hive.exec.mode.local.auto</name>
          <value>false</value>
          <description>Let hive determine whether to run in local mode automatically</description>
    </property>
    <property>
          <name>hive.metastore.warehouse.dir</name>
          <value>cfs:///user/hive/warehouse</value>
          <description>location of default database for the warehouse</description>
    </property>
    <property>
           <name>hive.hwi.war.file</name>
           <value>lib/hive-hwi.war</value>
           <description>This sets the path to the HWI war file, relative to${HIVE_HOME}    </description>
     </property>
     <property>
            <name>hive.metastore.rawstore.impl</name>
          <value>com.datastax.bdp.hadoop.hive.metastore.CassandraHiveMetaStore</value>
            <description>Use the Apache Cassandra Hive RawStore implementation</description>
      </property>
      <property>
             <name>hadoop.bin.path</name>
             <value>${dse.bin}/dse hadoop</value>
       </property>
       <!-- Set this to true to enable auto-creation of Cassandra keyspaces as Hive Databases -->
      <property>
            <name>cassandra.autoCreateHiveSchema</name>
           <value>true</value>
       </property>
  </configuration>

任何人都可以提出建议。 我错了或什么都没想到的地方。

WSO2 DSS 3.2.2+中存在libthrift版本冲突。 它们在其部署中包括一个libthrift jar,并且在您将其放入components \\ lib目录中之前先进行加载。 他们具有正确接口的更新的libthrift。 补救步骤:

重新安装。 安装路径在此文档中将称为$ home

http://maven.wso2.org/nexus/content/groups/wso2-public/libthrift/wso2/libthrift/0.8.0.wso2v1/下载libthrift-0.8.0.wso2v1.jar。

将Windows环境变量CLASSPATH设置为$ \\ home \\ repository \\ components \\ lib(这可能不是必需的)

将libthrift 0.8复制到$ home \\ repository \\ components \\ plugins。 取出libthift 0.7罐

编辑$ home \\ repository \\ components \\ features \\ org.wso2.carbon.logging.mgt.server_4.2.1。 对于libthrift行,使其以version =“ 0.8.0.wso2v2”结尾

编辑$ home \\ repository \\ components \\ features \\ org.wso2.carbon.databridge.commons.thrift.server_4.2.0。 在那里也更改libthrift的版本。

我认为您缺少一些HiveServer2 JDBC jar文件和依赖项。 链接https://cwiki.apache.org/confluence/display/Hive/HiveServer2+Clients提供了有关您需要添加到类路径中的jar文件的文档。

暂无
暂无

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM