[英]How to store JavaDStream<String> to Hive table in spark using Java?
I am receiving the messages from Kafka server into JavaDStream object. 我正在从Kafka服务器接收消息到JavaDStream对象中。 Now i want to store the messages into hive table. 现在我想将消息存储到配置单元表中。 How can i do that? 我怎样才能做到这一点?
Code snippet for receiving messages from Kafka stream: 用于从Kafka流接收消息的代码段:
JavaPairReceiverInputDStream<String, String> kafkaStream = KafkaUtils.createStream(jssc, zookeperIp, consumerGroup,
topic);
JavaDStream<String> messageData = kafkaStream.map(new Function<Tuple2<String, String>, String>() {
@Override
public String call(Tuple2<String, String> message) throws Exception {
// TODO Auto-generated method stub
return message._2;
}
});
Schema for table in Hive: Hive中表的架构:
create external table sample(value STRING);
I am using spark 2.1.0 version. 我正在使用spark 2.1.0版本。
please take a look at these official programming guides. 请查看这些官方编程指南。
You can easily use DataFrames and SQL operations on streaming data 您可以轻松地对流数据使用DataFrames和SQL操作
and Spark SQL also supports reading and writing data stored in Apache Hive 并且Spark SQL还支持读写存储在Apache Hive中的数据
声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.