Can I read messages from Kafka without Spark Streaming? I mean only with Spark Core library for the batch processing purpose. If yes can you please show some examples how to do it. I am using HDP 2.4, Kafka 0.9 and Spark 1.6.
There is a class called KafkaUtils
in spark streaming kafka api.
From this class you can use a method createRDD
, which is basically expecting offsets and it is useful only for non-streaming applications.
Dependency jar:
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-streaming-kafka_2.10</artifactId>
<version>1.6.0</version>
</dependency>
Also, check Kafka Connect, for example you want to read Kafka topic data and populate the data in HDFS, its very simple using Kafka Connect.
http://docs.confluent.io/3.0.0/connect/ http://www.confluent.io/product/connectors/
The technical post webpages of this site follow the CC BY-SA 4.0 protocol. If you need to reprint, please indicate the site URL or the original address.Any question please contact:yoyou2525@163.com.