Here the Flume acts as Consumer and stores in HDFS.īin/kafka-server-start.sh config/server.propertiesģ. And integration of both is needed to stream the data in Kafka topic with high speed to different Sinks. In Kafka, the Flume is integrated for streaming a high volume of data logs from Source to Destination for Storing data in HDFS.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |