Answer Posted / Madhumita Lalwani
Apache Flume is a distributed, reliable, and scalable data collection system. It gathers, aggregates, and moves large amounts of log data from various sources to Hadoop for processing. Flume uses agents, channels, sources, sinks, and channel selectors to accomplish this. Agents are the basic units that process data, while sources extract data from specific sources such as files or web servers. Channels store data temporarily, and sinks write the data into HDFS or other storage systems. Channel selectors control how data is routed between channels.
| Is This Answer Correct ? | 0 Yes | 0 No |
Post New Answer View All Answers