Apache Flume Tutorial
- Apache Flume - Data FlowFlume is a framework which is used to move log data into HDFS. Generally events and log data are generated by the log servers and these servers have Flume agents running on them. These agents receive the data from the data generators.
- Discuss Apache FlumeFlume is a standard, simple, robust, flexible, and extensible tool for data ingestion from various data producers (webservers) into Hadoop. In this tutorial, we will be using simple and illustrative example to explain the basics of Apache Flume and how to use it in practice.
- Apache Flume - Useful ResourcesThe following resources contain additional information on Apache Flume. Please use them to get more in-depth knowledge on this.
- Apache Flume - Quick GuideApache Flume is a tool/service/data ingestion mechanism for collecting aggregating and transporting large amounts of streaming data such as log files, events (etc...) from various sources to a centralized data store.
- Apache Flume - NetCat SourceThis chapter takes an example to explain how you can generate events and subsequently log them into the console. For this, we are using the NetCat source and the logger sink.
- Apache Flume - Sequence Generator SourceIn the previous chapter, we have seen how to fetch data from twitter source to HDFS. This chapter explains how to fetch data from Sequence generator.
Page 1 / 4Next