</>
Tutorials

Apache Flume Tutorial

  • Apache Flume - Data Flow
    Flume is a framework which is used to move log data into HDFS. Generally events and log data are generated by the log servers and these servers have Flume agents running on them. These agents receive the data from the data generators.
  • Discuss Apache Flume
    Flume is a standard, simple, robust, flexible, and extensible tool for data ingestion from various data producers (webservers) into Hadoop. In this tutorial, we will be using simple and illustrative example to explain the basics of Apache Flume and how to use it in practice.
  • Apache Flume - Useful Resources
    The following resources contain additional information on Apache Flume. Please use them to get more in-depth knowledge on this.
  • Apache Flume - Quick Guide
    Apache Flume is a tool/service/data ingestion mechanism for collecting aggregating and transporting large amounts of streaming data such as log files, events (etc...) from various sources to a centralized data store.
  • Apache Flume - NetCat Source
    This chapter takes an example to explain how you can generate events and subsequently log them into the console. For this, we are using the NetCat source and the logger sink.
  • Apache Flume - Sequence Generator Source
    In the previous chapter, we have seen how to fetch data from twitter source to HDFS. This chapter explains how to fetch data from Sequence generator.
Page 1 / 4Next