Apache Flume Tutorial PDF Version Quick Guide Resources Job Search Discussion Flume is a standard, simple, robust, flexible, and extensible tool for data ingestion from various data producers (webservers) into Hadoop. In this tutorial, we will be using simple and illustrative example to explain the basics of Apache Flume and how to use it in practice. Audience This tutorial is meant for all those professionals who would like to learn the process of transferring log and streaming data from various webservers to HDFS or HBase using Apache Flume. Prerequisites To make the most of this tutorial, you should have a good understanding of the basics of Hadoop and HDFS commands. Print Page Previous Next Advertisements ”;
Category: apache Flume
Data Transfer in Hadoop
Apache Flume – Data Transfer In Hadoop ”; Previous Next Big Data, as we know, is a collection of large datasets that cannot be processed using traditional computing techniques. Big Data, when analyzed, gives valuable results. Hadoop is an open-source framework that allows to store and process Big Data in a distributed environment across clusters of computers using simple programming models. Streaming / Log Data Generally, most of the data that is to be analyzed will be produced by various data sources like applications servers, social networking sites, cloud servers, and enterprise servers. This data will be in the form of log files and events. Log file − In general, a log file is a file that lists events/actions that occur in an operating system. For example, web servers list every request made to the server in the log files. On harvesting such log data, we can get information about − the application performance and locate various software and hardware failures. the user behavior and derive better business insights. The traditional method of transferring data into the HDFS system is to use the put command. Let us see how to use the put command. HDFS put Command The main challenge in handling the log data is in moving these logs produced by multiple servers to the Hadoop environment. Hadoop File System Shell provides commands to insert data into Hadoop and read from it. You can insert data into Hadoop using the put command as shown below. $ Hadoop fs –put /path of the required file /path in HDFS where to save the file Problem with put Command We can use the put command of Hadoop to transfer data from these sources to HDFS. But, it suffers from the following drawbacks − Using put command, we can transfer only one file at a time while the data generators generate data at a much higher rate. Since the analysis made on older data is less accurate, we need to have a solution to transfer data in real time. If we use put command, the data is needed to be packaged and should be ready for the upload. Since the webservers generate data continuously, it is a very difficult task. What we need here is a solutions that can overcome the drawbacks of put command and transfer the “streaming data” from data generators to centralized stores (especially HDFS) with less delay. Problem with HDFS In HDFS, the file exists as a directory entry and the length of the file will be considered as zero till it is closed. For example, if a source is writing data into HDFS and the network was interrupted in the middle of the operation (without closing the file), then the data written in the file will be lost. Therefore we need a reliable, configurable, and maintainable system to transfer the log data into HDFS. Note − In POSIX file system, whenever we are accessing a file (say performing write operation), other programs can still read this file (at least the saved portion of the file). This is because the file exists on the disc before it is closed. Available Solutions To send streaming data (log files, events etc..,) from various sources to HDFS, we have the following tools available at our disposal − Facebook’s Scribe Scribe is an immensely popular tool that is used to aggregate and stream log data. It is designed to scale to a very large number of nodes and be robust to network and node failures. Apache Kafka Kafka has been developed by Apache Software Foundation. It is an open-source message broker. Using Kafka, we can handle feeds with high-throughput and low-latency. Apache Flume Apache Flume is a tool/service/data ingestion mechanism for collecting aggregating and transporting large amounts of streaming data such as log data, events (etc…) from various webserves to a centralized data store. It is a highly reliable, distributed, and configurable tool that is principally designed to transfer streaming data from various sources to HDFS. In this tutorial, we will discuss in detail how to use Flume with some examples. Print Page Previous Next Advertisements ”;