#Get a copy of sample flume conf from common data hadoop fs -copyToLocal /data/flume/conf # Change the port if needed and location in HDFS nano conf/flume.properties #Launch the flume agent flume-ng agent --conf conf --conf-file conf/flume.properties --name a1 Dflume.root.logger=INFO,console # Open a new console and Connect to the same port that you defined in config nc localhost 44443 # Generate some data Type something in the console #Open a new console and Check in hdfs using hadoop fs -ls flume_webdata hadoop fs -cat 'flume_webdata/FlumeData*'
Let's do a hands-on exercise in Flume. We will read the data from the port and push it to HDFS. Login to the CloudxLab Linux console on two different terminals. On the first terminal, we will run flume-agent and on the second terminal, we will run a server from which we will read the data.
Copy flume configuration from HDFS to the Linux console. It is located at /data/flume/conf on HDFS. Open flume.properties. We've defined configurations for agent a1 in this file. We can define configuration for multiple agents a1, a2, a3 in the same file. While running flume, we can specify the name of the agent which we want to run on that machine.
We have specified source type as netcat. netcat is a good way to quickly create a server which listens on a specified port. Let's change the port number to 44444. While running flume-agent if port 44444 is used by any other user, it will throw up an "Address already in use" error. In that case please change the port to some other number like 44445 or 44446 in flume configuration file.
Sink type is HDFS. Change HDFS sink path to your home directory in HDFS.
Also please note that we are specifying the channel type as memory which will buffer events in memory. Bind the source and sink to the channel.
Let's run the flume-agent on the first terminal. Please note that we are specifying the agent name as a1. Port 44444 is used by another process. Let's change the port to 44445 and run the flume agent again. Port 44445 is also used by another process. Change the port to 44443. Run the agent again and this time it is started successfully
Now let’s produce some data. Go to the second terminal and type nc localhost 44443. Type in some data and see if it gets pushed to HDFS in the sink path.
Login to Hue and check the path in the File browser. Here we can see the data in binary format.
In this video, we have covered Flume introduction and its use case. we have also discussed flume agents and demonstrated steps on how to use Flume on CloudxLab.
Hope you enjoyed the video. Happy learning!
Taking you to the next exercise in seconds...