Apache Flume: Distributed Log Collection for Hadoop
If you are a Hadoop programmer who wants to learn about Flume to be able to move datasets into Hadoop in a timely and replicable manner, then this book is ideal for you. No prior knowledge about Apach
English
Created by Last updated Fri, 02-Feb-2024
+ View more
Course overview
If you are a Hadoop programmer who wants to learn about Flume to be able to move datasets into Hadoop in a timely and replicable manner, then this book is ideal for you. No prior knowledge about Apache Flume is necessary, but a basic knowledge of Hadoop and the Hadoop File System (HDFS) is assumed.
What will i learn?
Understand the Flume architecture, and also how to download and install open source Flume from Apache
Follow along a detailed example of transporting weblogs in Near Real Time (NRT) to Kibana/Elasticsearch and archival in HDFS
Learn tips and tricks for transporting logs and data in your production environment
Understand and configure the Hadoop File System (HDFS) Sink
Use a morphlinebacked Sink to feed data into Solr
Create redundant data flows using sink groups
Configure and use various sources to ingest data
Inspect data records and move them between multiple destinations based on payload content
Transform data enroute to Hadoop and monitor your data flows
Requirements
Curriculum for this course
1 Lessons
5 hrs 56 mins
Apache Flume: Distributed Log Collection for Hadoop
1 Lessons
05:56:00 Hours
Apache Flume: Distributed Log Collection for Hadoop
Write a public review