talend bigData components

The Talend Big Data Components training course provides a development environment that allows for interaction with big data sources and targets without the need to learn and write complicated code. Running 100% natively on Hadoop, Big Data provides massive scalability.

This course is an introduction to the components that ships several products for interacting with Big Data systems.

  • Attend training on Talend Studio Basics or possess equivalent knowledge

  • Anyone who wants to create Jobs using Talend Big Data Components

By attending Talend Big Data Components workshop, Participants will learn to:

  • Connect to a Hadoop cluster from a Talend Job
  • Store a raw Web log file to HDFS
  • Write text data files to HDFS
  • Read text files from HDFS
  • Read data from a SQL database and write it to HDFS
  • List a folder’s contents and operate on each file separately
  • Move, copy, append, delete, and rename HDFS files
  • Read selected file attributes from HDFS files
  • Conditionally operate on HDFS files
  • Connect to a Hive database from a Talend Job
  • Use a Talend Job to load data from HDFS into a Hive table
  • Use a Talend Job to read data from a Hive table and use it in a Job
  • Execute Hive commands iteratively in a Talend Job, based on variable inputs
  • Develop and run Pig Jobs using Talend components
  • Sort, join, and aggregate data using Pig components
  • Filter data in multiple ways using Pig components
  • Replicate Pig data streams
  • Run Talend Jobs with the Apache Oozie Job Manager