This 3-day big data integration course teaches students how to use Informatica to design and develop data integration mappings that run on Hadoop. The course includes Hadoop ingestion, offloading and execution of data integration mappings natively in Hadoop. This Informatica big data training also shows how to optimize data warehouse processing in Hadoop environments.
The labs in this Informatica big data training take you from using PowerCenter to Developer tool to populate Hadoop data stores, to running those mappings in Hadoop. Complex data files are also defined, and parsed in Hadoop.
After successfully completing this training, students should be able to:
- Describe data warehouse optimization in Hadoop environments
- Describe license capabilities included in Informatica Big Data Editions
- Offload data on Hadoop using Informatica PowerExchange for Hadoop
- Offload processing (workloads) into Hadoop using Informatica Developer Tool
- Process file types in Hadoop that cannot be processed in traditional DataWarehouse setting. Process complex binary files E.G Web Logs and Call Detail Records
- Describe optimum mapping design methods when executing Informatica mappings in Hadoop
- Read and Write MongoDB both in Relational and JSON forms
- Data integration / ETL developers
- Hadoop developers who would like to learn data integration / ETL
- Data integration / ETL project managers
- Team Leads