Data Engineering Integration: Administration

Training Summary

Establish a live DEI environment by performing numerous administrative tasks such as Hadoop integration, Databricks integration, security mechanism setup, Data Engineering recovery, monitoring, & performance tuning. Understand to integrate the Informatica domain with the Hadoop & Databricks ecosystem leveraging Hadoop’s lightning processing capability, & Databricks’ analytics cloud platform technology to churn substantial data sets. Applicable for users of software version 10.5. 

Audience profile

Administrators

Learning Objectives

After completing this training, professionals must be able to:

  • Describing DEI Architecture
  • Listing DEI Components
  • Listing the steps to enable SAML on the Domain
  • Creating Cluster Configuration Object for Hadoop integration
  • Setting up Informatica Security that includes different Authentication & Authorization mechanisms
  • Tune the performance of the system
  • Monitor, view, & troubleshoot DEI logs

Content Outline

  • Data Engineering & the role of DEI in the Big Data ecosystem
  • DEI Components
  • DEI architecture
  • Roles & responsibilities of Informatica DEI Administrator
  • DEI engines: Blaze, Spark, & Databricks
  • DEI 10.5 features
  • SAML Overview
  • SAML authentication in a domain
  • Steps to enable SAML on an existing Informatica domain
  • Cluster Integration Overview
  • Data Engineering Integration Component Architecture
  • Prerequisites for Hadoop integration
  • HDP integration tasks
  • Create a Cluster Configuration
  • Integration with Hadoop

Lab:

  • Create Cluster Configuration Object
  • Explore Cluster Configuration Views
  • Cluster Configuration Privileges & Permissions
  • DEI security
  • Security aspects
  • Authentication overview
  • Authorization Overview
  • Kerberos Authentication
  • Ranger Authorization
  • Pre-steps to run mapping in a Kerberos-Enabled Hadoop Environment
  • Run mapping on a cluster with Kerberos authentication & Ranger authorization

Lab: 

  • Executing Pre-steps for Running Mappings in a Kerberos-Enabled Hadoop Environment
  • Run Mapping in a Kerberos-Enabled Hadoop Environment
  • Operating System profiles for Data Integration Service
  • Operating System profile components
  • Configure system permissions for the Operating System profile users
  • Enable the Data Integration Service to use Operating System profiles
  • Execute a mapping using OS profiles

Lab: 

  • Execute a mapping using OS profiles
  • Authorization
  • HDFS permissions
  • Fine-Grained Authorization

Lab: 

  • Access Directories with HDFS Permissions
  • Run a Mapping with HDFS Permissions
  • Restrict Ranger Permissions for Hive Tables & Columns
  • Run a Mapping with Fine-Grained Authorization
  • DIS processing overview
  • DIS Queuing
  • Execution Pools
  • Data Engineering recovery
  • Monitor recovered jobs

Lab:

  • Recover DIS & execute a Mapping using Data Engineering Recovery
  • DEI Deployment types
  • Sizing recommendations
  • Hadoop cluster Hardware tuning
  • Tune Spark performance
  • infacmd autotune command

Lab: 

  • Tune DIS & MRS using the informed Autotune command
  • Hadoop Environment Logs
  • Spark Engine Monitoring
  • Blaze Engine Monitoring
  • Cloud File Management Utility
  • Log Aggregation
  • Log Packer
  • File Watcher
  • Customer pain points & solutions
  • Databricks Overview
  • Steps to configure Databricks
  • Databricks clusters
  • Notebooks, Jobs, & Data
  • Delta Lakes
  • Sequence generator for Databricks
  • Databricks warm pool
  • Databricks Integration
  • Components of the Informatica & the Databricks environments
  • A run-time process on the Databricks Spark Engine
  • Databricks Integration Task Flow
  • Prerequisites for Databricks integration

FAQs

A: Informatica Data Engineering Integration provides optimized run-time processing & simplified monitoring across multiple engines for faster, more flexible, & repeatable development & processing.

A: Hadoop & big data are almost synonymous terms. With the rising big data, Hadoop, a framework specialized in big data operations also become popular. For professionals to use big data & help businesses to make decisions, the framework can be used.

A: Radiant has highly intensive selection criteria for Technology Trainers & Consultants who deliver training programs. Our trainers & consultants undergo rigorous technical & behavioral interviews & assessment processes before they are onboarded in the company.

Our Technology experts/trainers & consultants carry deep-dive knowledge in the technical subject & are certified by the OEM.

Our training programs are practically oriented with 70% – 80% hands-on training technology tools. Our training program focuses on one-on-one interaction with each participant, the latest content in the curriculum, real-time projects & case studies during the training program.

Our faculty will provide you the knowledge of each training from a fundamental level in an easy way & you are free to ask your doubts any time from your respective faculty.

Our trainers have the patience & ability to explain difficult concepts in a simplistic way with depth & width of knowledge.

To ensure quality learning, we provide support sessions even after the training program.

A: Radiant Techlearning offers a training program on weekdays, weekends & a combination of weekdays & weekends. You can always choose the schedule that best suits your need.

A: We would always recommend you attend the live session to practice & clarify the doubts instantly & get more value from your investment. However, due to some contingency, if you have to skip the class, Radiant Techlearning will help you with the recorded session of that particular day. However, those recorded sessions are not meant only for personal consumption & NOT for distribution or any commercial use.

A: Radiant Techlearning has a data center containing the Virtual Training environment for the purpose of participant hand-on-practice. 

Participants can easily access these labs over Cloud with the help of a remote desktop connection. 

Radiant virtual labs provide you the flexibility to learn from anywhere in the world & in any time zone. 

Send a Message.


  • Enroll