Skip to main content

Home

Integrating with Informatica Big Data Management

Informatica Big Data Management (BDM) suite enables enterprises to deploy advanced data management capabilities, including data ingestion, data quality, data masking, and stream processing. BDM provides a graphical user interface for generating ETL mappings (jobs) and workflows for various frameworks, such as Spark, Hive on MapReduce, or Hive on Tez. You install BDM on an edge node in your Hadoop cluster and use it to create jobs and workflows that ingest data into your cluster. The BDM interface allows you to construct jobs graphically, by connecting data sources to mappings to data destinations; BDM creates Hive or Spark queries for you, which removes the need for you to know Hive or Spark programming. However, while a non-programmatic interface is convenient, the jobs BDM generates might need tweaking to improve their performance or to meet your SLAs. Optimizing these BDM-generated data workflows can be complex and a resource drain for many customers. This is where Unravel comes in. The Unravel UI shows more information for each BDM job or workflow than YARN or your cluster manager does.

This topic explains how to connect Unravel to your BDM jobs and workflows so that Unravel can provide operational insights, recommendations, and automation to maximize their performance.

  • You've already installed Informatica BDM on an edge node in your cluster.

  • You've already installed Unravel Server on an edge node in your cluster.

  • Your cluster is on-premises CDH/HDP/MapR, or cloud-based Amazon EMR or Microsoft Azure HDInsight.

Setting properties in each mapping and workflow

These steps enable Unravel UI to display your BDM workflows and jobs.

For each BDM mapping (job):
  1. Log into the BDM graphical user interface.

  2. Select a mapping.

  3. Open the mapping's Properties view.

  4. In the Parameters tab, create two parameters:

    1. Create a parameter named UTCTimeStamp of Type string, with Precision set to 1000, Scale set to 0 and Default Value set to Default.

    2. Create a parameter named workflowName of Type string, with Precision set to 1000, Scale set to 0 and Default Value set to Default.

    informatica_bdm_job_parameters1.png
  5. Set up workflow tags.

    In the Run-time tab, map the two parameters you created to Unravel's workflow properties:

    Expand the Hadoop label and edit Runtime Properties:

    1. For Hive-on-Spark, create the Unravel property spark.unravel.workflow.name; for Hive-on-MR or Hive-on-Tez, create unravel.workflow.name. Set the Unavel property's value to the BDM job's workflowName parameter.

    2. For Hive-on-Spark, create the Unravel property spark.unravel.workflow.utctimestamp; for Hive-on-MR or Hive-on-Tez, create unravel.workflow.utctimestamp. Set the Unavel property's value to the BDM job's UTCTimeStamp parameter.

    informatica_bdm_job_parameters2.png
For each BDM workflow:
  1. Log into the BDM graphical user interface.

  2. Select a workflow.

  3. Open the workflow's Properties view.

  4. In the Input tab, set values for the two parameters you created in each mapping:

    1. Click Mapping Parameter Inputs | Unravel.

    2. Set UTCTimeStamp to the system variable sys.InstanceID.

    3. Set workflowName to the system variable sys.WorkflowName.

    informatica_bdm_workflow_parameters1.png
Enabling additional instrumentation from BDM

To get even more instrumentation from BDM, follow these steps on the BDM host --in other words, the edge node that's running BDM.

  1. Log into your BDM console as an administrator.

  2. In the BDM console, select Manage | Connections.

  3. Put Unravel's Hive Hook JAR in AUX_CLASSPATH.

    1. In Domain Navigator, expand | Domain | ClusterConfigurations | HADOOP_hdp_cco.

    2. Edit Hive Pushdown Configuration:

      Set Engine Type to Hive or Tez.
      Set Advanced Properties to:
      AUX_CLASSPATH=${AUX_CLASSPATH}:/usr/local/unravel_client/unravel-hive-1.2.0-hook.jar
  4. Deploy Unravel's Spark JAR.

    1. In Domain Navigator, expand | Domain | ClusterConfigurations | HADOOP_hdp_cco.

    2. Under Spark Configuration, set Spark Event Log Directory to

      hdfs:///spark2-history/
    3. Under Spark Configuration, append Unravel's Spark sensors to the following properties in Advanced Properties:

      spark.unravel.server.hostport=unravel-host:4043
      
      spark.driver.extraJavaOptions=-javaagent:path-to-btrace-agent.jar=config=driver,libs=spark-version
      
      spark.executor.extraJavaOptions=-javaagent:path-to-btrace-agent.jar=config=executor,libs=spark-version
  5. Connect to the Hive service to enable Unravel's Hive hooks.

    1. In Domain Navigator, expand | Domain | ClusterConfigurations | your deployment, such as HDP_CCO for HDP| hive_site_xml, and click Edit.

    2. In the Edit hive_site_xml dialog, set the following properties:

      com.unraveldata.hive.hook.tcp = true
      com.unraveldata.host = unravel-host
      hive.exec.pre.hooks = com.unraveldata.dataflow.hive.hook.unravel
      hive.exec.post.hooks = com.unraveldata.dataflow.hive.hook.unravel
      hive.exec.failure.hooks = com.unraveldata.dataflow.hive.hook.unravel
    3. In Domain Navigator, expand | Domain | ClusterConfigurations | your deployment, such as HDP_CCO for HDP| tez_site_xml, and click Edit.

    4. In tez.am.launch.cmd-opts, add a space and append this line to the existing values:

      -javaagent:/usr/local/unravel-agent/jars/btrace-agent.jar=libs=mr,config=tez -Dunravel.server.hostport=unravel-host:4043
    5. In tez.task.launch.cmd-opts, add a space, and append this line to the existing values:

      -javaagent:/usr/local/unravel-agent/jars/btrace-agent.jar=libs=mr,config=tez -Dunravel.server.hostport=unravel-host:4043
Viewing BDM jobs and workflows in Unravel UI

In Unravel UI, look at Applications | Workflows. For more information, refer to the product documentation.

informatica_workflow_in_unravel_ui.png