Skip to main content

Home

Databricks FAQ

  1. Sign in to Unravel.

  2. Click Manager > Workspace and check if the corresponding Databricks workspace is shown in the Workspace list.

  1. Check if Spark conf, logging, and init scripts are present.

    For Spark conf, check if each and every property is correct.

  2. Refer Unravel > Workspace Manager > Cluster Configurations.

  1. Check if content in Spark conf, logging, and init scripts are correct.

  2. Refer Unravel > Workspace Manager > Cluster Configurations.

  1. Refer Unravel > Workspace Manager > Cluster Configurations.

  2. Check if spark.unravel.server.hostport is a valid address.

  1. Check if the host/IP is accessible from the Databricks notebook.

    %sh
    nslookup 10.2.1.4
  2. Check if port 4043 is open for outgoing traffic on Databricks.

    %sh
    nc -zv 10.2.1.4 4043
  3. Check if port 4043 is open for incoming traffic on Unravel.

    <see administrator>

Check if port 443 is open for outgoing traffic on Unravel

curl -X GET -H "Authorization: Bearer <token-here>" 'https://<instance-name-here>/api/2.0/dbfs/list?path=dbfs:/'
  1. Access the unravel.properties file in Unravel and get the workspace token.

  2. Run the following to check if the token is valid and works:

    curl -X GET -H "Authorization: Bearer <token>" 'https://<instance-name>/api/2.0/dbfs/list?path=dbfs:/'
    
  1. Run the following to get workspace token from unravel_db.properties file from DBFS.

    dbfs cat  
    dbfs:/databricks/unravel/unravel-db-sensor-archive/etc/unravel_db.properties
    

    Following is a sample of the output:

    #Thu Sep 30 20:06:42 UTC 2021
    unravel-server=01.0.0.1\:4043
    databricks-instance=https\://abc-0000000000000.18.azuredatabricks.net
    databricks-workspace-name=DBW-xxx-yyy
    databricks-workspace-id=<workspaceid>
    ssl_enabled=False
    insecure_ssl=True
    debug=False
    sleep-sec=30
    databricks-token=<databricks-token>
  2. Check if the token is valid using the following command:

    curl -X GET -H "Authorization: Bearer <token>" 'https://<instance-name>/api/2.0/dbfs/list?path=dbfs:/'
  3. If the token is invalid, you can regenerate the token and update the workspace from Unravel UI > Manage > Workspace.

You can register a workspace in Unravel from the command line with the manager command.

  1. Stop Unravel

    <Unravel installation directory>/unravel/manager stop
    
  2. Switch to Unravel user.

  3. Add the workspace details using the manager command as follows from the Unravel installation directory:

    source <path-to-python3-virtual environment-dir>/bin/activate
    <Unravel_installation_directory>/unravel/manager config databricks add --id <workspace-id> --name <workspace-name> --instance <workspace-instance> --access-token <workspace-token> --tier <tier_option>
    
    ##For example:
    /opt/unravel/manager config databricks add --id 0000000000000000 --name myworkspacename --instance https://adb-0000000000000000.16.azuredatabricks.net --access-token xxxx --tier premium
    
  4. Apply the changes.

    <Unravel installation directory>/unravel/manager config apply
    
  5. Start Unravel

    <Unravel installation directory>/unravel/manager start

Global init script applies the Unravel configurations to all clusters in a workspace. Do the following to set up Unravel configuration as Global init scripts.

    Global init script applies the Unravel configurations to all clusters in a workspace. Do the following to set up Unravel configuration as Global init scripts.

    1. On Databricks, go to Workspace > Settings > Admin Console > Global init scripts.

    2. Click +Add and set the following:

      Item

      Settings

      Name

      Set to unravel_cluster_init.

      Script

      Copy the contents from unravel_cluster_init.sh.

      Enabled

      Set to True.

    3. Click Add to save the settings.

    4. Click +Add again and set the following:

      Item

      Settings

      Name

      Set to unravel_spark_init.

      Script

      Copy the contents from unravel_spark_init.sh.

      Note

      This script supports up to Databricks 10.4 (latest at the time). In case a new DBR is available, test and update the condition as applicable.

      Enabled

      Set to True.

    5. Click Add to save the settings.

    Cluster init script applies the Unravel configurations for each cluster. To setup cluster init scripts from the cluster UI, do the following:

    1. Go to Unravel UI, click Manage > Workspaces > Cluster configuration to get the configuration details.

    2. Follow the instructions and update each cluster (Automated /Interactive) that you want to monitor with Unravel

      cluster-config-setup-databricks1.png
      cluster-config-setup-databricks2.png
      cluster-config-setup-databricks3.png
      cluster-config-setup-databricks4.png

    To add Unravel configurations to job clusters via API, use the JSON format as follows:

    {
        "settings": {
            "new_cluster": {
                "spark_conf": {
                    // Note: If extraJavaOptions is already in use, prepend the Unravel values. Also, for Databricks Runtime with spark 2.x.x, replace "spark-3.0" with "spark-2.4"
                    "spark.executor.extraJavaOptions": "-javaagent:/dbfs/databricks/unravel/unravel-agent-pack-bin/btrace-agent.jar=config=executor,libs=spark-3.0",
                    "spark.driver.extraJavaOptions": "-javaagent:/dbfs/databricks/unravel/unravel-agent-pack-bin/btrace-agent.jar=config=driver,script=StreamingProbe.btclass,libs=spark-3.0",
     
                    // rest of your spark properties here ...
                    ... 
                },
                "init_scripts": [
                    {
                        "dbfs": {
                            "destination": "dbfs:/databricks/unravel/unravel-db-sensor-archive/dbin/install-unravel.sh"
                        }
                    },
     
                    // rest of your init scripts here ...
                    ...
                ],
                "cluster_log_conf": {
                    "dbfs": {
                        "destination": "dbfs:/cluster-logs"
                    }
                },
     
                // rest of your cluster properties here ...
                ...
            },
            ...g
        }
    }
    

    Follow the instructions in this file to update instances and prices.

    You can configure discounted prices for VMs and DBUs in Databricks using Unravel properties. Add the properties to set the discount as a percentage value.

    1. Stop Unravel.

      <Unravel installation directory>/unravel/manager stop
      
    2. From the installation directory, set the following properties as follows:

      <Unravel installation directory>/unravel/manager config properties set com.unraveldata.databricks.vm.discount.percentage 
      <Unravel installation directory>/unravel/manager config properties set com.unraveldata.databricks.dbu.discount.percentage 
      ##Example: 
      /opt/unravel/manager config properties set com.unraveldata.databricks.vm.discount.percentage 10
      /opt/unravel/manager config properties set com.unraveldata.databricks.dbu.discount.percentage 20.5
      
    3. Apply the changes.

      <Unravel installation directory>/unravel/manager config apply
      
    4. Start Unravel.

      <Unravel installation directory>/unravel/manager start