Skip to main content

Home

v4.7.9.3 Release notes

Software version

Release date: January 25, 2024

See v4.7.9.3 for download information.v4.7.9.3

See also Unity App release notes

Software upgrade support

The following upgrade paths are supported:

  • 4.7.9.24.7.9.3

  • 4.7.8.0 Hotfix → 4.7.9.3

  • 4.7.8.0 4.7.9.3

  • 4.7.x (Databricks)4.7.9.3

For instructions to upgrade to Unravel v 4.7.9.3, see Upgrade to Unravel 4793

For fresh installations, see Deploy Unravel

Announcements

  • End of Support Announcement for RHEL 6

    Red Hat Enterprise Linux 6 (RHEL 6) is no longer supported with Unravel. If you are currently using RHEL 6, Unravel recommends that you plan an upgrade to a supported operating system to continue receiving updates and support. Contact support for any further assistance.

  • CPU Speed host metrics collection is not supported

    Starting from the 4793 release, we have deprecated the collection of CPU Speed host metrics.

Certified platforms

The following platforms are tested and certified in this release:

  • Databricks (Azure, AWS)

Review your platform's compatibility matrix before you install Unravel.Compatibility Matrix

Updates to Unravel's configuration properties

Updates to upgrading Unravel to v4.7.9.3

  1. Go to {unravel_install_dir}/versions/{unravel_version}/core/etc/dbx/cost

  2. Copy the following files:

    • prices_workload_tier_aws.tsv

    • prices_workload_tier_azure.tsv

  3. Paste the copied files and replace the existing files in this location:

    {unravel_install_dir}/data/conf/cost

The insight_upgrade.sh script be run after the upgrade. This script performs the following tasks:

  • Deletes older RealTimeLightProcessorEvent entries from the database and elasticsearch index.

  • Regenerates new NodeRightSizing events for certain clusters.

  1. Go to {unravel_install_dir}/unravel/services/insights_worker_1_1

  2. Run the insights_upgrade.sh script.

New features

  • Healthcheck ROI report

    A new Healthcheck ROI report is launched and is available as an App Store app. The app provides a comprehensive view of the Databricks environment, focusing on performance, costs, and potential savings. With this app, you can get insights into daily costs, hierarchical cost distribution, user, workspace, cluster, and job metrics. You can identify opportunities for workload optimization, worker resource classification, and migration savings through detailed analytics and recommendations. You can also have a holistic view of cluster metrics, including session costs, wastage analysis, and potential migration savings.

  • Support for Databricks 13.x and above

    Databricks Runtime 13.x and above is supported from this release.

    Note

    Databricks does not provide Ganglia metrics for Databricks Runtime 13 and above. Unravel now gathers all host-level metrics in real time from the /proc filesystem. There might be variations in the metrics collection approach of Unravel and Databricks itself.

  • Observability on Databricks SaaS is available with Standard (Free) tier

    Unravel has introduced observability on Databricks SaaS for free. You can now access essential observability features at no cost, allowing you to monitor your Databricks environment without incurring additional charges.

Improvements and enhancements

  • Improved On-demand Insights

    The on-demand Insights feature is now significantly faster, providing users with access to the most recent and relevant insights within 10 seconds, and enabling an intuitive comparison of resources that facilitates quick decision making. This update improves the user experience by streamlining the process of obtaining valuable insights.

  • Python upgrade

    In this release, Python is upgraded to version 3.8.12.

  • Backend updates to improve performance

    This release includes significant backend improvements aimed at enhancing overall system performance. These updates contribute to a more responsive and efficient system, ensuring a smoother experience.

The following table contains key issues addressed in the 4.7.9.3 release.

ID

Description

App Store

IMP-1089

Incorrect duration values are noted for the Interesting Apps data in specific applications.

Compute

IMP-1239

Modify the parsing logic for driver host metrics in the Spark Details page of Compute.

Insights

DT-1519

The Nodedownsizing event recommends a $0 cost saving for a successful job.

IMP-1217

Streaming applications are incorrectly generating RealtimeLightProcessor insights.

IMP-1272

An exception occurs while fetching feature data from the feature store.

Jobs

PIPELINE-1982

In a Spark application, there is a discrepancy in the displayed name on the Jobs page.

Kafka

CPLANE-2649

The Refresh Kafka command failed to start Zookeeper before initiating Kafka, resulting in an incomplete initialization.

Security

CUSTOMER-2584

Bind password is exposed in plain text within the AutoAction (AA) logs.

Sensor

CPLANE-3427

In the Unravel sensor logs, there is an occurrence of java.lang.NumberFormatException.

Workflows

CUSTOMER-2544

The sort functionality in the cost filter under the Workflow tab is not functioning as expected.

PIPELINE-2021

The cost filter under the Workflow tab is not functioning as expected.

The upcoming releases will include the following key fixes to enhance user experience. It is important to note that while these issues exist, there is no immediate critical impact on using the product, and users can continue to utilize its functionality with confidence.

ID

Description

Cost

DT-1879, DT-1871, DT-1853

The following issues are observed in the Budget page.

  • Redirection to the Compute page displays empty data for selected tags.

  • Redirection to the Chargeback page fails to populate scope filters correctly.

  • An incorrect date range is selected when redirecting to other pages from the Budget page.

UIX-6305

The Others category is displayed twice in legends when the number of clusters exceeds 1000 in the Chargeback page.

Compute

DT-2094

The cluster ID is displayed instead of the Cluster name for certain clusters.

DT-2079

The cluster cost displayed does not match the Azure billing report in some scenarios.

DT-2083

The Total Allocated Key Performance Indicators (KPIs) for Vcore and memory are not visible in the Compute > Trends page.

UIX-6321

All jobs in the running status are displayed in the Finished tab under Job Runs instead of showing only the finished jobs.

Insights

DT-2006

Recommendations are provided for a failed pipeline when users utilize multiple tasks with shared job clusters, and one of the tasks fails.

DT-2125

The UI shows a cost discrepancy for the Executor Idle time detected insight in the Databricks version 14.2 with Photon enabled.

Reports

DT-1841

The TopX Report displays an incorrect count of events.

Spark

DT-1742

The timezone for the NodeRightSizing insight event is inconsistent in the Spark details page.

DT-2012

Incorrect details are displayed on the AppSummary > Job Run page when a user repairs a previously failed job. The displayed information may not accurately reflect the repaired job's details.

/DT-2029

Applications in a success state may inaccurately display an associated job in a running state instead of transitioning to a failed state.

DT-2141

Clicking the line number in the Program tab does not highlight the actual line of code in the Spark details App Summary page

UIX-6523

The Sort by Write feature is currently not functioning as expected in the Spark details page.

SaaS (Free)

DT-2037

In the Databricks Standard (free) environment, there is an issue where the User Flow badge obstructs pagination.

Workflows

DT-2104

Sorting is incorrect when the list contains both strings starting with capital and small case letters,

  • Compute

    Jobs by status graphs in the Trends tab display spark application details and not the job details. Our development is actively looking into this design limitation and efforts are underway to address this in future updates to enhance the product's capabilities. (DT-2008)

  • Workflows

    The current implementation has a limitation where the wrong run count is displayed for the job ID when sorting by run count in the Workflows > Jobs section. This discrepancy is currently under investigation by our development team, and active efforts are being taken to resolve this issue. (UIX-6526)

Our development team is actively investigating the following Known issues and are working towards resolving them. It's important to note that while these issues exist, there is no immediate critical impact on using the product, and users can continue to utilize its functionality with confidence.

Bug ID

Description

Workaround

App Store

APP-614

App Store tasks fail to start with SSL enabled on the MySQL database.

Workaround

Compute

PIPELINE-1636

Inconsistent data is displayed for the cluster Duration and Start Time on the Compute page.

NA

Cost

UIX-5624

Data is not displayed when you click the Optimize button corresponding to OTHERS for the Cost > Chargeback results shown in the table.

NA

DT-1094

The No data available message is displayed on the Compute page after navigating from the  Trends and  Chargeback  pages with  Tag  filters.

NA

UIX-6310

On the Chargeback page, when no tag is provided for any application, NULL is displayed. Upon redirection from the Optimize link for the NULL tag, the Compute page shows all applications for the selected duration instead of specifically displaying applications with no tags.

NA

Datapage

DATAPAGE-502

If tables are created with the same name, accessed, deleted, and re-created, and if those tables are re-accessed, then their query and app count do not match.

NA

DATAPAGE-740

The query to fetch tableDailyKPIs is getting timed out when dealing with a huge table partition of 27 million records. From a threshold perspective, it has been verified that the API functions without issues for partition sizes up to 18 million.

NA

DATAPAGE-473

For Hive metastore 3.1.0 or earlier versions, the creation time of partitions is not captured if a partition is created dynamically.  Therefore, the Last Day KPI for the partition section is not shown in Unravel.

NA

Insights

DT-1987

There is a mismatch in the computation of costs for fleet and spot instances in Databricks clusters. This issue arises due to the unavailability of the exact node type in the cluster info response.

NA

UIX-5127, INSIGHTS-324,UIX-4176

Link re-direction issues, such as incorrect data filters for viewing Top Groups by Cost and Top Clusters by Cost, as well as missing re-direction links in the App Acceleration section.

NA

Performance

PIPELINE-1926

The Insight Worker daemon is experiencing performance lag, causing delays in processing insights and data analytics tasks.

NA

ASI-933

In the Lag setup, the Duration is not updated for running applications. The Duration should be updated every 15 minutes.

NA

ASI-936

In the Lag setup, the App Time data is missing in the Timing tab of many applications.

NA

Spark

DT-1404

Jobs created for the PySpark application using User-Defined Functions on a job cluster fail after applying the recommendations for node downsizing.

Workaround

PIPELINE-1616

If the Spark job is not running for Databricks, the values for the Duration and End time fields are not updated on the Databricks Run Details page.

NA

PLATFORM-2764

You can see a lag for SQL Streaming applications.

NA

UX-632

The timeline histogram needs to be generated correctly on the Spark application details page.

NA

PIPELINE-626

For PySpark applications, the processCPUTime and processCPULoad values are not captured properly.

NA

UI

UIX-5581

The job run count displayed on the Chargeback page differs from the job count shown on the Workflow page.

NA

PIPELINE-1935

In the Pipeline details page, when you select the data for a specific date, all instances are displayed instead of displaying only the instances within a selected date.

NA

UIX-6281

The cost comparison for all the instances is not displayed on the Pipeline detail page.

NA

PIPELINE-1934

On the Pipeline details page, the arrows must point only to the latest run instead of all the runs.

NA

UIX-6321

In the Workflow section, instead of displaying only jobs completed within the selected time frame, it currently displays jobs running within the selected duration.

NA

UIX-6263

The WhiteCross.png cross button on the Pipeline details page does not close the detail page when you click the bars inside the Gantt chart.

NA

UIX-3536

In the App summary page for Impala, the Query> Operator view is visible after scrolling down.

NA

Workflows

DT-1461, PIPELINE-1939, PIPELINE-1940, DT-1093, UIX-6274, PIPELINE-1924

The UI and data exhibit inconsistencies, including problems with job run details, issues related to multiple workflow runs and UTC timestamps , empty content in workflow job pages and issues with filter values and duration display.

NA

PIPELINE-1626, PIPELINE-1946

The Unravel UI has the issue of missing some Azure Databricks jobs and duplicate entries in Databricks workflow in certain scenarios.

NA

App Store

App Store tasks fail to start with SSL enabled on the MySQL database. (APP-614)

  1. Stop Unravel.

    <Unravel installation directory>/unravel/manager stop
  2. Use an editor to open <Installation_directory>/unravel/data/conf/unravel.yaml file.

  3. In the unravel.yaml file, under the database > advanced > python_flags block, enter the path to the trusted certificates. For example, if Unravel is installed at /opt/unravel, you must edit the unravel.yaml file as follows:

    unravel:
    ...snip... database:
    ...snip... advanced: python_flags: ssl_ca: /opt/unravel/data/certificates/trusted_certs.pem
  4. Use the manager utility to upload the certificates.

    <Unravel installation director>/manager config tls trust add --pem /path/to/certificate

    For example: /opt/unravel/manager config tls trust add --pem /path/to/certificate

  5. Enable the Truststore.

    <Unravel installation directory>/manager config tls trust enable
  6. Apply the changes and restart Unravel.

    <Unravel installation directory>/unravel/manager config apply --restart
Spark

Jobs created for the PySpark application using User-Defined Functions on a job cluster fail after applying the recommendations for node downsizing. (DT-1404)

  1. In your Databricks workspace, go to Configure Cluster > Advanced Options > Spark config .

  2. Add and set the following property to true for spark.driver.extraJavaOptions and spark.executor.extraJavaOptions spark configurations:

    • Dcom.unraveldata.metrics.proctree.enable=true

      For example:

      spark.executor.extraJavaOptions -Dcom.unraveldata.metrics.proctree.enable=true -javaagent:/dbfs/databricks/unravel/unravel-agent-pack-bin/btrace-agent.jar=config=executor,libs=spark-3.0 spark.driver.extraJavaOptions -Dcom.unraveldata.metrics.proctree.enable=true -javaagent:/dbfs/databricks/unravel/unravel-agent-pack-bin/btrace-agent.jar=config=driver,script=StreamingProbe.btclass,libs=spark-3.0