Add or Edit the Unity app reports
Use this API to create or edit the configuration for a new report. You can edit the configuration of a report while triggering the report as well.
PUT http://<unity_one_url>/api/reports/<report_name>
For example: PUT http://xyz.unraveldata.com:8111/api/reports/app_catalog_comparison
Parameters
Each report in the Unity app has common parameters as well as report-specific parameters.
The common parameters, which are used in all the reports, are described in the following table:
Name | Type | Description |
|---|---|---|
| boolean | To schedule the report generation at a fixed interval. Value can be true/false |
| integer | The number of days to keep the report. |
| boolean | To get detailed info on memory usage. Value can be true/false |
| string | Type of the report. You can specify any of the following as
|
| string | To get email notifications. The value should be comma-separated email IDs. |
The following sections provide details about report-specific parameters.
{
"enabled": false,
"retention_days": 50,
"profile_memory": false,
"report_type": "app_catalog_comparison",
"params": {
"kind": "spark",
"baseline_start_date": "2022-11-01T06:47:57.140Z",
"baseline_end_date": "2022-11-08T06:47:57.140Z",
"baseline_days": null,
"target_start_date": "2022-11-08T06:48:12.199Z",
"target_end_date": "2022-11-15T06:48:12.199Z",
"target_days": null,
"feature_filters": {
"null": null
},
"baseline_feature_filters": {
"null": null
}
}Name | Type | Description |
|---|---|---|
| string | Application kind. Value can be hive, impala, spark, mr |
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| list | Reports to be generated. Reports can be io, memory seconds,cpu time, duration, cost |
| integer | Number of topk apps to be included in the report. |
| Dict | Key-value pairs to filter the data. { ‘clusterUid’: ‘clusterUid From Unravel’, ‘clusterId’: ‘clusterId’,
} |
| list | Filter the results by given users. |
| list | Filter the results by given queues. |
| list | Filter the results by given clusters |
{
"start_date": null,
"end_date": null,
"days": 12,
"topk": 20,
"all_filters": {
"null": null
}
} Name | Type | Description |
|---|---|---|
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| integer | The number of topk apps to be included in the report. |
| dict | Features to filter the result. |
{
"start_date": "2023-03-01T11:57:25.429Z",
"end_date": "2023-03-27T11:57:25.429747Z",
"days": null,
"topk": 10,
"job_cost_threshold": 10,
"user_cost_threshold": 10,
"cluster_cost_threshold": 10,
"tag_filters": {
"Creator": null,
"RunName": null
},
"tag_cost_threshold": 10,
"billing_file": null,
"azure_account_type": "PAYG"
}Name | Type | Description |
|---|---|---|
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| string | Filter the results by given users. |
| integer | Number to analyze the topk entities. |
| integer | The number, which denotes the job cost threshold. |
| integer | The number, which denotes the user cost threshold. |
| integer | The number, which denotes cluster cost threshold. |
| string | The options are PAYG or EA. |
{
"kind": "spark",
"start_date": null,
"end_date": null,
"days": 100,
"reports": [
"app_dependencies",
"app_catalog"
],
"topk": 20,
"topk_options": [
"io",
"memorySeconds",
"cpuTime",
"duration",
"cost"
],
"feature_filters": {
"null": null
}
},Name | Type | Description |
|---|---|---|
| string | Application kind. Value can be hive, impala, spark, mr |
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| list | Specifies which report to run, report to include app_dependencies and app_catalog. |
| integer | The number of topk apps. |
| list | Options can be any of these or all these. [ "io","memorySeconds", "cpuTime","duration", "cost"] |
| dict | Key-value pairs to filter the data. |
{
"kind": "impala",
"target_system": "teradata",
"start_date": null,
"end_date": null,
"days": 12,
"users": [
"hive"
],
"queues": [
"default"
],
"clusters": [
"PG-CDP"
],
"tag_names": [
"RealUser"
],
"tag_values": [
"unravel"
],
"feature_filters": {
"clusterId": [
"PG-CDP"
]
}
} Name | Type | Description |
|---|---|---|
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| string | Application kind. Value can be hive, impala, spark, mr |
| string | Value can be any of the following target systems: teradata, snowflake, bigquery, athena, redshift |
| list | Filter the results by given users. |
| list | Filter the results by given queues. |
| list | Filter the results by given clusters. |
| list | Filter the results by given tag names. |
| list | Filter the results by given tag values. |
| dict | Key-value pairs to filter the data. |
{
"start_date": null,
"end_date": null,
"days": 10,
"interval": "10sec",
"queue_config_filepath": "",
"adhoc_cutoff": 20,
"users": [
"hive"
],
"workspace": [
"default"
],
"clusters": [
"PG-CDP"
],
"tag_names": [
"dept"
],
"tag_values": [
"Operations"
]
}, Name | Type | Description |
|---|---|---|
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| list | Filter the results by given users. |
| list | Filter the results by the given workspace. |
| list | Filter the results by given clusters. |
| list | Filter the results by given tag names. |
| list | Filter the results by given tag values. |
| list | Cost in dollars. |
{
"start_date": null,
"end_date": null,
"days": 14,
"reports": [
"NDSE",
"SQL_INEFF"
],
"feature_filters": {
"null": null
}
}
Name | Type | Description |
|---|---|---|
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| list | Reports to be generated. Reports can be "NDSE": "Apps with NodeDownsizingEvent", "CDE_NDSE": "Apps with NodeDownsizingEvent and ContendedDriverEvent", "SQL_INEFF": "SQL Apps with High Impact", "PNPE_NDSE": "Apps with NodeDownsizingEvent and PartitionsNotPrunedEvent", "DSE_NDSE": "Apps with NodeDownsizingEvent and DataSkewEvent", "IJE_NDSE": "Apps with NodeDownsizingEvent and InefficientJoinEvent", "IJCE_NDSE": "Apps with NodeDownsizingEvent and InefficientJoinConditionEvent", "SSFE_NDSE": "Apps with NodeDownsizingEvent and ScanSmallFilesEvent", "SSOE_NDSE": "Apps with NodeDownsizingEvent and SlowSQLOperatorEvent" |
| dict | Key-value pairs to filter the data. |
"params": { : Report parameters
"start_date": null, : Start date, to set exact date time.
"end_date": null, : End date, to set exact date time.
"days": 300, : No. of days to look back
} Name | Type | Description |
|---|---|---|
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| list | Filter the results by given users. |
| list | NFilter the results by given workspaces. |
| list | Filter the results by given clusters |
{
"target_days": 7,
"threshold": 0.001,
"featherbolt_path": "/opt/unravel/tmp/ondemand_fsimage/featherbolt_files/",
"report_values": "tables_small",
"project_name": "project",
"tenant_name": "playbook",
"database_name": "sys",
"table_name": "dag_meta"
} Name | Type | Description |
|---|---|---|
| integer | Number of days to look back. |
| integer | Threshold limit (MB ) value based on the report generated. |
| string | Featherbolt files path once fs image processed. |
| string | “tables_small” for Data Tables small files report. “hdfs_space” for HDFS space utilization threshold. |
| string | project name if hdfs_space gave as input otherwise it is empty. |
| string | tenant name if hdfs_space gave as input otherwise it is empty. |
| string | database name if tables_small gave as input otherwise it is empty. |
| string | table name if tables_small gave as input otherwise it is empty. |
{
"start_date": null,
"end_date": null,
"days": 10,
"interval": "10sec",
"queue_config_filepath": "",
"users": [
"hive"
],
"clusters": [
"PG-CDP"
],
"tag_names": [
"dept"
],
"tag_values": [
"Operations"
],
“Pools”:[
“root.default”
],
}, Name | Type | Description |
|---|---|---|
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| list | Filter the results by given users. |
| list | Filter the results by given clusters |
| list | Filter the results by given tag names. |
| list | Filter the results by given tag values. |
| list | Filter the results by given pools. |
{
"start_date": null,
"end_date": null,
"days": 10,
"interval": "10sec",
"queue_config_filepath": "",
"users": [
"hive"
],
"queues": [
"default"
],
"clusters": [
"PG-CDP"
],
"tag_names": [
"dept"
],
"tag_values": [
"Operations"
]
} Name | Type | Description |
|---|---|---|
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| list | Filter the results by given users. |
| list | Filter the results by given queues. |
| list | Filter the results by given clusters. |
| list | Filter the results by given tag names. |
| list | Filter the results by given tag values. |
"params": {
"kind": "hive",
"start_date": null,
"end_date": null,
"days": 300,
"event": "All"
}Name | Type | Description |
|---|---|---|
| string | Application kind. Value can be hive, impala, spark, mr |
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| list | Specific event. Events can be All, Application Failure, Resource Utilization, Speedup, Informational, Cost Savings |
| list | Filter the results by given users. |
| list | NFilter the results by given queues. |
| list | Filter the results by given clusters |
{
"start_date": null,
"end_date": null,
"days": 100,
"users": null,
"pools": null,
"memory_spilled": null,,
"Rows_produced": null,
"duration": null,
"est_per_node_peak_memory": null,
"per_node_peak_memory": null,
"aggregate_peak_memory": null,
"admission_wait_time": null,
"hdfs_remote_bytes_read": null,
"statistics_corrupt_or_missing": "False"
} Name | Type | Description |
|---|---|---|
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| list | Specific event. Events can be All, Application Failure, Resource Utilization, Speedup, Informational, Cost Savings |
| list | Filter the results by given pools. |
| integer | Memory spilled in bytes threshold values. |
| integer | Rows produced threshold value. |
| integer | Query duration in seconds threshold value |
| boolean | Boolean value. By default, False is given. |
| integer | Estimated Per Node Peak memory threshold value. |
| integer | Aggregate Peak Memory threshold value. |
| integer | Admission wait time threshold value. |
| integer | HDFS Remote Bytes threshold value. |
{
"start_date": null,
"end_date": null,
"days": 10,
"interval": "10sec",
"queue_config_filepath": "",
"adhoc_cutoff": 20,
"users": [
"hive"
],
"queues": [
"default"
],
"clusters": [
"PG-CDP"
],
"tag_names": [
"dept"
],
"tag_values": [
"Operations"
]
},
Name | Type | Description |
|---|---|---|
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| list | Filter the results by given users. |
| list | Filter the results by given queues. |
| list | Filter the results by given clusters |
| integer | Adhoc cut-off value. |
| list | Filter the results by given tag names. |
| list | Filter the results by given tag values. |
{
"start_date": null,
"end_date": null,
"days": 100,
"topk": 5
}Name | Type | Description |
|---|---|---|
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| integer | The number of topk apps to be included in the report. |
{
"baseline_start_date": null,
"baseline_end_date": null,
"baseline_days": 12,
"baseline_pipelines": [
"Recommendation_Cluster",
"Covid_Data_Processing"
],
"target_start_date": null,
"target_end_date": null,
"target_days": 12,
"target_pipelines": [
"Recommendation_Cluster",
"Covid_Data_Processing"
],
"improved_vcore_seconds_threshold": 5,
"improved_duration_threshold": 25,
"improved_io_threshold": 5,
"improved_memory_seconds_threshold": 5,
"degraded_vcore_seconds_threshold": 5,
"degraded_duration_threshold": 25,
"degraded_io_threshold": 5,
"degraded_memory_seconds_threshold": 5
},
Name | Type | Description |
|---|---|---|
| string | Baseline start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | Baseline end date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | target start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | target end date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of target days to look back. |
| list | The list should contain target pipeline names that need to be compared. |
| integer | Improved vcores threshold value should be specified here. |
| integer | The improved duration threshold value should be specified here. |
| integer | Improved IO threshold value should be specified here. |
| integer | Improved memory threshold value should be specified here. |
| integer | The degraded vcores threshold value should be specified here. |
| integer | The degraded duration threshold value should be specified here. |
| integer | The degraded IO threshold value should be specified here. |
| integer | The degraded memory threshold value should be specified here. |
"params": {
"start_date": null,
"end_date": null,
"days": 12,
"clusters": "default",
"resource_scheduler_port": "http://sd11.unraveldata.com:8088",
"resource_scheduler_config_path": null,
"queues": "root.*"
}Name | Type | Description |
|---|---|---|
| string | Application kind. Value can be hive, impala, spark, mr |
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| integer | Specific event. Events can be All, Application Failure, Resource Utilization, Speedup, Informational, Cost Savings |
| list | Filter the results by given users. |
| list | NFilter the results by given queues. |
| list | Filter the results by given clusters |
{
"kind": "hive",
"start_date": null,
"end_date": null,
"retention_days": 5,
"days": 300,
"reports": [
"io"
],
"topk": 10,
"feature_filters": {
"null": null
}
} Name | Type | Description |
|---|---|---|
| string | Application kind. Value can be hive, impala, spark, mr |
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| list | Reports to be generated. Reports can be io, memory seconds,cpu time, duration, cost |
| integer | The number of topk apps to be included in the report. |
| Dict | Key-value pairs to filter the data. { ‘clusterUid’: ‘clusterUid From Unravel’, ‘clusterId’: ‘clusterId’,
} |
| list | Filter the results by given users. |
| list | Filter the results by given queues. |
| list | Filter the results by given clusters |
{
"kinds": [
"hive",
"impala",
"spark",
"mr"
],
"start_date": null,
"end_date": null,
"days": 12,
"resource_metric": "memorySeconds",
"user_db_filepath": "/path/to/csv.csv",
"group_by_columns": "column1,column2,column3",
"country_column": "country",
"join_column": "id"
},Name | Type | Description |
|---|---|---|
| string | Start date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| string | End date, to set exact date time. Format: 2022-12-07T07:32:09.629946Z |
| integer | Number of days to look back. |
| int | Kinds should be any/all of these values. [ "hive","impala","spark","mr"] |
| integer | The number of topk apps to be included in the report. |
| string | Should be any of these values: memorySeconds, cpu |
| String | Path to user db. |
| string | Group by columns. |
| string | Country column name. |
| string | Join the column by ID. |
Sample request
PUT http://xyz.unraveldata.com:8111/api/reports/topk-apps-spark
{
"enabled": false,
"retention_days": 50,
"profile_memory": false,
"report_type": "topkapps",
"params": {
"kind": "spark",
"start_date": null,
"end_date": null,
"days": 100,
"ldap_conf": false,
"reports": [
"io",
"cost",
"cpuTime",
"duration",
"memorySeconds"
],
"topk": 10,
"feature_filters": {
"null": null
}
},
"notifications": {}
}Status codes
Code | Description |
|---|---|
200 | Successful operation |