From 69de75de362bdf86f01ceeee90ee188485d8d332 Mon Sep 17 00:00:00 2001 From: Divyansh Vijayvergia <171924202+Divyansh-db@users.noreply.github.com> Date: Tue, 6 Jan 2026 09:38:36 +0100 Subject: [PATCH] Update SDK to c0e83a0de9da6aaad4cd6924fb65b1496bfdedcb --- .codegen/_openapi_sha | 2 +- .gitattributes | 30 +- NEXT_CHANGELOG.md | 26 +- .../com/databricks/sdk/service/apps/App.java | 19 - .../sdk/service/apps/AppDeployment.java | 57 +-- .../AppManifestAppResourceExperimentSpec.java | 45 --- ...rceExperimentSpecExperimentPermission.java | 12 - .../apps/AppManifestAppResourceSpec.java | 17 - ...eUcSecurableSpecUcSecurablePermission.java | 2 - ...esourceUcSecurableSpecUcSecurableType.java | 2 - .../sdk/service/apps/AppResource.java | 16 - .../service/apps/AppResourceExperiment.java | 59 ---- ...esourceExperimentExperimentPermission.java | 12 - ...ourceUcSecurableUcSecurablePermission.java | 3 - ...AppResourceUcSecurableUcSecurableType.java | 3 - .../sdk/service/apps/AppUpdate.java | 24 +- .../databricks/sdk/service/apps/EnvVar.java | 77 ---- .../sdk/service/apps/GitRepository.java | 59 ---- .../sdk/service/apps/GitSource.java | 128 ------- .../catalog/AccessRequestDestinations.java | 68 +--- .../sdk/service/catalog/ConnectionType.java | 2 +- .../sdk/service/catalog/SchemaInfo.java | 2 +- .../sdk/service/catalog/SecurableKind.java | 2 +- .../sdk/service/catalog/TableInfo.java | 3 +- .../sdk/service/catalog/TablesAPI.java | 3 - .../sdk/service/catalog/TablesService.java | 3 - .../dataquality/AnomalyDetectionConfig.java | 24 +- .../databricks/sdk/service/jobs/JobRunAs.java | 22 +- .../sdk/service/ml/FeatureEngineeringAPI.java | 25 +- .../service/ml/FeatureEngineeringImpl.java | 3 +- .../service/ml/FeatureEngineeringService.java | 25 +- .../pipelines/TableSpecificConfig.java | 19 - .../sdk/service/postgres/Branch.java | 247 ++++++++++++- .../sdk/service/postgres/BranchSpec.java | 108 ------ ...ranchStatusState.java => BranchState.java} | 2 +- .../sdk/service/postgres/BranchStatus.java | 177 ---------- .../service/postgres/CreateRoleOperation.java | 161 --------- .../service/postgres/CreateRoleRequest.java | 82 ----- .../service/postgres/DeleteRoleOperation.java | 161 --------- .../service/postgres/DeleteRoleRequest.java | 71 ---- .../sdk/service/postgres/Endpoint.java | 331 ++++++++++++++++-- .../sdk/service/postgres/EndpointSpec.java | 146 -------- ...intStatusState.java => EndpointState.java} | 2 +- .../sdk/service/postgres/EndpointStatus.java | 246 ------------- .../sdk/service/postgres/GetRoleRequest.java | 44 --- .../service/postgres/ListRolesRequest.java | 79 ----- .../service/postgres/ListRolesResponse.java | 59 ---- .../sdk/service/postgres/PostgresAPI.java | 44 --- .../sdk/service/postgres/PostgresImpl.java | 57 --- .../sdk/service/postgres/PostgresService.java | 12 - .../sdk/service/postgres/Project.java | 231 ++++++++++-- .../sdk/service/postgres/ProjectSpec.java | 110 ------ .../sdk/service/postgres/ProjectStatus.java | 160 --------- .../databricks/sdk/service/postgres/Role.java | 123 ------- .../sdk/service/postgres/RoleAuthMethod.java | 13 - .../service/postgres/RoleIdentityType.java | 16 - .../postgres/RoleOperationMetadata.java | 28 -- .../sdk/service/postgres/RoleRoleSpec.java | 66 ---- .../sdk/service/postgres/RoleRoleStatus.java | 59 ---- .../AnomalyDetectionConfig.java | 21 +- .../sdk/service/sharing/SharedDataObject.java | 55 +-- .../databricks/sdk/service/sharing/Table.java | 37 +- .../vectorsearch/QueryVectorIndexRequest.java | 9 +- 63 files changed, 818 insertions(+), 2933 deletions(-) delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceExperimentSpec.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceExperimentSpecExperimentPermission.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceExperiment.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceExperimentExperimentPermission.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/EnvVar.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GitRepository.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GitSource.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchSpec.java rename databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/{BranchStatusState.java => BranchState.java} (89%) delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchStatus.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateRoleOperation.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateRoleRequest.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/DeleteRoleOperation.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/DeleteRoleRequest.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointSpec.java rename databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/{EndpointStatusState.java => EndpointState.java} (88%) delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointStatus.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/GetRoleRequest.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListRolesRequest.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListRolesResponse.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ProjectSpec.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ProjectStatus.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Role.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleAuthMethod.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleIdentityType.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleOperationMetadata.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleRoleSpec.java delete mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleRoleStatus.java diff --git a/.codegen/_openapi_sha b/.codegen/_openapi_sha index 528b0ddb0..57c80b78c 100644 --- a/.codegen/_openapi_sha +++ b/.codegen/_openapi_sha @@ -1 +1 @@ -44cbc832f1b070c47544ff470fd8498853d24cf3 \ No newline at end of file +c0e83a0de9da6aaad4cd6924fb65b1496bfdedcb \ No newline at end of file diff --git a/.gitattributes b/.gitattributes index ef3a43d21..31bcf3f71 100644 --- a/.gitattributes +++ b/.gitattributes @@ -20,8 +20,6 @@ /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppDeploymentState.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppDeploymentStatus.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifest.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceExperimentSpec.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceExperimentSpecExperimentPermission.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceJobSpec.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceJobSpecJobPermission.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceSecretSpec.java linguist-generated=true @@ -42,8 +40,6 @@ /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResource.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceDatabase.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceDatabaseDatabasePermission.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceExperiment.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceExperimentExperimentPermission.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceGenieSpace.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceGenieSpaceGenieSpacePermission.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceJob.java linguist-generated=true @@ -78,7 +74,6 @@ /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/CustomTemplate.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/DeleteAppRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/DeleteCustomTemplateRequest.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/EnvVar.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GetAppDeploymentRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GetAppPermissionLevelsRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GetAppPermissionLevelsResponse.java linguist-generated=true @@ -86,8 +81,6 @@ /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GetAppRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GetAppUpdateRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GetCustomTemplateRequest.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GitRepository.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GitSource.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/ListAppDeploymentsRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/ListAppDeploymentsResponse.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/ListAppsRequest.java linguist-generated=true @@ -2074,45 +2067,34 @@ /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/pipelines/UpdateStateInfoState.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Branch.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchOperationMetadata.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchSpec.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchStatus.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchStatusState.java linguist-generated=true +/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchState.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateBranchOperation.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateBranchRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateEndpointOperation.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateEndpointRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateProjectOperation.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateProjectRequest.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateRoleOperation.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateRoleRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/DatabricksServiceExceptionWithDetailsProto.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/DeleteBranchRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/DeleteEndpointRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/DeleteProjectRequest.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/DeleteRoleOperation.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/DeleteRoleRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Endpoint.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointOperationMetadata.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointPoolerMode.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointSettings.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointSpec.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointStatus.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointStatusState.java linguist-generated=true +/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointState.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointType.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ErrorCode.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/GetBranchRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/GetEndpointRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/GetOperationRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/GetProjectRequest.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/GetRoleRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListBranchesRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListBranchesResponse.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListEndpointsRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListEndpointsResponse.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListProjectsRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListProjectsResponse.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListRolesRequest.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListRolesResponse.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Operation.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/PostgresAPI.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/PostgresImpl.java linguist-generated=true @@ -2121,14 +2103,6 @@ /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ProjectDefaultEndpointSettings.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ProjectOperationMetadata.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ProjectSettings.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ProjectSpec.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ProjectStatus.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Role.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleAuthMethod.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleIdentityType.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleOperationMetadata.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleRoleSpec.java linguist-generated=true -/home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleRoleStatus.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/UpdateBranchOperation.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/UpdateBranchRequest.java linguist-generated=true /home/ubuntu/workspace/databricks-sdk-java/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/UpdateEndpointOperation.java linguist-generated=true diff --git a/NEXT_CHANGELOG.md b/NEXT_CHANGELOG.md index 3540fbd3f..dcac63ef3 100644 --- a/NEXT_CHANGELOG.md +++ b/NEXT_CHANGELOG.md @@ -38,4 +38,28 @@ * Add `command` and `envVars` fields for `com.databricks.sdk.service.apps.AppDeployment`. * Add `fullName` and `securableType` fields for `com.databricks.sdk.service.catalog.AccessRequestDestinations`. * [Breaking] Change `deleteKafkaConfig()` method for `workspaceClient.featureEngineering()` service . Method path has changed. -* [Breaking] Change long-running operation configuration for `workspaceClient.postgres().deleteRole()` method . Long running operation response type changed to `Void class`. \ No newline at end of file +* [Breaking] Change long-running operation configuration for `workspaceClient.postgres().deleteRole()` method . Long running operation response type changed to `Void class`. +* Add `currentState`, `default`, `effectiveDefault`, `effectiveIsProtected`, `effectiveSourceBranch`, `effectiveSourceBranchLsn`, `effectiveSourceBranchTime`, `isProtected`, `logicalSizeBytes`, `pendingState`, `sourceBranch`, `sourceBranchLsn`, `sourceBranchTime` and `stateChangeTime` fields for `com.databricks.sdk.service.postgres.Branch`. +* Add `autoscalingLimitMaxCu`, `autoscalingLimitMinCu`, `currentState`, `disabled`, `effectiveAutoscalingLimitMaxCu`, `effectiveAutoscalingLimitMinCu`, `effectiveDisabled`, `effectivePoolerMode`, `effectiveSettings`, `effectiveSuspendTimeoutDuration`, `host`, `lastActiveTime`, `pendingState`, `poolerMode`, `settings`, `startTime`, `suspendTime`, `suspendTimeoutDuration` and `endpointType` fields for `com.databricks.sdk.service.postgres.Endpoint`. +* [Breaking] Add `endpointType` field for `com.databricks.sdk.service.postgres.Endpoint`. +* Add `branchLogicalSizeLimitBytes`, `computeLastActiveTime`, `defaultEndpointSettings`, `displayName`, `effectiveDefaultEndpointSettings`, `effectiveDisplayName`, `effectiveHistoryRetentionDuration`, `effectivePgVersion`, `effectiveSettings`, `historyRetentionDuration`, `pgVersion`, `settings` and `syntheticStorageSizeBytes` fields for `com.databricks.sdk.service.postgres.Project`. +* [Breaking] Change `deleteKafkaConfig()` method for `workspaceClient.featureEngineering()` service . Method path has changed. +* [Breaking] Remove `createRole()`, `deleteRole()`, `getRole()` and `listRoles()` methods for `workspaceClient.postgres()` service. +* [Breaking] Remove `gitRepository` field for `com.databricks.sdk.service.apps.App`. +* [Breaking] Remove `command`, `envVars` and `gitSource` fields for `com.databricks.sdk.service.apps.AppDeployment`. +* [Breaking] Remove `experimentSpec` field for `com.databricks.sdk.service.apps.AppManifestAppResourceSpec`. +* [Breaking] Remove `experiment` field for `com.databricks.sdk.service.apps.AppResource`. +* [Breaking] Remove `gitRepository` field for `com.databricks.sdk.service.apps.AppUpdate`. +* [Breaking] Remove `destinationSourceSecurable`, `fullName` and `securableType` fields for `com.databricks.sdk.service.catalog.AccessRequestDestinations`. +* [Breaking] Remove `excludedTableFullNames` field for `com.databricks.sdk.service.dataquality.AnomalyDetectionConfig`. +* [Breaking] Remove `groupName` field for `com.databricks.sdk.service.jobs.JobRunAs`. +* [Breaking] Remove `rowFilter` field for `com.databricks.sdk.service.pipelines.TableSpecificConfig`. +* [Breaking] Remove `spec` and `status` fields for `com.databricks.sdk.service.postgres.Branch`. +* [Breaking] Remove `spec` and `status` fields for `com.databricks.sdk.service.postgres.Endpoint`. +* [Breaking] Remove `spec` and `status` fields for `com.databricks.sdk.service.postgres.Project`. +* [Breaking] Remove `excludedTableFullNames` field for `com.databricks.sdk.service.qualitymonitorv2.AnomalyDetectionConfig`. +* [Breaking] Remove `accessModes` and `storageLocation` fields for `com.databricks.sdk.service.sharing.Table`. +* [Breaking] Remove `EXECUTE` and `USE_CONNECTION` enum values for `com.databricks.sdk.service.apps.AppManifestAppResourceUcSecurableSpecUcSecurablePermission`. +* [Breaking] Remove `FUNCTION` and `CONNECTION` enum values for `com.databricks.sdk.service.apps.AppManifestAppResourceUcSecurableSpecUcSecurableType`. +* [Breaking] Remove `SELECT`, `EXECUTE` and `USE_CONNECTION` enum values for `com.databricks.sdk.service.apps.AppResourceUcSecurableUcSecurablePermission`. +* [Breaking] Remove `TABLE`, `FUNCTION` and `CONNECTION` enum values for `com.databricks.sdk.service.apps.AppResourceUcSecurableUcSecurableType`. \ No newline at end of file diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/App.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/App.java index d403abc7f..d7ef3d38a 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/App.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/App.java @@ -64,13 +64,6 @@ public class App { @JsonProperty("effective_user_api_scopes") private Collection effectiveUserApiScopes; - /** - * Git repository configuration for app deployments. When specified, deployments can reference - * code from this repository by providing only the git reference (branch, tag, or commit). - */ - @JsonProperty("git_repository") - private GitRepository gitRepository; - /** The unique identifier of the app. */ @JsonProperty("id") private String id; @@ -241,15 +234,6 @@ public Collection getEffectiveUserApiScopes() { return effectiveUserApiScopes; } - public App setGitRepository(GitRepository gitRepository) { - this.gitRepository = gitRepository; - return this; - } - - public GitRepository getGitRepository() { - return gitRepository; - } - public App setId(String id) { this.id = id; return this; @@ -393,7 +377,6 @@ public boolean equals(Object o) { && Objects.equals(effectiveBudgetPolicyId, that.effectiveBudgetPolicyId) && Objects.equals(effectiveUsagePolicyId, that.effectiveUsagePolicyId) && Objects.equals(effectiveUserApiScopes, that.effectiveUserApiScopes) - && Objects.equals(gitRepository, that.gitRepository) && Objects.equals(id, that.id) && Objects.equals(name, that.name) && Objects.equals(oauth2AppClientId, that.oauth2AppClientId) @@ -425,7 +408,6 @@ public int hashCode() { effectiveBudgetPolicyId, effectiveUsagePolicyId, effectiveUserApiScopes, - gitRepository, id, name, oauth2AppClientId, @@ -457,7 +439,6 @@ public String toString() { .add("effectiveBudgetPolicyId", effectiveBudgetPolicyId) .add("effectiveUsagePolicyId", effectiveUsagePolicyId) .add("effectiveUserApiScopes", effectiveUserApiScopes) - .add("gitRepository", gitRepository) .add("id", id) .add("name", name) .add("oauth2AppClientId", oauth2AppClientId) diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppDeployment.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppDeployment.java index bf9640846..0961135b1 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppDeployment.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppDeployment.java @@ -5,18 +5,10 @@ import com.databricks.sdk.support.Generated; import com.databricks.sdk.support.ToStringer; import com.fasterxml.jackson.annotation.JsonProperty; -import java.util.Collection; import java.util.Objects; @Generated public class AppDeployment { - /** - * The command with which to run the app. This will override the command specified in the app.yaml - * file. - */ - @JsonProperty("command") - private Collection command; - /** The creation time of the deployment. Formatted timestamp in ISO 6801. */ @JsonProperty("create_time") private String createTime; @@ -33,17 +25,6 @@ public class AppDeployment { @JsonProperty("deployment_id") private String deploymentId; - /** - * The environment variables to set in the app runtime environment. This will override the - * environment variables specified in the app.yaml file. - */ - @JsonProperty("env_vars") - private Collection envVars; - - /** Git repository to use as the source for the app deployment. */ - @JsonProperty("git_source") - private GitSource gitSource; - /** The mode of which the deployment will manage the source code. */ @JsonProperty("mode") private AppDeploymentMode mode; @@ -66,15 +47,6 @@ public class AppDeployment { @JsonProperty("update_time") private String updateTime; - public AppDeployment setCommand(Collection command) { - this.command = command; - return this; - } - - public Collection getCommand() { - return command; - } - public AppDeployment setCreateTime(String createTime) { this.createTime = createTime; return this; @@ -111,24 +83,6 @@ public String getDeploymentId() { return deploymentId; } - public AppDeployment setEnvVars(Collection envVars) { - this.envVars = envVars; - return this; - } - - public Collection getEnvVars() { - return envVars; - } - - public AppDeployment setGitSource(GitSource gitSource) { - this.gitSource = gitSource; - return this; - } - - public GitSource getGitSource() { - return gitSource; - } - public AppDeployment setMode(AppDeploymentMode mode) { this.mode = mode; return this; @@ -170,13 +124,10 @@ public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; AppDeployment that = (AppDeployment) o; - return Objects.equals(command, that.command) - && Objects.equals(createTime, that.createTime) + return Objects.equals(createTime, that.createTime) && Objects.equals(creator, that.creator) && Objects.equals(deploymentArtifacts, that.deploymentArtifacts) && Objects.equals(deploymentId, that.deploymentId) - && Objects.equals(envVars, that.envVars) - && Objects.equals(gitSource, that.gitSource) && Objects.equals(mode, that.mode) && Objects.equals(sourceCodePath, that.sourceCodePath) && Objects.equals(status, that.status) @@ -186,13 +137,10 @@ public boolean equals(Object o) { @Override public int hashCode() { return Objects.hash( - command, createTime, creator, deploymentArtifacts, deploymentId, - envVars, - gitSource, mode, sourceCodePath, status, @@ -202,13 +150,10 @@ public int hashCode() { @Override public String toString() { return new ToStringer(AppDeployment.class) - .add("command", command) .add("createTime", createTime) .add("creator", creator) .add("deploymentArtifacts", deploymentArtifacts) .add("deploymentId", deploymentId) - .add("envVars", envVars) - .add("gitSource", gitSource) .add("mode", mode) .add("sourceCodePath", sourceCodePath) .add("status", status) diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceExperimentSpec.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceExperimentSpec.java deleted file mode 100644 index 2df07f818..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceExperimentSpec.java +++ /dev/null @@ -1,45 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.apps; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import java.util.Objects; - -@Generated -public class AppManifestAppResourceExperimentSpec { - /** */ - @JsonProperty("permission") - private AppManifestAppResourceExperimentSpecExperimentPermission permission; - - public AppManifestAppResourceExperimentSpec setPermission( - AppManifestAppResourceExperimentSpecExperimentPermission permission) { - this.permission = permission; - return this; - } - - public AppManifestAppResourceExperimentSpecExperimentPermission getPermission() { - return permission; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - AppManifestAppResourceExperimentSpec that = (AppManifestAppResourceExperimentSpec) o; - return Objects.equals(permission, that.permission); - } - - @Override - public int hashCode() { - return Objects.hash(permission); - } - - @Override - public String toString() { - return new ToStringer(AppManifestAppResourceExperimentSpec.class) - .add("permission", permission) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceExperimentSpecExperimentPermission.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceExperimentSpecExperimentPermission.java deleted file mode 100644 index 27db3de37..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceExperimentSpecExperimentPermission.java +++ /dev/null @@ -1,12 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.apps; - -import com.databricks.sdk.support.Generated; - -@Generated -public enum AppManifestAppResourceExperimentSpecExperimentPermission { - CAN_EDIT, - CAN_MANAGE, - CAN_READ, -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceSpec.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceSpec.java index 70c87f518..f4487f6e5 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceSpec.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceSpec.java @@ -17,10 +17,6 @@ public class AppManifestAppResourceSpec { @JsonProperty("description") private String description; - /** */ - @JsonProperty("experiment_spec") - private AppManifestAppResourceExperimentSpec experimentSpec; - /** */ @JsonProperty("job_spec") private AppManifestAppResourceJobSpec jobSpec; @@ -54,16 +50,6 @@ public String getDescription() { return description; } - public AppManifestAppResourceSpec setExperimentSpec( - AppManifestAppResourceExperimentSpec experimentSpec) { - this.experimentSpec = experimentSpec; - return this; - } - - public AppManifestAppResourceExperimentSpec getExperimentSpec() { - return experimentSpec; - } - public AppManifestAppResourceSpec setJobSpec(AppManifestAppResourceJobSpec jobSpec) { this.jobSpec = jobSpec; return this; @@ -127,7 +113,6 @@ public boolean equals(Object o) { if (o == null || getClass() != o.getClass()) return false; AppManifestAppResourceSpec that = (AppManifestAppResourceSpec) o; return Objects.equals(description, that.description) - && Objects.equals(experimentSpec, that.experimentSpec) && Objects.equals(jobSpec, that.jobSpec) && Objects.equals(name, that.name) && Objects.equals(secretSpec, that.secretSpec) @@ -140,7 +125,6 @@ public boolean equals(Object o) { public int hashCode() { return Objects.hash( description, - experimentSpec, jobSpec, name, secretSpec, @@ -153,7 +137,6 @@ public int hashCode() { public String toString() { return new ToStringer(AppManifestAppResourceSpec.class) .add("description", description) - .add("experimentSpec", experimentSpec) .add("jobSpec", jobSpec) .add("name", name) .add("secretSpec", secretSpec) diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceUcSecurableSpecUcSecurablePermission.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceUcSecurableSpecUcSecurablePermission.java index a5c0749fa..1a614264f 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceUcSecurableSpecUcSecurablePermission.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceUcSecurableSpecUcSecurablePermission.java @@ -6,10 +6,8 @@ @Generated public enum AppManifestAppResourceUcSecurableSpecUcSecurablePermission { - EXECUTE, MANAGE, READ_VOLUME, SELECT, - USE_CONNECTION, WRITE_VOLUME, } diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceUcSecurableSpecUcSecurableType.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceUcSecurableSpecUcSecurableType.java index d8d452c05..45cccb2d5 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceUcSecurableSpecUcSecurableType.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppManifestAppResourceUcSecurableSpecUcSecurableType.java @@ -6,8 +6,6 @@ @Generated public enum AppManifestAppResourceUcSecurableSpecUcSecurableType { - CONNECTION, - FUNCTION, TABLE, VOLUME, } diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResource.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResource.java index e46cb6e8e..2761c1651 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResource.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResource.java @@ -17,10 +17,6 @@ public class AppResource { @JsonProperty("description") private String description; - /** */ - @JsonProperty("experiment") - private AppResourceExperiment experiment; - /** */ @JsonProperty("genie_space") private AppResourceGenieSpace genieSpace; @@ -67,15 +63,6 @@ public String getDescription() { return description; } - public AppResource setExperiment(AppResourceExperiment experiment) { - this.experiment = experiment; - return this; - } - - public AppResourceExperiment getExperiment() { - return experiment; - } - public AppResource setGenieSpace(AppResourceGenieSpace genieSpace) { this.genieSpace = genieSpace; return this; @@ -146,7 +133,6 @@ public boolean equals(Object o) { AppResource that = (AppResource) o; return Objects.equals(database, that.database) && Objects.equals(description, that.description) - && Objects.equals(experiment, that.experiment) && Objects.equals(genieSpace, that.genieSpace) && Objects.equals(job, that.job) && Objects.equals(name, that.name) @@ -161,7 +147,6 @@ public int hashCode() { return Objects.hash( database, description, - experiment, genieSpace, job, name, @@ -176,7 +161,6 @@ public String toString() { return new ToStringer(AppResource.class) .add("database", database) .add("description", description) - .add("experiment", experiment) .add("genieSpace", genieSpace) .add("job", job) .add("name", name) diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceExperiment.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceExperiment.java deleted file mode 100644 index 6f194ec5c..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceExperiment.java +++ /dev/null @@ -1,59 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.apps; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import java.util.Objects; - -@Generated -public class AppResourceExperiment { - /** */ - @JsonProperty("experiment_id") - private String experimentId; - - /** */ - @JsonProperty("permission") - private AppResourceExperimentExperimentPermission permission; - - public AppResourceExperiment setExperimentId(String experimentId) { - this.experimentId = experimentId; - return this; - } - - public String getExperimentId() { - return experimentId; - } - - public AppResourceExperiment setPermission(AppResourceExperimentExperimentPermission permission) { - this.permission = permission; - return this; - } - - public AppResourceExperimentExperimentPermission getPermission() { - return permission; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - AppResourceExperiment that = (AppResourceExperiment) o; - return Objects.equals(experimentId, that.experimentId) - && Objects.equals(permission, that.permission); - } - - @Override - public int hashCode() { - return Objects.hash(experimentId, permission); - } - - @Override - public String toString() { - return new ToStringer(AppResourceExperiment.class) - .add("experimentId", experimentId) - .add("permission", permission) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceExperimentExperimentPermission.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceExperimentExperimentPermission.java deleted file mode 100644 index db12ec2b0..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceExperimentExperimentPermission.java +++ /dev/null @@ -1,12 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.apps; - -import com.databricks.sdk.support.Generated; - -@Generated -public enum AppResourceExperimentExperimentPermission { - CAN_EDIT, - CAN_MANAGE, - CAN_READ, -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceUcSecurableUcSecurablePermission.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceUcSecurableUcSecurablePermission.java index 6ed176b3c..551de2d21 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceUcSecurableUcSecurablePermission.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceUcSecurableUcSecurablePermission.java @@ -6,9 +6,6 @@ @Generated public enum AppResourceUcSecurableUcSecurablePermission { - EXECUTE, READ_VOLUME, - SELECT, - USE_CONNECTION, WRITE_VOLUME, } diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceUcSecurableUcSecurableType.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceUcSecurableUcSecurableType.java index c97224882..b05b5435e 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceUcSecurableUcSecurableType.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppResourceUcSecurableUcSecurableType.java @@ -6,8 +6,5 @@ @Generated public enum AppResourceUcSecurableUcSecurableType { - CONNECTION, - FUNCTION, - TABLE, VOLUME, } diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppUpdate.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppUpdate.java index ff4970a5a..b34c390e7 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppUpdate.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/AppUpdate.java @@ -22,10 +22,6 @@ public class AppUpdate { @JsonProperty("description") private String description; - /** */ - @JsonProperty("git_repository") - private GitRepository gitRepository; - /** */ @JsonProperty("resources") private Collection resources; @@ -69,15 +65,6 @@ public String getDescription() { return description; } - public AppUpdate setGitRepository(GitRepository gitRepository) { - this.gitRepository = gitRepository; - return this; - } - - public GitRepository getGitRepository() { - return gitRepository; - } - public AppUpdate setResources(Collection resources) { this.resources = resources; return this; @@ -122,7 +109,6 @@ public boolean equals(Object o) { return Objects.equals(budgetPolicyId, that.budgetPolicyId) && Objects.equals(computeSize, that.computeSize) && Objects.equals(description, that.description) - && Objects.equals(gitRepository, that.gitRepository) && Objects.equals(resources, that.resources) && Objects.equals(status, that.status) && Objects.equals(usagePolicyId, that.usagePolicyId) @@ -132,14 +118,7 @@ public boolean equals(Object o) { @Override public int hashCode() { return Objects.hash( - budgetPolicyId, - computeSize, - description, - gitRepository, - resources, - status, - usagePolicyId, - userApiScopes); + budgetPolicyId, computeSize, description, resources, status, usagePolicyId, userApiScopes); } @Override @@ -148,7 +127,6 @@ public String toString() { .add("budgetPolicyId", budgetPolicyId) .add("computeSize", computeSize) .add("description", description) - .add("gitRepository", gitRepository) .add("resources", resources) .add("status", status) .add("usagePolicyId", usagePolicyId) diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/EnvVar.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/EnvVar.java deleted file mode 100644 index 87ea4d481..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/EnvVar.java +++ /dev/null @@ -1,77 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.apps; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import java.util.Objects; - -@Generated -public class EnvVar { - /** The name of the environment variable. */ - @JsonProperty("name") - private String name; - - /** The value for the environment variable. */ - @JsonProperty("value") - private String value; - - /** - * The name of an external Databricks resource that contains the value, such as a secret or a - * database table. - */ - @JsonProperty("value_from") - private String valueFrom; - - public EnvVar setName(String name) { - this.name = name; - return this; - } - - public String getName() { - return name; - } - - public EnvVar setValue(String value) { - this.value = value; - return this; - } - - public String getValue() { - return value; - } - - public EnvVar setValueFrom(String valueFrom) { - this.valueFrom = valueFrom; - return this; - } - - public String getValueFrom() { - return valueFrom; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - EnvVar that = (EnvVar) o; - return Objects.equals(name, that.name) - && Objects.equals(value, that.value) - && Objects.equals(valueFrom, that.valueFrom); - } - - @Override - public int hashCode() { - return Objects.hash(name, value, valueFrom); - } - - @Override - public String toString() { - return new ToStringer(EnvVar.class) - .add("name", name) - .add("value", value) - .add("valueFrom", valueFrom) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GitRepository.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GitRepository.java deleted file mode 100644 index bd523a790..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GitRepository.java +++ /dev/null @@ -1,59 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.apps; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import java.util.Objects; - -/** Git repository configuration specifying the location of the repository. */ -@Generated -public class GitRepository { - /** - * Git provider. Case insensitive. Supported values: gitHub, gitHubEnterprise, bitbucketCloud, - * bitbucketServer, azureDevOpsServices, gitLab, gitLabEnterpriseEdition, awsCodeCommit. - */ - @JsonProperty("provider") - private String provider; - - /** URL of the Git repository. */ - @JsonProperty("url") - private String url; - - public GitRepository setProvider(String provider) { - this.provider = provider; - return this; - } - - public String getProvider() { - return provider; - } - - public GitRepository setUrl(String url) { - this.url = url; - return this; - } - - public String getUrl() { - return url; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - GitRepository that = (GitRepository) o; - return Objects.equals(provider, that.provider) && Objects.equals(url, that.url); - } - - @Override - public int hashCode() { - return Objects.hash(provider, url); - } - - @Override - public String toString() { - return new ToStringer(GitRepository.class).add("provider", provider).add("url", url).toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GitSource.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GitSource.java deleted file mode 100644 index 6b9728c1b..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/apps/GitSource.java +++ /dev/null @@ -1,128 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.apps; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import java.util.Objects; - -/** Complete git source specification including repository location and reference. */ -@Generated -public class GitSource { - /** Git branch to checkout. */ - @JsonProperty("branch") - private String branch; - - /** Git commit SHA to checkout. */ - @JsonProperty("commit") - private String commit; - - /** Git repository configuration. Populated from the app's git_repository configuration. */ - @JsonProperty("git_repository") - private GitRepository gitRepository; - - /** - * The resolved commit SHA that was actually used for the deployment. This is populated by the - * system after resolving the reference (branch, tag, or commit). If commit is specified directly, - * this will match commit. If a branch or tag is specified, this contains the commit SHA that the - * branch or tag pointed to at deployment time. - */ - @JsonProperty("resolved_commit") - private String resolvedCommit; - - /** - * Relative path to the app source code within the Git repository. If not specified, the root of - * the repository is used. - */ - @JsonProperty("source_code_path") - private String sourceCodePath; - - /** Git tag to checkout. */ - @JsonProperty("tag") - private String tag; - - public GitSource setBranch(String branch) { - this.branch = branch; - return this; - } - - public String getBranch() { - return branch; - } - - public GitSource setCommit(String commit) { - this.commit = commit; - return this; - } - - public String getCommit() { - return commit; - } - - public GitSource setGitRepository(GitRepository gitRepository) { - this.gitRepository = gitRepository; - return this; - } - - public GitRepository getGitRepository() { - return gitRepository; - } - - public GitSource setResolvedCommit(String resolvedCommit) { - this.resolvedCommit = resolvedCommit; - return this; - } - - public String getResolvedCommit() { - return resolvedCommit; - } - - public GitSource setSourceCodePath(String sourceCodePath) { - this.sourceCodePath = sourceCodePath; - return this; - } - - public String getSourceCodePath() { - return sourceCodePath; - } - - public GitSource setTag(String tag) { - this.tag = tag; - return this; - } - - public String getTag() { - return tag; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - GitSource that = (GitSource) o; - return Objects.equals(branch, that.branch) - && Objects.equals(commit, that.commit) - && Objects.equals(gitRepository, that.gitRepository) - && Objects.equals(resolvedCommit, that.resolvedCommit) - && Objects.equals(sourceCodePath, that.sourceCodePath) - && Objects.equals(tag, that.tag); - } - - @Override - public int hashCode() { - return Objects.hash(branch, commit, gitRepository, resolvedCommit, sourceCodePath, tag); - } - - @Override - public String toString() { - return new ToStringer(GitSource.class) - .add("branch", branch) - .add("commit", commit) - .add("gitRepository", gitRepository) - .add("resolvedCommit", resolvedCommit) - .add("sourceCodePath", sourceCodePath) - .add("tag", tag) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/AccessRequestDestinations.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/AccessRequestDestinations.java index bb3858506..79eca7689 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/AccessRequestDestinations.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/AccessRequestDestinations.java @@ -17,36 +17,14 @@ public class AccessRequestDestinations { @JsonProperty("are_any_destinations_hidden") private Boolean areAnyDestinationsHidden; - /** - * The source securable from which the destinations are inherited. Either the same value as - * securable (if destination is set directly on the securable) or the nearest parent securable - * with destinations set. - */ - @JsonProperty("destination_source_securable") - private Securable destinationSourceSecurable; - /** The access request destinations for the securable. */ @JsonProperty("destinations") private Collection destinations; - /** - * The full name of the securable. Redundant with the name in the securable object, but necessary - * for Terraform integration - */ - @JsonProperty("full_name") - private String fullName; - - /** The securable for which the access request destinations are being modified or read. */ + /** The securable for which the access request destinations are being retrieved. */ @JsonProperty("securable") private Securable securable; - /** - * The type of the securable. Redundant with the type in the securable object, but necessary for - * Terraform integration - */ - @JsonProperty("securable_type") - private String securableType; - public AccessRequestDestinations setAreAnyDestinationsHidden(Boolean areAnyDestinationsHidden) { this.areAnyDestinationsHidden = areAnyDestinationsHidden; return this; @@ -56,16 +34,6 @@ public Boolean getAreAnyDestinationsHidden() { return areAnyDestinationsHidden; } - public AccessRequestDestinations setDestinationSourceSecurable( - Securable destinationSourceSecurable) { - this.destinationSourceSecurable = destinationSourceSecurable; - return this; - } - - public Securable getDestinationSourceSecurable() { - return destinationSourceSecurable; - } - public AccessRequestDestinations setDestinations( Collection destinations) { this.destinations = destinations; @@ -76,15 +44,6 @@ public Collection getDestinations() { return destinations; } - public AccessRequestDestinations setFullName(String fullName) { - this.fullName = fullName; - return this; - } - - public String getFullName() { - return fullName; - } - public AccessRequestDestinations setSecurable(Securable securable) { this.securable = securable; return this; @@ -94,48 +53,27 @@ public Securable getSecurable() { return securable; } - public AccessRequestDestinations setSecurableType(String securableType) { - this.securableType = securableType; - return this; - } - - public String getSecurableType() { - return securableType; - } - @Override public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; AccessRequestDestinations that = (AccessRequestDestinations) o; return Objects.equals(areAnyDestinationsHidden, that.areAnyDestinationsHidden) - && Objects.equals(destinationSourceSecurable, that.destinationSourceSecurable) && Objects.equals(destinations, that.destinations) - && Objects.equals(fullName, that.fullName) - && Objects.equals(securable, that.securable) - && Objects.equals(securableType, that.securableType); + && Objects.equals(securable, that.securable); } @Override public int hashCode() { - return Objects.hash( - areAnyDestinationsHidden, - destinationSourceSecurable, - destinations, - fullName, - securable, - securableType); + return Objects.hash(areAnyDestinationsHidden, destinations, securable); } @Override public String toString() { return new ToStringer(AccessRequestDestinations.class) .add("areAnyDestinationsHidden", areAnyDestinationsHidden) - .add("destinationSourceSecurable", destinationSourceSecurable) .add("destinations", destinations) - .add("fullName", fullName) .add("securable", securable) - .add("securableType", securableType) .toString(); } } diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/ConnectionType.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/ConnectionType.java index 859b57ee1..933db59ac 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/ConnectionType.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/ConnectionType.java @@ -4,7 +4,7 @@ import com.databricks.sdk.support.Generated; -/** Next Id: 53 */ +/** Next Id: 52 */ @Generated public enum ConnectionType { BIGQUERY, diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SchemaInfo.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SchemaInfo.java index 44c7b1bc9..04137ed3e 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SchemaInfo.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SchemaInfo.java @@ -8,7 +8,7 @@ import java.util.Map; import java.util.Objects; -/** Next ID: 44 */ +/** Next ID: 43 */ @Generated public class SchemaInfo { /** diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SecurableKind.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SecurableKind.java index 41969e227..537945deb 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SecurableKind.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SecurableKind.java @@ -4,7 +4,7 @@ import com.databricks.sdk.support.Generated; -/** Latest kind: CONNECTION_ONELAKE = 289; Next id:290 */ +/** Latest kind: CONNECTION_GOOGLE_ADS_OAUTH_U2M_WITH_DT = 284; Next id:285 */ @Generated public enum SecurableKind { TABLE_DB_STORAGE, diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TableInfo.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TableInfo.java index cbe5577c2..182c3d0f0 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TableInfo.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TableInfo.java @@ -160,8 +160,7 @@ public class TableInfo { * View dependencies (when table_type == **VIEW** or **MATERIALIZED_VIEW**, **STREAMING_TABLE**) - * when DependencyList is None, the dependency is not provided; - when DependencyList is an empty * list, the dependency is provided but is empty; - when DependencyList is not an empty list, - * dependencies are provided and recorded. Note: this field is not set in the output of the - * __listTables__ API. + * dependencies are provided and recorded. */ @JsonProperty("view_dependencies") private DependencyList viewDependencies; diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TablesAPI.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TablesAPI.java index 5db6a605a..ac41bdf5a 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TablesAPI.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TablesAPI.java @@ -117,9 +117,6 @@ public Iterable list(String catalogName, String schemaName) { * privilege on the parent catalog and the **USE_SCHEMA** privilege on the parent schema. There is * no guarantee of a specific ordering of the elements in the array. * - *

NOTE: **view_dependencies** and **table_constraints** are not returned by ListTables - * queries. - * *

NOTE: we recommend using max_results=0 to use the paginated version of this API. Unpaginated * calls will be deprecated soon. * diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TablesService.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TablesService.java index 1efde490f..34f2b78ed 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TablesService.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TablesService.java @@ -79,9 +79,6 @@ public interface TablesService { * privilege on the parent catalog and the **USE_SCHEMA** privilege on the parent schema. There is * no guarantee of a specific ordering of the elements in the array. * - *

NOTE: **view_dependencies** and **table_constraints** are not returned by ListTables - * queries. - * *

NOTE: we recommend using max_results=0 to use the paginated version of this API. Unpaginated * calls will be deprecated soon. * diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/dataquality/AnomalyDetectionConfig.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/dataquality/AnomalyDetectionConfig.java index 3230194c5..6fd6e421c 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/dataquality/AnomalyDetectionConfig.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/dataquality/AnomalyDetectionConfig.java @@ -4,44 +4,26 @@ import com.databricks.sdk.support.Generated; import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import java.util.Collection; import java.util.Objects; /** Anomaly Detection Configurations. */ @Generated public class AnomalyDetectionConfig { - /** List of fully qualified table names to exclude from anomaly detection. */ - @JsonProperty("excluded_table_full_names") - private Collection excludedTableFullNames; - - public AnomalyDetectionConfig setExcludedTableFullNames( - Collection excludedTableFullNames) { - this.excludedTableFullNames = excludedTableFullNames; - return this; - } - - public Collection getExcludedTableFullNames() { - return excludedTableFullNames; - } @Override public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; - AnomalyDetectionConfig that = (AnomalyDetectionConfig) o; - return Objects.equals(excludedTableFullNames, that.excludedTableFullNames); + return true; } @Override public int hashCode() { - return Objects.hash(excludedTableFullNames); + return Objects.hash(); } @Override public String toString() { - return new ToStringer(AnomalyDetectionConfig.class) - .add("excludedTableFullNames", excludedTableFullNames) - .toString(); + return new ToStringer(AnomalyDetectionConfig.class).toString(); } } diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/JobRunAs.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/JobRunAs.java index dcbf39048..53dc83891 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/JobRunAs.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/JobRunAs.java @@ -16,13 +16,6 @@ */ @Generated public class JobRunAs { - /** - * Group name of an account group assigned to the workspace. Setting this field requires being a - * member of the group. - */ - @JsonProperty("group_name") - private String groupName; - /** * Application ID of an active service principal. Setting this field requires the * `servicePrincipal/user` role. @@ -37,15 +30,6 @@ public class JobRunAs { @JsonProperty("user_name") private String userName; - public JobRunAs setGroupName(String groupName) { - this.groupName = groupName; - return this; - } - - public String getGroupName() { - return groupName; - } - public JobRunAs setServicePrincipalName(String servicePrincipalName) { this.servicePrincipalName = servicePrincipalName; return this; @@ -69,20 +53,18 @@ public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; JobRunAs that = (JobRunAs) o; - return Objects.equals(groupName, that.groupName) - && Objects.equals(servicePrincipalName, that.servicePrincipalName) + return Objects.equals(servicePrincipalName, that.servicePrincipalName) && Objects.equals(userName, that.userName); } @Override public int hashCode() { - return Objects.hash(groupName, servicePrincipalName, userName); + return Objects.hash(servicePrincipalName, userName); } @Override public String toString() { return new ToStringer(JobRunAs.class) - .add("groupName", groupName) .add("servicePrincipalName", servicePrincipalName) .add("userName", userName) .toString(); diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/FeatureEngineeringAPI.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/FeatureEngineeringAPI.java index 64b85bd5d..14ebda0b2 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/FeatureEngineeringAPI.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/FeatureEngineeringAPI.java @@ -35,10 +35,7 @@ public Feature createFeature(CreateFeatureRequest request) { return impl.createFeature(request); } - /** - * Create a Kafka config. During PrPr, Kafka configs can be read and used when creating features - * under the entire metastore. Only the creator of the Kafka config can delete it. - */ + /** Create a Kafka config. */ public KafkaConfig createKafkaConfig(CreateKafkaConfigRequest request) { return impl.createKafkaConfig(request); } @@ -61,10 +58,7 @@ public void deleteKafkaConfig(String name) { deleteKafkaConfig(new DeleteKafkaConfigRequest().setName(name)); } - /** - * Delete a Kafka config. During PrPr, Kafka configs can be read and used when creating features - * under the entire metastore. Only the creator of the Kafka config can delete it. - */ + /** Delete a Kafka config. */ public void deleteKafkaConfig(DeleteKafkaConfigRequest request) { impl.deleteKafkaConfig(request); } @@ -92,10 +86,7 @@ public KafkaConfig getKafkaConfig(String name) { return getKafkaConfig(new GetKafkaConfigRequest().setName(name)); } - /** - * Get a Kafka config. During PrPr, Kafka configs can be read and used when creating features - * under the entire metastore. Only the creator of the Kafka config can delete it. - */ + /** Get a Kafka config. */ public KafkaConfig getKafkaConfig(GetKafkaConfigRequest request) { return impl.getKafkaConfig(request); } @@ -125,10 +116,7 @@ public Iterable listFeatures(ListFeaturesRequest request) { }); } - /** - * List Kafka configs. During PrPr, Kafka configs can be read and used when creating features - * under the entire metastore. Only the creator of the Kafka config can delete it. - */ + /** List Kafka configs. */ public Iterable listKafkaConfigs(ListKafkaConfigsRequest request) { return new Paginator<>( request, @@ -164,10 +152,7 @@ public Feature updateFeature(UpdateFeatureRequest request) { return impl.updateFeature(request); } - /** - * Update a Kafka config. During PrPr, Kafka configs can be read and used when creating features - * under the entire metastore. Only the creator of the Kafka config can delete it. - */ + /** Update a Kafka config. */ public KafkaConfig updateKafkaConfig(UpdateKafkaConfigRequest request) { return impl.updateKafkaConfig(request); } diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/FeatureEngineeringImpl.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/FeatureEngineeringImpl.java index 5a8cf37e7..4b26f4703 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/FeatureEngineeringImpl.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/FeatureEngineeringImpl.java @@ -95,7 +95,8 @@ public void deleteFeature(DeleteFeatureRequest request) { @Override public void deleteKafkaConfig(DeleteKafkaConfigRequest request) { String path = - String.format("/api/2.0/feature-engineering/features/kafka-configs/%s", request.getName()); + String.format( + "/api/2.0/feature-engineering/features/kafka-configs/kafka/%s", request.getName()); try { Request req = new Request("DELETE", path); diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/FeatureEngineeringService.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/FeatureEngineeringService.java index e99d67822..5a6d11387 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/FeatureEngineeringService.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/FeatureEngineeringService.java @@ -19,10 +19,7 @@ BatchCreateMaterializedFeaturesResponse batchCreateMaterializedFeatures( /** Create a Feature. */ Feature createFeature(CreateFeatureRequest createFeatureRequest); - /** - * Create a Kafka config. During PrPr, Kafka configs can be read and used when creating features - * under the entire metastore. Only the creator of the Kafka config can delete it. - */ + /** Create a Kafka config. */ KafkaConfig createKafkaConfig(CreateKafkaConfigRequest createKafkaConfigRequest); /** Create a materialized feature. */ @@ -32,10 +29,7 @@ MaterializedFeature createMaterializedFeature( /** Delete a Feature. */ void deleteFeature(DeleteFeatureRequest deleteFeatureRequest); - /** - * Delete a Kafka config. During PrPr, Kafka configs can be read and used when creating features - * under the entire metastore. Only the creator of the Kafka config can delete it. - */ + /** Delete a Kafka config. */ void deleteKafkaConfig(DeleteKafkaConfigRequest deleteKafkaConfigRequest); /** Delete a materialized feature. */ @@ -44,10 +38,7 @@ MaterializedFeature createMaterializedFeature( /** Get a Feature. */ Feature getFeature(GetFeatureRequest getFeatureRequest); - /** - * Get a Kafka config. During PrPr, Kafka configs can be read and used when creating features - * under the entire metastore. Only the creator of the Kafka config can delete it. - */ + /** Get a Kafka config. */ KafkaConfig getKafkaConfig(GetKafkaConfigRequest getKafkaConfigRequest); /** Get a materialized feature. */ @@ -57,10 +48,7 @@ MaterializedFeature getMaterializedFeature( /** List Features. */ ListFeaturesResponse listFeatures(ListFeaturesRequest listFeaturesRequest); - /** - * List Kafka configs. During PrPr, Kafka configs can be read and used when creating features - * under the entire metastore. Only the creator of the Kafka config can delete it. - */ + /** List Kafka configs. */ ListKafkaConfigsResponse listKafkaConfigs(ListKafkaConfigsRequest listKafkaConfigsRequest); /** List materialized features. */ @@ -70,10 +58,7 @@ ListMaterializedFeaturesResponse listMaterializedFeatures( /** Update a Feature. */ Feature updateFeature(UpdateFeatureRequest updateFeatureRequest); - /** - * Update a Kafka config. During PrPr, Kafka configs can be read and used when creating features - * under the entire metastore. Only the creator of the Kafka config can delete it. - */ + /** Update a Kafka config. */ KafkaConfig updateKafkaConfig(UpdateKafkaConfigRequest updateKafkaConfigRequest); /** Update a materialized feature (pause/resume). */ diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/pipelines/TableSpecificConfig.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/pipelines/TableSpecificConfig.java index 5a1f380e3..4e8dc6013 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/pipelines/TableSpecificConfig.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/pipelines/TableSpecificConfig.java @@ -37,13 +37,6 @@ public class TableSpecificConfig { private IngestionPipelineDefinitionTableSpecificConfigQueryBasedConnectorConfig queryBasedConnectorConfig; - /** - * (Optional, Immutable) The row filter condition to be applied to the table. It must not contain - * the WHERE keyword, only the actual filter condition. It must be in DBSQL format. - */ - @JsonProperty("row_filter") - private String rowFilter; - /** * If true, formula fields defined in the table are included in the ingestion. This setting is * only valid for the Salesforce connector @@ -105,15 +98,6 @@ public TableSpecificConfig setQueryBasedConnectorConfig( return queryBasedConnectorConfig; } - public TableSpecificConfig setRowFilter(String rowFilter) { - this.rowFilter = rowFilter; - return this; - } - - public String getRowFilter() { - return rowFilter; - } - public TableSpecificConfig setSalesforceIncludeFormulaFields( Boolean salesforceIncludeFormulaFields) { this.salesforceIncludeFormulaFields = salesforceIncludeFormulaFields; @@ -161,7 +145,6 @@ public boolean equals(Object o) { && Objects.equals(includeColumns, that.includeColumns) && Objects.equals(primaryKeys, that.primaryKeys) && Objects.equals(queryBasedConnectorConfig, that.queryBasedConnectorConfig) - && Objects.equals(rowFilter, that.rowFilter) && Objects.equals(salesforceIncludeFormulaFields, that.salesforceIncludeFormulaFields) && Objects.equals(scdType, that.scdType) && Objects.equals(sequenceBy, that.sequenceBy) @@ -175,7 +158,6 @@ public int hashCode() { includeColumns, primaryKeys, queryBasedConnectorConfig, - rowFilter, salesforceIncludeFormulaFields, scdType, sequenceBy, @@ -189,7 +171,6 @@ public String toString() { .add("includeColumns", includeColumns) .add("primaryKeys", primaryKeys) .add("queryBasedConnectorConfig", queryBasedConnectorConfig) - .add("rowFilter", rowFilter) .add("salesforceIncludeFormulaFields", salesforceIncludeFormulaFields) .add("scdType", scdType) .add("sequenceBy", sequenceBy) diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Branch.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Branch.java index ee3f7d02d..9bdb82f5d 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Branch.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Branch.java @@ -14,6 +14,49 @@ public class Branch { @JsonProperty("create_time") private Timestamp createTime; + /** The branch's state, indicating if it is initializing, ready for use, or archived. */ + @JsonProperty("current_state") + private BranchState currentState; + + /** + * Whether the branch is the project's default branch. This field is only returned on + * create/update responses. See effective_default for the value that is actually applied to the + * branch. + */ + @JsonProperty("default") + private Boolean defaultValue; + + /** Whether the branch is the project's default branch. */ + @JsonProperty("effective_default") + private Boolean effectiveDefault; + + /** Whether the branch is protected. */ + @JsonProperty("effective_is_protected") + private Boolean effectiveIsProtected; + + /** + * The name of the source branch from which this branch was created. Format: + * projects/{project_id}/branches/{branch_id} + */ + @JsonProperty("effective_source_branch") + private String effectiveSourceBranch; + + /** The Log Sequence Number (LSN) on the source branch from which this branch was created. */ + @JsonProperty("effective_source_branch_lsn") + private String effectiveSourceBranchLsn; + + /** The point in time on the source branch from which this branch was created. */ + @JsonProperty("effective_source_branch_time") + private Timestamp effectiveSourceBranchTime; + + /** Whether the branch is protected. */ + @JsonProperty("is_protected") + private Boolean isProtected; + + /** The logical size of the branch. */ + @JsonProperty("logical_size_bytes") + private Long logicalSizeBytes; + /** The resource name of the branch. Format: projects/{project_id}/branches/{branch_id} */ @JsonProperty("name") private String name; @@ -22,13 +65,28 @@ public class Branch { @JsonProperty("parent") private String parent; - /** The desired state of a Branch. */ - @JsonProperty("spec") - private BranchSpec spec; + /** The pending state of the branch, if a state transition is in progress. */ + @JsonProperty("pending_state") + private BranchState pendingState; - /** The current status of a Branch. */ - @JsonProperty("status") - private BranchStatus status; + /** + * The name of the source branch from which this branch was created. Format: + * projects/{project_id}/branches/{branch_id} + */ + @JsonProperty("source_branch") + private String sourceBranch; + + /** The Log Sequence Number (LSN) on the source branch from which this branch was created. */ + @JsonProperty("source_branch_lsn") + private String sourceBranchLsn; + + /** The point in time on the source branch from which this branch was created. */ + @JsonProperty("source_branch_time") + private Timestamp sourceBranchTime; + + /** A timestamp indicating when the `current_state` began. */ + @JsonProperty("state_change_time") + private Timestamp stateChangeTime; /** System generated unique ID for the branch. */ @JsonProperty("uid") @@ -47,6 +105,87 @@ public Timestamp getCreateTime() { return createTime; } + public Branch setCurrentState(BranchState currentState) { + this.currentState = currentState; + return this; + } + + public BranchState getCurrentState() { + return currentState; + } + + public Branch setDefault(Boolean defaultValue) { + this.defaultValue = defaultValue; + return this; + } + + public Boolean getDefault() { + return defaultValue; + } + + public Branch setEffectiveDefault(Boolean effectiveDefault) { + this.effectiveDefault = effectiveDefault; + return this; + } + + public Boolean getEffectiveDefault() { + return effectiveDefault; + } + + public Branch setEffectiveIsProtected(Boolean effectiveIsProtected) { + this.effectiveIsProtected = effectiveIsProtected; + return this; + } + + public Boolean getEffectiveIsProtected() { + return effectiveIsProtected; + } + + public Branch setEffectiveSourceBranch(String effectiveSourceBranch) { + this.effectiveSourceBranch = effectiveSourceBranch; + return this; + } + + public String getEffectiveSourceBranch() { + return effectiveSourceBranch; + } + + public Branch setEffectiveSourceBranchLsn(String effectiveSourceBranchLsn) { + this.effectiveSourceBranchLsn = effectiveSourceBranchLsn; + return this; + } + + public String getEffectiveSourceBranchLsn() { + return effectiveSourceBranchLsn; + } + + public Branch setEffectiveSourceBranchTime(Timestamp effectiveSourceBranchTime) { + this.effectiveSourceBranchTime = effectiveSourceBranchTime; + return this; + } + + public Timestamp getEffectiveSourceBranchTime() { + return effectiveSourceBranchTime; + } + + public Branch setIsProtected(Boolean isProtected) { + this.isProtected = isProtected; + return this; + } + + public Boolean getIsProtected() { + return isProtected; + } + + public Branch setLogicalSizeBytes(Long logicalSizeBytes) { + this.logicalSizeBytes = logicalSizeBytes; + return this; + } + + public Long getLogicalSizeBytes() { + return logicalSizeBytes; + } + public Branch setName(String name) { this.name = name; return this; @@ -65,22 +204,49 @@ public String getParent() { return parent; } - public Branch setSpec(BranchSpec spec) { - this.spec = spec; + public Branch setPendingState(BranchState pendingState) { + this.pendingState = pendingState; + return this; + } + + public BranchState getPendingState() { + return pendingState; + } + + public Branch setSourceBranch(String sourceBranch) { + this.sourceBranch = sourceBranch; + return this; + } + + public String getSourceBranch() { + return sourceBranch; + } + + public Branch setSourceBranchLsn(String sourceBranchLsn) { + this.sourceBranchLsn = sourceBranchLsn; + return this; + } + + public String getSourceBranchLsn() { + return sourceBranchLsn; + } + + public Branch setSourceBranchTime(Timestamp sourceBranchTime) { + this.sourceBranchTime = sourceBranchTime; return this; } - public BranchSpec getSpec() { - return spec; + public Timestamp getSourceBranchTime() { + return sourceBranchTime; } - public Branch setStatus(BranchStatus status) { - this.status = status; + public Branch setStateChangeTime(Timestamp stateChangeTime) { + this.stateChangeTime = stateChangeTime; return this; } - public BranchStatus getStatus() { - return status; + public Timestamp getStateChangeTime() { + return stateChangeTime; } public Branch setUid(String uid) { @@ -107,27 +273,70 @@ public boolean equals(Object o) { if (o == null || getClass() != o.getClass()) return false; Branch that = (Branch) o; return Objects.equals(createTime, that.createTime) + && Objects.equals(currentState, that.currentState) + && Objects.equals(defaultValue, that.defaultValue) + && Objects.equals(effectiveDefault, that.effectiveDefault) + && Objects.equals(effectiveIsProtected, that.effectiveIsProtected) + && Objects.equals(effectiveSourceBranch, that.effectiveSourceBranch) + && Objects.equals(effectiveSourceBranchLsn, that.effectiveSourceBranchLsn) + && Objects.equals(effectiveSourceBranchTime, that.effectiveSourceBranchTime) + && Objects.equals(isProtected, that.isProtected) + && Objects.equals(logicalSizeBytes, that.logicalSizeBytes) && Objects.equals(name, that.name) && Objects.equals(parent, that.parent) - && Objects.equals(spec, that.spec) - && Objects.equals(status, that.status) + && Objects.equals(pendingState, that.pendingState) + && Objects.equals(sourceBranch, that.sourceBranch) + && Objects.equals(sourceBranchLsn, that.sourceBranchLsn) + && Objects.equals(sourceBranchTime, that.sourceBranchTime) + && Objects.equals(stateChangeTime, that.stateChangeTime) && Objects.equals(uid, that.uid) && Objects.equals(updateTime, that.updateTime); } @Override public int hashCode() { - return Objects.hash(createTime, name, parent, spec, status, uid, updateTime); + return Objects.hash( + createTime, + currentState, + defaultValue, + effectiveDefault, + effectiveIsProtected, + effectiveSourceBranch, + effectiveSourceBranchLsn, + effectiveSourceBranchTime, + isProtected, + logicalSizeBytes, + name, + parent, + pendingState, + sourceBranch, + sourceBranchLsn, + sourceBranchTime, + stateChangeTime, + uid, + updateTime); } @Override public String toString() { return new ToStringer(Branch.class) .add("createTime", createTime) + .add("currentState", currentState) + .add("defaultValue", defaultValue) + .add("effectiveDefault", effectiveDefault) + .add("effectiveIsProtected", effectiveIsProtected) + .add("effectiveSourceBranch", effectiveSourceBranch) + .add("effectiveSourceBranchLsn", effectiveSourceBranchLsn) + .add("effectiveSourceBranchTime", effectiveSourceBranchTime) + .add("isProtected", isProtected) + .add("logicalSizeBytes", logicalSizeBytes) .add("name", name) .add("parent", parent) - .add("spec", spec) - .add("status", status) + .add("pendingState", pendingState) + .add("sourceBranch", sourceBranch) + .add("sourceBranchLsn", sourceBranchLsn) + .add("sourceBranchTime", sourceBranchTime) + .add("stateChangeTime", stateChangeTime) .add("uid", uid) .add("updateTime", updateTime) .toString(); diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchSpec.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchSpec.java deleted file mode 100644 index 28830ce4f..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchSpec.java +++ /dev/null @@ -1,108 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import com.google.protobuf.Timestamp; -import java.util.Objects; - -@Generated -public class BranchSpec { - /** Whether the branch is the project's default branch. */ - @JsonProperty("default") - private Boolean defaultValue; - - /** Whether the branch is protected. */ - @JsonProperty("is_protected") - private Boolean isProtected; - - /** - * The name of the source branch from which this branch was created. Format: - * projects/{project_id}/branches/{branch_id} - */ - @JsonProperty("source_branch") - private String sourceBranch; - - /** The Log Sequence Number (LSN) on the source branch from which this branch was created. */ - @JsonProperty("source_branch_lsn") - private String sourceBranchLsn; - - /** The point in time on the source branch from which this branch was created. */ - @JsonProperty("source_branch_time") - private Timestamp sourceBranchTime; - - public BranchSpec setDefault(Boolean defaultValue) { - this.defaultValue = defaultValue; - return this; - } - - public Boolean getDefault() { - return defaultValue; - } - - public BranchSpec setIsProtected(Boolean isProtected) { - this.isProtected = isProtected; - return this; - } - - public Boolean getIsProtected() { - return isProtected; - } - - public BranchSpec setSourceBranch(String sourceBranch) { - this.sourceBranch = sourceBranch; - return this; - } - - public String getSourceBranch() { - return sourceBranch; - } - - public BranchSpec setSourceBranchLsn(String sourceBranchLsn) { - this.sourceBranchLsn = sourceBranchLsn; - return this; - } - - public String getSourceBranchLsn() { - return sourceBranchLsn; - } - - public BranchSpec setSourceBranchTime(Timestamp sourceBranchTime) { - this.sourceBranchTime = sourceBranchTime; - return this; - } - - public Timestamp getSourceBranchTime() { - return sourceBranchTime; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - BranchSpec that = (BranchSpec) o; - return Objects.equals(defaultValue, that.defaultValue) - && Objects.equals(isProtected, that.isProtected) - && Objects.equals(sourceBranch, that.sourceBranch) - && Objects.equals(sourceBranchLsn, that.sourceBranchLsn) - && Objects.equals(sourceBranchTime, that.sourceBranchTime); - } - - @Override - public int hashCode() { - return Objects.hash(defaultValue, isProtected, sourceBranch, sourceBranchLsn, sourceBranchTime); - } - - @Override - public String toString() { - return new ToStringer(BranchSpec.class) - .add("defaultValue", defaultValue) - .add("isProtected", isProtected) - .add("sourceBranch", sourceBranch) - .add("sourceBranchLsn", sourceBranchLsn) - .add("sourceBranchTime", sourceBranchTime) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchStatusState.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchState.java similarity index 89% rename from databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchStatusState.java rename to databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchState.java index 5fa60cdd7..0a0cec938 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchStatusState.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchState.java @@ -6,7 +6,7 @@ /** The state of the database branch. */ @Generated -public enum BranchStatusState { +public enum BranchState { ARCHIVED, IMPORTING, INIT, diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchStatus.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchStatus.java deleted file mode 100644 index 041beb73e..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/BranchStatus.java +++ /dev/null @@ -1,177 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import com.google.protobuf.Timestamp; -import java.util.Objects; - -@Generated -public class BranchStatus { - /** The branch's state, indicating if it is initializing, ready for use, or archived. */ - @JsonProperty("current_state") - private BranchStatusState currentState; - - /** Whether the branch is the project's default branch. */ - @JsonProperty("default") - private Boolean defaultValue; - - /** Whether the branch is protected. */ - @JsonProperty("is_protected") - private Boolean isProtected; - - /** The logical size of the branch. */ - @JsonProperty("logical_size_bytes") - private Long logicalSizeBytes; - - /** The pending state of the branch, if a state transition is in progress. */ - @JsonProperty("pending_state") - private BranchStatusState pendingState; - - /** - * The name of the source branch from which this branch was created. Format: - * projects/{project_id}/branches/{branch_id} - */ - @JsonProperty("source_branch") - private String sourceBranch; - - /** The Log Sequence Number (LSN) on the source branch from which this branch was created. */ - @JsonProperty("source_branch_lsn") - private String sourceBranchLsn; - - /** The point in time on the source branch from which this branch was created. */ - @JsonProperty("source_branch_time") - private Timestamp sourceBranchTime; - - /** A timestamp indicating when the `current_state` began. */ - @JsonProperty("state_change_time") - private Timestamp stateChangeTime; - - public BranchStatus setCurrentState(BranchStatusState currentState) { - this.currentState = currentState; - return this; - } - - public BranchStatusState getCurrentState() { - return currentState; - } - - public BranchStatus setDefault(Boolean defaultValue) { - this.defaultValue = defaultValue; - return this; - } - - public Boolean getDefault() { - return defaultValue; - } - - public BranchStatus setIsProtected(Boolean isProtected) { - this.isProtected = isProtected; - return this; - } - - public Boolean getIsProtected() { - return isProtected; - } - - public BranchStatus setLogicalSizeBytes(Long logicalSizeBytes) { - this.logicalSizeBytes = logicalSizeBytes; - return this; - } - - public Long getLogicalSizeBytes() { - return logicalSizeBytes; - } - - public BranchStatus setPendingState(BranchStatusState pendingState) { - this.pendingState = pendingState; - return this; - } - - public BranchStatusState getPendingState() { - return pendingState; - } - - public BranchStatus setSourceBranch(String sourceBranch) { - this.sourceBranch = sourceBranch; - return this; - } - - public String getSourceBranch() { - return sourceBranch; - } - - public BranchStatus setSourceBranchLsn(String sourceBranchLsn) { - this.sourceBranchLsn = sourceBranchLsn; - return this; - } - - public String getSourceBranchLsn() { - return sourceBranchLsn; - } - - public BranchStatus setSourceBranchTime(Timestamp sourceBranchTime) { - this.sourceBranchTime = sourceBranchTime; - return this; - } - - public Timestamp getSourceBranchTime() { - return sourceBranchTime; - } - - public BranchStatus setStateChangeTime(Timestamp stateChangeTime) { - this.stateChangeTime = stateChangeTime; - return this; - } - - public Timestamp getStateChangeTime() { - return stateChangeTime; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - BranchStatus that = (BranchStatus) o; - return Objects.equals(currentState, that.currentState) - && Objects.equals(defaultValue, that.defaultValue) - && Objects.equals(isProtected, that.isProtected) - && Objects.equals(logicalSizeBytes, that.logicalSizeBytes) - && Objects.equals(pendingState, that.pendingState) - && Objects.equals(sourceBranch, that.sourceBranch) - && Objects.equals(sourceBranchLsn, that.sourceBranchLsn) - && Objects.equals(sourceBranchTime, that.sourceBranchTime) - && Objects.equals(stateChangeTime, that.stateChangeTime); - } - - @Override - public int hashCode() { - return Objects.hash( - currentState, - defaultValue, - isProtected, - logicalSizeBytes, - pendingState, - sourceBranch, - sourceBranchLsn, - sourceBranchTime, - stateChangeTime); - } - - @Override - public String toString() { - return new ToStringer(BranchStatus.class) - .add("currentState", currentState) - .add("defaultValue", defaultValue) - .add("isProtected", isProtected) - .add("logicalSizeBytes", logicalSizeBytes) - .add("pendingState", pendingState) - .add("sourceBranch", sourceBranch) - .add("sourceBranchLsn", sourceBranchLsn) - .add("sourceBranchTime", sourceBranchTime) - .add("stateChangeTime", stateChangeTime) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateRoleOperation.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateRoleOperation.java deleted file mode 100644 index fb229fe50..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateRoleOperation.java +++ /dev/null @@ -1,161 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.core.DatabricksException; -import com.databricks.sdk.core.utils.SerDeUtils; -import com.databricks.sdk.service.common.lro.LroOptions; -import com.databricks.sdk.support.Generated; -import com.fasterxml.jackson.core.JsonProcessingException; -import com.fasterxml.jackson.databind.JsonNode; -import com.fasterxml.jackson.databind.ObjectMapper; -import java.time.Duration; -import java.util.Optional; -import java.util.concurrent.TimeoutException; -import org.slf4j.Logger; -import org.slf4j.LoggerFactory; - -/** - * Wrapper for interacting with a long-running createRole operation. Provides methods to wait for - * completion, check status, cancel, and access metadata. - */ -@Generated -public class CreateRoleOperation { - private static final Logger LOG = LoggerFactory.getLogger(CreateRoleOperation.class); - - private final PostgresService impl; - private Operation operation; - private final ObjectMapper objectMapper; - - public CreateRoleOperation(PostgresService impl, Operation operation) { - this.impl = impl; - this.operation = operation; - this.objectMapper = SerDeUtils.createMapper(); - } - - /** - * Wait for the operation to complete and return the resulting Role. Waits indefinitely if no - * timeout is specified. - * - * @return the created Role - * @throws TimeoutException if the operation doesn't complete within the timeout - * @throws DatabricksException if the operation fails - */ - public Role waitForCompletion() throws TimeoutException { - return waitForCompletion(Optional.empty()); - } - - /** - * Wait for the operation to complete and return the resulting Role. - * - * @param options the options for configuring the wait behavior, can be empty for defaults - * @return the created Role - * @throws TimeoutException if the operation doesn't complete within the timeout - * @throws DatabricksException if the operation fails - */ - public Role waitForCompletion(Optional options) throws TimeoutException { - Optional timeout = options.flatMap(LroOptions::getTimeout); - long deadline = - timeout.isPresent() - ? System.currentTimeMillis() + timeout.get().toMillis() - : Long.MAX_VALUE; - String statusMessage = "polling operation..."; - int attempt = 1; - - while (System.currentTimeMillis() < deadline) { - // Refresh the operation state - refreshOperation(); - - if (operation.getDone() != null && operation.getDone()) { - // Operation completed, check for success or failure - if (operation.getError() != null) { - String errorMsg = "unknown error"; - if (operation.getError().getMessage() != null - && !operation.getError().getMessage().isEmpty()) { - errorMsg = operation.getError().getMessage(); - } - - if (operation.getError().getErrorCode() != null) { - errorMsg = String.format("[%s] %s", operation.getError().getErrorCode(), errorMsg); - } - - throw new DatabricksException("Operation failed: " + errorMsg); - } - - // Operation completed successfully, unmarshal response - if (operation.getResponse() == null) { - throw new DatabricksException("Operation completed but no response available"); - } - - try { - JsonNode responseJson = objectMapper.valueToTree(operation.getResponse()); - return objectMapper.treeToValue(responseJson, Role.class); - } catch (JsonProcessingException e) { - throw new DatabricksException("Failed to unmarshal role response: " + e.getMessage(), e); - } - } - - // Operation still in progress, wait before polling again - String prefix = String.format("operation=%s", operation.getName()); - int sleep = Math.min(attempt, 10); // sleep 10s max per attempt - LOG.info("{}: operation in progress (sleeping ~{}s)", prefix, sleep); - - try { - Thread.sleep((long) (sleep * 1000L + Math.random() * 1000)); - } catch (InterruptedException e) { - Thread.currentThread().interrupt(); - throw new DatabricksException("Current thread was interrupted", e); - } - attempt++; - } - - String timeoutMessage = - timeout.isPresent() - ? String.format("Operation timed out after %s: %s", timeout.get(), statusMessage) - : String.format("Operation timed out: %s", statusMessage); - throw new TimeoutException(timeoutMessage); - } - - /** - * Get the operation name. - * - * @return the operation name - */ - public String getName() { - return operation.getName(); - } - - /** - * Get the operation metadata. - * - * @return the operation metadata, or null if not available - * @throws DatabricksException if the metadata cannot be deserialized - */ - public RoleOperationMetadata getMetadata() { - if (operation.getMetadata() == null) { - return null; - } - - try { - JsonNode metadataJson = objectMapper.valueToTree(operation.getMetadata()); - return objectMapper.treeToValue(metadataJson, RoleOperationMetadata.class); - } catch (JsonProcessingException e) { - throw new DatabricksException("Failed to unmarshal operation metadata: " + e.getMessage(), e); - } - } - - /** - * Check if the operation is done. This method refreshes the operation state before checking. - * - * @return true if the operation is complete, false otherwise - * @throws DatabricksException if the status check fails - */ - public boolean isDone() { - refreshOperation(); - return operation.getDone() != null && operation.getDone(); - } - - /** Refresh the operation state by polling the server. */ - private void refreshOperation() { - operation = impl.getOperation(new GetOperationRequest().setName(operation.getName())); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateRoleRequest.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateRoleRequest.java deleted file mode 100644 index de2f68cf9..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/CreateRoleRequest.java +++ /dev/null @@ -1,82 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.QueryParam; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonIgnore; -import com.fasterxml.jackson.annotation.JsonProperty; -import java.util.Objects; - -@Generated -public class CreateRoleRequest { - /** The Branch where this Role is created. Format: projects/{project_id}/branches/{branch_id} */ - @JsonIgnore private String parent; - - /** The desired specification of a Role. */ - @JsonProperty("role") - private Role role; - - /** - * The ID to use for the Role, which will become the final component of the branch's resource - * name. This ID becomes the role in postgres. - * - *

This value should be 4-63 characters, and only use characters available in DNS names, as - * defined by RFC-1123 - */ - @JsonIgnore - @QueryParam("role_id") - private String roleId; - - public CreateRoleRequest setParent(String parent) { - this.parent = parent; - return this; - } - - public String getParent() { - return parent; - } - - public CreateRoleRequest setRole(Role role) { - this.role = role; - return this; - } - - public Role getRole() { - return role; - } - - public CreateRoleRequest setRoleId(String roleId) { - this.roleId = roleId; - return this; - } - - public String getRoleId() { - return roleId; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - CreateRoleRequest that = (CreateRoleRequest) o; - return Objects.equals(parent, that.parent) - && Objects.equals(role, that.role) - && Objects.equals(roleId, that.roleId); - } - - @Override - public int hashCode() { - return Objects.hash(parent, role, roleId); - } - - @Override - public String toString() { - return new ToStringer(CreateRoleRequest.class) - .add("parent", parent) - .add("role", role) - .add("roleId", roleId) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/DeleteRoleOperation.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/DeleteRoleOperation.java deleted file mode 100644 index 368527abd..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/DeleteRoleOperation.java +++ /dev/null @@ -1,161 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.core.DatabricksException; -import com.databricks.sdk.core.utils.SerDeUtils; -import com.databricks.sdk.service.common.lro.LroOptions; -import com.databricks.sdk.support.Generated; -import com.fasterxml.jackson.core.JsonProcessingException; -import com.fasterxml.jackson.databind.JsonNode; -import com.fasterxml.jackson.databind.ObjectMapper; -import java.time.Duration; -import java.util.Optional; -import java.util.concurrent.TimeoutException; -import org.slf4j.Logger; -import org.slf4j.LoggerFactory; - -/** - * Wrapper for interacting with a long-running deleteRole operation. Provides methods to wait for - * completion, check status, cancel, and access metadata. - */ -@Generated -public class DeleteRoleOperation { - private static final Logger LOG = LoggerFactory.getLogger(DeleteRoleOperation.class); - - private final PostgresService impl; - private Operation operation; - private final ObjectMapper objectMapper; - - public DeleteRoleOperation(PostgresService impl, Operation operation) { - this.impl = impl; - this.operation = operation; - this.objectMapper = SerDeUtils.createMapper(); - } - - /** - * Wait for the operation to complete and return the resulting . Waits indefinitely if no timeout - * is specified. - * - * @return the created - * @throws TimeoutException if the operation doesn't complete within the timeout - * @throws DatabricksException if the operation fails - */ - public void waitForCompletion() throws TimeoutException { - waitForCompletion(Optional.empty()); - } - - /** - * Wait for the operation to complete and return the resulting . - * - * @param options the options for configuring the wait behavior, can be empty for defaults - * @return the created - * @throws TimeoutException if the operation doesn't complete within the timeout - * @throws DatabricksException if the operation fails - */ - public void waitForCompletion(Optional options) throws TimeoutException { - Optional timeout = options.flatMap(LroOptions::getTimeout); - long deadline = - timeout.isPresent() - ? System.currentTimeMillis() + timeout.get().toMillis() - : Long.MAX_VALUE; - String statusMessage = "polling operation..."; - int attempt = 1; - - while (System.currentTimeMillis() < deadline) { - // Refresh the operation state - refreshOperation(); - - if (operation.getDone() != null && operation.getDone()) { - // Operation completed, check for success or failure - if (operation.getError() != null) { - String errorMsg = "unknown error"; - if (operation.getError().getMessage() != null - && !operation.getError().getMessage().isEmpty()) { - errorMsg = operation.getError().getMessage(); - } - - if (operation.getError().getErrorCode() != null) { - errorMsg = String.format("[%s] %s", operation.getError().getErrorCode(), errorMsg); - } - - throw new DatabricksException("Operation failed: " + errorMsg); - } - - // Operation completed successfully, unmarshal response - if (operation.getResponse() == null) { - throw new DatabricksException("Operation completed but no response available"); - } - - try { - JsonNode responseJson = objectMapper.valueToTree(operation.getResponse()); - objectMapper.treeToValue(responseJson, Void.class); - } catch (JsonProcessingException e) { - throw new DatabricksException("Failed to unmarshal response: " + e.getMessage(), e); - } - } - - // Operation still in progress, wait before polling again - String prefix = String.format("operation=%s", operation.getName()); - int sleep = Math.min(attempt, 10); // sleep 10s max per attempt - LOG.info("{}: operation in progress (sleeping ~{}s)", prefix, sleep); - - try { - Thread.sleep((long) (sleep * 1000L + Math.random() * 1000)); - } catch (InterruptedException e) { - Thread.currentThread().interrupt(); - throw new DatabricksException("Current thread was interrupted", e); - } - attempt++; - } - - String timeoutMessage = - timeout.isPresent() - ? String.format("Operation timed out after %s: %s", timeout.get(), statusMessage) - : String.format("Operation timed out: %s", statusMessage); - throw new TimeoutException(timeoutMessage); - } - - /** - * Get the operation name. - * - * @return the operation name - */ - public String getName() { - return operation.getName(); - } - - /** - * Get the operation metadata. - * - * @return the operation metadata, or null if not available - * @throws DatabricksException if the metadata cannot be deserialized - */ - public RoleOperationMetadata getMetadata() { - if (operation.getMetadata() == null) { - return null; - } - - try { - JsonNode metadataJson = objectMapper.valueToTree(operation.getMetadata()); - return objectMapper.treeToValue(metadataJson, RoleOperationMetadata.class); - } catch (JsonProcessingException e) { - throw new DatabricksException("Failed to unmarshal operation metadata: " + e.getMessage(), e); - } - } - - /** - * Check if the operation is done. This method refreshes the operation state before checking. - * - * @return true if the operation is complete, false otherwise - * @throws DatabricksException if the status check fails - */ - public boolean isDone() { - refreshOperation(); - return operation.getDone() != null && operation.getDone(); - } - - /** Refresh the operation state by polling the server. */ - private void refreshOperation() { - operation = impl.getOperation(new GetOperationRequest().setName(operation.getName())); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/DeleteRoleRequest.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/DeleteRoleRequest.java deleted file mode 100644 index bf2373fa6..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/DeleteRoleRequest.java +++ /dev/null @@ -1,71 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.QueryParam; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonIgnore; -import java.util.Objects; - -@Generated -public class DeleteRoleRequest { - /** - * The resource name of the postgres role. Format: - * projects/{project_id}/branch/{branch_id}/roles/{role_id} - */ - @JsonIgnore private String name; - - /** - * Reassign objects. If this is set, all objects owned by the role are reassigned to the role - * specified in this parameter. - * - *

NOTE: setting this requires spinning up a compute to succeed, since it involves running SQL - * queries. - * - *

TODO: #LKB-7187 implement reassign_owned_to on LBM side. This might end-up being a - * synchronous query when this parameter is used. - */ - @JsonIgnore - @QueryParam("reassign_owned_to") - private String reassignOwnedTo; - - public DeleteRoleRequest setName(String name) { - this.name = name; - return this; - } - - public String getName() { - return name; - } - - public DeleteRoleRequest setReassignOwnedTo(String reassignOwnedTo) { - this.reassignOwnedTo = reassignOwnedTo; - return this; - } - - public String getReassignOwnedTo() { - return reassignOwnedTo; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - DeleteRoleRequest that = (DeleteRoleRequest) o; - return Objects.equals(name, that.name) && Objects.equals(reassignOwnedTo, that.reassignOwnedTo); - } - - @Override - public int hashCode() { - return Objects.hash(name, reassignOwnedTo); - } - - @Override - public String toString() { - return new ToStringer(DeleteRoleRequest.class) - .add("name", name) - .add("reassignOwnedTo", reassignOwnedTo) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Endpoint.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Endpoint.java index 925c3766e..b8adfb44f 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Endpoint.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Endpoint.java @@ -5,15 +5,79 @@ import com.databricks.sdk.support.Generated; import com.databricks.sdk.support.ToStringer; import com.fasterxml.jackson.annotation.JsonProperty; +import com.google.protobuf.Duration; import com.google.protobuf.Timestamp; import java.util.Objects; @Generated public class Endpoint { + /** The maximum number of Compute Units. */ + @JsonProperty("autoscaling_limit_max_cu") + private Double autoscalingLimitMaxCu; + + /** The minimum number of Compute Units. */ + @JsonProperty("autoscaling_limit_min_cu") + private Double autoscalingLimitMinCu; + /** A timestamp indicating when the compute endpoint was created. */ @JsonProperty("create_time") private Timestamp createTime; + /** */ + @JsonProperty("current_state") + private EndpointState currentState; + + /** + * Whether to restrict connections to the compute endpoint. Enabling this option schedules a + * suspend compute operation. A disabled compute endpoint cannot be enabled by a connection or + * console action. + */ + @JsonProperty("disabled") + private Boolean disabled; + + /** The maximum number of Compute Units. */ + @JsonProperty("effective_autoscaling_limit_max_cu") + private Double effectiveAutoscalingLimitMaxCu; + + /** The minimum number of Compute Units. */ + @JsonProperty("effective_autoscaling_limit_min_cu") + private Double effectiveAutoscalingLimitMinCu; + + /** + * Whether to restrict connections to the compute endpoint. Enabling this option schedules a + * suspend compute operation. A disabled compute endpoint cannot be enabled by a connection or + * console action. + */ + @JsonProperty("effective_disabled") + private Boolean effectiveDisabled; + + /** */ + @JsonProperty("effective_pooler_mode") + private EndpointPoolerMode effectivePoolerMode; + + /** */ + @JsonProperty("effective_settings") + private EndpointSettings effectiveSettings; + + /** Duration of inactivity after which the compute endpoint is automatically suspended. */ + @JsonProperty("effective_suspend_timeout_duration") + private Duration effectiveSuspendTimeoutDuration; + + /** The endpoint type. There could be only one READ_WRITE endpoint per branch. */ + @JsonProperty("endpoint_type") + private EndpointType endpointType; + + /** + * The hostname of the compute endpoint. This is the hostname specified when connecting to a + * database. + */ + @JsonProperty("host") + private String host; + + /** A timestamp indicating when the compute endpoint was last active. */ + @JsonProperty("last_active_time") + private Timestamp lastActiveTime; + /** * The resource name of the endpoint. Format: * projects/{project_id}/branches/{branch_id}/endpoints/{endpoint_id} @@ -25,13 +89,29 @@ public class Endpoint { @JsonProperty("parent") private String parent; - /** The desired state of an Endpoint. */ - @JsonProperty("spec") - private EndpointSpec spec; + /** */ + @JsonProperty("pending_state") + private EndpointState pendingState; + + /** */ + @JsonProperty("pooler_mode") + private EndpointPoolerMode poolerMode; + + /** */ + @JsonProperty("settings") + private EndpointSettings settings; - /** The current status of an Endpoint. */ - @JsonProperty("status") - private EndpointStatus status; + /** A timestamp indicating when the compute endpoint was last started. */ + @JsonProperty("start_time") + private Timestamp startTime; + + /** A timestamp indicating when the compute endpoint was last suspended. */ + @JsonProperty("suspend_time") + private Timestamp suspendTime; + + /** Duration of inactivity after which the compute endpoint is automatically suspended. */ + @JsonProperty("suspend_timeout_duration") + private Duration suspendTimeoutDuration; /** System generated unique ID for the endpoint. */ @JsonProperty("uid") @@ -41,6 +121,24 @@ public class Endpoint { @JsonProperty("update_time") private Timestamp updateTime; + public Endpoint setAutoscalingLimitMaxCu(Double autoscalingLimitMaxCu) { + this.autoscalingLimitMaxCu = autoscalingLimitMaxCu; + return this; + } + + public Double getAutoscalingLimitMaxCu() { + return autoscalingLimitMaxCu; + } + + public Endpoint setAutoscalingLimitMinCu(Double autoscalingLimitMinCu) { + this.autoscalingLimitMinCu = autoscalingLimitMinCu; + return this; + } + + public Double getAutoscalingLimitMinCu() { + return autoscalingLimitMinCu; + } + public Endpoint setCreateTime(Timestamp createTime) { this.createTime = createTime; return this; @@ -50,6 +148,105 @@ public Timestamp getCreateTime() { return createTime; } + public Endpoint setCurrentState(EndpointState currentState) { + this.currentState = currentState; + return this; + } + + public EndpointState getCurrentState() { + return currentState; + } + + public Endpoint setDisabled(Boolean disabled) { + this.disabled = disabled; + return this; + } + + public Boolean getDisabled() { + return disabled; + } + + public Endpoint setEffectiveAutoscalingLimitMaxCu(Double effectiveAutoscalingLimitMaxCu) { + this.effectiveAutoscalingLimitMaxCu = effectiveAutoscalingLimitMaxCu; + return this; + } + + public Double getEffectiveAutoscalingLimitMaxCu() { + return effectiveAutoscalingLimitMaxCu; + } + + public Endpoint setEffectiveAutoscalingLimitMinCu(Double effectiveAutoscalingLimitMinCu) { + this.effectiveAutoscalingLimitMinCu = effectiveAutoscalingLimitMinCu; + return this; + } + + public Double getEffectiveAutoscalingLimitMinCu() { + return effectiveAutoscalingLimitMinCu; + } + + public Endpoint setEffectiveDisabled(Boolean effectiveDisabled) { + this.effectiveDisabled = effectiveDisabled; + return this; + } + + public Boolean getEffectiveDisabled() { + return effectiveDisabled; + } + + public Endpoint setEffectivePoolerMode(EndpointPoolerMode effectivePoolerMode) { + this.effectivePoolerMode = effectivePoolerMode; + return this; + } + + public EndpointPoolerMode getEffectivePoolerMode() { + return effectivePoolerMode; + } + + public Endpoint setEffectiveSettings(EndpointSettings effectiveSettings) { + this.effectiveSettings = effectiveSettings; + return this; + } + + public EndpointSettings getEffectiveSettings() { + return effectiveSettings; + } + + public Endpoint setEffectiveSuspendTimeoutDuration(Duration effectiveSuspendTimeoutDuration) { + this.effectiveSuspendTimeoutDuration = effectiveSuspendTimeoutDuration; + return this; + } + + public Duration getEffectiveSuspendTimeoutDuration() { + return effectiveSuspendTimeoutDuration; + } + + public Endpoint setEndpointType(EndpointType endpointType) { + this.endpointType = endpointType; + return this; + } + + public EndpointType getEndpointType() { + return endpointType; + } + + public Endpoint setHost(String host) { + this.host = host; + return this; + } + + public String getHost() { + return host; + } + + public Endpoint setLastActiveTime(Timestamp lastActiveTime) { + this.lastActiveTime = lastActiveTime; + return this; + } + + public Timestamp getLastActiveTime() { + return lastActiveTime; + } + public Endpoint setName(String name) { this.name = name; return this; @@ -68,22 +265,58 @@ public String getParent() { return parent; } - public Endpoint setSpec(EndpointSpec spec) { - this.spec = spec; + public Endpoint setPendingState(EndpointState pendingState) { + this.pendingState = pendingState; + return this; + } + + public EndpointState getPendingState() { + return pendingState; + } + + public Endpoint setPoolerMode(EndpointPoolerMode poolerMode) { + this.poolerMode = poolerMode; + return this; + } + + public EndpointPoolerMode getPoolerMode() { + return poolerMode; + } + + public Endpoint setSettings(EndpointSettings settings) { + this.settings = settings; + return this; + } + + public EndpointSettings getSettings() { + return settings; + } + + public Endpoint setStartTime(Timestamp startTime) { + this.startTime = startTime; + return this; + } + + public Timestamp getStartTime() { + return startTime; + } + + public Endpoint setSuspendTime(Timestamp suspendTime) { + this.suspendTime = suspendTime; return this; } - public EndpointSpec getSpec() { - return spec; + public Timestamp getSuspendTime() { + return suspendTime; } - public Endpoint setStatus(EndpointStatus status) { - this.status = status; + public Endpoint setSuspendTimeoutDuration(Duration suspendTimeoutDuration) { + this.suspendTimeoutDuration = suspendTimeoutDuration; return this; } - public EndpointStatus getStatus() { - return status; + public Duration getSuspendTimeoutDuration() { + return suspendTimeoutDuration; } public Endpoint setUid(String uid) { @@ -109,28 +342,86 @@ public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; Endpoint that = (Endpoint) o; - return Objects.equals(createTime, that.createTime) + return Objects.equals(autoscalingLimitMaxCu, that.autoscalingLimitMaxCu) + && Objects.equals(autoscalingLimitMinCu, that.autoscalingLimitMinCu) + && Objects.equals(createTime, that.createTime) + && Objects.equals(currentState, that.currentState) + && Objects.equals(disabled, that.disabled) + && Objects.equals(effectiveAutoscalingLimitMaxCu, that.effectiveAutoscalingLimitMaxCu) + && Objects.equals(effectiveAutoscalingLimitMinCu, that.effectiveAutoscalingLimitMinCu) + && Objects.equals(effectiveDisabled, that.effectiveDisabled) + && Objects.equals(effectivePoolerMode, that.effectivePoolerMode) + && Objects.equals(effectiveSettings, that.effectiveSettings) + && Objects.equals(effectiveSuspendTimeoutDuration, that.effectiveSuspendTimeoutDuration) + && Objects.equals(endpointType, that.endpointType) + && Objects.equals(host, that.host) + && Objects.equals(lastActiveTime, that.lastActiveTime) && Objects.equals(name, that.name) && Objects.equals(parent, that.parent) - && Objects.equals(spec, that.spec) - && Objects.equals(status, that.status) + && Objects.equals(pendingState, that.pendingState) + && Objects.equals(poolerMode, that.poolerMode) + && Objects.equals(settings, that.settings) + && Objects.equals(startTime, that.startTime) + && Objects.equals(suspendTime, that.suspendTime) + && Objects.equals(suspendTimeoutDuration, that.suspendTimeoutDuration) && Objects.equals(uid, that.uid) && Objects.equals(updateTime, that.updateTime); } @Override public int hashCode() { - return Objects.hash(createTime, name, parent, spec, status, uid, updateTime); + return Objects.hash( + autoscalingLimitMaxCu, + autoscalingLimitMinCu, + createTime, + currentState, + disabled, + effectiveAutoscalingLimitMaxCu, + effectiveAutoscalingLimitMinCu, + effectiveDisabled, + effectivePoolerMode, + effectiveSettings, + effectiveSuspendTimeoutDuration, + endpointType, + host, + lastActiveTime, + name, + parent, + pendingState, + poolerMode, + settings, + startTime, + suspendTime, + suspendTimeoutDuration, + uid, + updateTime); } @Override public String toString() { return new ToStringer(Endpoint.class) + .add("autoscalingLimitMaxCu", autoscalingLimitMaxCu) + .add("autoscalingLimitMinCu", autoscalingLimitMinCu) .add("createTime", createTime) + .add("currentState", currentState) + .add("disabled", disabled) + .add("effectiveAutoscalingLimitMaxCu", effectiveAutoscalingLimitMaxCu) + .add("effectiveAutoscalingLimitMinCu", effectiveAutoscalingLimitMinCu) + .add("effectiveDisabled", effectiveDisabled) + .add("effectivePoolerMode", effectivePoolerMode) + .add("effectiveSettings", effectiveSettings) + .add("effectiveSuspendTimeoutDuration", effectiveSuspendTimeoutDuration) + .add("endpointType", endpointType) + .add("host", host) + .add("lastActiveTime", lastActiveTime) .add("name", name) .add("parent", parent) - .add("spec", spec) - .add("status", status) + .add("pendingState", pendingState) + .add("poolerMode", poolerMode) + .add("settings", settings) + .add("startTime", startTime) + .add("suspendTime", suspendTime) + .add("suspendTimeoutDuration", suspendTimeoutDuration) .add("uid", uid) .add("updateTime", updateTime) .toString(); diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointSpec.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointSpec.java deleted file mode 100644 index 74fe3672b..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointSpec.java +++ /dev/null @@ -1,146 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import com.google.protobuf.Duration; -import java.util.Objects; - -@Generated -public class EndpointSpec { - /** The maximum number of Compute Units. */ - @JsonProperty("autoscaling_limit_max_cu") - private Double autoscalingLimitMaxCu; - - /** The minimum number of Compute Units. */ - @JsonProperty("autoscaling_limit_min_cu") - private Double autoscalingLimitMinCu; - - /** - * Whether to restrict connections to the compute endpoint. Enabling this option schedules a - * suspend compute operation. A disabled compute endpoint cannot be enabled by a connection or - * console action. - */ - @JsonProperty("disabled") - private Boolean disabled; - - /** The endpoint type. A branch can only have one READ_WRITE endpoint. */ - @JsonProperty("endpoint_type") - private EndpointType endpointType; - - /** */ - @JsonProperty("pooler_mode") - private EndpointPoolerMode poolerMode; - - /** */ - @JsonProperty("settings") - private EndpointSettings settings; - - /** Duration of inactivity after which the compute endpoint is automatically suspended. */ - @JsonProperty("suspend_timeout_duration") - private Duration suspendTimeoutDuration; - - public EndpointSpec setAutoscalingLimitMaxCu(Double autoscalingLimitMaxCu) { - this.autoscalingLimitMaxCu = autoscalingLimitMaxCu; - return this; - } - - public Double getAutoscalingLimitMaxCu() { - return autoscalingLimitMaxCu; - } - - public EndpointSpec setAutoscalingLimitMinCu(Double autoscalingLimitMinCu) { - this.autoscalingLimitMinCu = autoscalingLimitMinCu; - return this; - } - - public Double getAutoscalingLimitMinCu() { - return autoscalingLimitMinCu; - } - - public EndpointSpec setDisabled(Boolean disabled) { - this.disabled = disabled; - return this; - } - - public Boolean getDisabled() { - return disabled; - } - - public EndpointSpec setEndpointType(EndpointType endpointType) { - this.endpointType = endpointType; - return this; - } - - public EndpointType getEndpointType() { - return endpointType; - } - - public EndpointSpec setPoolerMode(EndpointPoolerMode poolerMode) { - this.poolerMode = poolerMode; - return this; - } - - public EndpointPoolerMode getPoolerMode() { - return poolerMode; - } - - public EndpointSpec setSettings(EndpointSettings settings) { - this.settings = settings; - return this; - } - - public EndpointSettings getSettings() { - return settings; - } - - public EndpointSpec setSuspendTimeoutDuration(Duration suspendTimeoutDuration) { - this.suspendTimeoutDuration = suspendTimeoutDuration; - return this; - } - - public Duration getSuspendTimeoutDuration() { - return suspendTimeoutDuration; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - EndpointSpec that = (EndpointSpec) o; - return Objects.equals(autoscalingLimitMaxCu, that.autoscalingLimitMaxCu) - && Objects.equals(autoscalingLimitMinCu, that.autoscalingLimitMinCu) - && Objects.equals(disabled, that.disabled) - && Objects.equals(endpointType, that.endpointType) - && Objects.equals(poolerMode, that.poolerMode) - && Objects.equals(settings, that.settings) - && Objects.equals(suspendTimeoutDuration, that.suspendTimeoutDuration); - } - - @Override - public int hashCode() { - return Objects.hash( - autoscalingLimitMaxCu, - autoscalingLimitMinCu, - disabled, - endpointType, - poolerMode, - settings, - suspendTimeoutDuration); - } - - @Override - public String toString() { - return new ToStringer(EndpointSpec.class) - .add("autoscalingLimitMaxCu", autoscalingLimitMaxCu) - .add("autoscalingLimitMinCu", autoscalingLimitMinCu) - .add("disabled", disabled) - .add("endpointType", endpointType) - .add("poolerMode", poolerMode) - .add("settings", settings) - .add("suspendTimeoutDuration", suspendTimeoutDuration) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointStatusState.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointState.java similarity index 88% rename from databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointStatusState.java rename to databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointState.java index 57714e82b..b0d141670 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointStatusState.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointState.java @@ -6,7 +6,7 @@ /** The state of the compute endpoint. */ @Generated -public enum EndpointStatusState { +public enum EndpointState { ACTIVE, IDLE, INIT, diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointStatus.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointStatus.java deleted file mode 100644 index 4781f087d..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/EndpointStatus.java +++ /dev/null @@ -1,246 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import com.google.protobuf.Duration; -import com.google.protobuf.Timestamp; -import java.util.Objects; - -@Generated -public class EndpointStatus { - /** The maximum number of Compute Units. */ - @JsonProperty("autoscaling_limit_max_cu") - private Double autoscalingLimitMaxCu; - - /** The minimum number of Compute Units. */ - @JsonProperty("autoscaling_limit_min_cu") - private Double autoscalingLimitMinCu; - - /** */ - @JsonProperty("current_state") - private EndpointStatusState currentState; - - /** - * Whether to restrict connections to the compute endpoint. Enabling this option schedules a - * suspend compute operation. A disabled compute endpoint cannot be enabled by a connection or - * console action. - */ - @JsonProperty("disabled") - private Boolean disabled; - - /** The endpoint type. A branch can only have one READ_WRITE endpoint. */ - @JsonProperty("endpoint_type") - private EndpointType endpointType; - - /** - * The hostname of the compute endpoint. This is the hostname specified when connecting to a - * database. - */ - @JsonProperty("host") - private String host; - - /** A timestamp indicating when the compute endpoint was last active. */ - @JsonProperty("last_active_time") - private Timestamp lastActiveTime; - - /** */ - @JsonProperty("pending_state") - private EndpointStatusState pendingState; - - /** */ - @JsonProperty("pooler_mode") - private EndpointPoolerMode poolerMode; - - /** */ - @JsonProperty("settings") - private EndpointSettings settings; - - /** A timestamp indicating when the compute endpoint was last started. */ - @JsonProperty("start_time") - private Timestamp startTime; - - /** A timestamp indicating when the compute endpoint was last suspended. */ - @JsonProperty("suspend_time") - private Timestamp suspendTime; - - /** Duration of inactivity after which the compute endpoint is automatically suspended. */ - @JsonProperty("suspend_timeout_duration") - private Duration suspendTimeoutDuration; - - public EndpointStatus setAutoscalingLimitMaxCu(Double autoscalingLimitMaxCu) { - this.autoscalingLimitMaxCu = autoscalingLimitMaxCu; - return this; - } - - public Double getAutoscalingLimitMaxCu() { - return autoscalingLimitMaxCu; - } - - public EndpointStatus setAutoscalingLimitMinCu(Double autoscalingLimitMinCu) { - this.autoscalingLimitMinCu = autoscalingLimitMinCu; - return this; - } - - public Double getAutoscalingLimitMinCu() { - return autoscalingLimitMinCu; - } - - public EndpointStatus setCurrentState(EndpointStatusState currentState) { - this.currentState = currentState; - return this; - } - - public EndpointStatusState getCurrentState() { - return currentState; - } - - public EndpointStatus setDisabled(Boolean disabled) { - this.disabled = disabled; - return this; - } - - public Boolean getDisabled() { - return disabled; - } - - public EndpointStatus setEndpointType(EndpointType endpointType) { - this.endpointType = endpointType; - return this; - } - - public EndpointType getEndpointType() { - return endpointType; - } - - public EndpointStatus setHost(String host) { - this.host = host; - return this; - } - - public String getHost() { - return host; - } - - public EndpointStatus setLastActiveTime(Timestamp lastActiveTime) { - this.lastActiveTime = lastActiveTime; - return this; - } - - public Timestamp getLastActiveTime() { - return lastActiveTime; - } - - public EndpointStatus setPendingState(EndpointStatusState pendingState) { - this.pendingState = pendingState; - return this; - } - - public EndpointStatusState getPendingState() { - return pendingState; - } - - public EndpointStatus setPoolerMode(EndpointPoolerMode poolerMode) { - this.poolerMode = poolerMode; - return this; - } - - public EndpointPoolerMode getPoolerMode() { - return poolerMode; - } - - public EndpointStatus setSettings(EndpointSettings settings) { - this.settings = settings; - return this; - } - - public EndpointSettings getSettings() { - return settings; - } - - public EndpointStatus setStartTime(Timestamp startTime) { - this.startTime = startTime; - return this; - } - - public Timestamp getStartTime() { - return startTime; - } - - public EndpointStatus setSuspendTime(Timestamp suspendTime) { - this.suspendTime = suspendTime; - return this; - } - - public Timestamp getSuspendTime() { - return suspendTime; - } - - public EndpointStatus setSuspendTimeoutDuration(Duration suspendTimeoutDuration) { - this.suspendTimeoutDuration = suspendTimeoutDuration; - return this; - } - - public Duration getSuspendTimeoutDuration() { - return suspendTimeoutDuration; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - EndpointStatus that = (EndpointStatus) o; - return Objects.equals(autoscalingLimitMaxCu, that.autoscalingLimitMaxCu) - && Objects.equals(autoscalingLimitMinCu, that.autoscalingLimitMinCu) - && Objects.equals(currentState, that.currentState) - && Objects.equals(disabled, that.disabled) - && Objects.equals(endpointType, that.endpointType) - && Objects.equals(host, that.host) - && Objects.equals(lastActiveTime, that.lastActiveTime) - && Objects.equals(pendingState, that.pendingState) - && Objects.equals(poolerMode, that.poolerMode) - && Objects.equals(settings, that.settings) - && Objects.equals(startTime, that.startTime) - && Objects.equals(suspendTime, that.suspendTime) - && Objects.equals(suspendTimeoutDuration, that.suspendTimeoutDuration); - } - - @Override - public int hashCode() { - return Objects.hash( - autoscalingLimitMaxCu, - autoscalingLimitMinCu, - currentState, - disabled, - endpointType, - host, - lastActiveTime, - pendingState, - poolerMode, - settings, - startTime, - suspendTime, - suspendTimeoutDuration); - } - - @Override - public String toString() { - return new ToStringer(EndpointStatus.class) - .add("autoscalingLimitMaxCu", autoscalingLimitMaxCu) - .add("autoscalingLimitMinCu", autoscalingLimitMinCu) - .add("currentState", currentState) - .add("disabled", disabled) - .add("endpointType", endpointType) - .add("host", host) - .add("lastActiveTime", lastActiveTime) - .add("pendingState", pendingState) - .add("poolerMode", poolerMode) - .add("settings", settings) - .add("startTime", startTime) - .add("suspendTime", suspendTime) - .add("suspendTimeoutDuration", suspendTimeoutDuration) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/GetRoleRequest.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/GetRoleRequest.java deleted file mode 100644 index 18ab6a7b0..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/GetRoleRequest.java +++ /dev/null @@ -1,44 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonIgnore; -import java.util.Objects; - -@Generated -public class GetRoleRequest { - /** - * The name of the Role to retrieve. Format: - * projects/{project_id}/branches/{branch_id}/roles/{role_id} - */ - @JsonIgnore private String name; - - public GetRoleRequest setName(String name) { - this.name = name; - return this; - } - - public String getName() { - return name; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - GetRoleRequest that = (GetRoleRequest) o; - return Objects.equals(name, that.name); - } - - @Override - public int hashCode() { - return Objects.hash(name); - } - - @Override - public String toString() { - return new ToStringer(GetRoleRequest.class).add("name", name).toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListRolesRequest.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListRolesRequest.java deleted file mode 100644 index 4b28a9b33..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListRolesRequest.java +++ /dev/null @@ -1,79 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.QueryParam; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonIgnore; -import java.util.Objects; - -@Generated -public class ListRolesRequest { - /** Upper bound for items returned. */ - @JsonIgnore - @QueryParam("page_size") - private Long pageSize; - - /** Pagination token to go to the next page of Roles. Requests first page if absent. */ - @JsonIgnore - @QueryParam("page_token") - private String pageToken; - - /** - * The Branch that owns this collection of roles. Format: - * projects/{project_id}/branches/{branch_id} - */ - @JsonIgnore private String parent; - - public ListRolesRequest setPageSize(Long pageSize) { - this.pageSize = pageSize; - return this; - } - - public Long getPageSize() { - return pageSize; - } - - public ListRolesRequest setPageToken(String pageToken) { - this.pageToken = pageToken; - return this; - } - - public String getPageToken() { - return pageToken; - } - - public ListRolesRequest setParent(String parent) { - this.parent = parent; - return this; - } - - public String getParent() { - return parent; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - ListRolesRequest that = (ListRolesRequest) o; - return Objects.equals(pageSize, that.pageSize) - && Objects.equals(pageToken, that.pageToken) - && Objects.equals(parent, that.parent); - } - - @Override - public int hashCode() { - return Objects.hash(pageSize, pageToken, parent); - } - - @Override - public String toString() { - return new ToStringer(ListRolesRequest.class) - .add("pageSize", pageSize) - .add("pageToken", pageToken) - .add("parent", parent) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListRolesResponse.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListRolesResponse.java deleted file mode 100644 index 1ca00fa00..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ListRolesResponse.java +++ /dev/null @@ -1,59 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import java.util.Collection; -import java.util.Objects; - -@Generated -public class ListRolesResponse { - /** Pagination token to request the next page of roles. */ - @JsonProperty("next_page_token") - private String nextPageToken; - - /** List of roles. */ - @JsonProperty("roles") - private Collection roles; - - public ListRolesResponse setNextPageToken(String nextPageToken) { - this.nextPageToken = nextPageToken; - return this; - } - - public String getNextPageToken() { - return nextPageToken; - } - - public ListRolesResponse setRoles(Collection roles) { - this.roles = roles; - return this; - } - - public Collection getRoles() { - return roles; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - ListRolesResponse that = (ListRolesResponse) o; - return Objects.equals(nextPageToken, that.nextPageToken) && Objects.equals(roles, that.roles); - } - - @Override - public int hashCode() { - return Objects.hash(nextPageToken, roles); - } - - @Override - public String toString() { - return new ToStringer(ListRolesResponse.class) - .add("nextPageToken", nextPageToken) - .add("roles", roles) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/PostgresAPI.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/PostgresAPI.java index b41027c27..b2fe70422 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/PostgresAPI.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/PostgresAPI.java @@ -42,12 +42,6 @@ public CreateProjectOperation createProject(CreateProjectRequest request) { return new CreateProjectOperation(impl, operation); } - /** Create a role for a branch. */ - public CreateRoleOperation createRole(CreateRoleRequest request) { - Operation operation = impl.createRole(request); - return new CreateRoleOperation(impl, operation); - } - public void deleteBranch(String name) { deleteBranch(new DeleteBranchRequest().setName(name)); } @@ -75,16 +69,6 @@ public void deleteProject(DeleteProjectRequest request) { impl.deleteProject(request); } - public DeleteRoleOperation deleteRole(String name) { - return deleteRole(new DeleteRoleRequest().setName(name)); - } - - /** Delete a role in a branch. */ - public DeleteRoleOperation deleteRole(DeleteRoleRequest request) { - Operation operation = impl.deleteRole(request); - return new DeleteRoleOperation(impl, operation); - } - public Branch getBranch(String name) { return getBranch(new GetBranchRequest().setName(name)); } @@ -121,15 +105,6 @@ public Project getProject(GetProjectRequest request) { return impl.getProject(request); } - public Role getRole(String name) { - return getRole(new GetRoleRequest().setName(name)); - } - - /** Get a Role. */ - public Role getRole(GetRoleRequest request) { - return impl.getRole(request); - } - public Iterable listBranches(String parent) { return listBranches(new ListBranchesRequest().setParent(parent)); } @@ -183,25 +158,6 @@ public Iterable listProjects(ListProjectsRequest request) { }); } - public Iterable listRoles(String parent) { - return listRoles(new ListRolesRequest().setParent(parent)); - } - - /** List Roles. */ - public Iterable listRoles(ListRolesRequest request) { - return new Paginator<>( - request, - impl::listRoles, - ListRolesResponse::getRoles, - response -> { - String token = response.getNextPageToken(); - if (token == null || token.isEmpty()) { - return null; - } - return request.setPageToken(token); - }); - } - /** Update a Branch. */ public UpdateBranchOperation updateBranch(UpdateBranchRequest request) { Operation operation = impl.updateBranch(request); diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/PostgresImpl.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/PostgresImpl.java index a91a37155..fe18af789 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/PostgresImpl.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/PostgresImpl.java @@ -61,21 +61,6 @@ public Operation createProject(CreateProjectRequest request) { } } - @Override - public Operation createRole(CreateRoleRequest request) { - String path = String.format("/api/2.0/postgres/%s/roles", request.getParent()); - try { - Request req = new Request("PATCH", path, apiClient.serialize(request.getRole())); - - ApiClient.setQuery(req, request); - req.withHeader("Accept", "application/json"); - req.withHeader("Content-Type", "application/json"); - return apiClient.execute(req, Operation.class); - } catch (IOException e) { - throw new DatabricksException("IO error: " + e.getMessage(), e); - } - } - @Override public void deleteBranch(DeleteBranchRequest request) { String path = String.format("/api/2.0/postgres/%s", request.getName()); @@ -118,20 +103,6 @@ public void deleteProject(DeleteProjectRequest request) { } } - @Override - public Operation deleteRole(DeleteRoleRequest request) { - String path = String.format("/api/2.0/postgres/%s", request.getName()); - try { - Request req = new Request("DELETE", path); - - ApiClient.setQuery(req, request); - req.withHeader("Accept", "application/json"); - return apiClient.execute(req, Operation.class); - } catch (IOException e) { - throw new DatabricksException("IO error: " + e.getMessage(), e); - } - } - @Override public Branch getBranch(GetBranchRequest request) { String path = String.format("/api/2.0/postgres/%s", request.getName()); @@ -188,20 +159,6 @@ public Project getProject(GetProjectRequest request) { } } - @Override - public Role getRole(GetRoleRequest request) { - String path = String.format("/api/2.0/postgres/%s", request.getName()); - try { - Request req = new Request("GET", path); - - ApiClient.setQuery(req, request); - req.withHeader("Accept", "application/json"); - return apiClient.execute(req, Role.class); - } catch (IOException e) { - throw new DatabricksException("IO error: " + e.getMessage(), e); - } - } - @Override public ListBranchesResponse listBranches(ListBranchesRequest request) { String path = String.format("/api/2.0/postgres/%s/branches", request.getParent()); @@ -244,20 +201,6 @@ public ListProjectsResponse listProjects(ListProjectsRequest request) { } } - @Override - public ListRolesResponse listRoles(ListRolesRequest request) { - String path = String.format("/api/2.0/postgres/%s/roles", request.getParent()); - try { - Request req = new Request("GET", path); - - ApiClient.setQuery(req, request); - req.withHeader("Accept", "application/json"); - return apiClient.execute(req, ListRolesResponse.class); - } catch (IOException e) { - throw new DatabricksException("IO error: " + e.getMessage(), e); - } - } - @Override public Operation updateBranch(UpdateBranchRequest request) { String path = String.format("/api/2.0/postgres/%s", request.getName()); diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/PostgresService.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/PostgresService.java index e5b7d832f..871532e59 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/PostgresService.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/PostgresService.java @@ -21,9 +21,6 @@ public interface PostgresService { /** Create a Project. */ Operation createProject(CreateProjectRequest createProjectRequest); - /** Create a role for a branch. */ - Operation createRole(CreateRoleRequest createRoleRequest); - /** Delete a Branch. */ void deleteBranch(DeleteBranchRequest deleteBranchRequest); @@ -33,9 +30,6 @@ public interface PostgresService { /** Delete a Project. */ void deleteProject(DeleteProjectRequest deleteProjectRequest); - /** Delete a role in a branch. */ - Operation deleteRole(DeleteRoleRequest deleteRoleRequest); - /** Get a Branch. */ Branch getBranch(GetBranchRequest getBranchRequest); @@ -48,9 +42,6 @@ public interface PostgresService { /** Get a Project. */ Project getProject(GetProjectRequest getProjectRequest); - /** Get a Role. */ - Role getRole(GetRoleRequest getRoleRequest); - /** List Branches. */ ListBranchesResponse listBranches(ListBranchesRequest listBranchesRequest); @@ -60,9 +51,6 @@ public interface PostgresService { /** List Projects. */ ListProjectsResponse listProjects(ListProjectsRequest listProjectsRequest); - /** List Roles. */ - ListRolesResponse listRoles(ListRolesRequest listRolesRequest); - /** Update a Branch. */ Operation updateBranch(UpdateBranchRequest updateBranchRequest); diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Project.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Project.java index 034c68aea..0f7827672 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Project.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Project.java @@ -5,26 +5,77 @@ import com.databricks.sdk.support.Generated; import com.databricks.sdk.support.ToStringer; import com.fasterxml.jackson.annotation.JsonProperty; +import com.google.protobuf.Duration; import com.google.protobuf.Timestamp; import java.util.Objects; @Generated public class Project { + /** The logical size limit for a branch. */ + @JsonProperty("branch_logical_size_limit_bytes") + private Long branchLogicalSizeLimitBytes; + + /** The most recent time when any endpoint of this project was active. */ + @JsonProperty("compute_last_active_time") + private Timestamp computeLastActiveTime; + /** A timestamp indicating when the project was created. */ @JsonProperty("create_time") private Timestamp createTime; + /** */ + @JsonProperty("default_endpoint_settings") + private ProjectDefaultEndpointSettings defaultEndpointSettings; + + /** Human-readable project name. */ + @JsonProperty("display_name") + private String displayName; + + /** */ + @JsonProperty("effective_default_endpoint_settings") + private ProjectDefaultEndpointSettings effectiveDefaultEndpointSettings; + + /** */ + @JsonProperty("effective_display_name") + private String effectiveDisplayName; + + /** */ + @JsonProperty("effective_history_retention_duration") + private Duration effectiveHistoryRetentionDuration; + + /** */ + @JsonProperty("effective_pg_version") + private Long effectivePgVersion; + + /** */ + @JsonProperty("effective_settings") + private ProjectSettings effectiveSettings; + + /** + * The number of seconds to retain the shared history for point in time recovery for all branches + * in this project. + */ + @JsonProperty("history_retention_duration") + private Duration historyRetentionDuration; + /** The resource name of the project. Format: projects/{project_id} */ @JsonProperty("name") private String name; - /** The desired state of a Project. */ - @JsonProperty("spec") - private ProjectSpec spec; + /** The major Postgres version number. */ + @JsonProperty("pg_version") + private Long pgVersion; + + /** */ + @JsonProperty("settings") + private ProjectSettings settings; - /** The current status of a Project. */ - @JsonProperty("status") - private ProjectStatus status; + /** + * The current space occupied by the project in storage. Synthetic storage size combines the + * logical data size and Write-Ahead Log (WAL) size for all branches in a project. + */ + @JsonProperty("synthetic_storage_size_bytes") + private Long syntheticStorageSizeBytes; /** System generated unique ID for the project. */ @JsonProperty("uid") @@ -34,6 +85,24 @@ public class Project { @JsonProperty("update_time") private Timestamp updateTime; + public Project setBranchLogicalSizeLimitBytes(Long branchLogicalSizeLimitBytes) { + this.branchLogicalSizeLimitBytes = branchLogicalSizeLimitBytes; + return this; + } + + public Long getBranchLogicalSizeLimitBytes() { + return branchLogicalSizeLimitBytes; + } + + public Project setComputeLastActiveTime(Timestamp computeLastActiveTime) { + this.computeLastActiveTime = computeLastActiveTime; + return this; + } + + public Timestamp getComputeLastActiveTime() { + return computeLastActiveTime; + } + public Project setCreateTime(Timestamp createTime) { this.createTime = createTime; return this; @@ -43,6 +112,80 @@ public Timestamp getCreateTime() { return createTime; } + public Project setDefaultEndpointSettings( + ProjectDefaultEndpointSettings defaultEndpointSettings) { + this.defaultEndpointSettings = defaultEndpointSettings; + return this; + } + + public ProjectDefaultEndpointSettings getDefaultEndpointSettings() { + return defaultEndpointSettings; + } + + public Project setDisplayName(String displayName) { + this.displayName = displayName; + return this; + } + + public String getDisplayName() { + return displayName; + } + + public Project setEffectiveDefaultEndpointSettings( + ProjectDefaultEndpointSettings effectiveDefaultEndpointSettings) { + this.effectiveDefaultEndpointSettings = effectiveDefaultEndpointSettings; + return this; + } + + public ProjectDefaultEndpointSettings getEffectiveDefaultEndpointSettings() { + return effectiveDefaultEndpointSettings; + } + + public Project setEffectiveDisplayName(String effectiveDisplayName) { + this.effectiveDisplayName = effectiveDisplayName; + return this; + } + + public String getEffectiveDisplayName() { + return effectiveDisplayName; + } + + public Project setEffectiveHistoryRetentionDuration(Duration effectiveHistoryRetentionDuration) { + this.effectiveHistoryRetentionDuration = effectiveHistoryRetentionDuration; + return this; + } + + public Duration getEffectiveHistoryRetentionDuration() { + return effectiveHistoryRetentionDuration; + } + + public Project setEffectivePgVersion(Long effectivePgVersion) { + this.effectivePgVersion = effectivePgVersion; + return this; + } + + public Long getEffectivePgVersion() { + return effectivePgVersion; + } + + public Project setEffectiveSettings(ProjectSettings effectiveSettings) { + this.effectiveSettings = effectiveSettings; + return this; + } + + public ProjectSettings getEffectiveSettings() { + return effectiveSettings; + } + + public Project setHistoryRetentionDuration(Duration historyRetentionDuration) { + this.historyRetentionDuration = historyRetentionDuration; + return this; + } + + public Duration getHistoryRetentionDuration() { + return historyRetentionDuration; + } + public Project setName(String name) { this.name = name; return this; @@ -52,22 +195,31 @@ public String getName() { return name; } - public Project setSpec(ProjectSpec spec) { - this.spec = spec; + public Project setPgVersion(Long pgVersion) { + this.pgVersion = pgVersion; + return this; + } + + public Long getPgVersion() { + return pgVersion; + } + + public Project setSettings(ProjectSettings settings) { + this.settings = settings; return this; } - public ProjectSpec getSpec() { - return spec; + public ProjectSettings getSettings() { + return settings; } - public Project setStatus(ProjectStatus status) { - this.status = status; + public Project setSyntheticStorageSizeBytes(Long syntheticStorageSizeBytes) { + this.syntheticStorageSizeBytes = syntheticStorageSizeBytes; return this; } - public ProjectStatus getStatus() { - return status; + public Long getSyntheticStorageSizeBytes() { + return syntheticStorageSizeBytes; } public Project setUid(String uid) { @@ -93,26 +245,65 @@ public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; Project that = (Project) o; - return Objects.equals(createTime, that.createTime) + return Objects.equals(branchLogicalSizeLimitBytes, that.branchLogicalSizeLimitBytes) + && Objects.equals(computeLastActiveTime, that.computeLastActiveTime) + && Objects.equals(createTime, that.createTime) + && Objects.equals(defaultEndpointSettings, that.defaultEndpointSettings) + && Objects.equals(displayName, that.displayName) + && Objects.equals(effectiveDefaultEndpointSettings, that.effectiveDefaultEndpointSettings) + && Objects.equals(effectiveDisplayName, that.effectiveDisplayName) + && Objects.equals(effectiveHistoryRetentionDuration, that.effectiveHistoryRetentionDuration) + && Objects.equals(effectivePgVersion, that.effectivePgVersion) + && Objects.equals(effectiveSettings, that.effectiveSettings) + && Objects.equals(historyRetentionDuration, that.historyRetentionDuration) && Objects.equals(name, that.name) - && Objects.equals(spec, that.spec) - && Objects.equals(status, that.status) + && Objects.equals(pgVersion, that.pgVersion) + && Objects.equals(settings, that.settings) + && Objects.equals(syntheticStorageSizeBytes, that.syntheticStorageSizeBytes) && Objects.equals(uid, that.uid) && Objects.equals(updateTime, that.updateTime); } @Override public int hashCode() { - return Objects.hash(createTime, name, spec, status, uid, updateTime); + return Objects.hash( + branchLogicalSizeLimitBytes, + computeLastActiveTime, + createTime, + defaultEndpointSettings, + displayName, + effectiveDefaultEndpointSettings, + effectiveDisplayName, + effectiveHistoryRetentionDuration, + effectivePgVersion, + effectiveSettings, + historyRetentionDuration, + name, + pgVersion, + settings, + syntheticStorageSizeBytes, + uid, + updateTime); } @Override public String toString() { return new ToStringer(Project.class) + .add("branchLogicalSizeLimitBytes", branchLogicalSizeLimitBytes) + .add("computeLastActiveTime", computeLastActiveTime) .add("createTime", createTime) + .add("defaultEndpointSettings", defaultEndpointSettings) + .add("displayName", displayName) + .add("effectiveDefaultEndpointSettings", effectiveDefaultEndpointSettings) + .add("effectiveDisplayName", effectiveDisplayName) + .add("effectiveHistoryRetentionDuration", effectiveHistoryRetentionDuration) + .add("effectivePgVersion", effectivePgVersion) + .add("effectiveSettings", effectiveSettings) + .add("historyRetentionDuration", historyRetentionDuration) .add("name", name) - .add("spec", spec) - .add("status", status) + .add("pgVersion", pgVersion) + .add("settings", settings) + .add("syntheticStorageSizeBytes", syntheticStorageSizeBytes) .add("uid", uid) .add("updateTime", updateTime) .toString(); diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ProjectSpec.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ProjectSpec.java deleted file mode 100644 index d823fbc78..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ProjectSpec.java +++ /dev/null @@ -1,110 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import com.google.protobuf.Duration; -import java.util.Objects; - -@Generated -public class ProjectSpec { - /** */ - @JsonProperty("default_endpoint_settings") - private ProjectDefaultEndpointSettings defaultEndpointSettings; - - /** Human-readable project name. */ - @JsonProperty("display_name") - private String displayName; - - /** - * The number of seconds to retain the shared history for point in time recovery for all branches - * in this project. - */ - @JsonProperty("history_retention_duration") - private Duration historyRetentionDuration; - - /** The major Postgres version number. */ - @JsonProperty("pg_version") - private Long pgVersion; - - /** */ - @JsonProperty("settings") - private ProjectSettings settings; - - public ProjectSpec setDefaultEndpointSettings( - ProjectDefaultEndpointSettings defaultEndpointSettings) { - this.defaultEndpointSettings = defaultEndpointSettings; - return this; - } - - public ProjectDefaultEndpointSettings getDefaultEndpointSettings() { - return defaultEndpointSettings; - } - - public ProjectSpec setDisplayName(String displayName) { - this.displayName = displayName; - return this; - } - - public String getDisplayName() { - return displayName; - } - - public ProjectSpec setHistoryRetentionDuration(Duration historyRetentionDuration) { - this.historyRetentionDuration = historyRetentionDuration; - return this; - } - - public Duration getHistoryRetentionDuration() { - return historyRetentionDuration; - } - - public ProjectSpec setPgVersion(Long pgVersion) { - this.pgVersion = pgVersion; - return this; - } - - public Long getPgVersion() { - return pgVersion; - } - - public ProjectSpec setSettings(ProjectSettings settings) { - this.settings = settings; - return this; - } - - public ProjectSettings getSettings() { - return settings; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - ProjectSpec that = (ProjectSpec) o; - return Objects.equals(defaultEndpointSettings, that.defaultEndpointSettings) - && Objects.equals(displayName, that.displayName) - && Objects.equals(historyRetentionDuration, that.historyRetentionDuration) - && Objects.equals(pgVersion, that.pgVersion) - && Objects.equals(settings, that.settings); - } - - @Override - public int hashCode() { - return Objects.hash( - defaultEndpointSettings, displayName, historyRetentionDuration, pgVersion, settings); - } - - @Override - public String toString() { - return new ToStringer(ProjectSpec.class) - .add("defaultEndpointSettings", defaultEndpointSettings) - .add("displayName", displayName) - .add("historyRetentionDuration", historyRetentionDuration) - .add("pgVersion", pgVersion) - .add("settings", settings) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ProjectStatus.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ProjectStatus.java deleted file mode 100644 index 3c6eb9012..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/ProjectStatus.java +++ /dev/null @@ -1,160 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import com.google.protobuf.Duration; -import com.google.protobuf.Timestamp; -import java.util.Objects; - -@Generated -public class ProjectStatus { - /** The logical size limit for a branch. */ - @JsonProperty("branch_logical_size_limit_bytes") - private Long branchLogicalSizeLimitBytes; - - /** The most recent time when any endpoint of this project was active. */ - @JsonProperty("compute_last_active_time") - private Timestamp computeLastActiveTime; - - /** The effective default endpoint settings. */ - @JsonProperty("default_endpoint_settings") - private ProjectDefaultEndpointSettings defaultEndpointSettings; - - /** The effective human-readable project name. */ - @JsonProperty("display_name") - private String displayName; - - /** The effective number of seconds to retain the shared history for point in time recovery. */ - @JsonProperty("history_retention_duration") - private Duration historyRetentionDuration; - - /** The effective major Postgres version number. */ - @JsonProperty("pg_version") - private Long pgVersion; - - /** The effective project settings. */ - @JsonProperty("settings") - private ProjectSettings settings; - - /** The current space occupied by the project in storage. */ - @JsonProperty("synthetic_storage_size_bytes") - private Long syntheticStorageSizeBytes; - - public ProjectStatus setBranchLogicalSizeLimitBytes(Long branchLogicalSizeLimitBytes) { - this.branchLogicalSizeLimitBytes = branchLogicalSizeLimitBytes; - return this; - } - - public Long getBranchLogicalSizeLimitBytes() { - return branchLogicalSizeLimitBytes; - } - - public ProjectStatus setComputeLastActiveTime(Timestamp computeLastActiveTime) { - this.computeLastActiveTime = computeLastActiveTime; - return this; - } - - public Timestamp getComputeLastActiveTime() { - return computeLastActiveTime; - } - - public ProjectStatus setDefaultEndpointSettings( - ProjectDefaultEndpointSettings defaultEndpointSettings) { - this.defaultEndpointSettings = defaultEndpointSettings; - return this; - } - - public ProjectDefaultEndpointSettings getDefaultEndpointSettings() { - return defaultEndpointSettings; - } - - public ProjectStatus setDisplayName(String displayName) { - this.displayName = displayName; - return this; - } - - public String getDisplayName() { - return displayName; - } - - public ProjectStatus setHistoryRetentionDuration(Duration historyRetentionDuration) { - this.historyRetentionDuration = historyRetentionDuration; - return this; - } - - public Duration getHistoryRetentionDuration() { - return historyRetentionDuration; - } - - public ProjectStatus setPgVersion(Long pgVersion) { - this.pgVersion = pgVersion; - return this; - } - - public Long getPgVersion() { - return pgVersion; - } - - public ProjectStatus setSettings(ProjectSettings settings) { - this.settings = settings; - return this; - } - - public ProjectSettings getSettings() { - return settings; - } - - public ProjectStatus setSyntheticStorageSizeBytes(Long syntheticStorageSizeBytes) { - this.syntheticStorageSizeBytes = syntheticStorageSizeBytes; - return this; - } - - public Long getSyntheticStorageSizeBytes() { - return syntheticStorageSizeBytes; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - ProjectStatus that = (ProjectStatus) o; - return Objects.equals(branchLogicalSizeLimitBytes, that.branchLogicalSizeLimitBytes) - && Objects.equals(computeLastActiveTime, that.computeLastActiveTime) - && Objects.equals(defaultEndpointSettings, that.defaultEndpointSettings) - && Objects.equals(displayName, that.displayName) - && Objects.equals(historyRetentionDuration, that.historyRetentionDuration) - && Objects.equals(pgVersion, that.pgVersion) - && Objects.equals(settings, that.settings) - && Objects.equals(syntheticStorageSizeBytes, that.syntheticStorageSizeBytes); - } - - @Override - public int hashCode() { - return Objects.hash( - branchLogicalSizeLimitBytes, - computeLastActiveTime, - defaultEndpointSettings, - displayName, - historyRetentionDuration, - pgVersion, - settings, - syntheticStorageSizeBytes); - } - - @Override - public String toString() { - return new ToStringer(ProjectStatus.class) - .add("branchLogicalSizeLimitBytes", branchLogicalSizeLimitBytes) - .add("computeLastActiveTime", computeLastActiveTime) - .add("defaultEndpointSettings", defaultEndpointSettings) - .add("displayName", displayName) - .add("historyRetentionDuration", historyRetentionDuration) - .add("pgVersion", pgVersion) - .add("settings", settings) - .add("syntheticStorageSizeBytes", syntheticStorageSizeBytes) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Role.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Role.java deleted file mode 100644 index c6fa0856f..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/Role.java +++ /dev/null @@ -1,123 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import com.google.protobuf.Timestamp; -import java.util.Objects; - -/** Role represents a Postgres role within a Branch. */ -@Generated -public class Role { - /** */ - @JsonProperty("create_time") - private Timestamp createTime; - - /** - * The resource name of the role. Format: projects/{project_id}/branch/{branch_id}/roles/{role_id} - */ - @JsonProperty("name") - private String name; - - /** The Branch where this Role exists. Format: projects/{project_id}/branches/{branch_id} */ - @JsonProperty("parent") - private String parent; - - /** The desired state of the Role. */ - @JsonProperty("spec") - private RoleRoleSpec spec; - - /** The observed state of the Role. */ - @JsonProperty("status") - private RoleRoleStatus status; - - /** */ - @JsonProperty("update_time") - private Timestamp updateTime; - - public Role setCreateTime(Timestamp createTime) { - this.createTime = createTime; - return this; - } - - public Timestamp getCreateTime() { - return createTime; - } - - public Role setName(String name) { - this.name = name; - return this; - } - - public String getName() { - return name; - } - - public Role setParent(String parent) { - this.parent = parent; - return this; - } - - public String getParent() { - return parent; - } - - public Role setSpec(RoleRoleSpec spec) { - this.spec = spec; - return this; - } - - public RoleRoleSpec getSpec() { - return spec; - } - - public Role setStatus(RoleRoleStatus status) { - this.status = status; - return this; - } - - public RoleRoleStatus getStatus() { - return status; - } - - public Role setUpdateTime(Timestamp updateTime) { - this.updateTime = updateTime; - return this; - } - - public Timestamp getUpdateTime() { - return updateTime; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - Role that = (Role) o; - return Objects.equals(createTime, that.createTime) - && Objects.equals(name, that.name) - && Objects.equals(parent, that.parent) - && Objects.equals(spec, that.spec) - && Objects.equals(status, that.status) - && Objects.equals(updateTime, that.updateTime); - } - - @Override - public int hashCode() { - return Objects.hash(createTime, name, parent, spec, status, updateTime); - } - - @Override - public String toString() { - return new ToStringer(Role.class) - .add("createTime", createTime) - .add("name", name) - .add("parent", parent) - .add("spec", spec) - .add("status", status) - .add("updateTime", updateTime) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleAuthMethod.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleAuthMethod.java deleted file mode 100644 index 7732b1a78..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleAuthMethod.java +++ /dev/null @@ -1,13 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; - -/** How the role is authenticated when connecting to Postgres. */ -@Generated -public enum RoleAuthMethod { - LAKEBASE_OAUTH_V1, - NO_LOGIN, - PG_PASSWORD_SCRAM_SHA_256, -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleIdentityType.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleIdentityType.java deleted file mode 100644 index 3908f1bf0..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleIdentityType.java +++ /dev/null @@ -1,16 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; - -/** - * The type of the Databricks managed identity that this Role represents. Leave empty if you wish to - * create a regular Postgres role not associated with a Databricks identity. - */ -@Generated -public enum RoleIdentityType { - GROUP, - SERVICE_PRINCIPAL, - USER, -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleOperationMetadata.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleOperationMetadata.java deleted file mode 100644 index f9849dc93..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleOperationMetadata.java +++ /dev/null @@ -1,28 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import java.util.Objects; - -@Generated -public class RoleOperationMetadata { - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - return true; - } - - @Override - public int hashCode() { - return Objects.hash(); - } - - @Override - public String toString() { - return new ToStringer(RoleOperationMetadata.class).toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleRoleSpec.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleRoleSpec.java deleted file mode 100644 index 417c91918..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleRoleSpec.java +++ /dev/null @@ -1,66 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import java.util.Objects; - -@Generated -public class RoleRoleSpec { - /** - * If auth_method is left unspecified, a meaningful authentication method is derived from the - * identity_type: * For the managed identities, OAUTH is used. * For the regular postgres roles, - * authentication based on postgres passwords is used. - * - *

NOTE: this is ignored for the Databricks identity type GROUP, and NO_LOGIN is implicitly - * assumed instead for the GROUP identity type. - */ - @JsonProperty("auth_method") - private RoleAuthMethod authMethod; - - /** The type of the role. */ - @JsonProperty("identity_type") - private RoleIdentityType identityType; - - public RoleRoleSpec setAuthMethod(RoleAuthMethod authMethod) { - this.authMethod = authMethod; - return this; - } - - public RoleAuthMethod getAuthMethod() { - return authMethod; - } - - public RoleRoleSpec setIdentityType(RoleIdentityType identityType) { - this.identityType = identityType; - return this; - } - - public RoleIdentityType getIdentityType() { - return identityType; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - RoleRoleSpec that = (RoleRoleSpec) o; - return Objects.equals(authMethod, that.authMethod) - && Objects.equals(identityType, that.identityType); - } - - @Override - public int hashCode() { - return Objects.hash(authMethod, identityType); - } - - @Override - public String toString() { - return new ToStringer(RoleRoleSpec.class) - .add("authMethod", authMethod) - .add("identityType", identityType) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleRoleStatus.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleRoleStatus.java deleted file mode 100644 index 76728b173..000000000 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/postgres/RoleRoleStatus.java +++ /dev/null @@ -1,59 +0,0 @@ -// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. - -package com.databricks.sdk.service.postgres; - -import com.databricks.sdk.support.Generated; -import com.databricks.sdk.support.ToStringer; -import com.fasterxml.jackson.annotation.JsonProperty; -import java.util.Objects; - -@Generated -public class RoleRoleStatus { - /** */ - @JsonProperty("auth_method") - private RoleAuthMethod authMethod; - - /** The type of the role. */ - @JsonProperty("identity_type") - private RoleIdentityType identityType; - - public RoleRoleStatus setAuthMethod(RoleAuthMethod authMethod) { - this.authMethod = authMethod; - return this; - } - - public RoleAuthMethod getAuthMethod() { - return authMethod; - } - - public RoleRoleStatus setIdentityType(RoleIdentityType identityType) { - this.identityType = identityType; - return this; - } - - public RoleIdentityType getIdentityType() { - return identityType; - } - - @Override - public boolean equals(Object o) { - if (this == o) return true; - if (o == null || getClass() != o.getClass()) return false; - RoleRoleStatus that = (RoleRoleStatus) o; - return Objects.equals(authMethod, that.authMethod) - && Objects.equals(identityType, that.identityType); - } - - @Override - public int hashCode() { - return Objects.hash(authMethod, identityType); - } - - @Override - public String toString() { - return new ToStringer(RoleRoleStatus.class) - .add("authMethod", authMethod) - .add("identityType", identityType) - .toString(); - } -} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/qualitymonitorv2/AnomalyDetectionConfig.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/qualitymonitorv2/AnomalyDetectionConfig.java index c3d49125e..5b14da636 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/qualitymonitorv2/AnomalyDetectionConfig.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/qualitymonitorv2/AnomalyDetectionConfig.java @@ -5,15 +5,10 @@ import com.databricks.sdk.support.Generated; import com.databricks.sdk.support.ToStringer; import com.fasterxml.jackson.annotation.JsonProperty; -import java.util.Collection; import java.util.Objects; @Generated public class AnomalyDetectionConfig { - /** List of fully qualified table names to exclude from anomaly detection. */ - @JsonProperty("excluded_table_full_names") - private Collection excludedTableFullNames; - /** Run id of the last run of the workflow */ @JsonProperty("last_run_id") private String lastRunId; @@ -22,16 +17,6 @@ public class AnomalyDetectionConfig { @JsonProperty("latest_run_status") private AnomalyDetectionRunStatus latestRunStatus; - public AnomalyDetectionConfig setExcludedTableFullNames( - Collection excludedTableFullNames) { - this.excludedTableFullNames = excludedTableFullNames; - return this; - } - - public Collection getExcludedTableFullNames() { - return excludedTableFullNames; - } - public AnomalyDetectionConfig setLastRunId(String lastRunId) { this.lastRunId = lastRunId; return this; @@ -55,20 +40,18 @@ public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; AnomalyDetectionConfig that = (AnomalyDetectionConfig) o; - return Objects.equals(excludedTableFullNames, that.excludedTableFullNames) - && Objects.equals(lastRunId, that.lastRunId) + return Objects.equals(lastRunId, that.lastRunId) && Objects.equals(latestRunStatus, that.latestRunStatus); } @Override public int hashCode() { - return Objects.hash(excludedTableFullNames, lastRunId, latestRunStatus); + return Objects.hash(lastRunId, latestRunStatus); } @Override public String toString() { return new ToStringer(AnomalyDetectionConfig.class) - .add("excludedTableFullNames", excludedTableFullNames) .add("lastRunId", lastRunId) .add("latestRunStatus", latestRunStatus) .toString(); diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sharing/SharedDataObject.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sharing/SharedDataObject.java index 071f11697..f6e090ef5 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sharing/SharedDataObject.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sharing/SharedDataObject.java @@ -58,29 +58,10 @@ public class SharedDataObject { private Collection partitions; /** - * A user-provided alias name for table-like data objects within the share. - * - * Use this field for table-like objects (for example: TABLE, VIEW, - * MATERIALIZED_VIEW, STREAMING_TABLE, FOREIGN_TABLE). For non-table objects - * (for example: VOLUME, MODEL, NOTEBOOK_FILE, FUNCTION), use - * `string_shared_as` instead. - * - * Important: For non-table objects, this field must be omitted entirely. - * - * Format: Must be a 2-part name `.` (e.g., - * "sales_schema.orders_table") - Both schema and table names must contain - * only alphanumeric characters and underscores - No periods, spaces, forward - * slashes, or control characters are allowed within each part - Do not - * include the catalog name (use 2 parts, not 3) - * - * Behavior: - If not provided, the service automatically generates the alias - * as `.` from the object's original name - If you don't want - * to specify this field, omit it entirely from the request (do not pass an - * empty string) - The `shared_as` name must be unique within the share - * - * Examples: - Valid: "analytics_schema.customer_view" - Invalid: - * "catalog.analytics_schema.customer_view" (3 parts not allowed) - Invalid: - * "analytics-schema.customer-view" (hyphens not allowed) + * A user-provided new name for the data object within the share. If this new + * name is not provided, the object's original name will be used as the + * `shared_as` name. The `shared_as` name must be unique within a share. For + * tables, the new name must follow the format of `.
`. */ @JsonProperty("shared_as") private String sharedAs; @@ -101,30 +82,10 @@ public class SharedDataObject { private SharedDataObjectStatus status; /** - * A user-provided alias name for non-table data objects within the share. - * - *

Use this field for non-table objects (for example: VOLUME, MODEL, NOTEBOOK_FILE, FUNCTION). - * For table-like objects (for example: TABLE, VIEW, MATERIALIZED_VIEW, STREAMING_TABLE, - * FOREIGN_TABLE), use `shared_as` instead. - * - *

Important: For table-like objects, this field must be omitted entirely. - * - *

Format: - For VOLUME: Must be a 2-part name `.` (e.g., - * "data_schema.ml_models") - For FUNCTION: Must be a 2-part name `.` - * (e.g., "udf_schema.calculate_tax") - For MODEL: Must be a 2-part name - * `.` (e.g., "models.prediction_model") - For NOTEBOOK_FILE: Should be - * the notebook file name (e.g., "analysis_notebook.py") - All names must contain only - * alphanumeric characters and underscores - No periods, spaces, forward slashes, or control - * characters are allowed within each part - * - *

Behavior: - If not provided, the service automatically generates the alias from the object's - * original name - If you don't want to specify this field, omit it entirely from the request (do - * not pass an empty string) - The `string_shared_as` name must be unique for objects of the same - * type within the share - * - *

Examples: - Valid for VOLUME: "data_schema.training_data" - Valid for FUNCTION: - * "analytics.calculate_revenue" - Invalid: "catalog.data_schema.training_data" (3 parts not - * allowed for volumes) - Invalid: "data-schema.training-data" (hyphens not allowed) + * A user-provided new name for the shared object within the share. If this new name is not not + * provided, the object's original name will be used as the `string_shared_as` name. The + * `string_shared_as` name must be unique for objects of the same type within a Share. For + * notebooks, the new name should be the new notebook file name. */ @JsonProperty("string_shared_as") private String stringSharedAs; diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sharing/Table.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sharing/Table.java index 61cff4487..119e8a050 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sharing/Table.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sharing/Table.java @@ -10,13 +10,6 @@ @Generated public class Table { - /** - * The access modes supported for this table (e.g., "url", "dir"). Used for open sharing to - * indicate how the table can be accessed. - */ - @JsonProperty("access_modes") - private Collection accessModes; - /** The comment of the table. */ @JsonProperty("comment") private String comment; @@ -53,23 +46,10 @@ public class Table { @JsonProperty("share_id") private String shareId; - /** The cloud storage location of the table for open sharing. */ - @JsonProperty("storage_location") - private String storageLocation; - /** The Tags of the table. */ @JsonProperty("tags") private Collection tags; - public Table setAccessModes(Collection accessModes) { - this.accessModes = accessModes; - return this; - } - - public Collection getAccessModes() { - return accessModes; - } - public Table setComment(String comment) { this.comment = comment; return this; @@ -151,15 +131,6 @@ public String getShareId() { return shareId; } - public Table setStorageLocation(String storageLocation) { - this.storageLocation = storageLocation; - return this; - } - - public String getStorageLocation() { - return storageLocation; - } - public Table setTags(Collection tags) { this.tags = tags; return this; @@ -174,8 +145,7 @@ public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; Table that = (Table) o; - return Objects.equals(accessModes, that.accessModes) - && Objects.equals(comment, that.comment) + return Objects.equals(comment, that.comment) && Objects.equals(id, that.id) && Objects.equals(internalAttributes, that.internalAttributes) && Objects.equals(materializationNamespace, that.materializationNamespace) @@ -184,14 +154,12 @@ public boolean equals(Object o) { && Objects.equals(schema, that.schema) && Objects.equals(share, that.share) && Objects.equals(shareId, that.shareId) - && Objects.equals(storageLocation, that.storageLocation) && Objects.equals(tags, that.tags); } @Override public int hashCode() { return Objects.hash( - accessModes, comment, id, internalAttributes, @@ -201,14 +169,12 @@ public int hashCode() { schema, share, shareId, - storageLocation, tags); } @Override public String toString() { return new ToStringer(Table.class) - .add("accessModes", accessModes) .add("comment", comment) .add("id", id) .add("internalAttributes", internalAttributes) @@ -218,7 +184,6 @@ public String toString() { .add("schema", schema) .add("share", share) .add("shareId", shareId) - .add("storageLocation", storageLocation) .add("tags", tags) .toString(); } diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/vectorsearch/QueryVectorIndexRequest.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/vectorsearch/QueryVectorIndexRequest.java index 62727e197..98d3df913 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/vectorsearch/QueryVectorIndexRequest.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/vectorsearch/QueryVectorIndexRequest.java @@ -53,14 +53,7 @@ public class QueryVectorIndexRequest { @JsonProperty("query_vector") private Collection queryVector; - /** - * If set, the top 50 results are reranked with the Databricks Reranker model before returning the - * `num_results` results to the user. The setting `columns_to_rerank` selects which columns are - * used for reranking. For each datapoint, the columns selected are concatenated before being sent - * to the reranking model. See - * https://docs.databricks.com/aws/en/vector-search/query-vector-search#rerank for more - * information. - */ + /** */ @JsonProperty("reranker") private RerankerConfig reranker;