From 0b4ef7db3c78720f56abf6daeefd9d87f5543b8b Mon Sep 17 00:00:00 2001 From: "databricks-ci-ghec-1[bot]" <184311507+databricks-ci-ghec-1[bot]@users.noreply.github.com> Date: Wed, 28 Jan 2026 09:51:28 +0000 Subject: [PATCH] Update SDK to 6a63813a81db5d1794c2b8cddd33db1cdbb85aa7 --- .codegen/_openapi_sha | 2 +- .gitattributes | 1 + NEXT_CHANGELOG.md | 5 ++ .../sdk/service/catalog/SchemaInfo.java | 2 +- .../sdk/service/ml/BackfillSource.java | 48 +++++++++++++++++++ .../sdk/service/ml/KafkaConfig.java | 29 ++++++++++- .../service/serving/ServedEntityInput.java | 22 ++++++++- .../service/serving/ServedEntityOutput.java | 22 ++++++++- .../sdk/service/serving/ServedModelInput.java | 22 ++++++++- .../service/serving/ServedModelOutput.java | 22 ++++++++- .../workspace/ListCredentialsRequest.java | 2 +- 11 files changed, 169 insertions(+), 8 deletions(-) create mode 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/BackfillSource.java mode change 100755 => 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/KafkaConfig.java mode change 100755 => 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedEntityInput.java mode change 100755 => 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedEntityOutput.java mode change 100755 => 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedModelInput.java mode change 100755 => 100644 databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedModelOutput.java diff --git a/.codegen/_openapi_sha b/.codegen/_openapi_sha index 2f7775fd6..56da24bf7 100644 --- a/.codegen/_openapi_sha +++ b/.codegen/_openapi_sha @@ -1 +1 @@ -5b781e770944bde168c283fd04bc94fb7d910254 \ No newline at end of file +6a63813a81db5d1794c2b8cddd33db1cdbb85aa7 \ No newline at end of file diff --git a/.gitattributes b/.gitattributes index bdc5111ef..0461697a6 100644 --- a/.gitattributes +++ b/.gitattributes @@ -1683,6 +1683,7 @@ databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/ActivityType.jav databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/ApproveTransitionRequest.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/ApproveTransitionRequestResponse.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/AuthConfig.java linguist-generated=true +databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/BackfillSource.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/BatchCreateMaterializedFeaturesRequest.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/BatchCreateMaterializedFeaturesResponse.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/ColumnIdentifier.java linguist-generated=true diff --git a/NEXT_CHANGELOG.md b/NEXT_CHANGELOG.md index 8310e3b7b..644884ac6 100644 --- a/NEXT_CHANGELOG.md +++ b/NEXT_CHANGELOG.md @@ -13,3 +13,8 @@ ### Internal Changes ### API Changes +* Add `backfillSource` field for `com.databricks.sdk.service.ml.KafkaConfig`. +* Add `burstScalingEnabled` field for `com.databricks.sdk.service.serving.ServedEntityInput`. +* Add `burstScalingEnabled` field for `com.databricks.sdk.service.serving.ServedEntityOutput`. +* Add `burstScalingEnabled` field for `com.databricks.sdk.service.serving.ServedModelInput`. +* Add `burstScalingEnabled` field for `com.databricks.sdk.service.serving.ServedModelOutput`. \ No newline at end of file diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SchemaInfo.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SchemaInfo.java index 44c7b1bc9..a56c67a82 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SchemaInfo.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SchemaInfo.java @@ -8,7 +8,7 @@ import java.util.Map; import java.util.Objects; -/** Next ID: 44 */ +/** Next ID: 45 */ @Generated public class SchemaInfo { /** diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/BackfillSource.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/BackfillSource.java new file mode 100644 index 000000000..c30adeb8d --- /dev/null +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/BackfillSource.java @@ -0,0 +1,48 @@ +// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. + +package com.databricks.sdk.service.ml; + +import com.databricks.sdk.support.Generated; +import com.databricks.sdk.support.ToStringer; +import com.fasterxml.jackson.annotation.JsonProperty; +import java.util.Objects; + +@Generated +public class BackfillSource { + /** + * The Delta table source containing the historic data to backfill. Only the delta table name is + * used for backfill, the entity columns and timeseries column are ignored as they are defined by + * the associated KafkaSource. + */ + @JsonProperty("delta_table_source") + private DeltaTableSource deltaTableSource; + + public BackfillSource setDeltaTableSource(DeltaTableSource deltaTableSource) { + this.deltaTableSource = deltaTableSource; + return this; + } + + public DeltaTableSource getDeltaTableSource() { + return deltaTableSource; + } + + @Override + public boolean equals(Object o) { + if (this == o) return true; + if (o == null || getClass() != o.getClass()) return false; + BackfillSource that = (BackfillSource) o; + return Objects.equals(deltaTableSource, that.deltaTableSource); + } + + @Override + public int hashCode() { + return Objects.hash(deltaTableSource); + } + + @Override + public String toString() { + return new ToStringer(BackfillSource.class) + .add("deltaTableSource", deltaTableSource) + .toString(); + } +} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/KafkaConfig.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/KafkaConfig.java old mode 100755 new mode 100644 index 4fc314969..700ad4c98 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/KafkaConfig.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/ml/KafkaConfig.java @@ -14,6 +14,15 @@ public class KafkaConfig { @JsonProperty("auth_config") private AuthConfig authConfig; + /** + * A user-provided and managed source for backfilling data. Historical data is used when creating + * a training set from streaming features linked to this Kafka config. In the future, a separate + * table will be maintained by Databricks for forward filling data. The schema for this source + * must match exactly that of the key and value schemas specified for this Kafka config. + */ + @JsonProperty("backfill_source") + private BackfillSource backfillSource; + /** A comma-separated list of host/port pairs pointing to Kafka cluster. */ @JsonProperty("bootstrap_servers") private String bootstrapServers; @@ -60,6 +69,15 @@ public AuthConfig getAuthConfig() { return authConfig; } + public KafkaConfig setBackfillSource(BackfillSource backfillSource) { + this.backfillSource = backfillSource; + return this; + } + + public BackfillSource getBackfillSource() { + return backfillSource; + } + public KafkaConfig setBootstrapServers(String bootstrapServers) { this.bootstrapServers = bootstrapServers; return this; @@ -120,6 +138,7 @@ public boolean equals(Object o) { if (o == null || getClass() != o.getClass()) return false; KafkaConfig that = (KafkaConfig) o; return Objects.equals(authConfig, that.authConfig) + && Objects.equals(backfillSource, that.backfillSource) && Objects.equals(bootstrapServers, that.bootstrapServers) && Objects.equals(extraOptions, that.extraOptions) && Objects.equals(keySchema, that.keySchema) @@ -131,13 +150,21 @@ public boolean equals(Object o) { @Override public int hashCode() { return Objects.hash( - authConfig, bootstrapServers, extraOptions, keySchema, name, subscriptionMode, valueSchema); + authConfig, + backfillSource, + bootstrapServers, + extraOptions, + keySchema, + name, + subscriptionMode, + valueSchema); } @Override public String toString() { return new ToStringer(KafkaConfig.class) .add("authConfig", authConfig) + .add("backfillSource", backfillSource) .add("bootstrapServers", bootstrapServers) .add("extraOptions", extraOptions) .add("keySchema", keySchema) diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedEntityInput.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedEntityInput.java old mode 100755 new mode 100644 index ca9ccf251..f73344a2a --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedEntityInput.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedEntityInput.java @@ -10,6 +10,14 @@ @Generated public class ServedEntityInput { + /** + * Whether burst scaling is enabled. When enabled (default), the endpoint can automatically scale + * up beyond provisioned capacity to handle traffic spikes. When disabled, the endpoint maintains + * fixed capacity at provisioned_model_units. + */ + @JsonProperty("burst_scaling_enabled") + private Boolean burstScalingEnabled; + /** * The name of the entity to be served. The entity may be a model in the Databricks Model * Registry, a model in the Unity Catalog (UC), or a function of type FEATURE_SPEC in the UC. If @@ -112,6 +120,15 @@ public class ServedEntityInput { @JsonProperty("workload_type") private ServingModelWorkloadType workloadType; + public ServedEntityInput setBurstScalingEnabled(Boolean burstScalingEnabled) { + this.burstScalingEnabled = burstScalingEnabled; + return this; + } + + public Boolean getBurstScalingEnabled() { + return burstScalingEnabled; + } + public ServedEntityInput setEntityName(String entityName) { this.entityName = entityName; return this; @@ -243,7 +260,8 @@ public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; ServedEntityInput that = (ServedEntityInput) o; - return Objects.equals(entityName, that.entityName) + return Objects.equals(burstScalingEnabled, that.burstScalingEnabled) + && Objects.equals(entityName, that.entityName) && Objects.equals(entityVersion, that.entityVersion) && Objects.equals(environmentVars, that.environmentVars) && Objects.equals(externalModel, that.externalModel) @@ -262,6 +280,7 @@ public boolean equals(Object o) { @Override public int hashCode() { return Objects.hash( + burstScalingEnabled, entityName, entityVersion, environmentVars, @@ -281,6 +300,7 @@ public int hashCode() { @Override public String toString() { return new ToStringer(ServedEntityInput.class) + .add("burstScalingEnabled", burstScalingEnabled) .add("entityName", entityName) .add("entityVersion", entityVersion) .add("environmentVars", environmentVars) diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedEntityOutput.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedEntityOutput.java old mode 100755 new mode 100644 index b2764b5c1..a5d27729c --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedEntityOutput.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedEntityOutput.java @@ -10,6 +10,14 @@ @Generated public class ServedEntityOutput { + /** + * Whether burst scaling is enabled. When enabled (default), the endpoint can automatically scale + * up beyond provisioned capacity to handle traffic spikes. When disabled, the endpoint maintains + * fixed capacity at provisioned_model_units. + */ + @JsonProperty("burst_scaling_enabled") + private Boolean burstScalingEnabled; + /** */ @JsonProperty("creation_timestamp") private Long creationTimestamp; @@ -128,6 +136,15 @@ public class ServedEntityOutput { @JsonProperty("workload_type") private ServingModelWorkloadType workloadType; + public ServedEntityOutput setBurstScalingEnabled(Boolean burstScalingEnabled) { + this.burstScalingEnabled = burstScalingEnabled; + return this; + } + + public Boolean getBurstScalingEnabled() { + return burstScalingEnabled; + } + public ServedEntityOutput setCreationTimestamp(Long creationTimestamp) { this.creationTimestamp = creationTimestamp; return this; @@ -295,7 +312,8 @@ public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; ServedEntityOutput that = (ServedEntityOutput) o; - return Objects.equals(creationTimestamp, that.creationTimestamp) + return Objects.equals(burstScalingEnabled, that.burstScalingEnabled) + && Objects.equals(creationTimestamp, that.creationTimestamp) && Objects.equals(creator, that.creator) && Objects.equals(entityName, that.entityName) && Objects.equals(entityVersion, that.entityVersion) @@ -318,6 +336,7 @@ public boolean equals(Object o) { @Override public int hashCode() { return Objects.hash( + burstScalingEnabled, creationTimestamp, creator, entityName, @@ -341,6 +360,7 @@ public int hashCode() { @Override public String toString() { return new ToStringer(ServedEntityOutput.class) + .add("burstScalingEnabled", burstScalingEnabled) .add("creationTimestamp", creationTimestamp) .add("creator", creator) .add("entityName", entityName) diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedModelInput.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedModelInput.java old mode 100755 new mode 100644 index 93b608063..4ba2bf39f --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedModelInput.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedModelInput.java @@ -10,6 +10,14 @@ @Generated public class ServedModelInput { + /** + * Whether burst scaling is enabled. When enabled (default), the endpoint can automatically scale + * up beyond provisioned capacity to handle traffic spikes. When disabled, the endpoint maintains + * fixed capacity at provisioned_model_units. + */ + @JsonProperty("burst_scaling_enabled") + private Boolean burstScalingEnabled; + /** * An object containing a set of optional, user-specified environment variable key-value pairs * used for serving this entity. Note: this is an experimental feature and subject to change. @@ -95,6 +103,15 @@ public class ServedModelInput { @JsonProperty("workload_type") private ServedModelInputWorkloadType workloadType; + public ServedModelInput setBurstScalingEnabled(Boolean burstScalingEnabled) { + this.burstScalingEnabled = burstScalingEnabled; + return this; + } + + public Boolean getBurstScalingEnabled() { + return burstScalingEnabled; + } + public ServedModelInput setEnvironmentVars(Map environmentVars) { this.environmentVars = environmentVars; return this; @@ -217,7 +234,8 @@ public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; ServedModelInput that = (ServedModelInput) o; - return Objects.equals(environmentVars, that.environmentVars) + return Objects.equals(burstScalingEnabled, that.burstScalingEnabled) + && Objects.equals(environmentVars, that.environmentVars) && Objects.equals(instanceProfileArn, that.instanceProfileArn) && Objects.equals(maxProvisionedConcurrency, that.maxProvisionedConcurrency) && Objects.equals(maxProvisionedThroughput, that.maxProvisionedThroughput) @@ -235,6 +253,7 @@ public boolean equals(Object o) { @Override public int hashCode() { return Objects.hash( + burstScalingEnabled, environmentVars, instanceProfileArn, maxProvisionedConcurrency, @@ -253,6 +272,7 @@ public int hashCode() { @Override public String toString() { return new ToStringer(ServedModelInput.class) + .add("burstScalingEnabled", burstScalingEnabled) .add("environmentVars", environmentVars) .add("instanceProfileArn", instanceProfileArn) .add("maxProvisionedConcurrency", maxProvisionedConcurrency) diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedModelOutput.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedModelOutput.java old mode 100755 new mode 100644 index dfdc57241..550c83383 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedModelOutput.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/serving/ServedModelOutput.java @@ -10,6 +10,14 @@ @Generated public class ServedModelOutput { + /** + * Whether burst scaling is enabled. When enabled (default), the endpoint can automatically scale + * up beyond provisioned capacity to handle traffic spikes. When disabled, the endpoint maintains + * fixed capacity at provisioned_model_units. + */ + @JsonProperty("burst_scaling_enabled") + private Boolean burstScalingEnabled; + /** */ @JsonProperty("creation_timestamp") private Long creationTimestamp; @@ -99,6 +107,15 @@ public class ServedModelOutput { @JsonProperty("workload_type") private ServingModelWorkloadType workloadType; + public ServedModelOutput setBurstScalingEnabled(Boolean burstScalingEnabled) { + this.burstScalingEnabled = burstScalingEnabled; + return this; + } + + public Boolean getBurstScalingEnabled() { + return burstScalingEnabled; + } + public ServedModelOutput setCreationTimestamp(Long creationTimestamp) { this.creationTimestamp = creationTimestamp; return this; @@ -230,7 +247,8 @@ public boolean equals(Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; ServedModelOutput that = (ServedModelOutput) o; - return Objects.equals(creationTimestamp, that.creationTimestamp) + return Objects.equals(burstScalingEnabled, that.burstScalingEnabled) + && Objects.equals(creationTimestamp, that.creationTimestamp) && Objects.equals(creator, that.creator) && Objects.equals(environmentVars, that.environmentVars) && Objects.equals(instanceProfileArn, that.instanceProfileArn) @@ -249,6 +267,7 @@ public boolean equals(Object o) { @Override public int hashCode() { return Objects.hash( + burstScalingEnabled, creationTimestamp, creator, environmentVars, @@ -268,6 +287,7 @@ public int hashCode() { @Override public String toString() { return new ToStringer(ServedModelOutput.class) + .add("burstScalingEnabled", burstScalingEnabled) .add("creationTimestamp", creationTimestamp) .add("creator", creator) .add("environmentVars", environmentVars) diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/workspace/ListCredentialsRequest.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/workspace/ListCredentialsRequest.java index 471dbb754..0ef45be2d 100644 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/workspace/ListCredentialsRequest.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/workspace/ListCredentialsRequest.java @@ -11,7 +11,7 @@ @Generated public class ListCredentialsRequest { /** - * The ID of the service principal whose credentials will be modified. Only service principal + * The ID of the service principal whose credentials will be listed. Only service principal * managers can perform this action. */ @JsonIgnore