Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion .codegen/_openapi_sha
Original file line number Diff line number Diff line change
@@ -1 +1 @@
2ea07a797146cab9b579355bb39c58296a27d44f
2d7aaedaf5d20f82b17cc1de48543b88e609f464
1 change: 1 addition & 0 deletions acceptance/bundle/migrate/basic/out.original_state.json
Original file line number Diff line number Diff line change
Expand Up @@ -79,6 +79,7 @@
"task": [
{
"clean_rooms_notebook_task": [],
"compute": [],
"condition_task": [],
"dashboard_task": [],
"dbt_cloud_task": [],
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -142,6 +142,7 @@
"task": [
{
"clean_rooms_notebook_task": [],
"compute": [],
"condition_task": [],
"dashboard_task": [],
"dbt_cloud_task": [],
Expand Down Expand Up @@ -198,6 +199,7 @@
},
{
"clean_rooms_notebook_task": [],
"compute": [],
"condition_task": [],
"dashboard_task": [],
"dbt_cloud_task": [],
Expand Down Expand Up @@ -264,6 +266,7 @@
},
{
"clean_rooms_notebook_task": [],
"compute": [],
"condition_task": [],
"dashboard_task": [],
"dbt_cloud_task": [],
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -79,6 +79,7 @@
"task": [
{
"clean_rooms_notebook_task": [],
"compute": [],
"condition_task": [],
"dashboard_task": [],
"dbt_cloud_task": [],
Expand Down
4 changes: 2 additions & 2 deletions acceptance/bundle/migrate/runas/out.create_requests.json
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
{
"headers": {
"User-Agent": [
"databricks-tf-provider/1.104.0 databricks-sdk-go/[SDK_VERSION] go/1.24.0 os/[OS] cli/[DEV_VERSION] terraform/1.5.5 sdk/sdkv2 resource/pipeline auth/pat"
"databricks-tf-provider/1.105.0 databricks-sdk-go/[SDK_VERSION] go/1.24.0 os/[OS] cli/[DEV_VERSION] terraform/1.5.5 sdk/sdkv2 resource/pipeline auth/pat"
]
},
"method": "POST",
Expand Down Expand Up @@ -32,7 +32,7 @@
{
"headers": {
"User-Agent": [
"databricks-tf-provider/1.104.0 databricks-sdk-go/[SDK_VERSION] go/1.24.0 os/[OS] cli/[DEV_VERSION] terraform/1.5.5 sdk/sdkv2 resource/permissions auth/pat"
"databricks-tf-provider/1.105.0 databricks-sdk-go/[SDK_VERSION] go/1.24.0 os/[OS] cli/[DEV_VERSION] terraform/1.5.5 sdk/sdkv2 resource/permissions auth/pat"
]
},
"method": "PUT",
Expand Down
11 changes: 11 additions & 0 deletions acceptance/bundle/refschema/out.fields.txt
Original file line number Diff line number Diff line change
Expand Up @@ -97,6 +97,7 @@ resources.apps.*.app_status.state apps.ApplicationState ALL
resources.apps.*.budget_policy_id string ALL
resources.apps.*.compute_size apps.ComputeSize ALL
resources.apps.*.compute_status *apps.ComputeStatus ALL
resources.apps.*.compute_status.active_instances int ALL
resources.apps.*.compute_status.message string ALL
resources.apps.*.compute_status.state apps.ComputeState ALL
resources.apps.*.config *resources.AppConfig INPUT
Expand Down Expand Up @@ -868,6 +869,8 @@ resources.jobs.*.tasks[*].clean_rooms_notebook_task.etag string ALL
resources.jobs.*.tasks[*].clean_rooms_notebook_task.notebook_base_parameters map[string]string ALL
resources.jobs.*.tasks[*].clean_rooms_notebook_task.notebook_base_parameters.* string ALL
resources.jobs.*.tasks[*].clean_rooms_notebook_task.notebook_name string ALL
resources.jobs.*.tasks[*].compute *jobs.Compute ALL
resources.jobs.*.tasks[*].compute.hardware_accelerator compute.HardwareAcceleratorType ALL
resources.jobs.*.tasks[*].condition_task *jobs.ConditionTask ALL
resources.jobs.*.tasks[*].condition_task.left string ALL
resources.jobs.*.tasks[*].condition_task.op jobs.ConditionTaskOp ALL
Expand Down Expand Up @@ -928,6 +931,8 @@ resources.jobs.*.tasks[*].for_each_task.task.clean_rooms_notebook_task.etag stri
resources.jobs.*.tasks[*].for_each_task.task.clean_rooms_notebook_task.notebook_base_parameters map[string]string ALL
resources.jobs.*.tasks[*].for_each_task.task.clean_rooms_notebook_task.notebook_base_parameters.* string ALL
resources.jobs.*.tasks[*].for_each_task.task.clean_rooms_notebook_task.notebook_name string ALL
resources.jobs.*.tasks[*].for_each_task.task.compute *jobs.Compute ALL
resources.jobs.*.tasks[*].for_each_task.task.compute.hardware_accelerator compute.HardwareAcceleratorType ALL
resources.jobs.*.tasks[*].for_each_task.task.condition_task *jobs.ConditionTask ALL
resources.jobs.*.tasks[*].for_each_task.task.condition_task.left string ALL
resources.jobs.*.tasks[*].for_each_task.task.condition_task.op jobs.ConditionTaskOp ALL
Expand Down Expand Up @@ -1602,6 +1607,7 @@ resources.model_serving_endpoints.*.config.auto_capture_config.schema_name strin
resources.model_serving_endpoints.*.config.auto_capture_config.table_name_prefix string INPUT STATE
resources.model_serving_endpoints.*.config.served_entities []serving.ServedEntityInput INPUT STATE
resources.model_serving_endpoints.*.config.served_entities[*] serving.ServedEntityInput INPUT STATE
resources.model_serving_endpoints.*.config.served_entities[*].burst_scaling_enabled bool INPUT STATE
resources.model_serving_endpoints.*.config.served_entities[*].entity_name string INPUT STATE
resources.model_serving_endpoints.*.config.served_entities[*].entity_version string INPUT STATE
resources.model_serving_endpoints.*.config.served_entities[*].environment_vars map[string]string INPUT STATE
Expand Down Expand Up @@ -1673,6 +1679,7 @@ resources.model_serving_endpoints.*.config.served_entities[*].workload_size stri
resources.model_serving_endpoints.*.config.served_entities[*].workload_type serving.ServingModelWorkloadType INPUT STATE
resources.model_serving_endpoints.*.config.served_models []serving.ServedModelInput INPUT STATE
resources.model_serving_endpoints.*.config.served_models[*] serving.ServedModelInput INPUT STATE
resources.model_serving_endpoints.*.config.served_models[*].burst_scaling_enabled bool INPUT STATE
resources.model_serving_endpoints.*.config.served_models[*].environment_vars map[string]string INPUT STATE
resources.model_serving_endpoints.*.config.served_models[*].environment_vars.* string INPUT STATE
resources.model_serving_endpoints.*.config.served_models[*].instance_profile_arn string INPUT STATE
Expand Down Expand Up @@ -1749,6 +1756,7 @@ resources.model_serving_endpoints.*.endpoint_details.config.auto_capture_config.
resources.model_serving_endpoints.*.endpoint_details.config.config_version int64 REMOTE
resources.model_serving_endpoints.*.endpoint_details.config.served_entities []serving.ServedEntityOutput REMOTE
resources.model_serving_endpoints.*.endpoint_details.config.served_entities[*] serving.ServedEntityOutput REMOTE
resources.model_serving_endpoints.*.endpoint_details.config.served_entities[*].burst_scaling_enabled bool REMOTE
resources.model_serving_endpoints.*.endpoint_details.config.served_entities[*].creation_timestamp int64 REMOTE
resources.model_serving_endpoints.*.endpoint_details.config.served_entities[*].creator string REMOTE
resources.model_serving_endpoints.*.endpoint_details.config.served_entities[*].entity_name string REMOTE
Expand Down Expand Up @@ -1830,6 +1838,7 @@ resources.model_serving_endpoints.*.endpoint_details.config.served_entities[*].w
resources.model_serving_endpoints.*.endpoint_details.config.served_entities[*].workload_type serving.ServingModelWorkloadType REMOTE
resources.model_serving_endpoints.*.endpoint_details.config.served_models []serving.ServedModelOutput REMOTE
resources.model_serving_endpoints.*.endpoint_details.config.served_models[*] serving.ServedModelOutput REMOTE
resources.model_serving_endpoints.*.endpoint_details.config.served_models[*].burst_scaling_enabled bool REMOTE
resources.model_serving_endpoints.*.endpoint_details.config.served_models[*].creation_timestamp int64 REMOTE
resources.model_serving_endpoints.*.endpoint_details.config.served_models[*].creator string REMOTE
resources.model_serving_endpoints.*.endpoint_details.config.served_models[*].environment_vars map[string]string REMOTE
Expand Down Expand Up @@ -1883,6 +1892,7 @@ resources.model_serving_endpoints.*.endpoint_details.pending_config.auto_capture
resources.model_serving_endpoints.*.endpoint_details.pending_config.config_version int REMOTE
resources.model_serving_endpoints.*.endpoint_details.pending_config.served_entities []serving.ServedEntityOutput REMOTE
resources.model_serving_endpoints.*.endpoint_details.pending_config.served_entities[*] serving.ServedEntityOutput REMOTE
resources.model_serving_endpoints.*.endpoint_details.pending_config.served_entities[*].burst_scaling_enabled bool REMOTE
resources.model_serving_endpoints.*.endpoint_details.pending_config.served_entities[*].creation_timestamp int64 REMOTE
resources.model_serving_endpoints.*.endpoint_details.pending_config.served_entities[*].creator string REMOTE
resources.model_serving_endpoints.*.endpoint_details.pending_config.served_entities[*].entity_name string REMOTE
Expand Down Expand Up @@ -1964,6 +1974,7 @@ resources.model_serving_endpoints.*.endpoint_details.pending_config.served_entit
resources.model_serving_endpoints.*.endpoint_details.pending_config.served_entities[*].workload_type serving.ServingModelWorkloadType REMOTE
resources.model_serving_endpoints.*.endpoint_details.pending_config.served_models []serving.ServedModelOutput REMOTE
resources.model_serving_endpoints.*.endpoint_details.pending_config.served_models[*] serving.ServedModelOutput REMOTE
resources.model_serving_endpoints.*.endpoint_details.pending_config.served_models[*].burst_scaling_enabled bool REMOTE
resources.model_serving_endpoints.*.endpoint_details.pending_config.served_models[*].creation_timestamp int64 REMOTE
resources.model_serving_endpoints.*.endpoint_details.pending_config.served_models[*].creator string REMOTE
resources.model_serving_endpoints.*.endpoint_details.pending_config.served_models[*].environment_vars map[string]string REMOTE
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,6 @@
"create_time": "[TIMESTAMP]",
"dashboard_id": "[DASHBOARD_ID]",
"display_name": "test dashboard",
"etag": [ETAG],
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

TF changes here likely stopped etags from being serialized in patch request. It was optional always and does not impact correctness for dashboards.

We never set etag in direct PATCH requests.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

"lifecycle_state": "ACTIVE",
"parent_path": "/Users/[USERNAME]/.bundle/change-serialized-dashboard-[UNIQUE_NAME]/default/resources",
"path": "/Users/[USERNAME]/.bundle/change-serialized-dashboard-[UNIQUE_NAME]/default/resources/test dashboard.lvdash.json",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,6 @@
"create_time": "[TIMESTAMP]",
"dashboard_id": "[DASHBOARD_ID]",
"display_name": "test-dashboard-[UNIQUE_NAME]",
"etag": [ETAG],
"lifecycle_state": "ACTIVE",
"parent_path": "/Users/[USERNAME]/.bundle/[UNIQUE_NAME]/resources",
"path": "/Users/[USERNAME]/.bundle/[UNIQUE_NAME]/resources/test-dashboard-[UNIQUE_NAME].lvdash.json",
Expand Down

This file was deleted.

This file was deleted.

Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
{
"method": "DELETE",
"path": "/api/2.0/postgres/[MAIN_ID]"
"path": "/api/2.0/postgres/[MAIN_ID_2]"
}
{
"method": "DELETE",
Expand Down
Original file line number Diff line number Diff line change
@@ -1,14 +1,17 @@
{
"method": "DELETE",
"path": "/api/2.0/postgres/[MAIN_ID]"
}
{
"body": {
"name": "[MAIN_ID]",
"parent": "projects/test-pg-proj-[UNIQUE_NAME]",
"spec": {
"no_expiry": true
}
},
"method": "PATCH",
"path": "/api/2.0/postgres/[MAIN_ID]",
"method": "POST",
"path": "/api/2.0/postgres/projects/test-pg-proj-[UNIQUE_NAME]/branches",
"q": {
"update_mask": "spec"
"branch_id": "new-branch-[UNIQUE_NAME]"
}
}
Original file line number Diff line number Diff line change
Expand Up @@ -34,15 +34,15 @@ mv databricks.yml.new databricks.yml
trace cat databricks.yml

# Plan output differs between engines (recreate vs update)
$CLI bundle plan > out.plan.$DATABRICKS_BUNDLE_ENGINE.txt 2>&1
$CLI bundle plan > out.plan.txt 2>&1
trace $CLI bundle deploy --auto-approve

print_requests update

title "Fetch new branch ID"
branch_id_2=`read_id.py main`
# Get output differs between engines (different branch IDs)
$CLI postgres get-branch $branch_id_2 | jq 'del(.create_time, .update_time, .status.logical_size_bytes)' > out.get_branch.$DATABRICKS_BUNDLE_ENGINE.txt
$CLI postgres get-branch $branch_id_2 | jq 'del(.create_time, .update_time, .status.logical_size_bytes)' > out.get_branch.txt

title "Destroy and verify cleanup"
trace $CLI bundle destroy --auto-approve
Expand Down
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@
Badness = "Terraform provider bug: missing ForceNew on branch_id causes PATCH instead of DELETE+POST when branch_id changes"

[[Repls]]
# Normalize TraceId in error messages (run before default NUMID replacement at Order 9-10)
Old = 'TraceId: [a-f0-9]+'
Expand Down

This file was deleted.

This file was deleted.

This file was deleted.

This file was deleted.

This file was deleted.

Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ project_id_1=`read_id.py my_project`

print_requests() {
local name=$1
jq --sort-keys 'select(.method != "GET" and (.path | contains("/postgres")))' < out.requests.txt > out.requests.${name}.$DATABRICKS_BUNDLE_ENGINE.txt
jq --sort-keys 'select(.method != "GET" and (.path | contains("/postgres")))' < out.requests.txt > out.requests.${name}.txt
rm -f out.requests.txt
}

Expand All @@ -34,15 +34,15 @@ mv databricks.yml.new databricks.yml
trace cat databricks.yml

# Plan output differs between engines (recreate vs update)
$CLI bundle plan > out.plan.$DATABRICKS_BUNDLE_ENGINE.txt 2>&1
$CLI bundle plan > out.plan.txt 2>&1
trace $CLI bundle deploy --auto-approve

print_requests update

title "Fetch new project ID"
project_id_2=`read_id.py my_project`
# Get output differs between engines (different project IDs)
$CLI postgres get-project $project_id_2 > out.get_project.$DATABRICKS_BUNDLE_ENGINE.txt
$CLI postgres get-project $project_id_2 > out.get_project.txt

title "Destroy and verify cleanup"
trace $CLI bundle destroy --auto-approve
Expand Down
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@
Badness = "Terraform provider bug: missing ForceNew on project_id causes PATCH instead of DELETE+POST when project_id changes"

[[Repls]]
# Normalize TraceId in error messages (run before default NUMID replacement at Order 9-10)
Old = 'TraceId: [a-f0-9]+'
Expand Down
4 changes: 2 additions & 2 deletions acceptance/bundle/state/state_present/output.txt
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ Updating deployment state...
Deployment complete!

>>> print_requests.py //api/2.1/unity-catalog/schemas
"databricks-tf-provider/1.104.0 databricks-sdk-go/[SDK_VERSION] go/1.24.0 os/[OS] cli/[DEV_VERSION] terraform/1.5.5 sdk/sdkv2 resource/schema auth/pat"
"databricks-tf-provider/1.105.0 databricks-sdk-go/[SDK_VERSION] go/1.24.0 os/[OS] cli/[DEV_VERSION] terraform/1.5.5 sdk/sdkv2 resource/schema auth/pat"

>>> DATABRICKS_BUNDLE_ENGINE= [CLI] bundle deploy
Uploading bundle files to /Workspace/Users/[USERNAME]/.bundle/test-bundle/default/files...
Expand All @@ -24,7 +24,7 @@ Updating deployment state...
Deployment complete!

>>> print_requests.py --get //api/2.1/unity-catalog/schemas
"databricks-tf-provider/1.104.0 databricks-sdk-go/[SDK_VERSION] go/1.24.0 os/[OS] cli/[DEV_VERSION] terraform/1.5.5 sdk/sdkv2 resource/schema auth/pat"
"databricks-tf-provider/1.105.0 databricks-sdk-go/[SDK_VERSION] go/1.24.0 os/[OS] cli/[DEV_VERSION] terraform/1.5.5 sdk/sdkv2 resource/schema auth/pat"

=== Adding resources.json with lower serial does not change anything
>>> DATABRICKS_BUNDLE_ENGINE=direct musterr [CLI] bundle plan
Expand Down
Loading