Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Adds databricks_volume as data source #3211

Merged
merged 17 commits into from
Jul 3, 2024
Merged
Prev Previous commit
Next Next commit
merge conflict
  • Loading branch information
karolusz committed Jul 2, 2024
commit 759d3870de2b9427c595f29f7388830f47ecdbd8
98 changes: 50 additions & 48 deletions provider/provider.go
Original file line number Diff line number Diff line change
Expand Up @@ -67,54 +67,56 @@ func init() {
func DatabricksProvider() *schema.Provider {
p := &schema.Provider{
DataSourcesMap: map[string]*schema.Resource{ // must be in alphabetical order
"databricks_aws_crossaccount_policy": aws.DataAwsCrossaccountPolicy().ToResource(),
"databricks_aws_assume_role_policy": aws.DataAwsAssumeRolePolicy().ToResource(),
"databricks_aws_bucket_policy": aws.DataAwsBucketPolicy().ToResource(),
"databricks_aws_unity_catalog_policy": aws.DataAwsUnityCatalogPolicy().ToResource(),
"databricks_cluster": clusters.DataSourceCluster().ToResource(),
"databricks_clusters": clusters.DataSourceClusters().ToResource(),
"databricks_cluster_policy": policies.DataSourceClusterPolicy().ToResource(),
"databricks_catalog": catalog.DataSourceCatalog().ToResource(),
"databricks_catalogs": catalog.DataSourceCatalogs().ToResource(),
"databricks_current_config": mws.DataSourceCurrentConfiguration().ToResource(),
"databricks_current_metastore": catalog.DataSourceCurrentMetastore().ToResource(),
"databricks_current_user": scim.DataSourceCurrentUser().ToResource(),
"databricks_dbfs_file": storage.DataSourceDbfsFile().ToResource(),
"databricks_dbfs_file_paths": storage.DataSourceDbfsFilePaths().ToResource(),
"databricks_directory": workspace.DataSourceDirectory().ToResource(),
"databricks_external_location": catalog.DataSourceExternalLocation().ToResource(),
"databricks_external_locations": catalog.DataSourceExternalLocations().ToResource(),
"databricks_group": scim.DataSourceGroup().ToResource(),
"databricks_instance_pool": pools.DataSourceInstancePool().ToResource(),
"databricks_instance_profiles": aws.DataSourceInstanceProfiles().ToResource(),
"databricks_jobs": jobs.DataSourceJobs().ToResource(),
"databricks_job": jobs.DataSourceJob().ToResource(),
"databricks_metastore": catalog.DataSourceMetastore().ToResource(),
"databricks_metastores": catalog.DataSourceMetastores().ToResource(),
"databricks_mlflow_model": mlflow.DataSourceModel().ToResource(),
"databricks_mws_credentials": mws.DataSourceMwsCredentials().ToResource(),
"databricks_mws_workspaces": mws.DataSourceMwsWorkspaces().ToResource(),
"databricks_node_type": clusters.DataSourceNodeType().ToResource(),
"databricks_notebook": workspace.DataSourceNotebook().ToResource(),
"databricks_notebook_paths": workspace.DataSourceNotebookPaths().ToResource(),
"databricks_pipelines": pipelines.DataSourcePipelines().ToResource(),
"databricks_schemas": catalog.DataSourceSchemas().ToResource(),
"databricks_service_principal": scim.DataSourceServicePrincipal().ToResource(),
"databricks_service_principals": scim.DataSourceServicePrincipals().ToResource(),
"databricks_share": sharing.DataSourceShare().ToResource(),
"databricks_shares": sharing.DataSourceShares().ToResource(),
"databricks_spark_version": clusters.DataSourceSparkVersion().ToResource(),
"databricks_sql_warehouse": sql.DataSourceWarehouse().ToResource(),
"databricks_sql_warehouses": sql.DataSourceWarehouses().ToResource(),
"databricks_storage_credential": catalog.DataSourceStorageCredential().ToResource(),
"databricks_storage_credentials": catalog.DataSourceStorageCredentials().ToResource(),
"databricks_table": catalog.DataSourceTable().ToResource(),
"databricks_tables": catalog.DataSourceTables().ToResource(),
"databricks_views": catalog.DataSourceViews().ToResource(),
"databricks_volume": catalog.DataSourceVolume().ToResource(),
"databricks_volumes": catalog.DataSourceVolumes().ToResource(),
"databricks_user": scim.DataSourceUser().ToResource(),
"databricks_zones": clusters.DataSourceClusterZones().ToResource(),
"databricks_aws_crossaccount_policy": aws.DataAwsCrossaccountPolicy().ToResource(),
"databricks_aws_assume_role_policy": aws.DataAwsAssumeRolePolicy().ToResource(),
"databricks_aws_bucket_policy": aws.DataAwsBucketPolicy().ToResource(),
"databricks_aws_unity_catalog_assume_role_policy": aws.DataAwsUnityCatalogAssumeRolePolicy().ToResource(),
"databricks_aws_unity_catalog_policy": aws.DataAwsUnityCatalogPolicy().ToResource(),
"databricks_cluster": clusters.DataSourceCluster().ToResource(),
"databricks_clusters": clusters.DataSourceClusters().ToResource(),
"databricks_cluster_policy": policies.DataSourceClusterPolicy().ToResource(),
"databricks_catalog": catalog.DataSourceCatalog().ToResource(),
"databricks_catalogs": catalog.DataSourceCatalogs().ToResource(),
"databricks_current_config": mws.DataSourceCurrentConfiguration().ToResource(),
"databricks_current_metastore": catalog.DataSourceCurrentMetastore().ToResource(),
"databricks_current_user": scim.DataSourceCurrentUser().ToResource(),
"databricks_dbfs_file": storage.DataSourceDbfsFile().ToResource(),
"databricks_dbfs_file_paths": storage.DataSourceDbfsFilePaths().ToResource(),
"databricks_directory": workspace.DataSourceDirectory().ToResource(),
"databricks_external_location": catalog.DataSourceExternalLocation().ToResource(),
"databricks_external_locations": catalog.DataSourceExternalLocations().ToResource(),
"databricks_group": scim.DataSourceGroup().ToResource(),
"databricks_instance_pool": pools.DataSourceInstancePool().ToResource(),
"databricks_instance_profiles": aws.DataSourceInstanceProfiles().ToResource(),
"databricks_jobs": jobs.DataSourceJobs().ToResource(),
"databricks_job": jobs.DataSourceJob().ToResource(),
"databricks_metastore": catalog.DataSourceMetastore().ToResource(),
"databricks_metastores": catalog.DataSourceMetastores().ToResource(),
"databricks_mlflow_experiment": mlflow.DataSourceExperiment().ToResource(),
"databricks_mlflow_model": mlflow.DataSourceModel().ToResource(),
"databricks_mws_credentials": mws.DataSourceMwsCredentials().ToResource(),
"databricks_mws_workspaces": mws.DataSourceMwsWorkspaces().ToResource(),
"databricks_node_type": clusters.DataSourceNodeType().ToResource(),
"databricks_notebook": workspace.DataSourceNotebook().ToResource(),
"databricks_notebook_paths": workspace.DataSourceNotebookPaths().ToResource(),
"databricks_pipelines": pipelines.DataSourcePipelines().ToResource(),
"databricks_schemas": catalog.DataSourceSchemas().ToResource(),
"databricks_service_principal": scim.DataSourceServicePrincipal().ToResource(),
"databricks_service_principals": scim.DataSourceServicePrincipals().ToResource(),
"databricks_share": sharing.DataSourceShare().ToResource(),
"databricks_shares": sharing.DataSourceShares().ToResource(),
"databricks_spark_version": clusters.DataSourceSparkVersion().ToResource(),
"databricks_sql_warehouse": sql.DataSourceWarehouse().ToResource(),
"databricks_sql_warehouses": sql.DataSourceWarehouses().ToResource(),
"databricks_storage_credential": catalog.DataSourceStorageCredential().ToResource(),
"databricks_storage_credentials": catalog.DataSourceStorageCredentials().ToResource(),
"databricks_table": catalog.DataSourceTable().ToResource(),
"databricks_tables": catalog.DataSourceTables().ToResource(),
"databricks_views": catalog.DataSourceViews().ToResource(),
"databricks_volume": catalog.DataSourceVolume().ToResource(),
"databricks_volumes": catalog.DataSourceVolumes().ToResource(),
"databricks_user": scim.DataSourceUser().ToResource(),
"databricks_zones": clusters.DataSourceClusterZones().ToResource(),
},
ResourcesMap: map[string]*schema.Resource{ // must be in alphabetical order
"databricks_access_control_rule_set": permissions.ResourceAccessControlRuleSet().ToResource(),
Expand Down
Loading
You are viewing a condensed version of this merge commit. You can view the full changes here.