May 2024

These features and Databricks platform improvements were released in May 2024.

Note

Releases are staged. Your Databricks account might not be updated until a week or more after the initial release date.

Databricks Runtime 15.0 series support ends

May 31, 2024

Support for Databricks Runtime 15.0 and Databricks Runtime 15.0 for Machine Learning ended on May 31. See Databricks support lifecycles.

Databricks Runtime 15.3 (Beta)

May 30, 2024

Databricks Runtime 15.3 and Databricks Runtime 15.3 ML are now available as Beta releases.

See Databricks Runtime 15.3 and Databricks Runtime 15.2 for Machine Learning.

The compute metrics UI is now available on all Databricks Runtime versions

May 30, 2024

The compute metrics UI has been rolled out to all Databricks Runtime versions. Previously, these metrics were available only on compute resources running on Databricks Runtime 13.3 and above. See View compute metrics.

Improved search and filtering in notebook and SQL editor results tables

May 28, 2024

You can now easily search columns and select filter values from a dropdown list of existing values in the results tables within notebooks and the SQL editor.

New dashboard helps Databricks Marketplace providers monitor listing usage

May 28, 2024

The new Provider Analytics Dashboard enables Databricks Marketplace providers to monitor listing views, requests, and installs. The dashboard pulls data from the Marketplace system tables. See Monitor listing usage metrics using dashboards.

Notebooks on SQL warehouses are now GA

May 28, 2024

Notebooks on SQL warehouses allow you to take advantage of fully managed, instant, and scalable compute for your SQL workloads within the rich, collaborative authoring environment of a notebook. For details, see Use a notebook with a SQL warehouse.

View system-generated federated queries in Query Profile

May 24, 2024

Lakehouse Federation now supports viewing system-generated federated queries and their metrics in the Query Profile. Click the federation scan node in the graph view to reveal the query pushed down into the data source. See View system-generated federated queries.

Compute plane outbound IP addresses must be added to a workspace IP allow list

May 24, 2024

If you configure IP access lists on your workspace, you must either add all public IPs that the compute plane uses to access the control plane to an allow list or configure back-end Private Service Connect. This change will impact all new workspaces on July 29 2024, and existing workspaces on August 26 2024. For more information, see the Databricks Community post.

For example, when you configure a customer-managed VPC, subnets must have outbound access to the public network using a Google Cloud NAT or use a similar approach. Those public IPs must be present in an allow list. See Egress requirements. Alternatively, if you use a Databricks-managed VPC and you configure the managed NAT gateway to access public IPs, those IPs must be present in an allow list.

See Configure IP access lists for workspaces.

OAuth is supported in Lakehouse Federation for Snowflake

May 24, 2024

Unity Catalog now allows you to create Snowflake connections using OAuth. See Run federated queries on Snowflake.

Bulk move and delete workspace objects from the workspace browser

May 24, 2024

You can now select multiple items in the workspace to move or delete. When multiple objects are selected, an action bar appears and has options to move or delete items. Additionally, you can select multiple items using your mouse and drag them to a new location. Existing permissions on objects still apply during bulk move and delete operations.

Unity Catalog objects are available in recents and favorites

May 23, 2024

You can now find Unity Catalog objects like catalogs and schemas in your recents list. You can also favorite Unity Catalog objects in Catalog Explorer and the schema browser and find favorited objects on the workspace homepage.

New dbt-databricks connector 1.8.0 adopts decoupled dbt architecture

May 23, 2024

DBT-Databricks connector 1.8.0 is the first version to adopt the new decoupled dbt architecture. Rather than depend on dbt-core to free customers from having to specify versions for both libraries, the connector now depends on a shared abstraction layer between the adapter and dbt-core. As a result, the connector no longer needs to match the Databricks feature version to that of dbt-core, and is free to adopt semantic versioning. This means that connector developers no longer need to release significant features like compute-per-model as patches.

This release also brings the following:

  • Improvements to the declaration and operation of materialized views and streaming tables, including the ability to schedule automatic refreshes.

  • Support for Unity Catalog securable object tags. To distinguish them from dbt tags, which are metadata that is often used for selecting models in a dbt operation, these tags are named databricks_tags in the model configuration.

  • Several improvements to metadata processing performance.

New compliance and security settings APIs (Public Preview)

May 23, 2024

Databricks has introduced new APIs for the compliance security profile, enhanced security monitoring, and automatic cluster update settings in workspaces. See:

Databricks Runtime 15.2 is GA

May 22, 2024

Databricks Runtime 15.2 and Databricks Runtime 15.2 ML are now generally available.

See Databricks Runtime 15.2 and Databricks Runtime 15.2 for Machine Learning.

New Tableau connector for Delta Sharing

May 22, 2024

The new Tableau Delta Sharing Connector simplifies Tableau Desktop access to data that was shared with you using the Delta Sharing open sharing protocol. See Tableau: Read shared data.

New deep learning recommendation model examples

May 22, 2024

Databricks has published two new examples illustrating modern deep learning recommendation models, including the two-tower model and Meta’s DLRM. To learn more about deep learning recommendation models, see Train recommender models.

Bind storage credentials and external locations to specific workspaces (Public Preview)

May 22, 2024

You can now bind storage credentials and external locations to specific workspaces, preventing access to those objects from other workspaces. This feature is especially helpful if you use workspaces to isolate user data access, for example, if you have separate production and development workspaces or a dedicated workspace for handling sensitive data.

For more information, see (Optional) Assign an external location to specific workspaces and (Optional) Assign a storage credential to specific workspaces.

Git folders are GA

May 22, 2024

Git folders are now generally available. See Git integration with Databricks Git folders. If you are a user of the former “Repos” feature, see What happened to Databricks Repos?.

Databricks Assistant autocomplete (Public Preview)

May 20, 2024

Databricks Assistant autocomplete provides AI-powered suggestions in real-time as you type in notebooks, queries, and files. To enable it, go to Settings > Developer > Experimental Features and toggle Databricks Assistant autocomplete. For details, see AI-based autocomplete.

New changes to Git folder UI

May 17, 2024

You may notice some changes to the user interface for Git folder interactions. We’ve added the following:

  • When you share a Git folder, you will see a new alert in a banner that prompts you to Copy link to create Git folder. When you click the button, a URL is copied to your local clipboard, which you can send to another user. When that recipient user loads that URL in a browser, the user is taken to the workspace where they can create their own Git folder cloned from the same remote Git repository. When the recipient accesses the URL, they will see a Create Git folder dialog in the UI that is pre-populated with the values taken from your Git folder.

    Click the Copy link to Git folder button the banner to share the Git repo configuration for the folder with another user in your Databricks organization
  • Similarly, a new button, Create Git folder, appears on a new alert banner when you view a Git folder created by another user. Click this button to create your own Git folder for the same Git repository, based on the pre-populated values in the Create Git folder dialog.

    When viewing another user's Git folder, click the **Create Git folder** button in the banner to make a copy of that folder in your own workspace

Attribute tag values for Unity Catalog objects can now be 1000 characters long (Public Preview)

May 8, 2024

Attribute tag values in Unity Catalog can now be up to 1000 characters long. The character limit for tag keys remains 255. See Apply tags to Unity Catalog securable objects.

New Previews page

May 8, 2024

Enable and manage access to Databricks previews on the new Previews page. See Manage Databricks Previews.

Compute cloning now clones any libraries installed on the original compute resource

May 8, 2024

When cloning compute, any libraries installed on the original compute resource are also cloned. Alternatively, you can create a clone without libraries. See Clone a compute.

Hive metastore table access controls are deprecated

May 7, 2024

Hive metastore table access controls are deprecated on Databricks Runtime 15.0 and support will be removed in an upcoming DBR version.

Upgrade to Unity Catalog to simplify the security and governance of your data by providing a central place to administer and audit data access across multiple workspaces in your account. See What is Unity Catalog?.

Databricks JDBC driver 2.6.38

May 6, 2024

We have released version 2.6.38 of the Databricks JDBC driver (download). This release adds the following new features and enhancements:

  • Native Parameterized Query support if the server uses SPARK_CLI_SERVICE_PROTOCOL_V8. The limit of the number of parameters in a query is 256 in the native query mode.

  • Data ingestion using a Unity Catalog volume support. See more about Unity Catalog volumes in Connect to cloud object storage using Unity Catalog. To use this, set UseNativeQuery to 1.

  • QueryProfile interface added to IHadoopStatement allows applications to retrieve a query’s query id. The query id can be be used to fetch the query’s metadata using Databricks REST APIs.

  • Async operations for metadata Thrift calls if the server uses SPARK_CLI_SERVICE_PROTOCOL_V9. To use this feature, set EnableAsyncModeForMetadataOperation property to 1.

  • JWT assertion support. The connector now supports JWT assertion OAuth using client credentials. To do this, set the UseJWTAssertion property to 1.

This release also resolves the following issues:

  • Jackson libraries updates. The connector now uses the following libraries for the Jackson JSON parser: jackson-annotations 2.16.0 (previously 2.15.2), jackson-core 2.16.0 (previously 2.15.2), jackson-databind-2.16.0 (previously 2.15.2)

  • The connector contains unshaded class files in META-INF directory.

Databricks Runtime 15.2 (Beta)

May 2, 2024

Databricks Runtime 15.2 and Databricks Runtime 15.2 ML are now available as Beta releases.

See Databricks Runtime 15.2 and Databricks Runtime 15.2 for Machine Learning.

Notebooks now detect and auto-complete column names for Spark Connect DataFrames

May 1, 2024

Databricks notebooks now automatically detect and display the column names in Spark Connect DataFrames and allow you to use auto-complete to select columns.