May 2024
These features and Databricks platform improvements were released in May 2024.
Note
Releases are staged. Your Databricks account might not be updated until a week or more after the initial release date.
Databricks Runtime 15.0 series support ends
May 31, 2024
Support for Databricks Runtime 15.0 and Databricks Runtime 15.0 for Machine Learning ended on May 31. See Databricks support lifecycles.
Databricks Runtime 15.3 (Beta)
May 30, 2024
Databricks Runtime 15.3 and Databricks Runtime 15.3 ML are now available as Beta releases.
See Databricks Runtime 15.3 and Databricks Runtime 15.2 for Machine Learning.
The compute metrics UI is now available on all Databricks Runtime versions
May 30, 2024
The compute metrics UI has been rolled out to all Databricks Runtime versions. Previously, these metrics were available only on compute resources running on Databricks Runtime 13.3 and above. See View compute metrics.
Improved search and filtering in notebook and SQL editor results tables
May 28, 2024
You can now easily search columns and select filter values from a dropdown list of existing values in the results tables within notebooks and the SQL editor.
New dashboard helps Databricks Marketplace providers monitor listing usage
May 28, 2024
The new Provider Analytics Dashboard enables Databricks Marketplace providers to monitor listing views, requests, and installs. The dashboard pulls data from the Marketplace system tables. See Monitor listing usage metrics using dashboards.
Notebooks on SQL warehouses are now GA
May 28, 2024
Notebooks on SQL warehouses allow you to take advantage of fully managed, instant, and scalable compute for your SQL workloads within the rich, collaborative authoring environment of a notebook. For details, see Use a notebook with a SQL warehouse.
View system-generated federated queries in Query Profile
May 24, 2024
Lakehouse Federation now supports viewing system-generated federated queries and their metrics in the Query Profile. Click the federation scan node in the graph view to reveal the query pushed down into the data source. See View system-generated federated queries.
Compute plane outbound IP addresses must be added to a workspace IP allow list
May 24, 2024
If you configure IP access lists on your workspace, you must either add all public IPs that the compute plane uses to access the control plane to an allow list or configure back-end Private Service Connect. This change will impact all new workspaces on July 29 2024, and existing workspaces on August 26 2024. For more information, see the Databricks Community post.
For example, when you configure a customer-managed VPC, subnets must have outbound access to the public network using a Google Cloud NAT or use a similar approach. Those public IPs must be present in an allow list. See Egress requirements. Alternatively, if you use a Databricks-managed VPC and you configure the managed NAT gateway to access public IPs, those IPs must be present in an allow list.
OAuth is supported in Lakehouse Federation for Snowflake
May 24, 2024
Unity Catalog now allows you to create Snowflake connections using OAuth. See Run federated queries on Snowflake.
Bulk move and delete workspace objects from the workspace browser
May 24, 2024
You can now select multiple items in the workspace to move or delete. When multiple objects are selected, an action bar appears and has options to move or delete items. Additionally, you can select multiple items using your mouse and drag them to a new location. Existing permissions on objects still apply during bulk move and delete operations.
Unity Catalog objects are available in recents and favorites
May 23, 2024
You can now find Unity Catalog objects like catalogs and schemas in your recents list. You can also favorite Unity Catalog objects in Catalog Explorer and the schema browser and find favorited objects on the workspace homepage.
New dbt-databricks connector 1.8.0 adopts decoupled dbt architecture
May 23, 2024
DBT-Databricks connector 1.8.0 is the first version to adopt the new decoupled dbt architecture. Rather than depend on dbt-core to free customers from having to specify versions for both libraries, the connector now depends on a shared abstraction layer between the adapter and dbt-core. As a result, the connector no longer needs to match the Databricks feature version to that of dbt-core, and is free to adopt semantic versioning. This means that connector developers no longer need to release significant features like compute-per-model as patches.
This release also brings the following:
Improvements to the declaration and operation of materialized views and streaming tables, including the ability to schedule automatic refreshes.
Support for Unity Catalog securable object tags. To distinguish them from dbt tags, which are metadata that is often used for selecting models in a dbt operation, these tags are named
databricks_tags
in the model configuration.Several improvements to metadata processing performance.
New compliance and security settings APIs (Public Preview)
May 23, 2024
Databricks has introduced new APIs for the compliance security profile, enhanced security monitoring, and automatic cluster update settings in workspaces. See:
Databricks Runtime 15.2 is GA
May 22, 2024
Databricks Runtime 15.2 and Databricks Runtime 15.2 ML are now generally available.
See Databricks Runtime 15.2 and Databricks Runtime 15.2 for Machine Learning.
New Tableau connector for Delta Sharing
May 22, 2024
The new Tableau Delta Sharing Connector simplifies Tableau Desktop access to data that was shared with you using the Delta Sharing open sharing protocol. See Tableau: Read shared data.
New deep learning recommendation model examples
May 22, 2024
Databricks has published two new examples illustrating modern deep learning recommendation models, including the two-tower model and Meta’s DLRM. To learn more about deep learning recommendation models, see Train recommender models.
Bind storage credentials and external locations to specific workspaces (Public Preview)
May 22, 2024
You can now bind storage credentials and external locations to specific workspaces, preventing access to those objects from other workspaces. This feature is especially helpful if you use workspaces to isolate user data access, for example, if you have separate production and development workspaces or a dedicated workspace for handling sensitive data.
For more information, see (Optional) Assign an external location to specific workspaces and (Optional) Assign a storage credential to specific workspaces.
Git folders are GA
May 22, 2024
Git folders are now generally available. See Git integration for Databricks Git folders. If you are a user of the former “Repos” feature, see What happened to Databricks Repos?.
Databricks Assistant autocomplete (Public Preview)
May 20, 2024
Databricks Assistant autocomplete provides AI-powered suggestions in real-time as you type in notebooks, queries, and files. To enable it, go to Settings > Developer > Experimental Features and toggle Databricks Assistant autocomplete. For details, see AI-based autocomplete.
New changes to Git folder UI
May 17, 2024
You may notice some changes to the user interface for Git folder interactions. We’ve added the following:
When you share a Git folder, you will see a new alert in a banner that prompts you to Copy link to create Git folder. When you click the button, a URL is copied to your local clipboard, which you can send to another user. When that recipient user loads that URL in a browser, the user is taken to the workspace where they can create their own Git folder cloned from the same remote Git repository. When the recipient accesses the URL, they will see a Create Git folder dialog in the UI that is pre-populated with the values taken from your Git folder.
Similarly, a new button, Create Git folder, appears on a new alert banner when you view a Git folder created by another user. Click this button to create your own Git folder for the same Git repository, based on the pre-populated values in the Create Git folder dialog.
Allow users to copy data to the clipboard from results table
May 9, 2024
Admins can now enable or disable the ability for users to copy data to their clipboard from results tables. Previously, this feature was limited to notebooks. Now, this setting applies to the following interfaces:
Notebooks
Dashboards
Genie Spaces
Catalog Explorer
File editor
SQL editor
Attribute tag values for Unity Catalog objects can now be 1000 characters long (Public Preview)
May 8, 2024
Attribute tag values in Unity Catalog can now be up to 1000 characters long. The character limit for tag keys remains 255. See Apply tags to Unity Catalog securable objects.
New Previews page
May 8, 2024
Enable and manage access to Databricks previews on the new Previews page. See Manage Databricks Previews.
Compute cloning now clones any libraries installed on the original compute resource
May 8, 2024
When cloning compute, any libraries installed on the original compute resource are also cloned. Alternatively, you can create a clone without libraries. See Clone a compute.
Hive metastore table access controls are deprecated
May 7, 2024
Hive metastore table access controls are deprecated on Databricks Runtime 15.0 and support will be removed in an upcoming DBR version.
Upgrade to Unity Catalog to simplify the security and governance of your data by providing a central place to administer and audit data access across multiple workspaces in your account. See What is Unity Catalog?.
Databricks JDBC driver 2.6.38
May 6, 2024
We have released version 2.6.38 of the Databricks JDBC driver (download). This release adds the following new features and enhancements:
Native Parameterized Query support if the server uses
SPARK_CLI_SERVICE_PROTOCOL_V8
. The limit of the number of parameters in a query is256
in the native query mode.Data ingestion using a
Unity Catalog
volume support. See more aboutUnity Catalog
volumes in Connect to cloud object storage using Unity Catalog. To use this, setUseNativeQuery
to1
.QueryProfile
interface added toIHadoopStatement
allows applications to retrieve a query’squery id
. Thequery id
can be be used to fetch the query’s metadata using Databricks REST APIs.Async operations for metadata Thrift calls if the server uses
SPARK_CLI_SERVICE_PROTOCOL_V9
. To use this feature, setEnableAsyncModeForMetadataOperation
property to1
.JWT assertion support. The connector now supports JWT assertion OAuth using client credentials. To do this, set the
UseJWTAssertion
property to1
.
This release also resolves the following issues:
Jackson libraries updates. The connector now uses the following libraries for the Jackson JSON parser: jackson-annotations 2.16.0 (previously 2.15.2), jackson-core 2.16.0 (previously 2.15.2), jackson-databind-2.16.0 (previously 2.15.2)
The connector contains unshaded class files in META-INF directory.
Databricks Runtime 15.2 (Beta)
May 2, 2024
Databricks Runtime 15.2 and Databricks Runtime 15.2 ML are now available as Beta releases.
See Databricks Runtime 15.2 and Databricks Runtime 15.2 for Machine Learning.