Delta Lake APIs

For most read and write operations on Delta tables, you can use Apache Spark reader and writer APIs. For examples, see Table batch reads and writes and Table streaming reads and writes.

However, there are some operations that are specific to Delta Lake and you must use Delta Lake APIs. For examples, see Table utility commands.


Some Delta Lake APIs are still evolving and are indicated with the Evolving qualifier in the API docs.

Databricks ensures binary compatibility between the Delta Lake project and Delta Lake in Databricks Runtime. To view the Delta Lake API version packaged in each Databricks Runtime version and links to the API documentation, see the Delta Lake API compatibility matrix.