Feature Store workflow overview

This page gives an overview of how to use Databricks Feature Store in a machine learning workflow.

The typical machine learning workflow using Feature Store follows this path:

  1. Write code to convert raw data into features and create a Spark DataFrame containing the desired features.

  2. Write the DataFrame as a feature table in Feature Store.

  3. Train a model using features from the feature store. When you do this, the model stores the specifications of features used for training. When the model is used for inference, it automatically joins features from the appropriate feature tables.

  4. Register model in Model Registry.

You can now use the model to make predictions on new data. The model automatically retrieves the features it needs from Feature Store.

Feature Store workflow for batch machine learning use cases.

Example notebook

The Feature Store taxi example notebook illustrates the process of creating features, updating them, and using them for model training and batch inference.

Feature Store taxi example notebook

Open notebook in new tab