Get started with MLflow experiments

This collection of quickstart notebooks demonstrate how you can get up and running with MLflow experiment runs.

MLflow components

MLflowis an open source platform for managing the end-to-end machine learning lifecycle. MLflow has three primary components:

  • Tracking

  • Models

  • Projects

The MLflow Tracking component lets you log and query machine model training sessions (runs) using the following APIs:

An MLflowrunis a collection of parameters, metrics, tags, and artifacts associated with a machine learning model training process.

What are experiments in MLflow?

Experimentsare the primary unit of organization in MLflow; all MLflow runs belong to an experiment. Each experiment lets you visualize, search, and compare runs, as well as download run artifacts or metadata for analysis in other tools. Experiments are maintained in a Databricks hosted MLflow tracking server.

Experiments are located in theworkspacefile tree. You manage experiments using the same tools you use to manage other workspace objects such as folders, notebooks, and libraries.

MLflow quickstart notebooks

The following quickstart notebooks demonstrate how to create and log to an MLflow run using the MLflow tracking APIs, as well how to use the experiment UI to view the run. These notebooks are available in Python, Scala, and R.

The Python and R notebooks use anotebook experiment. The Scala notebook creates an experiment in theSharedfolder.

Note

With Databricks Runtime 10.3 ML and above,Databricks Autologgingis enabled by default for Python notebooks.