Training a ML.NET Model with Azure ML

Becca McHenry

Model Builder makes it easy to get started with Machine Learning and create your first model. As you gather more data over time, you may want to continuously refine or retrain your model. Using a combination of CLI and Azure tooling, you can train a new ML.NET model and integrate the training into a pipeline. This blog post shows an example of a training pipeline that can be easily rerun using Azure.

We’re going to use Azure Machine Learning Datasets to track data and an Azure ML Pipeline to train a new model. This retraining pipeline can then be triggered by Azure DevOps. In this post, we will cover:

  1. Creating an Azure Machine Learning Dataset
  2. Training a ML.NET model via the Azure Machine Learning CLI (v2)
  3. Creating a pipeline in Azure DevOps for re-training


  1. Azure Machine Learning Workspace
  2. Compute Cluster in the workspace

Creating an Azure Machine Learning Dataset

  1. Open the workspace in the Microsoft Azure Machine Learning Studio.

  2. We need to create a file dataset. Navigate to Datasets. Click + Create Dataset.

  3. Choose the datasource. We will upload a copy of this song popularity dataset available from Kaggle. It’s a fairly large dataset that I don’t want to maintain locally.

  4. Give the dataset a unique name and make sure to choose “File” as the Dataset type. Screenshot of Azure ML Create Dataset Basic Info step

  5. Upload from a local file to the default workspaceblobstore. Take note of the file name. Screenshot of Azure ML Create Dataset Datastore and file selection step

  6. When the data upload finishes, create the dataset.

  7. Click on the completed dataset to view it. Confirm the preview available in the Explore tab looks correct.

  8. Make note of the dataset name, file name, and if you uploaded multiple versions, the version number. We will use these values in the next step.

Training a ML.NET model via Azure Machine Learning

Now that we have a dataset uploaded to Azure ML we can create an Azure ML training pipeline, and use Azure CLI v2 to run it. The pipeline below will create a Docker container with a ML.NET CLI instance that will conduct the training.

  1. Create the Dockerfile and save it in a new folder for this experiment. If not familiar with Dockerfiles, these file types don’t have an extension. The file should be called “Dockerfile” with no extension, and contain the following:

    RUN dotnet tool install -g microsoft.mlnet-linux-x64
    ENV PATH="$PATH:/root/.dotnet/tools"
  2. We will need to figure out our ML.NET CLI command to train our model. If needed, see installation instructions for the ML.NET CLI. ML.NET CLI Regression command information

    We’re doing regression and will specify a dataset and label column. Text classification and recommendation are also supported for tabular files. Check the command information or ML.NET CLI docs for more details on other training scenarios.

    Make sure to include the option --verbosity q, as some of the CLI features can cause problems in the Linux environment.

    mlnet regression --dataset <YOUR_DATA_FILE_NAME> --label-col <YOUR_LABEL> --output outputs --log-file-path outputs/logs --verbosity q

  3. Create the AzureTrain.yml file in the same folder as the Dockerfile. This is what will be passed to the Azure CLI. By using input data in the pipeline, Azure ML will download the file dataset to our compute. The training file can then be referenced directly. We just need to specify the path in the command to the ML.NET CLI. Do the following:

    • Replace with the unique dataset name, and with the version number (likely 1). Both values are visible in the Dataset tab. In this example the value is dataset: azureml:song_popularity:1.
    • Replace command with the local ML.NET CLI command. Instead of the local file path, we’ll use {} to tell the pipeline to use the download path on the Azure compute. Add the data file name. In this example it is --dataset {}/song_data.csv.
    • Replace the compute with our compute name. The available compute clusters in the workspace are visible under Computes -> Compute clusters.

For more information see command job YAML schema documentation.

    dataset: azureml:<DATASET_NAME>:<VERSION>
    mode: download
experiment_name: mldotnet-training
  local_path: .
command: mlnet regression --dataset {}/<YOUR_DATA_FILE_NAME> --label-col <YOUR_LABEL_COLUMN> --output outputs --log-file-path outputs/logs --verbosity q
compute: azureml:<YOUR-COMPUTE-NAME>
    local_path: .
    dockerfile_path: Dockerfile

Run manually

To kick off training from a local machine, or just test the functionality of the run, we can install and setup the Azure CLI (v2) with ML extension. In these instructions I’m running ml extension version 2.0.7.

  1. Machine learning subcommands require the –workspace/-w and –resource-group/-g parameters. Configure the defaults for the group and workspace of the dataset.
    az configure –-defaults group=<YOUR_RESOURCE_GROUP> workspace=<YOUR_WORKSPACE>

  2. Run the retraining pipeline created in the previous step.
    az ml job create –-file AzureTrain.yml

  3. Check the results of the run online in the Azure Machine Learning Studio under Experiments -> mldotnet-training

Automate training with Azure DevOps Services pipelines

We can run the Azure ML training via Azure DevOps Pipelines. This allows the use of any trigger, including time based or file changes.

Below are the steps to get the Azure ML pipeline running. For more details, see step-by-step instructions for setting up Azure DevOps and Azure ML.

  1. Check the Dockerfile and AzureTrain.yml into source control. It is best to create a new subfolder to put these files into. Azure CLI will upload the whole containing folder when running the experiment.
  2. Create a service connection between Azure ML and Azure DevOps. In Azure DevOps:
    1. Go to Project settings. Select Pipelines -> Service connections
    2. Create a new connection of type Azure Resource Manager
    3. Select Service principal (automatic) and Scope Level Machine Learning Workspace. Configure it to the Resource Group of your Machine Learning workspace. Name it aml-ws.
  3. In Azure DevOps create a new pipeline, using the following file as a template. Replace the variables and trigger (if applicable). The ml-ws-connection is the connection created in step 2. Depending on where the file is checked in, add the AzureTrain.yml file path to the ‘Create training job’ step.
  ml-ws-connection: 'aml-ws' # Workspace Service Connection name
  ml-ws: '<YOUR_VALUE>' # AML Workspace name
  ml-rg: '<YOUR_VALUE>' # AML resource Group name


  vmImage: ubuntu-latest


- task: AzureCLI@2
  displayName: 'Set config functionality'
    azureSubscription: $(ml-ws-connection)
    scriptLocation: inlineScript
    scriptType: 'bash'
    inlineScript: 'az config set extension.use_dynamic_install=yes_without_prompt'

- task: AzureCLI@2
  displayName: 'Install AML CLI (azureml-v2-preview)'
    azureSubscription: $(ml-ws-connection)
    scriptLocation: inlineScript
    scriptType: 'bash'
    inlineScript: 'az extension add -n ml'

- task: AzureCLI@2
  displayName: 'Setup default config values'
    azureSubscription: $(ml-ws-connection)
    scriptLocation: inlineScript
    scriptType: 'bash'
    inlineScript: 'az configure --defaults group=$(ml-rg) workspace=$(ml-ws)'

- task: AzureCLI@2
  displayName: 'Create training job'
    azureSubscription: $(ml-ws-connection)
    scriptLocation: inlineScript
    scriptType: 'bash'
    inlineScript: 'az ml job create --file <YOUR_PATH>/AzureTrain.yml'

Running the Azure CLI job either locally or from Azure DevOps will create an output model in Azure. To see the model, go to Microsoft Azure Machine Learning Studio and navigate to your ML workspace. Click on Experiments -> mldotnet-training. Toggle “View only my runs” to see runs started by the Azure Pipelines Service Principal. The completed training run should be visible. The trained model, and example code, is generated in the Outputs + Logs section, in the outputs folder.

In this post, we’ve created a flexible way to track our data and model via Azure ML. The Azure ML Dataset can be added to and updated while maintaining historical data. This Azure ML retraining pipeline can be run manually or automatically in Azure DevOps. Once your model is trained, you can deploy it using Azure ML custom containers.

Set up your own ML.NET retraining pipeline with Azure Machine Learning Datasets and Azure DevOps? Let us know of any issues, feature requests, or general feedback by filing an issue in the ML.NET Tooling (Model Builder & ML.NET CLI) GitHub repo.


Comments are closed. Login to edit/delete your existing comments

  • Shaun Sullivan 0

    Thus far, I have really enjoyed what (to me) is a critical aspect of ML.NET; which was it had no dependencies or coupling with cloud services. I hope the functionality outlined in this post is not a harbinger of things to come where ML.NET gets sucked into the Azure maelstrom.

    Great product, and I am a huge fan, this is just a vote to make sure that moving forward, we have the “local” DIY options as much as is technically possible. Thanks for all you and the team do!

    electric plum

    • Luis QuintanillaMicrosoft employee 0

      Hi Shaun,

      Thanks for your feedback and support! Training ML.NET models in Azure ML is in no way intended to replace current practices in ML.NET. ML.NET is open-source and we want to give you the choice of how and where you train and deploy your machine learning models to. The intent here was to showcase and gather feedback on alternate automated training patterns if you happen to already be using Azure ML or Azure DevOps as part of your workflows.

  • Sam Wheat 0

    Thanks for posting this article. I hope more developers discover ML.Net – it really is easy to use and very powerful.
    With that said, Microsoft, if you want developers to use to your products you have to support them. I posted a question here and after more than three months it remains unanswered as do many of the other questions posted in the repo. This just looks bad – like Microsoft has abandoned ML.Net.

    • Luis QuintanillaMicrosoft employee 0

      Hi Sam,

      Apologies for that question falling through the cracks. The machine learning samples repo is intended to be a place to hold sample applications for a variety of ML.NET scenarios, but issues aren’t actively monitored. Thank you for calling that out. I think we can do better in communicating that in the repo to avoid confusion and will do so in the future. For future reference, I’d suggest posting questions and issues in the dotnet/machinelearning repository where we are actively monitoring issues. That being said, by no means are thinking of abandoning ML.NET and are making investments in several areas such as enabling more deep learning scenarios, bringing the same AutoML experiences you have in the tooling to the framework for code-first use, and others which you can read more about in our roadmap. Thanks again for your support and feedback.

Feedback usabilla icon