title | description | services | author | ms.service | ms.author | manager | ms.custom | ms.topic | ms.date |
---|---|---|---|---|---|---|---|---|---|
include file |
include file |
machine-learning |
sdgilley |
machine-learning |
sgilley |
cgronlund |
include file |
include |
08/23/2019 |
Compute target | Used for | GPU support | FPGA support | Description |
---|---|---|---|---|
Local web service | Testing/debugging | Use for limited testing and troubleshooting. Hardware acceleration depends on use of libraries in the local system. | ||
Notebook VM web service | Testing/debugging | Use for limited testing and troubleshooting. | ||
Azure Kubernetes Service (AKS) | Real-time inference | Yes (web service deployment) | Yes | Use for high-scale production deployments. Provides fast response time and autoscaling of the deployed service. Cluster autoscaling isn't supported through the Azure Machine Learning SDK. To change the nodes in the AKS cluster, use the UI for your AKS cluster in the Azure portal. AKS is the only option available for the visual interface. |
Azure Container Instances | Testing or development | Use for low-scale CPU-based workloads that require less than 48 GB of RAM. | ||
Azure Machine Learning Compute | (Preview) Batch inference | Yes (machine learning pipeline) | Run batch scoring on serverless compute. Supports normal and low-priority VMs. | |
Azure IoT Edge | (Preview) IoT module | Deploy and serve ML models on IoT devices. | ||
Azure Data Box Edge | Via IoT Edge | Yes | Deploy and serve ML models on IoT devices. |
Note
Although compute targets like local, Notebook VM, and Azure Machine Learning Compute support GPU for training and experimentation, using GPU for inference when deployed as a web service is supported only on Azure Kubernetes Service.
Using a GPU for inference when scoring with a machine learning pipeline is supported only on Azure Machine Learning Compute.