Best Practices for Routing Requests within Inference Clusters

Lawrence Wong 26 Reputation points
2020-09-25T14:14:19.4+00:00

Hi,

I have a Kubernetes Service attached as an inference cluster to an azure machine learning workspace. I have deployed multiple models to that the AKS service, each with their own endpoints. I plan to configure this such that I just need to send the request to one main endpoint, which after applying some conditions, will redirect the request to one of the endpoints (e.g. redirect the request to the appropriate model). Are there any best practices to approach this problem?

There seems to be an Azure ML router using azureml-fe that does something similar, but I cannot find any documentation about it.

Thanks,
Lawrence

Azure Machine Learning
Azure Machine Learning
An Azure machine learning service for building and deploying models.
3,137 questions
{count} votes

Accepted answer
  1. Sethu 76 Reputation points Microsoft Employee
    2020-10-05T21:49:22.66+00:00

    Hello @Lawrence Wong ,
    We do have a solution for this in private preview (called Many Models solution accelerator).
    Please send your email id to AzCommunity[at]microsoft[dot]com). Include title and link to this thread in the email (and reply here once you do for faster response) and we can take the conversation from there.


0 additional answers

Sort by: Most helpful

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.