Azure ML Assisted Labeling typically uses CPU resources for the labeling process, as it involves human-in-the-loop tasks like data annotation and review. The time required for labeling 500 text files (with 20% pre-labeled) depends on the complexity of the text and the number of annotators, but it could take several hours to a few days.
Costs are primarily driven by the compute instance used for labeling and the storage of labeled data. Azure ML pricing for compute instances varies based on the VM size, but you can expect costs to be relatively low for CPU-based labeling.
I am not expert in this subject but what I read about thr NER model training in Azure AutoML, I recall that a GPU is generally recommended due to the computational intensity of NLP tasks.
Azure offers GPU options like the NC-series ( NC6 or NC12) for training so the training time depends on the dataset size and model complexity but could range from a few hours to a day. Costs will depend on the GPU type and training duration, with hourly rates for GPUs being higher than CPUs. You can always estimate costs using Azure pricing calculator.
You can deploy your NER model to a CPU-based endpoint for inference because it is suitable for real-time predictions if the model is optimized and the request volume is moderate (for example 1,000 daily requests). For higher volumes or latency-sensitive applications, a GPU endpoint might be more appropriate.
To auto-label new files from Blob Storage, you can use Azure ML data labeling capabilities or a pre-trained model to generate labels. These new labeled files can be appended to your existing dataset. Retraining on the updated dataset should ideally use a GPU for efficiency, especially if the dataset grows significantly. Azure ML can automate model replacement in the endpoint, but you’ll need to configure the pipeline to update the endpoint with the latest model.
For automation, a combination of Azure Functions, Event Grid, and Logic Apps is effective. Azure Functions can trigger labeling and retraining workflows, Event Grid can monitor Blob Storage for new files, and Logic Apps can orchestrate the pipeline. Azure ML Pipelines can also be used to automate the entire workflow, including dataset updates, retraining, and endpoint deployment.
GPU charges apply only when the resources are actively used. If you provision a GPU for weekly labeling and retraining, you’ll be charged only for the hours the GPU is in use. For your scenario, with 100 new files daily and weekly retraining, GPU costs will be incurred weekly during the retraining process.