Error when deploying Auto ML model for image classification

Chris (Proudback) 20 Reputation points
2025-11-27T03:40:05.33+00:00

Hi community,

we created a simple image classification (single class) model using auto ml. The process seems to finish fine and the model (based on seresnext) is available for deploy.

When using the automated deploy mechanism to a real-time endpoint, system uses generated scoring script, conda files etc. to create the environment.

Creation of the endpoint succeeds but deployment fails with the following error log:

LibMambaUnsatisfiableError: Encountered problems while solving:

2025-11-27T02:19:59: #10 29.79 - nothing provides _python_rc needed by python-3.14.0rc1-h4dad89b_2_cp314t

Then comes a long list of incompatible libraries:

Could not solve for environment specs

2025-11-27T02:19:59: #10 29.79 The following packages are incompatible

2025-11-27T02:19:59: #10 29.79 ├─ botocore =1.23.19 * is installable and it requires

2025-11-27T02:19:59: #10 29.79 │ └─ urllib3 >=1.25.4,<1.27 * with the potential options

2025-11-27T02:19:59: #10 29.79 │ ├─ urllib3 [1.25.5|1.25.6|1.25.7] would require

2025-11-27T02:19:59: #10 29.79 │ │ └─ python >=2.7,<2.8.0a0 * with the potential options

2025-11-27T02:19:59: #10 29.79 │ │ ├─ python [2.7.18|3.10.10|...|3.8.20] would require

2025-11-27T02:19:59: #10 29.79 │ │ │ ├─ libffi >=3.4,<4.0a0 *, which can be installed;

2025-11-27T02:19:59: #10 29.79 │ │ │ └─ pypy3.8 [=7.3.11 *|=7.3.9 *] with the potential options

2025-11-27T02:19:59: #10 29.79 │ │ │ ├─ pypy3.8 [7.3.11|7.3.9] would require

2025-11-27T02:19:59: #10 29.79 │ │ │ │ └─ libffi >=3.4,<4.0a0 *, which can be installed;

2025-11-27T02:19:59: #10 29.79 │ │ │ └─ pypy3.8 [7.3.8|7.3.9] would require

2025-11-27T02:19:59: #10 29.79 │ │ │ └─ libffi >=3.4.2,<3.5.0a0 *, which can be installed;

2025-11-27T02:19:59: #10 29.79 │ │ ├─ python [2.7.13|2.7.14|...|3.8.6] would require

2025-11-27T02:19:59: #10 29.79 │ │ │ ├─ libffi [=3.2 *|>=3.2.1,<3.3.0a0 *|>=3.2.1,<3.3a0 *], which can be installed;

And so on, leading to

ERROR: process "/bin/sh -c ldconfig /usr/local/cuda/lib64/stubs && conda env create -p /azureml-envs/azureml_75ba4dda9b29345db018f712dXXXXXX -f azureml-environment-setup/mutated_conda_dependencies.yml && rm -rf "$HOME/.cache/pip" && conda clean -aqy && CONDA_ROOT_DIR=$(conda info --root) && rm -rf "$CONDA_ROOT_DIR/pkgs" && find "$CONDA_ROOT_DIR" -type d -name pycache -exec rm -rf {} + && ldconfig" did not complete successfully: exit code: 1

Creating the right dependencies is the whole point of the auto creation of environments or is it required to go into the conda files?

Or is there a recommended curated environment for computer vision (i.e. image classification) models?

Thanks!

Chris

Azure Machine Learning
{count} votes

2 answers

Sort by: Most helpful
  1. Anshika Varshney 3,795 Reputation points Microsoft External Staff Moderator
    2025-11-27T04:38:08.03+00:00

    Hi Chris (Proudback),

    Thanks for the question.

    The error you’re seeing usually happens when the AutoML deployment tries to build an environment with package versions that cannot be resolved. In your case, the logs show that the environment is trying to use a release-candidate Python version (python-3.14.0rc1), which causes dependency conflicts during the image-build step.

    AutoML often auto-generates a Conda environment, but sometimes the selected Python version or library versions are incompatible. When that happens, the deployment fails before the model is even packaged. This is a common issue with computer-vision models because they depend on several heavy libraries.

    A good next step is to download the generated conda file from your AutoML run and replace the Python version with a stable one (for example Python 3.8 or 3.10). After updating the file, you can redeploy the model using a custom environment. This typically resolves the dependency conflict.

    You can also test the environment locally or manually rebuild the environment in Azure ML by specifying your own conda YAML. Many users find that creating a simple custom environment avoids these “unsatisfiable dependency” errors completely.

    If the issue continues even after updating the environment, let me know or if there are any remaining questions or additional details, I can help with, I’ll be glad to provide further clarification or guidance.

    Thankyou!


  2. Sina Salam 26,661 Reputation points Volunteer Moderator
    2025-11-27T16:25:29.63+00:00

    Hello Chris (Proudback),

    Welcome to the Microsoft Q&A and thank you for posting your questions here.

    I understand that you are having error when deploying Auto ML model for image classification.

    To avoid any misconception, do the followings:

    1. Get logs (replace names):
    # if you deployed via older az ml service:
    az ml service get-logs --workspace-name MY_WS --name MY_SERVICE --verbose
    # or check portal deployment per your SDK/CLI version:
    az ml online-endpoint get-logs --endpoint-name MY_ENDPOINT --deployment-name MY_DEPLOYMENT
    

    Read logs for either:

    • Explicit provider registration error
    • Stack traces from init()
    • Scheduler/0/3 nodes are available: Insufficient nvidia.com/gpu → run Step C
    1. If provider error:
    az provider register --namespace Microsoft.MachineLearningServices
    az provider show --namespace Microsoft.MachineLearningServices --query registrationState
    
    1. If container crash: locate image name in logs, then:
       docker pull <image>
       docker run -p 8000:5001 <image>
       # run your sample POST to /score to reproduce init/run errors
    
    
    1. If dependent package missing: create a conda YAML with needed packages and use that environment in the deploy config (or use a custom Dockerfile). See AutoML docs for how to supply custom environment. https://learn.microsoft.com/en-us/azure/machine-learning/how-to-auto-train-image-models?view=azureml-api-2

    To fix:

    1. Container crashed because scoring init() throws, requires reading container logs and fixing code/dependencies, not just re-registering providers.
    2. Dependency mismatches between train and inference environment, require environment changes / rebuild or use the correct curated inference image; not solved by region/provider changes. https://learn.microsoft.com/en-us/azure/machine-learning/how-to-auto-train-image-models?view=azureml-api-2

    I hope this is helpful! Do not hesitate to let me know if you have any other questions or clarifications.


    Please don't forget to close up the thread here by upvoting and accept it as an answer if it is helpful.

    0 comments No comments

Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.