How to load model programmatically? #2360
Replies: 3 comments
-
Hello @velociraptor111, As of now, the Python SDK doesn't have the ability to reference a defined model in the SageMaker platform. I have added an item in the backlog (MLFW-2709), as I believe implementing this ability in conjunction with the existing Model class will take some time. It is possible to do this using the AWS SDK or boto3 directly, however that won't be as convenient. If you do go this route, you would specify the model name when generating the endpoint configuration.
Let me reach out to the corresponding team and get back to you on that. Thanks! |
Beta Was this translation helpful? Give feedback.
-
Just wanted to circle back on this:
The free tier limit for model deployment is 125 hours of m4.xlarge or m5.xlarge for real-time inference and batch transform, and this usage can be for a single model or aggregated across multiple models. |
Beta Was this translation helpful? Give feedback.
-
@ChoiByungWook is there any update on loading a model from SageMaker by its name? |
Beta Was this translation helpful? Give feedback.
-
I notice that whenever I run a new batch transform jobs, it will create a new model and saves it.
I can see all the models from my batch transform jobs in my AWS Sagemaker Dashboard/inference/models
Here is the script that I run
I've looked into the source code for declaration of class MXNetModel
But I am not seeing anywhere where I can simply load the MXNetModel object using a URL Endpoint to the models in my dashboard.
If I go to console and click one of those models, I can see a button for Create batch transform job, so I know internally this is possible. But I can't find anything on the docs to do it programmatically.
Also as a side question:
How many models does the Free tier provide? In the free tier page: https://aws.amazon.com/sagemaker/pricing/ it just says the number of hours, but not necessarily the number of models
Beta Was this translation helpful? Give feedback.
All reactions