[英]Serveless inference over multi-model endpoint - Amazon Sagemaker
I created a model on Sagemaker using the following two options.我使用以下两个选项在 Sagemaker 上创建了一个 model。 I also specified the URI for the custom container under ECR as well as the root path for the model archives.我还指定了 ECR 下自定义容器的 URI 以及 model 档案的根路径。
I am able to successfully created provisioned endpoint configuration however, in case of serverless, the following message showed up.我能够成功创建配置的端点配置,但是,在无服务器的情况下,会显示以下消息。 Does this mean that it is absolutely not possible on Sagemaker to have a serverless multimodel endpoint?这是否意味着 Sagemaker 上绝对不可能拥有无服务器多模型端点?
Does this mean that it is absolutely not possible on Sagemaker to have a serverless multimodel endpoint?这是否意味着 Sagemaker 上绝对不可能拥有无服务器多模型端点? Basically with serverless you can deploy each model as a different endpoint and its cost effective as you pay only for usage.基本上使用无服务器,您可以将每个 model 部署为不同的端点,并且它具有成本效益,因为您只需为使用付费。 To answer your question technically you can't deploy multiple models on a serverless endpoint like you do with Multi model endpoints.要从技术上回答您的问题,您不能像使用 Multi model 端点那样在无服务器端点上部署多个模型。
声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.