ModelMesh serving not working with Kserve Client APIs
abhisheksinghblr opened this issue · comments
I used the KServe Client API to serve the model in modelmesh serving.
name='iris-minio'
_namespace = utils.get_default_target_namespace()
default_model_spec = V1beta1InferenceServiceSpec(predictor=V1beta1PredictorSpec(sklearn=V1beta1SKLearnSpec(storage_uri='s3://model-serving-scikit/v4/model.pkl')))
isvc = V1beta1InferenceService(api_version=constants.KSERVE_V1BETA1,kind=constants.KSERVE_KIND, metadata=client.V1ObjectMeta(name=name, namespace=_namespace, annotations={'serving.kserve.io/deploymentMode': 'ModelMesh', 'serving.kserve.io/secretKey': 'localMinIO'}), spec=default_model_spec)
KServe = KServeClient()
KServe.create(isvc)
It created in isvc
but then it does not yet ready:
Model Status:
Copies:
Failed Copies: 0
Last Failure Info:
Message: No ServingRuntime supports specified model type and/or protocol
Model Revision Name: iris-minio__isvc-b789ab6f4a
Reason: NoSupportingRuntime
Time: 2022-12-22T09:45:59Z
States:
Active Model State: FailedToLoad
Target Model State:
Please check, looks to be a bug.
Closing since you opened the same issue in the main repo kserve/modelmesh-serving#293.