Got an error on SageMaker "GET was unable to find an engine to execute this computation"
#1
by
aljeshi
- opened
Hello,
I have followed all the instructions deploy_llava.ipynb
notebook and it all went without issues or errors. Until I reached the part where the inference happens:
data = {
"image" : 'https://raw.githubusercontent.com/haotian-liu/LLaVA/main/images/llava_logo.png',
"question" : "Describe the image and color details.",
# "max_new_tokens" : 1024,
# "temperature" : 0.2,
# "stop_str" : "###"
}
# request
output = predictor.predict(data)
print(output)
I got the following error:
ModelError: An error occurred (ModelError) when calling the InvokeEndpoint operation: Received client error (400) from primary with message "{
"code": 400,
"type": "InternalServerException",
"message": "GET was unable to find an engine to execute this computation"
}
I tried using the endpoint directly, but it returned the same error.
Any idea on how I can overcome this?
Hi, I have the same issue. I think its because the underlying hardware requirements needed to run this doesnt work as is with sagemaker. I am very confused, but one suggestion someone said was to use Djl containers instead