please refer to the issue detailed here; /static-proxy?url=https%3A%2F%2Fdiscuss.huggingface.co%2Ft%2Fdeploying-open-ais-whisper-on-sagemaker%2F24761%2F54%3Fu%3Dmllife%3C%2Fa%3E%3C%2Fp%3E basically, the currently sagemaker have a max payload set to 5 MB, and there is no way around it on how the current code for whisper is streaming the file to the end-point using just audio_path as input; which makes this model nearly useless for sagemaker deployment. if someone has done custom inference.py which loads file from s3_path at the endpoint itself and later processes it, please share
basically, the currently sagemaker have a max payload set to 5 MB, and there is no way around it on how the current code for whisper is streaming the file to the end-point using just audio_path as input; which makes this model nearly useless for sagemaker deployment.
if someone has done custom inference.py which loads file from s3_path at the endpoint itself and later processes it, please share
· Sign up or log in to comment