I have created a custom container for prediction and successfully uploaded the model to Vertex AI. I was also able to deploy the model to an endpoint and successfully request predictions from the endpoint. Within the custom container code, I use the parameters
field as described here, which I then supply later on when making an online prediction request.
My questions are regarding requesting batch predictions from a custom container for prediction.
I cannot find any documentation that describes what happens when I request a batch prediction. Say, for example, I use the
my_model.batch_predict
function from the Python SDK and set theinstances_format
to "csv" and provide thegcs_source
. Now, I have setup my custom container to expect prediction requests at/predict
as described in this documentation. Does Vertex AI make a POST request to this path, converting the cvs data into the appropriate POST body?How do I specify the
parameters
field for batch prediction as I did for online prediction?