I'd like to specify a minimum number of workers for my job that autoscaling will not go below (akin to how it works for max_num_workers
). Is this possible? My reason is that sometimes the worker startup takes long enough that the autoscaling decides to drop the number of workers to one, even though doing so is not optimal for my job. I'd still like to use autoscaling in case the job is larger than my estimated minimum.
Minimum number of workers is not yet supported. Could file a ticket with job details so that it support can take a look to understand why it downscales to too few workers?
According to the Autoscaling documentation, you could specify the maximum number of workers in the --maxNumWorkers
option and the --numWorkers
as the initial number of workers. You could find a description of these options in this document
Try using flag 'minNumWorkers' to configure the minimum number of workers for dataflow jobs with autoscaling.
--experiments=minNumWorkers=N
This flag is yet not GA, though the GCP product team is working on it to make it so soon.
You can specify minimum number of workers using --numWorkers
which is the initial number of workers to be used when application is deployed.
You can specify maximum number of workers using --maxNumWorkers
which is when dataflow autoscales then maximum how many workers it can use.
© 2022 - 2024 — McMap. All rights reserved.