Hello,
I recently updated my GCP SDK to 294.0.0 and received the following error when starting a cluster with hailctl.
wma13-a4c:~ 250 1 $ hailctl dataproc start mw --properties=spark:spark.speculation=true --num-preemptible-workers=2 --worker-machine-type=n1-highmem-8 --max-idle=120m --packages slackclient==2.0.0,websocket-client,sklearn,statsmodels,scikit-learn,hdbscan,matplotlib,google-cloud-bigquery,gnomad
Your active configuration is: [seqr-work]
gcloud beta dataproc clusters create
mw
âimage-version=1.4-debian9
âproperties=spark:spark.task.maxFailures=20,spark:spark.driver.extraJavaOptions=-Xss4M,spark:spark.executor.extraJavaOptions=-Xss4M,hdfs:dfs.replication=1,dataproc:dataproc.logging.stackdriver.enable=false,dataproc:dataproc.monitoring.stackdriver.enable=false,spark:spark.speculation=true,spark:spark.driver.memory=41g
âinitialization-actions=gs://hail-common/hailctl/dataproc/0.2.33/init_notebook.py
âmetadata=^|||^WHEEL=gs://hail-common/hailctl/dataproc/0.2.33/hail-0.2.33-py3-none-any.whl|||PKGS=aiohttp>=3.6,<3.7|aiohttp_session>=2.7,<2.8|asyncinit>=0.2.4,<0.3|bokeh>1.1,<1.3|decorator<5|gcsfs==0.2.1|humanize==1.0.0|hurry.filesize==0.9|nest_asyncio|numpy<2|pandas>0.24,<0.26|parsimonious<0.9|PyJWT|python-json-logger==0.1.11|requests>=2.21.0,<2.21.1|scipy>1.2,<1.4|tabulate==0.8.3|tqdm==4.42.1|slackclient==2.0.0|websocket-client|sklearn|statsmodels|scikit-learn|hdbscan|matplotlib|google-cloud-bigquery|gnomad
âmaster-machine-type=n1-highmem-8
âmaster-boot-disk-size=100GB
ânum-master-local-ssds=0
ânum-preemptible-workers=2
ânum-worker-local-ssds=0
ânum-workers=2
âpreemptible-worker-boot-disk-size=40GB
âworker-boot-disk-size=40GB
âworker-machine-type=n1-highmem-8
âzone=us-central1-b
âinitialization-action-timeout=20m
âlabels=creator=mwilson_broadinstitute_org
âmax-idle=120m
Starting cluster âmwââŚ
WARNING: The--num-preemptible-workers
flag is deprecated. Use the--num-secondary-workers
flag instead.
WARNING: The--preemptible-worker-boot-disk-size
flag is deprecated. Use the--secondary-worker-boot-disk-size
flag instead.
ERROR: (gcloud.beta.dataproc.clusters.create) Error parsing [cluster].
The [cluster] resource is not properly specified.
Failed to find attribute [region]. The attribute can be set in the following ways:
- provide the argument [âregion] on the command line
- set the property [dataproc/region]
Traceback (most recent call last):
File â/Users/mwilson/anaconda3/bin/hailctlâ, line 8, in
sys.exit(main())
File â/Users/mwilson/anaconda3/lib/python3.6/site-packages/hailtop/hailctl/main.pyâ, line 94, in main
cli.main(args)
File â/Users/mwilson/anaconda3/lib/python3.6/site-packages/hailtop/hailctl/dataproc/cli.pyâ, line 107, in main
jmp[args.module].main(args, pass_through_args)
File â/Users/mwilson/anaconda3/lib/python3.6/site-packages/hailtop/hailctl/dataproc/start.pyâ, line 202, in main
sp.check_call(cmd)
File â/Users/mwilson/anaconda3/lib/python3.6/subprocess.pyâ, line 311, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command â[âgcloudâ, âbetaâ, âdataprocâ, âclustersâ, âcreateâ, âmwâ, ââimage-version=1.4-debian9â, ââproperties=spark:spark.task.maxFailures=20,spark:spark.driver.extraJavaOptions=-Xss4M,spark:spark.executor.extraJavaOptions=-Xss4M,hdfs:dfs.replication=1,dataproc:dataproc.logging.stackdriver.enable=false,dataproc:dataproc.monitoring.stackdriver.enable=false,spark:spark.speculation=true,spark:spark.driver.memory=41gâ, ââinitialization-actions=gs://hail-common/hailctl/dataproc/0.2.33/init_notebook.pyâ, ââmetadata=^|||^WHEEL=gs://hail-common/hailctl/dataproc/0.2.33/hail-0.2.33-py3-none-any.whl|||PKGS=aiohttp>=3.6,<3.7|aiohttp_session>=2.7,<2.8|asyncinit>=0.2.4,<0.3|bokeh>1.1,<1.3|decorator<5|gcsfs==0.2.1|humanize==1.0.0|hurry.filesize==0.9|nest_asyncio|numpy<2|pandas>0.24,<0.26|parsimonious<0.9|PyJWT|python-json-logger==0.1.11|requests>=2.21.0,<2.21.1|scipy>1.2,<1.4|tabulate==0.8.3|tqdm==4.42.1|slackclient==2.0.0|websocket-client|sklearn|statsmodels|scikit-learn|hdbscan|matplotlib|google-cloud-bigquery|gnomadâ, ââmaster-machine-type=n1-highmem-8â, ââmaster-boot-disk-size=100GBâ, âânum-master-local-ssds=0â, âânum-preemptible-workers=2â, âânum-worker-local-ssds=0â, âânum-workers=2â, ââpreemptible-worker-boot-disk-size=40GBâ, ââworker-boot-disk-size=40GBâ, ââworker-machine-type=n1-highmem-8â, ââzone=us-central1-bâ, ââinitialization-action-timeout=20mâ, ââlabels=creator=mwilson_broadinstitute_orgâ, ââmax-idle=120mâ]â returned non-zero exit status 1.
It looks like google updated the flags for preemptibles. I reverted back to 259.0.0 but figured Iâd let you guys know. Considering I havenât updated the google SDK in awhile Iâm not sure when this change happened, I couldnât find any documentation on it in a brief search.