I have already set the region to us-east4 and zone to us-east4-c, but still have issue in starting the issue of not be able to create the dataproc.
hailctl dataproc: Creating a cluster with workers of machine type n1-standard-8.
Allocating 14592 MB of memory per executor (4 cores),
with at least 8755 MB for Hail off-heap values and 5837 MB for the JVM. Using a maximum Hail memory reservation of 3648 MB per core.
gcloud dataproc clusters create exome10 \
--image-version=2.0.29-debian10 \
--properties=^|||^spark:spark.task.maxFailures=20|||spark:spark.driver.extraJavaOptions=-Xss4M|||spark:spark.executor.extraJavaOptions=-Xss4M|||spark:spark.speculation=true|||hdfs:dfs.replication=1|||dataproc:dataproc.logging.stackdriver.enable=false|||dataproc:dataproc.monitoring.stackdriver.enable=false|||spark:spark.driver.memory=41g|||yarn:yarn.nodemanager.resource.memory-mb=29184|||yarn:yarn.scheduler.maximum-allocation-mb=14592|||spark:spark.executor.cores=4|||spark:spark.executor.memory=5837m|||spark:spark.executor.memoryOverhead=8755m|||spark:spark.memory.storageFraction=0.2|||spark:spark.executorEnv.HAIL_WORKER_OFF_HEAP_MEMORY_PER_CORE_MB=3648 \
--initialization-actions=gs://hail-common/hailctl/dataproc/0.2.95/init_notebook.py \
--metadata=^|||^WHEEL=gs://hail-common/hailctl/dataproc/0.2.95/hail-0.2.95-py3-none-any.whl|||PKGS=aiohttp==3.8.1|aiohttp_session>=2.7,<2.8|asyncinit>=0.2.4,<0.3|avro>=1.10,<1.12|azure-identity==1.6.0|azure-storage-blob==12.11.0|bokeh>1.3,<2.0|boto3>=1.17,<2.0|botocore>=1.20,<2.0|decorator<5|Deprecated>=1.2.10,<1.3|dill>=0.3.1.1,<0.4|google-auth==1.27.0|google-cloud-storage==1.25.*|humanize==1.0.0|hurry.filesize==0.9|janus>=0.6,<1.1|Jinja2==3.0.3|nest_asyncio==1.5.4|numpy<2|orjson==3.6.4|pandas>=1.3.0,<1.5.0|parsimonious<0.9|plotly>=5.5.0,<5.6|PyJWT|python-json-logger==2.0.2|requests==2.25.1|scipy>1.2,<1.8|sortedcontainers==2.4.0|tabulate==0.8.9|tqdm==4.*|uvloop==0.16.0; sys_platform != 'win32' \
--master-machine-type=n1-highmem-8 \
--master-boot-disk-size=100GB \
--num-master-local-ssds=0 \
--num-secondary-workers=0 \
--num-worker-local-ssds=0 \
--num-workers=2 \
--secondary-worker-boot-disk-size=40GB \
--worker-boot-disk-size=40GB \
--worker-machine-type=n1-standard-8 \
--initialization-action-timeout=20m \
--labels=creator=hufengzhou_g_harvard_edu \
--autoscaling-policy=max-50
Starting cluster 'exome10'...
ERROR: (gcloud.dataproc.clusters.create) INVALID_ARGUMENT: Multiple validation errors:
- 'us-east1' violates constraint 'constraints/gcp.resourceLocations'.
- Zone 'gsp-ccdg-f3/us-east4-c' resides in unsupported region 'https://www.googleapis.com/compute/v1/projects/gsp-ccdg-f3/regions/us-east4'. Supported regions: [us-east1]
Traceback (most recent call last):
File "/n/home05/zhouhufeng/anaconda3/bin/hailctl", line 8, in <module>
sys.exit(main())
File "/n/home05/zhouhufeng/anaconda3/lib/python3.9/site-packages/hailtop/hailctl/__main__.py", line 107, in main
cli.main(args)
File "/n/home05/zhouhufeng/anaconda3/lib/python3.9/site-packages/hailtop/hailctl/dataproc/cli.py", line 123, in main
asyncio.get_event_loop().run_until_complete(
File "/n/home05/zhouhufeng/anaconda3/lib/python3.9/asyncio/base_events.py", line 647, in run_until_complete
return future.result()
File "/n/home05/zhouhufeng/anaconda3/lib/python3.9/site-packages/hailtop/hailctl/dataproc/start.py", line 429, in main
gcloud.run(cmd[1:])
File "/n/home05/zhouhufeng/anaconda3/lib/python3.9/site-packages/hailtop/hailctl/dataproc/gcloud.py", line 9, in run
return subprocess.check_call(["gcloud"] + command)
File "/n/home05/zhouhufeng/anaconda3/lib/python3.9/subprocess.py", line 373, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['gcloud', 'dataproc', 'clusters', 'create', 'exome10', '--image-version=2.0.29-debian10', '--properties=^|||^spark:spark.task.maxFailures=20|||spark:spark.driver.extraJavaOptions=-Xss4M|||spark:spark.executor.extraJavaOptions=-Xss4M|||spark:spark.speculation=true|||hdfs:dfs.replication=1|||dataproc:dataproc.logging.stackdriver.enable=false|||dataproc:dataproc.monitoring.stackdriver.enable=false|||spark:spark.driver.memory=41g|||yarn:yarn.nodemanager.resource.memory-mb=29184|||yarn:yarn.scheduler.maximum-allocation-mb=14592|||spark:spark.executor.cores=4|||spark:spark.executor.memory=5837m|||spark:spark.executor.memoryOverhead=8755m|||spark:spark.memory.storageFraction=0.2|||spark:spark.executorEnv.HAIL_WORKER_OFF_HEAP_MEMORY_PER_CORE_MB=3648', '--initialization-actions=gs://hail-common/hailctl/dataproc/0.2.95/init_notebook.py', "--metadata=^|||^WHEEL=gs://hail-common/hailctl/dataproc/0.2.95/hail-0.2.95-py3-none-any.whl|||PKGS=aiohttp==3.8.1|aiohttp_session>=2.7,<2.8|asyncinit>=0.2.4,<0.3|avro>=1.10,<1.12|azure-identity==1.6.0|azure-storage-blob==12.11.0|bokeh>1.3,<2.0|boto3>=1.17,<2.0|botocore>=1.20,<2.0|decorator<5|Deprecated>=1.2.10,<1.3|dill>=0.3.1.1,<0.4|google-auth==1.27.0|google-cloud-storage==1.25.*|humanize==1.0.0|hurry.filesize==0.9|janus>=0.6,<1.1|Jinja2==3.0.3|nest_asyncio==1.5.4|numpy<2|orjson==3.6.4|pandas>=1.3.0,<1.5.0|parsimonious<0.9|plotly>=5.5.0,<5.6|PyJWT|python-json-logger==2.0.2|requests==2.25.1|scipy>1.2,<1.8|sortedcontainers==2.4.0|tabulate==0.8.9|tqdm==4.*|uvloop==0.16.0; sys_platform != 'win32'", '--master-machine-type=n1-highmem-8', '--master-boot-disk-size=100GB', '--num-master-local-ssds=0', '--num-secondary-workers=0', '--num-worker-local-ssds=0', '--num-workers=2', '--secondary-worker-boot-disk-size=40GB', '--worker-boot-disk-size=40GB', '--worker-machine-type=n1-standard-8', '--initialization-action-timeout=20m', '--labels=creator=hufengzhou_g_harvard_edu', '--autoscaling-policy=max-50']' returned non-zero exit status 1.