kevgeo commented on issue #38127:
URL: https://github.com/apache/airflow/issues/38127#issuecomment-2046155880

   Hi @Tchopane, I was trying to reproduce your issue, both in airflow composer 
and locally from dev environment with the following config where I copied your 
software_config values.
   
   ``` 
   CLUSTER_CONFIG = {
       "master_config": {
           "num_instances": 1,
           "machine_type_uri": "n1-standard-4",
           "disk_config": {"boot_disk_type": "pd-standard", 
"boot_disk_size_gb": 32},
       },
       "worker_config": {
           "num_instances": 2,
           "machine_type_uri": "n1-standard-4",
           "disk_config": {"boot_disk_type": "pd-standard", 
"boot_disk_size_gb": 32},
       },
       "secondary_worker_config": {
           "num_instances": 1,
           "machine_type_uri": "n1-standard-4",
           "disk_config": {
               "boot_disk_type": "pd-standard",
               "boot_disk_size_gb": 32,
           },
           "is_preemptible": True,
           "preemptibility": "PREEMPTIBLE",
       },
       "software_config": {
           "image_version": "2.0.95-debian10",
           "properties": {
               
"capacity-scheduler:yarn.scheduler.capacity.maximum-am-resource-percent": "0.5",
               
"capacity-scheduler:yarn.scheduler.capacity.root.default.ordering-policy": 
"fair",
               "core:fs.gs.block.size": "134217728",
               "core:fs.gs.metadata.cache.enable": "false",
               "core:hadoop.ssl.enabled.protocols": "TLSv1,TLSv1.1,TLSv1.2",
               "dataproc:dataproc.allow.zero.workers": "true",
               "dataproc:dataproc.conscrypt.provider.enable": "false",
               "dataproc:dataproc.logging.stackdriver.enable": "false",
               "dataproc:dataproc.monitoring.stackdriver.enable": "false",
               "distcp:mapreduce.map.java.opts": "-Xmx576m",
               "distcp:mapreduce.map.memory.mb": "768",
               "distcp:mapreduce.reduce.java.opts": "-Xmx576m",
               "distcp:mapreduce.reduce.memory.mb": "768",
               "hadoop-env:HADOOP_DATANODE_OPTS": "-Xmx512m",
               "hdfs:dfs.datanode.address": "0.0.0.0:9866",
               "hdfs:dfs.datanode.http.address": "0.0.0.0:9864",
               "hdfs:dfs.datanode.https.address": "0.0.0.0:9865",
               "hdfs:dfs.datanode.ipc.address": "0.0.0.0:9867",
               "hdfs:dfs.namenode.handler.count": "20",
               "hdfs:dfs.namenode.http-address": "0.0.0.0:9870",
               "hdfs:dfs.namenode.https-address": "0.0.0.0:9871",
               "hdfs:dfs.namenode.secondary.http-address": "0.0.0.0:9868",
               "hdfs:dfs.namenode.secondary.https-address": "0.0.0.0:9869",
               "hdfs:dfs.namenode.service.handler.count": "10",
               "hive:hive.fetch.task.conversion": "none",
               "mapred-env:HADOOP_JOB_HISTORYSERVER_HEAPSIZE": "2048",
               "mapred:mapreduce.job.maps": "9",
               "mapred:mapreduce.job.reduce.slowstart.completedmaps": "0.95",
               "mapred:mapreduce.job.reduces": "3",
               "mapred:mapreduce.jobhistory.recovery.store.class": 
"org.apache.hadoop.mapreduce.v2.hs.HistoryServerLeveldbStateStoreService",
               "mapred:mapreduce.map.cpu.vcores": "1",
               "mapred:mapreduce.map.java.opts": "-Xmx2621m",
               "mapred:mapreduce.map.memory.mb": "3277",
               "mapred:mapreduce.reduce.cpu.vcores": "1",
               "mapred:mapreduce.reduce.java.opts": "-Xmx2621m",
               "mapred:mapreduce.reduce.memory.mb": "3277",
               "mapred:mapreduce.task.io.sort.mb": "256",
               "mapred:yarn.app.mapreduce.am.command-opts": "-Xmx2621m",
               "mapred:yarn.app.mapreduce.am.resource.cpu-vcores": "1",
               "mapred:yarn.app.mapreduce.am.resource.mb": "3277",
               "spark-env:SPARK_DAEMON_MEMORY": "2048m",
               "spark:spark.default.parallelism": "32",
               "spark:spark.driver.maxResultSize": "1024m",
               "spark:spark.driver.memory": "2048m",
               "spark:spark.executor.cores": "1",
               "spark:spark.executor.instances": "2",
               "spark:spark.executor.memory": "2893m",
               "spark:spark.executorEnv.OPENBLAS_NUM_THREADS": "1",
               "spark:spark.metrics.namespace": "spark",
               "spark:spark.scheduler.mode": "FAIR",
               
"spark:spark.sql.adaptive.coalescePartitions.initialPartitionNum": "128",
               "spark:spark.sql.cbo.enabled": "true",
               "spark:spark.ui.port": "0",
               "spark:spark.yarn.am.memory": "640m",
               "yarn-env:YARN_NODEMANAGER_HEAPSIZE": "819",
               "yarn-env:YARN_RESOURCEMANAGER_HEAPSIZE": "2048",
               "yarn-env:YARN_TIMELINESERVER_HEAPSIZE": "2048",
               "yarn:yarn.nodemanager.address": "0.0.0.0:8026",
               "yarn:yarn.nodemanager.delete.debug-delay-sec": "86400",
               "yarn:yarn.nodemanager.pmem-check-enabled": "false",
               "yarn:yarn.nodemanager.resource.cpu-vcores": "2",
               "yarn:yarn.nodemanager.resource.memory-mb": "6554",
               "yarn:yarn.nodemanager.vmem-check-enabled": "false",
               
"yarn:yarn.resourcemanager.decommissioning-nodes-watcher.decommission-if-no-shuffle-data":
 "true",
               
"yarn:yarn.resourcemanager.nodemanager-graceful-decommission-timeout-secs": 
"86400",
               "yarn:yarn.scheduler.maximum-allocation-mb": "6554",
               "yarn:yarn.scheduler.minimum-allocation-mb": "1"
           }
       }
   }
   ```
   
   I am not getting any error when creating the Dataproc cluster. Also looking 
at the underlying code logic of the 
[operator](https://github.com/apache/airflow/blob/main/airflow/providers/google/cloud/operators/dataproc.py#L278),
 I do not see a reason why the string values would be converted to int.
   
   Please let me know if i have correctly understood the issue.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to