This is an automated email from the ASF dual-hosted git repository. lfrolov pushed a commit to branch DATALAB-2587 in repository https://gitbox.apache.org/repos/asf/incubator-datalab.git
commit cc5c80ba8239d51f1f598fcc8b4009fe39d17d6e Author: leonidfrolov <[email protected]> AuthorDate: Fri Nov 19 12:07:15 2021 +0200 [DATALAB-2587]: added host ip to spark zeppelin configuration --- .../src/general/scripts/os/zeppelin_dataengine_create_configs.py | 5 +++-- .../src/general/templates/os/dataengine_interpreter_spark.json | 7 +++++++ 2 files changed, 10 insertions(+), 2 deletions(-) diff --git a/infrastructure-provisioning/src/general/scripts/os/zeppelin_dataengine_create_configs.py b/infrastructure-provisioning/src/general/scripts/os/zeppelin_dataengine_create_configs.py index 1f2abcc..8f1a98b 100644 --- a/infrastructure-provisioning/src/general/scripts/os/zeppelin_dataengine_create_configs.py +++ b/infrastructure-provisioning/src/general/scripts/os/zeppelin_dataengine_create_configs.py @@ -56,7 +56,7 @@ spark_link = "https://archive.apache.org/dist/spark/spark-" + spark_version + "/ "-bin-hadoop" + hadoop_version + ".tgz" -def configure_zeppelin_dataengine_interpreter(cluster_name, cluster_dir, os_user, multiple_clusters, spark_master, python_version): +def configure_zeppelin_dataengine_interpreter(cluster_name, cluster_dir, os_user, multiple_clusters, spark_master, python_version, notebook_ip): try: port_number_found = False zeppelin_restarted = False @@ -127,6 +127,7 @@ def configure_zeppelin_dataengine_interpreter(cluster_name, cluster_dir, os_user fr = open(template_file, 'r+') text = fr.read() text = text.replace('CLUSTERNAME', cluster_name) + text = text.replace('HOST_IP', notebook_ip) text = text.replace('PYTHONVERSION', python_version[:3]) text = text.replace('PYTHONVER_FULL', python_version) text = text.replace('SPARK_HOME', cluster_dir + 'spark/') @@ -175,5 +176,5 @@ if __name__ == "__main__": if args.multiple_clusters == 'true': install_remote_livy(args) configure_zeppelin_dataengine_interpreter(args.cluster_name, cluster_dir, args.os_user, - args.multiple_clusters, args.spark_master, args.python_version) + args.multiple_clusters, args.spark_master, args.python_version, args.notebook_ip) update_zeppelin_interpreters(args.multiple_clusters, args.r_enabled) \ No newline at end of file diff --git a/infrastructure-provisioning/src/general/templates/os/dataengine_interpreter_spark.json b/infrastructure-provisioning/src/general/templates/os/dataengine_interpreter_spark.json index ded07c1..eaa85e7 100644 --- a/infrastructure-provisioning/src/general/templates/os/dataengine_interpreter_spark.json +++ b/infrastructure-provisioning/src/general/templates/os/dataengine_interpreter_spark.json @@ -37,6 +37,13 @@ "description": "Spark master uri. ex) spark://masterhost:7077", "type": "string" }, + "spark.driver.host":{ + "envName": "spark.driver.host", + "propertyName": "spark.driver.host", + "value": "HOST_IP", + "description": "driver host ip (zeppelin)", + "type": "string" + }, "zeppelin.spark.concurrentSQL": { "envName": "ZEPPELIN_SPARK_CONCURRENTSQL", "propertyName": "zeppelin.spark.concurrentSQL", --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
