This is an automated email from the ASF dual-hosted git repository.

mykolabodnar pushed a commit to branch DLAB-1985
in repository https://gitbox.apache.org/repos/asf/incubator-dlab.git

commit e6e7a15798beafcca7599c1d5639a955f5199f06
Author: bodnarmykola <bodnarmyk...@gmail.com>
AuthorDate: Fri Aug 21 09:34:16 2020 +0300

    [DLAB-1985] - Got rid of previous logic for allocation of Spark executor 
memory
    [DLAB-1979] - Deep Learning creation fixed
---
 infrastructure-provisioning/src/general/conf/dlab.ini                 | 4 ++--
 infrastructure-provisioning/src/general/lib/aws/actions_lib.py        | 4 ++--
 infrastructure-provisioning/src/general/lib/azure/actions_lib.py      | 4 ++--
 infrastructure-provisioning/src/general/lib/gcp/actions_lib.py        | 4 ++--
 infrastructure-provisioning/src/general/lib/os/fab.py                 | 2 +-
 .../src/general/scripts/aws/dataengine-service_create.py              | 2 +-
 .../src/general/scripts/aws/dataengine-service_prepare.py             | 2 +-
 .../src/general/scripts/os/jupyter_install_dataengine_kernels.py      | 4 ++--
 8 files changed, 13 insertions(+), 13 deletions(-)

diff --git a/infrastructure-provisioning/src/general/conf/dlab.ini 
b/infrastructure-provisioning/src/general/conf/dlab.ini
index 597b260..ed6ff5c 100644
--- a/infrastructure-provisioning/src/general/conf/dlab.ini
+++ b/infrastructure-provisioning/src/general/conf/dlab.ini
@@ -262,7 +262,7 @@ cntk2_version = 2.3.1
 ### CNTK library version for DeepLearning notebook for python3
 cntk_version = 2.7
 ### MXNet library version for DeepLearning notebook for python
-mxnet_version = 1.6.0
+mxnet_version = 1.6.0.post0
 ### Keras library version for Tensor/DeepLearning notebook
 keras_version = 2.1.6
 ### Theano library version for Tensor/DeepLearning notebook
@@ -278,7 +278,7 @@ cudnn_file_name = cudnn-10.1-linux-x64-v7.6.5.32.tgz
 ### R enabled on Jupyter/Zeppelin notebook
 r_enabled = true
 ### Temporary fixed python libraries due to dependencies
-tornado_version = 4.5.3
+tornado_version = 5.1.1
 ipykernel_version = 4.8.2
 ### Version of ungit if previous needed. Use latest as default.
 ungit_version = 1.4.36
diff --git a/infrastructure-provisioning/src/general/lib/aws/actions_lib.py 
b/infrastructure-provisioning/src/general/lib/aws/actions_lib.py
index 8b9ffab..6437a93 100644
--- a/infrastructure-provisioning/src/general/lib/aws/actions_lib.py
+++ b/infrastructure-provisioning/src/general/lib/aws/actions_lib.py
@@ -1649,8 +1649,8 @@ def configure_local_spark(jars_dir, templates_dir, 
memory_type='driver'):
              '/tmp/notebook_spark-defaults_local.conf')
         if not exists('/opt/spark/conf/spark-env.sh'):
             sudo('mv /opt/spark/conf/spark-env.sh.template 
/opt/spark/conf/spark-env.sh')
-            java_home = run("update-alternatives --query java | grep -o 
\'/.*/java-8.*/jre\'").splitlines()[0]
-            sudo("echo 'export JAVA_HOME=\'{}\'' >> 
/opt/spark/conf/spark-env.sh".format(java_home))
+        java_home = run("update-alternatives --query java | grep -o 
--color=never \'/.*/java-8.*/jre\'").splitlines()[0]
+        sudo("echo 'export JAVA_HOME=\'{}\'' >> 
/opt/spark/conf/spark-env.sh".format(java_home))
         if os.environ['application'] == 'zeppelin':
             sudo('echo \"spark.jars $(ls -1 ' + jars_dir + '* | tr \'\\n\' 
\',\')\" >> '
                                                            
'/tmp/notebook_spark-defaults_local.conf')
diff --git a/infrastructure-provisioning/src/general/lib/azure/actions_lib.py 
b/infrastructure-provisioning/src/general/lib/azure/actions_lib.py
index 761be9c..4f2057a 100644
--- a/infrastructure-provisioning/src/general/lib/azure/actions_lib.py
+++ b/infrastructure-provisioning/src/general/lib/azure/actions_lib.py
@@ -1128,8 +1128,8 @@ def configure_local_spark(jars_dir, templates_dir, 
memory_type='driver'):
                                                                                
               spark_memory))
         if not exists('/opt/spark/conf/spark-env.sh'):
             sudo('mv /opt/spark/conf/spark-env.sh.template 
/opt/spark/conf/spark-env.sh')
-            java_home = run("update-alternatives --query java | grep -o 
\'/.*/java-8.*/jre\'").splitlines()[0]
-            sudo("echo 'export JAVA_HOME=\'{}\'' >> 
/opt/spark/conf/spark-env.sh".format(java_home))
+        java_home = run("update-alternatives --query java | grep -o 
--color=never \'/.*/java-8.*/jre\'").splitlines()[0]
+        sudo("echo 'export JAVA_HOME=\'{}\'' >> 
/opt/spark/conf/spark-env.sh".format(java_home))
         if 'spark_configurations' in os.environ:
             dlab_header = sudo('cat /tmp/notebook_spark-defaults_local.conf | 
grep "^#"')
             spark_configurations = 
ast.literal_eval(os.environ['spark_configurations'])
diff --git a/infrastructure-provisioning/src/general/lib/gcp/actions_lib.py 
b/infrastructure-provisioning/src/general/lib/gcp/actions_lib.py
index e550d9d..50d0c4e 100644
--- a/infrastructure-provisioning/src/general/lib/gcp/actions_lib.py
+++ b/infrastructure-provisioning/src/general/lib/gcp/actions_lib.py
@@ -1410,8 +1410,8 @@ def configure_local_spark(jars_dir, templates_dir, 
memory_type='driver'):
                                                                                
               spark_memory))
         if not exists('/opt/spark/conf/spark-env.sh'):
             sudo('mv /opt/spark/conf/spark-env.sh.template 
/opt/spark/conf/spark-env.sh')
-            java_home = run("update-alternatives --query java | grep -o 
\'/.*/java-8.*/jre\'").splitlines()[0]
-            sudo("echo 'export JAVA_HOME=\'{}\'' >> 
/opt/spark/conf/spark-env.sh".format(java_home))
+        java_home = run("update-alternatives --query java | grep -o 
--color=never \'/.*/java-8.*/jre\'").splitlines()[0]
+        sudo("echo 'export JAVA_HOME=\'{}\'' >> 
/opt/spark/conf/spark-env.sh".format(java_home))
         if 'spark_configurations' in os.environ:
             dlab_header = sudo('cat /tmp/notebook_spark-defaults_local.conf | 
grep "^#"')
             spark_configurations = 
ast.literal_eval(os.environ['spark_configurations'])
diff --git a/infrastructure-provisioning/src/general/lib/os/fab.py 
b/infrastructure-provisioning/src/general/lib/os/fab.py
index df8cbe1..7563096 100644
--- a/infrastructure-provisioning/src/general/lib/os/fab.py
+++ b/infrastructure-provisioning/src/general/lib/os/fab.py
@@ -215,7 +215,7 @@ def configure_jupyter(os_user, jupyter_conf_file, 
templates_dir, jupyter_version
             #    http_proxy))
             #sudo('sed -i \'/\[Service\]/ a\Environment=\"HTTPS_PROXY={}\"\'  
/tmp/jupyter-notebook.service'.format(
             #    https_proxy))
-            java_home = run("update-alternatives --query java | grep -o 
\'/.*/java-8.*/jre\'").splitlines()[0]
+            java_home = run("update-alternatives --query java | grep -o 
--color=never \'/.*/java-8.*/jre\'").splitlines()[0]
             sudo('sed -i \'/\[Service\]/ a\Environment=\"JAVA_HOME={}\"\'  
/tmp/jupyter-notebook.service'.format(
                 java_home))
             sudo('\cp /tmp/jupyter-notebook.service 
/etc/systemd/system/jupyter-notebook.service')
diff --git 
a/infrastructure-provisioning/src/general/scripts/aws/dataengine-service_create.py
 
b/infrastructure-provisioning/src/general/scripts/aws/dataengine-service_create.py
index 6017968..0cdba1f 100644
--- 
a/infrastructure-provisioning/src/general/scripts/aws/dataengine-service_create.py
+++ 
b/infrastructure-provisioning/src/general/scripts/aws/dataengine-service_create.py
@@ -40,7 +40,7 @@ parser = argparse.ArgumentParser()
 parser.add_argument('--dry_run', action='store_true', help='Print all 
variables')
 parser.add_argument('--name', type=str, default='', help='Name to be applied 
to Cluster ( MANDATORY !!! )')
 parser.add_argument('--applications', type=str, default='',
-                    help='Set of applications to be installed on EMR (Default 
are: "Hadoop Hive Hue Spark")')
+                    help='Set of applications to be installed on EMR (Default 
are: "Hadoop Hive Hue Spark Livy")')
 parser.add_argument('--master_instance_type', type=str, default='', help='EC2 
instance size for Master-Node (Default: m3.xlarge)')
 parser.add_argument('--slave_instance_type', type=str, default='', help='EC2 
instance size for Worker-Nodes (Default: m3.xlarge)')
 parser.add_argument('--instance_count', type=int, default='',
diff --git 
a/infrastructure-provisioning/src/general/scripts/aws/dataengine-service_prepare.py
 
b/infrastructure-provisioning/src/general/scripts/aws/dataengine-service_prepare.py
index 7dd94d9..3a08ba4 100644
--- 
a/infrastructure-provisioning/src/general/scripts/aws/dataengine-service_prepare.py
+++ 
b/infrastructure-provisioning/src/general/scripts/aws/dataengine-service_prepare.py
@@ -78,7 +78,7 @@ if __name__ == "__main__":
             emr_conf['computational_name'] = os.environ['computational_name']
         else:
             emr_conf['computational_name'] = ''
-        emr_conf['apps'] = 'Hadoop Hive Hue Spark'
+        emr_conf['apps'] = 'Hadoop Hive Hue Spark Livy'
         emr_conf['tag_name'] = '{0}-tag'.format(emr_conf['service_base_name'])
         emr_conf['key_name'] = os.environ['conf_key_name']
         emr_conf['endpoint_tag'] = emr_conf['endpoint_name']
diff --git 
a/infrastructure-provisioning/src/general/scripts/os/jupyter_install_dataengine_kernels.py
 
b/infrastructure-provisioning/src/general/scripts/os/jupyter_install_dataengine_kernels.py
index 01afe7b..33e29a6 100644
--- 
a/infrastructure-provisioning/src/general/scripts/os/jupyter_install_dataengine_kernels.py
+++ 
b/infrastructure-provisioning/src/general/scripts/os/jupyter_install_dataengine_kernels.py
@@ -55,8 +55,8 @@ def configure_notebook(keyfile, hoststring):
     #put(templates_dir + 'run_template.sh', 
'/tmp/{}/run_template.sh'.format(args.cluster_name))
     put(templates_dir + 'notebook_spark-defaults_local.conf', 
'/tmp/{}/notebook_spark-defaults_local.conf'.format(args.cluster_name))
     spark_master_ip = args.spark_master.split('//')[1].split(':')[0]
-    spark_memory = get_spark_memory(True, args.os_user, spark_master_ip, 
keyfile)
-    run('echo "spark.executor.memory {0}m" >> 
/tmp/{1}/notebook_spark-defaults_local.conf'.format(spark_memory, 
args.cluster_name))
+    #spark_memory = get_spark_memory(True, args.os_user, spark_master_ip, 
keyfile)
+    #run('echo "spark.executor.memory {0}m" >> 
/tmp/{1}/notebook_spark-defaults_local.conf'.format(spark_memory, 
args.cluster_name))
     if not exists('/usr/local/bin/jupyter_dataengine_create_configs.py'):
         put(scripts_dir + 'jupyter_dataengine_create_configs.py', 
'/usr/local/bin/jupyter_dataengine_create_configs.py', use_sudo=True)
         sudo('chmod 755 /usr/local/bin/jupyter_dataengine_create_configs.py')


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@dlab.apache.org
For additional commands, e-mail: commits-h...@dlab.apache.org

Reply via email to