Repository: ambari Updated Branches: refs/heads/trunk 02fcea6c3 -> 93a2106a6
http://git-wip-us.apache.org/repos/asf/ambari/blob/93a2106a/ambari-server/src/test/python/stacks/2.2/SPARK/test_job_history_server.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/test/python/stacks/2.2/SPARK/test_job_history_server.py b/ambari-server/src/test/python/stacks/2.2/SPARK/test_job_history_server.py index 6d9081c..4b87531 100644 --- a/ambari-server/src/test/python/stacks/2.2/SPARK/test_job_history_server.py +++ b/ambari-server/src/test/python/stacks/2.2/SPARK/test_job_history_server.py @@ -62,6 +62,7 @@ class TestJobHistoryServer(RMFTestCase): kinit_path_local = '/usr/bin/kinit', principal_name = UnknownConfigurationMock(), user = 'hdfs', + dfs_type = '', action = ['execute'], hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf', ) @@ -125,6 +126,7 @@ class TestJobHistoryServer(RMFTestCase): kinit_path_local='/usr/bin/kinit', principal_name=UnknownConfigurationMock(), security_enabled=True, + dfs_type = '', user=UnknownConfigurationMock() ) @@ -172,6 +174,7 @@ class TestJobHistoryServer(RMFTestCase): kinit_path_local = '/usr/bin/kinit', principal_name = UnknownConfigurationMock(), user = 'hdfs', + dfs_type = '', owner = 'spark', hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf', type = 'directory', @@ -187,6 +190,7 @@ class TestJobHistoryServer(RMFTestCase): kinit_path_local = '/usr/bin/kinit', principal_name = UnknownConfigurationMock(), user = 'hdfs', + dfs_type = '', action = ['execute'], hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf', ) @@ -237,6 +241,7 @@ class TestJobHistoryServer(RMFTestCase): user = UnknownConfigurationMock(), owner = 'spark', hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf', + dfs_type = '', type = 'directory', action = ['create_on_execute'], mode = 0775, @@ -250,6 +255,7 @@ class TestJobHistoryServer(RMFTestCase): kinit_path_local = '/usr/bin/kinit', principal_name = UnknownConfigurationMock(), user = UnknownConfigurationMock(), + dfs_type = '', action = ['execute'], hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf', ) http://git-wip-us.apache.org/repos/asf/ambari/blob/93a2106a/ambari-server/src/test/python/stacks/2.3/MAHOUT/test_mahout_service_check.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/test/python/stacks/2.3/MAHOUT/test_mahout_service_check.py b/ambari-server/src/test/python/stacks/2.3/MAHOUT/test_mahout_service_check.py index 3c44d94..c9f376b 100644 --- a/ambari-server/src/test/python/stacks/2.3/MAHOUT/test_mahout_service_check.py +++ b/ambari-server/src/test/python/stacks/2.3/MAHOUT/test_mahout_service_check.py @@ -46,6 +46,7 @@ class TestMahoutClient(RMFTestCase): keytab = UnknownConfigurationMock(), kinit_path_local = '/usr/bin/kinit', user = 'hdfs', + dfs_type = '', action = ['delete_on_execute'], hdfs_site=self.getConfig()['configurations']['hdfs-site'], principal_name=UnknownConfigurationMock(), default_fs='hdfs://c6401.ambari.apache.org:8020', hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf', type = 'directory', @@ -56,6 +57,7 @@ class TestMahoutClient(RMFTestCase): keytab = UnknownConfigurationMock(), kinit_path_local = '/usr/bin/kinit', user = 'hdfs', + dfs_type = '', owner = 'ambari-qa', hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf', type = 'directory', @@ -68,6 +70,7 @@ class TestMahoutClient(RMFTestCase): kinit_path_local = '/usr/bin/kinit', source = '/tmp/sample-mahout-test.txt', user = 'hdfs', + dfs_type = '', owner = 'ambari-qa', hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf', type = 'file', @@ -79,6 +82,7 @@ class TestMahoutClient(RMFTestCase): keytab = UnknownConfigurationMock(), kinit_path_local = '/usr/bin/kinit', user = 'hdfs', + dfs_type = '', action = ['execute'], hdfs_site=self.getConfig()['configurations']['hdfs-site'], principal_name=UnknownConfigurationMock(), default_fs='hdfs://c6401.ambari.apache.org:8020', hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf', ) http://git-wip-us.apache.org/repos/asf/ambari/blob/93a2106a/ambari-web/app/data/HDP2/site_properties.js ---------------------------------------------------------------------- diff --git a/ambari-web/app/data/HDP2/site_properties.js b/ambari-web/app/data/HDP2/site_properties.js index a3b9451..0cda85c 100644 --- a/ambari-web/app/data/HDP2/site_properties.js +++ b/ambari-web/app/data/HDP2/site_properties.js @@ -3118,6 +3118,21 @@ var hdp2properties = [ "filename": "hbase-env.xml", "category": "Advanced hbase-env" }, + //***************************************** ECS stack******************************************** + { + "id": "puppet var", + "name": "hdfs_log_dir_prefix", + "displayName": "Hadoop Log Dir Prefix", + "description": "The parent directory for Hadoop log files. The HDFS log directory will be ${hadoop_log_dir_prefix} / ${hdfs_user} and the MapReduce log directory will be ${hadoop_log_dir_prefix} / ${mapred_user}.", + "recommendedValue": "/var/log/hadoop", + "isReconfigurable": false, + "displayType": "directory", + "isOverridable": false, + "isVisible": false, + "serviceName": "ECS", + "filename": "hadoop-env.xml", + "category": "General Hadoop" + }, //***************************************** GLUSTERFS stack******************************************** { "id": "site property", http://git-wip-us.apache.org/repos/asf/ambari/blob/93a2106a/contrib/fast-hdfs-resource/src/main/java/org/apache/ambari/fast_hdfs_resource/Runner.java ---------------------------------------------------------------------- diff --git a/contrib/fast-hdfs-resource/src/main/java/org/apache/ambari/fast_hdfs_resource/Runner.java b/contrib/fast-hdfs-resource/src/main/java/org/apache/ambari/fast_hdfs_resource/Runner.java index 291a2d9..0c10ca7 100644 --- a/contrib/fast-hdfs-resource/src/main/java/org/apache/ambari/fast_hdfs_resource/Runner.java +++ b/contrib/fast-hdfs-resource/src/main/java/org/apache/ambari/fast_hdfs_resource/Runner.java @@ -82,8 +82,12 @@ public class Runner { dfs.delete(pathHadoop, true); } } - - } finally { + } + catch(Exception e) { + System.out.println("Exception occurred, Reason: " + e.getMessage()); + e.printStackTrace(); + } + finally { dfs.close(); }
