Dmytro Grinenko created AMBARI-19588:
----------------------------------------

             Summary: Step7 installer config validation request failed
                 Key: AMBARI-19588
                 URL: https://issues.apache.org/jira/browse/AMBARI-19588
             Project: Ambari
          Issue Type: Bug
    Affects Versions: trunk
            Reporter: Dmytro Grinenko
            Assignee: Dmytro Grinenko
            Priority: Critical


* try install cluster using trunk-build
  1. 3 nodes
  2. HDFS, ZK, AMS, SS
* Proceed to step7
* Click "Next"

Validations request failed with stderr:

{noformat}
Traceback (most recent call last):
  File "/var/lib/ambari-server/resources/scripts/stack_advisor.py", line 158, 
in <module>
    main(sys.argv)
  File "/var/lib/ambari-server/resources/scripts/stack_advisor.py", line 115, 
in main
    result = stackAdvisor.validateConfigurations(services, hosts)
  File "/var/lib/ambari-server/resources/scripts/../stacks/stack_advisor.py", 
line 862, in validateConfigurations
    validationItems = self.getConfigurationsValidationItems(services, hosts)
  File "/var/lib/ambari-server/resources/scripts/../stacks/stack_advisor.py", 
line 943, in getConfigurationsValidationItems
    
items.extend(self.getConfigurationsValidationItemsForService(configurations, 
recommendedDefaults, service, services, hosts))
  File "/var/lib/ambari-server/resources/scripts/../stacks/stack_advisor.py", 
line 967, in getConfigurationsValidationItemsForService
    resultItems = self.validateConfigurationsForSite(configurations, 
recommendedDefaults, services, hosts, siteName, method)
  File "/var/lib/ambari-server/resources/scripts/../stacks/stack_advisor.py", 
line 958, in validateConfigurationsForSite
    return method(siteProperties, siteRecommendations, configurations, 
services, hosts)
  File 
"/var/lib/ambari-server/resources/scripts/./../stacks/HDP/2.0.6/services/stack_advisor.py",
 line 1420, in validateAmsHbaseEnvConfigurations
    requiredMemory = getMemorySizeRequired(hostComponents, configurations)
NameError: global name 'getMemorySizeRequired' is not defined
{noformat}


Stdout:
{noformat}
StackAdvisor implementation for stack HDP, version 2.0.6 was loaded
StackAdvisor implementation for stack HDP, version 2.1 was loaded
StackAdvisor implementation for stack HDP, version 2.2 was loaded
StackAdvisor implementation for stack HDP, version 2.3 was loaded
StackAdvisor implementation for stack HDP, version 2.4 was loaded
StackAdvisor implementation for stack HDP, version 2.5 was loaded
StackAdvisor implementation for stack HDP, version 2.6 was loaded
Returning HDP26StackAdvisor implementation
max_inmemory_regions: -1.15
Processing file: 
/var/lib/ambari-server/resources/stacks/HDP/2.6/services/../../../../common-services/AMBARI_METRICS/0.1.0/package/files/service-metrics/HDFS.txt
Processing file: 
/var/lib/ambari-server/resources/stacks/HDP/2.6/services/../../../../common-services/AMBARI_METRICS/0.1.0/package/files/service-metrics/AMBARI_METRICS.txt
Processing file: 
/var/lib/ambari-server/resources/stacks/HDP/2.6/services/../../../../common-services/AMBARI_METRICS/0.1.0/package/files/service-metrics/HOST.txt
metrics length: 133
2017-01-11 08:54:20,489 - Calculating Hadoop Proxy User recommendations for 
HDFS service.
2017-01-11 08:54:20,489 - Calculating Hadoop Proxy User recommendations for 
YARN service.
2017-01-11 08:54:20,489 - Calculating Hadoop Proxy User recommendations for 
HIVE service.
2017-01-11 08:54:20,489 - Calculating Hadoop Proxy User recommendations for 
OOZIE service.
2017-01-11 08:54:20,489 - Calculating Hadoop Proxy User recommendations for 
FALCON service.
2017-01-11 08:54:20,490 - Calculating Hadoop Proxy User recommendations for 
SPARK service.
2017-01-11 08:54:20,490 - Updated hadoop.proxyuser.hdfs.hosts as : *
2017-01-11 08:54:20,492 - ServiceAdvisor implementation for service SMARTSENSE 
was loaded
SiteName: ams-env, method: validateAmsEnvConfigurations
Site properties: {'ambari_metrics_user': 'ams', 'metrics_monitor_log_dir': 
'/var/log/ambari-metrics-monitor', 'metrics_collector_log_dir': 
'/var/log/ambari-metrics-collector', 'metrics_monitor_pid_dir': 
'/var/run/ambari-metrics-monitor', 'metrics_collector_heapsize': '512', 
'content': '\n# Set environment variables here.\n\n# AMS instance name\nexport 
AMS_INSTANCE_NAME={{hostname}}\n\n# The java implementation to use. Java 1.6 
required.\nexport JAVA_HOME={{java64_home}}\n\n# Collector Log directory for 
log4j\nexport AMS_COLLECTOR_LOG_DIR={{ams_collector_log_dir}}\n\n# Monitor Log 
directory for outfile\nexport AMS_MONITOR_LOG_DIR={{ams_monitor_log_dir}}\n\n# 
Collector pid directory\nexport 
AMS_COLLECTOR_PID_DIR={{ams_collector_pid_dir}}\n\n# Monitor pid 
directory\nexport AMS_MONITOR_PID_DIR={{ams_monitor_pid_dir}}\n\n# AMS HBase 
pid directory\nexport AMS_HBASE_PID_DIR={{hbase_pid_dir}}\n\n# AMS Collector 
heapsize\nexport AMS_COLLECTOR_HEAPSIZE={{metrics_collector_heapsize}}\n\n# 
HBase Tables Initialization check enabled\nexport 
AMS_HBASE_INIT_CHECK_ENABLED={{ams_hbase_init_check_enabled}}\n\n# AMS 
Collector options\nexport 
AMS_COLLECTOR_OPTS="-Djava.library.path=/usr/lib/ams-hbase/lib/hadoop-native"\n{%
 if security_enabled %}\nexport AMS_COLLECTOR_OPTS="$AMS_COLLECTOR_OPTS 
-Djava.security.auth.login.config={{ams_collector_jaas_config_file}}"\n{% endif 
%}\n\n# AMS Collector GC options\nexport 
AMS_COLLECTOR_GC_OPTS="-XX:+UseConcMarkSweepGC -verbose:gc -XX:+PrintGCDetails 
-XX:+PrintGCDateStamps -Xloggc:{{ams_collector_log_dir}}/collector-gc.log-`date 
+\'%Y%m%d%H%M\'`"\nexport AMS_COLLECTOR_OPTS="$AMS_COLLECTOR_OPTS 
$AMS_COLLECTOR_GC_OPTS"\n\n# Metrics collector host will be blacklisted for 
specified number of seconds if metric monitor failed to connect to it.\nexport 
AMS_FAILOVER_STRATEGY_BLACKLISTED_INTERVAL={{failover_strategy_blacklisted_interval}}',
 'metrics_collector_pid_dir': '/var/run/ambari-metrics-collector', 
'timeline.metrics.skip.disk.metrics.patterns': 'true', 
'failover_strategy_blacklisted_interval': '300'}
Recommendations: {'metrics_collector_heapsize': '512'}
SiteName: ams-hbase-env, method: validateAmsHbaseEnvConfigurations
Site properties: {'hbase_pid_dir': '/var/run/ambari-metrics-collector/', 
'hbase_classpath_additional': '', 'regionserver_xmn_size': '128', 
'max_open_files_limit': '32768', 'hbase_master_maxperm_size': '128', 
'hbase_regionserver_xmn_ratio': '0.2', 'hbase_master_heapsize': '640', 
'content': '\n# Set environment variables here.\n\n# The java implementation to 
use. Java 1.6+ required.\nexport JAVA_HOME={{java64_home}}\n\n# HBase 
Configuration directory\nexport 
HBASE_CONF_DIR=${HBASE_CONF_DIR:-{{hbase_conf_dir}}}\n\n# Extra Java CLASSPATH 
elements. Optional.\nadditional_cp={{hbase_classpath_additional}}\nif [  -n 
"$additional_cp" ];\nthen\n  export 
HBASE_CLASSPATH=${HBASE_CLASSPATH}:$additional_cp\nelse\n  export 
HBASE_CLASSPATH=${HBASE_CLASSPATH}\nfi\n\n# The maximum amount of heap to use 
for hbase shell.\nexport HBASE_SHELL_OPTS="-Xmx256m"\n\n# Extra Java runtime 
options.\n# Below are what we set by default. May only work with SUN JVM.\n# 
For more on why as well as other possible settings,\n# see 
http://wiki.apache.org/hadoop/PerformanceTuning\nexport 
HBASE_OPTS="-XX:+UseConcMarkSweepGC 
-XX:ErrorFile={{hbase_log_dir}}/hs_err_pid%p.log 
-Djava.io.tmpdir={{hbase_tmp_dir}}"\nexport SERVER_GC_OPTS="-verbose:gc 
-XX:+PrintGCDetails -XX:+PrintGCDateStamps 
-Xloggc:{{hbase_log_dir}}/gc.log-`date +\'%Y%m%d%H%M\'`"\n# Uncomment below to 
enable java garbage collection logging.\n# export HBASE_OPTS="$HBASE_OPTS 
-verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStamps 
-Xloggc:$HBASE_HOME/logs/gc-hbase.log"\n\n# Uncomment and adjust to enable JMX 
exporting\n# See jmxremote.password and jmxremote.access in 
$JRE_HOME/lib/management to configure remote password access.\n# More details 
at: 
http://java.sun.com/javase/6/docs/technotes/guides/management/agent.html\n#\n# 
export HBASE_JMX_BASE="-Dcom.sun.management.jmxremote.ssl=false 
-Dcom.sun.management.jmxremote.authenticate=false"\n\n{% if java_version < 8 
%}\nexport HBASE_MASTER_OPTS=" -XX:PermSize=64m 
-XX:MaxPermSize={{hbase_master_maxperm_size}} -Xms{{hbase_heapsize}} 
-Xmx{{hbase_heapsize}} -Xmn{{hbase_master_xmn_size}} 
-XX:CMSInitiatingOccupancyFraction=70 
-XX:+UseCMSInitiatingOccupancyOnly"\nexport 
HBASE_REGIONSERVER_OPTS="-XX:MaxPermSize=128m -Xmn{{regionserver_xmn_size}} 
-XX:CMSInitiatingOccupancyFraction=70 -XX:+UseCMSInitiatingOccupancyOnly 
-Xms{{regionserver_heapsize}} -Xmx{{regionserver_heapsize}}"\n{% else 
%}\nexport HBASE_MASTER_OPTS=" -Xms{{hbase_heapsize}} -Xmx{{hbase_heapsize}} 
-Xmn{{hbase_master_xmn_size}} -XX:CMSInitiatingOccupancyFraction=70 
-XX:+UseCMSInitiatingOccupancyOnly"\nexport HBASE_REGIONSERVER_OPTS=" 
-Xmn{{regionserver_xmn_size}} -XX:CMSInitiatingOccupancyFraction=70 
-XX:+UseCMSInitiatingOccupancyOnly -Xms{{regionserver_heapsize}} 
-Xmx{{regionserver_heapsize}}"\n{% endif %}\n\n\n# export 
HBASE_THRIFT_OPTS="$HBASE_JMX_BASE 
-Dcom.sun.management.jmxremote.port=10103"\n# export 
HBASE_ZOOKEEPER_OPTS="$HBASE_JMX_BASE 
-Dcom.sun.management.jmxremote.port=10104"\n\n# File naming hosts on which 
HRegionServers will run. $HBASE_HOME/conf/regionservers by default.\nexport 
HBASE_REGIONSERVERS=${HBASE_CONF_DIR}/regionservers\n\n# Extra ssh options. 
Empty by default.\n# export HBASE_SSH_OPTS="-o ConnectTimeout=1 -o 
SendEnv=HBASE_CONF_DIR"\n\n# Where log files are stored. $HBASE_HOME/logs by 
default.\nexport HBASE_LOG_DIR={{hbase_log_dir}}\n\n# A string representing 
this instance of hbase. $USER by default.\n# export 
HBASE_IDENT_STRING=$USER\n\n# The scheduling priority for daemon processes. See 
\'man nice\'.\n# export HBASE_NICENESS=10\n\n# The directory where pid files 
are stored. /tmp by default.\nexport HBASE_PID_DIR={{hbase_pid_dir}}\n\n# 
Seconds to sleep between slave commands. Unset by default. This\n# can be 
useful in large clusters, where, e.g., slave rsyncs can\n# otherwise arrive 
faster than the master can service them.\n# export HBASE_SLAVE_SLEEP=0.1\n\n# 
Tell HBase whether it should manage it\'s own instance of Zookeeper or 
not.\nexport HBASE_MANAGES_ZK=false\n\n{% if security_enabled %}\nexport 
HBASE_OPTS="$HBASE_OPTS 
-Djava.security.auth.login.config={{client_jaas_config_file}}"\nexport 
HBASE_MASTER_OPTS="$HBASE_MASTER_OPTS 
-Djava.security.auth.login.config={{master_jaas_config_file}}"\nexport 
HBASE_REGIONSERVER_OPTS="$HBASE_REGIONSERVER_OPTS 
-Djava.security.auth.login.config={{regionserver_jaas_config_file}}"\nexport 
HBASE_ZOOKEEPER_OPTS="$HBASE_ZOOKEEPER_OPTS 
-Djava.security.auth.login.config={{ams_zookeeper_jaas_config_file}}"\n{% endif 
%}\n\n# use embedded native 
libs\n_HADOOP_NATIVE_LIB="/usr/lib/ams-hbase/lib/hadoop-native/"\nexport 
HBASE_OPTS="$HBASE_OPTS -Djava.library.path=${_HADOOP_NATIVE_LIB}"\n\n# Unset 
HADOOP_HOME to avoid importing HADOOP installed cluster related configs like: 
/usr/hdp/2.2.0.0-2041/hadoop/conf/\nexport 
HADOOP_HOME={{ams_hbase_home_dir}}\n\n# Explicitly Setting HBASE_HOME for AMS 
HBase so that there is no conflict\nexport HBASE_HOME={{ams_hbase_home_dir}}', 
'hbase_regionserver_shutdown_timeout': '30', 'hbase_regionserver_heapsize': 
'768', 'hbase_log_dir': '/var/log/ambari-metrics-collector', 
'hbase_master_xmn_size': '192'}
Recommendations: {'hbase_master_heapsize': '640', 
'hbase_regionserver_heapsize': '768', 'hbase_log_dir': 
'/var/log/ambari-metrics-collector', 'hbase_master_xmn_size': '192'}
Error occured in stack advisor.
Error details: global name 'getMemorySizeRequired' is not defined
{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to