Nick Allen created METRON-2162:
----------------------------------

             Summary: Ambari client exception occurred: No JSON object could be 
decoded
                 Key: METRON-2162
                 URL: https://issues.apache.org/jira/browse/METRON-2162
             Project: Metron
          Issue Type: Sub-task
            Reporter: Nick Allen
            Assignee: Nick Allen


{code:java}
TASK [ambari_config : Deploy cluster with Ambari; http://node1:8080] ***********

The full traceback is:
File "/tmp/ansible_LTBkrV/ansible_module_ambari_cluster_state.py", line 211, in 
main
if not blueprint_exists(ambari_url, username, password, blueprint_name):
File "/tmp/ansible_LTBkrV/ansible_module_ambari_cluster_state.py", line 334, in 
blueprint_exists
blueprints = get_blueprints(ambari_url, user, password)
File "/tmp/ansible_LTBkrV/ansible_module_ambari_cluster_state.py", line 315, in 
get_blueprints
services = json.loads(r.content)
File "/usr/lib64/python2.7/json/__init__.py", line 338, in loads
return _default_decoder.decode(s)
File "/usr/lib64/python2.7/json/decoder.py", line 366, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/usr/lib64/python2.7/json/decoder.py", line 384, in raw_decode
raise ValueError("No JSON object could be decoded")

fatal: [node1]: FAILED! => {
"changed": false,
"invocation": {
"module_args": {
"blueprint_name": "metron_blueprint",
"blueprint_var": {
"groups": [
{
"cardinality": 1,
"components": [
{
"name": "NAMENODE"
},
{
"name": "SECONDARY_NAMENODE"
},
{
"name": "RESOURCEMANAGER"
},
{
"name": "HISTORYSERVER"
},
{
"name": "ZOOKEEPER_SERVER"
},
{
"name": "NIMBUS"
},
{
"name": "STORM_UI_SERVER"
},
{
"name": "DRPC_SERVER"
},
{
"name": "HBASE_MASTER"
},
{
"name": "HBASE_CLIENT"
},
{
"name": "APP_TIMELINE_SERVER"
},
{
"name": "DATANODE"
},
{
"name": "HDFS_CLIENT"
},
{
"name": "NODEMANAGER"
},
{
"name": "YARN_CLIENT"
},
{
"name": "MAPREDUCE2_CLIENT"
},
{
"name": "ZOOKEEPER_CLIENT"
},
{
"name": "SUPERVISOR"
},
{
"name": "KAFKA_BROKER"
},
{
"name": "HBASE_REGIONSERVER"
},
{
"name": "KIBANA_MASTER"
},
{
"name": "METRON_INDEXING"
},
{
"name": "METRON_PROFILER"
},
{
"name": "METRON_PCAP"
},
{
"name": "METRON_ENRICHMENT_MASTER"
},
{
"name": "METRON_PARSERS"
},
{
"name": "METRON_REST"
},
{
"name": "METRON_MANAGEMENT_UI"
},
{
"name": "METRON_ALERTS_UI"
},
{
"name": "ES_MASTER"
}
],
"configurations": [],
"name": "host_group_1"
}
],
"required_configurations": [
{
"metron-env": {
"es_hosts": "node1",
"solr_zookeeper_url": "node1:9983",
"storm_rest_addr": "http://node1:8744";,
"zeppelin_server_url": "node1:9995"
}
},
{
"metron-rest-env": {
"metron_jdbc_driver": "org.h2.Driver",
"metron_jdbc_password": "root",
"metron_jdbc_platform": "h2",
"metron_jdbc_url": "jdbc:h2:file:~/metrondb",
"metron_jdbc_username": "root"
}
},
{
"kibana-env": {
"kibana_default_application": "dashboard/AV-YpDmwdXwc6Ua9Muh9",
"kibana_es_url": "http://node1:9200";,
"kibana_log_dir": "/var/log/kibana",
"kibana_pid_dir": "/var/run/kibana",
"kibana_server_host": "0.0.0.0",
"kibana_server_port": 5000
}
}
],
"stack_name": "HDP",
"stack_version": "3.1"
},
"cluster_name": "metron_cluster",
"cluster_state": "present",
"configurations": [
{
"zoo.cfg": {
"dataDir": "/hadoop/zookeeper"
}
},
{
"hadoop-env": {
"dtnode_heapsize": 512,
"hadoop_heapsize": 1024,
"namenode_heapsize": 2048,
"namenode_opt_permsize": "128m"
}
},
{
"hbase-env": {
"hbase_master_heapsize": 512,
"hbase_regionserver_heapsize": 512,
"hbase_regionserver_xmn_max": 512
}
},
{
"hdfs-site": {
"dfs.datanode.data.dir": "/hadoop/hdfs/data",
"dfs.journalnode.edits.dir": "/hadoop/hdfs/journalnode",
"dfs.namenode.checkpoint.dir": "/hadoop/hdfs/namesecondary",
"dfs.namenode.name.dir": "/hadoop/hdfs/namenode",
"dfs.replication": 1
}
},
{
"yarn-env": {
"apptimelineserver_heapsize": 512,
"min_user_id": 500,
"nodemanager_heapsize": 512,
"resourcemanager_heapsize": 1024,
"yarn_heapsize": 512
}
},
{
"mapred-env": {
"jobhistory_heapsize": 256
}
},
{
"mapred-site": {
"mapreduce.jobhistory.recovery.store.leveldb.path": "/hadoop/mapreduce/jhs",
"mapreduce.map.java.opts": "-Xmx1024m",
"mapreduce.map.memory.mb": 1229,
"mapreduce.reduce.java.opts": "-Xmx1024m",
"mapreduce.reduce.memory.mb": 1229
}
},
{
"yarn-site": {
"yarn.nodemanager.local-dirs": "/hadoop/yarn/local",
"yarn.nodemanager.log-dirs": "/hadoop/yarn/log",
"yarn.nodemanager.resource.memory-mb": 4096,
"yarn.timeline-service.leveldb-state-store.path": "/hadoop/yarn/timeline",
"yarn.timeline-service.leveldb-timeline-store.path": "/hadoop/yarn/timeline"
}
},
{
"storm-site": {
"nimbus.childopts": "-Xmx1024m _JAAS_PLACEHOLDER",
"storm.cluster.metrics.consumer.register": "[{\"class\": 
\"org.apache.storm.metric.LoggingMetricsConsumer\"}]",
"storm.local.dir": "/hadoop/storm",
"supervisor.childopts": "-Xmx256m _JAAS_PLACEHOLDER",
"supervisor.slots.ports": "[6700, 6701, 6702, 6703, 6704, 6705]",
"topology.classpath": "/etc/hbase/conf:/etc/hadoop/conf",
"topology.metrics.consumer.register": "[{\"class\": 
\"org.apache.storm.metric.LoggingMetricsConsumer\", \"parallelism.hint\": 1, 
\"whitelist\": [\"kafkaOffset\\\\..+/\", \"__complete-latency\", 
\"__process-latency\", \"__receive\\\\.population$\", 
\"__sendqueue\\\\.population$\", \"__execute-count\", \"__emit-count\", 
\"__ack-count\", \"__fail-count\", \"memory/heap\\\\.usedBytes$\", 
\"memory/nonHeap\\\\.usedBytes$\", \"GC/.+\\\\.count$\", 
\"GC/.+\\\\.timeMs$\"]}]",
"worker.childopts": "-Xmx768m _JAAS_PLACEHOLDER"
}
},
{
"kafka-env": {
"content": "\n#!/bin/bash\n\n# Set KAFKA specific environment variables 
here.\n\n# The java implementation to use.\nexport KAFKA_HEAP_OPTS=\"-Xms256M 
-Xmx256M\"\nexport KAFKA_JVM_PERFORMANCE_OPTS=\"-server -XX:+UseG1GC 
-XX:+DisableExplicitGC -Djava.awt.headless=true\"\nexport 
JAVA_HOME={{java64_home}}\nexport PATH=$PATH:$JAVA_HOME/bin\nexport 
PID_DIR={{kafka_pid_dir}}\nexport LOG_DIR={{kafka_log_dir}}\nexport 
KAFKA_KERBEROS_PARAMS={{kafka_kerberos_params}}\n# Add kafka sink to classpath 
and related depenencies\nif [ -e 
\"/usr/lib/ambari-metrics-kafka-sink/ambari-metrics-kafka-sink.jar\" ]; then\n 
export 
CLASSPATH=$CLASSPATH:/usr/lib/ambari-metrics-kafka-sink/ambari-metrics-kafka-sink.jar\n
 export CLASSPATH=$CLASSPATH:/usr/lib/ambari-metrics-kafka-sink/lib/*\nfi\nif [ 
-f /etc/kafka/conf/kafka-ranger-env.sh ]; then\n . 
/etc/kafka/conf/kafka-ranger-env.sh\nfi"
}
},
{
"kafka-broker": {
"delete.topic.enable": "true",
"log.dirs": "/kafka-log",
"offsets.topic.replication.factor": 1
}
},
{
"metron-rest-env": {
"metron_spring_profiles_active": "dev"
}
},
{
"metron-parsers-env": {
"parsers": "\"bro,snort,yaf\""
}
},
{
"elastic-site": {
"gateway_recover_after_data_nodes": 1,
"index_number_of_replicas": 0,
"index_number_of_shards": 1,
"masters_also_are_datanodes": "1",
"network_host": "[ _local_, _site_ ]",
"zen_discovery_ping_unicast_hosts": "[ node1 ]"
}
}
],
"host": "node1",
"password": "admin",
"port": 8080,
"username": "admin",
"wait_for_complete": true
}
},
"msg": "Ambari client exception occurred: No JSON object could be decoded"
}
to retry, use: --limit 
@/Users/nallen/Development/metron/metron-deployment/development/centos7/ansible/playbook.retry

PLAY RECAP *********************************************************************
node1 : ok=83 changed=16 unreachable=0 failed=1

[WARNING]: Module did not set no_log for password

Ansible failed to complete successfully. Any error output should be
visible above. Please fix these errors and try again.
{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to