Missed one important point. You actually choose the data center here [2] in the configuration file.
[2] https://github.com/apache/metron/blob/master/metron-deployment/amazon-ec2/conf/defaults.yml#L43 On Fri, May 19, 2017 at 1:52 PM, Nick Allen <n...@nickallen.org> wrote: > I am not sure what the issue is. Nothing is jumping out at me. > > Weird issues can arise when the machine that you are running the Ansible > deployment from has intermittent connectivity issues to the AWS data center. > > Is Amazon's us-west-2 data center the one closest to you? You might try a > different/closer data center. You can alter the data center here [1]. > > [1] https://github.com/apache/metron/blob/master/metron- > deployment/amazon-ec2/conf/defaults.yml#L27-L39 > > Other than that, you could try standing up a small node in AWS and then > use that to run the Metron deployment from. That always works much more > smoothly. > > > > > On Wed, May 17, 2017 at 4:41 PM, Laurens Vets <laur...@daemon.be> wrote: > >> Also, it's only 1 out of 10 hosts that failing it seems. >> >> >> On 2017-05-17 13:25, Laurens Vets wrote: >> >>> When you say Metron host, do you mean the host I'm pushing from to ec2 >>> or the master ambari server in ec2? >>> >>> On 2017-05-17 11:51, Ryan Merriman wrote: >>> >>>> That happens when you don't have the zookeeper url configured correctly. >>>> Can you check the contents of the /etc/sysconfig/metron file on the >>>> Metron >>>> host? >>>> >>>> On Wed, May 17, 2017 at 1:36 PM, Laurens Vets <laur...@daemon.be> >>>> wrote: >>>> >>>> For testing purposes, I decided to spin up the default Metron AWS >>>>> config. >>>>> This resulted in a hang from ansible here: >>>>> >>>>> TASK [librdkafka : include] ****************************** >>>>> ********************** >>>>> task path: /home/laurens/SAPSource/metron/metron-deployment/roles/ >>>>> librdkafka/tasks/main.yml:18 >>>>> included: /home/laurens/SAPSource/metron/metron-deployment/roles/ >>>>> librdkafka/tasks/dependencies.yml for ec2-34-210-194-189.us-west-2.c >>>>> ompute.amazonaws.com >>>>> >>>>> TASK [librdkafka : Install prerequisites] >>>>> ****************************** >>>>> ******** >>>>> task path: /home/laurens/SAPSource/metron/metron-deployment/roles/ >>>>> librdkafka/tasks/dependencies.yml:18 >>>>> <ec2-34-210-194-189.us-west-2.compute.amazonaws.com> ESTABLISH >>>>> CONNECTION >>>>> FOR USER: centos on PORT 22 TO ec2-34-210-194-189.us-west-2.c >>>>> ompute.amazonaws.com >>>>> /usr/lib/python2.7/dist-packages/Crypto/Cipher/blockalgo.py:141: >>>>> FutureWarning: CTR mode needs counter parameter, not IV >>>>> self._cipher = factory.new(key, *args, **kwargs) >>>>> <ec2-34-210-194-189.us-west-2.compute.amazonaws.com> EXEC ( umask 22 >>>>> && >>>>> mkdir -p "$( echo $HOME/.ansible/tmp/ansible-tmp >>>>> -1495041091.74-92163853889508 >>>>> )" && echo "$( echo $HOME/.ansible/tmp/ansible-tmp >>>>> -1495041091.74-92163853889508 >>>>> )" ) >>>>> <ec2-34-210-194-189.us-west-2.compute.amazonaws.com> PUT >>>>> /tmp/tmpwnH61y >>>>> TO /home/centos/.ansible/tmp/ansible-tmp-1495041091.74-92163853 >>>>> 889508/yum >>>>> <ec2-34-210-194-189.us-west-2.compute.amazonaws.com> EXEC /bin/sh -c >>>>> 'sudo -H -S -n -u root /bin/sh -c '"'"'echo >>>>> BECOME-SUCCESS-rmswjjyhfdywqvwtvqwcmbsqpsbohvxh; >>>>> LANG=en_CA.UTF-8 LC_ALL=en_CA.UTF-8 LC_MESSAGES=en_CA.UTF-8 >>>>> /usr/bin/python >>>>> -tt /home/centos/.ansible/tmp/ansible-tmp-1495041091.74-92163853 >>>>> 889508/yum; >>>>> rm -rf "/home/centos/.ansible/tmp/ansible-tmp-1495041091.74-9216385 >>>>> 3889508/" >>>>> > /dev/null 2>&1'"'"'' >>>>> >>>>> Looking in the machine logs, I see the following for Kafka and Metron >>>>> REST: >>>>> >>>>> Kafka: >>>>> [2017-05-17 17:03:14,831] INFO KafkaConfig values: >>>>> advertised.host.name = null >>>>> metric.reporters = [] >>>>> quota.producer.default = 9223372036854775807 >>>>> offsets.topic.num.partitions = 50 >>>>> log.flush.interval.messages = 9223372036854775807 >>>>> auto.create.topics.enable = true >>>>> controller.socket.timeout.ms = 30000 >>>>> log.flush.interval.ms = null >>>>> principal.builder.class = class org.apache.kafka.common.securi >>>>> ty.auth.DefaultPrincipalBuilder >>>>> replica.socket.receive.buffer.bytes = 65536 >>>>> min.insync.replicas = 1 >>>>> replica.fetch.wait.max.ms = 500 >>>>> num.recovery.threads.per.data.dir = 1 >>>>> ssl.keystore.type = JKS >>>>> sasl.mechanism.inter.broker.protocol = GSSAPI >>>>> default.replication.factor = 1 >>>>> ssl.truststore.password = null >>>>> log.preallocate = false >>>>> sasl.kerberos.principal.to.local.rules = [DEFAULT] >>>>> fetch.purgatory.purge.interval.requests = 10000 >>>>> ssl.endpoint.identification.algorithm = null >>>>> replica.socket.timeout.ms = 30000 >>>>> message.max.bytes = 1000000 >>>>> num.io.threads = 8 >>>>> offsets.commit.required.acks = -1 >>>>> log.flush.offset.checkpoint.interval.ms = 60000 >>>>> delete.topic.enable = false >>>>> quota.window.size.seconds = 1 >>>>> ssl.truststore.type = JKS >>>>> offsets.commit.timeout.ms = 5000 >>>>> quota.window.num = 11 >>>>> zookeeper.connect = ec2-34-223-200-113.us-west-2.c >>>>> ompute.amazonaws.com:2181 >>>>> authorizer.class.name = >>>>> num.replica.fetchers = 1 >>>>> log.retention.ms = null >>>>> log.roll.jitter.hours = 0 >>>>> log.cleaner.enable = true >>>>> offsets.load.buffer.size = 5242880 >>>>> log.cleaner.delete.retention.ms = 86400000 >>>>> ssl.client.auth = none >>>>> controlled.shutdown.max.retries = 3 >>>>> queued.max.requests = 500 >>>>> offsets.topic.replication.factor = 3 >>>>> log.cleaner.threads = 1 >>>>> sasl.kerberos.service.name = null >>>>> sasl.kerberos.ticket.renew.jitter = 0.05 >>>>> socket.request.max.bytes = 104857600 >>>>> ssl.trustmanager.algorithm = PKIX >>>>> zookeeper.session.timeout.ms = 30000 >>>>> log.retention.bytes = -1 >>>>> log.message.timestamp.type = CreateTime >>>>> sasl.kerberos.min.time.before.relogin = 60000 >>>>> zookeeper.set.acl = false >>>>> connections.max.idle.ms = 600000 >>>>> offsets.retention.minutes = 86400000 >>>>> replica.fetch.backoff.ms = 1000 >>>>> inter.broker.protocol.version = 0.10.0-IV1 >>>>> log.retention.hours = 168 >>>>> num.partitions = 1 >>>>> broker.id.generation.enable = true >>>>> listeners = PLAINTEXT://ec2-34-209-53-166. >>>>> us-west-2.compute.amazonaws.com:6667 >>>>> ssl.provider = null >>>>> ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] >>>>> log.roll.ms = null >>>>> log.flush.scheduler.interval.ms = 9223372036854775807 >>>>> ssl.cipher.suites = null >>>>> log.index.size.max.bytes = 10485760 >>>>> ssl.keymanager.algorithm = SunX509 >>>>> security.inter.broker.protocol = PLAINTEXT >>>>> replica.fetch.max.bytes = 1048576 >>>>> advertised.port = null >>>>> log.cleaner.dedupe.buffer.size = 134217728 >>>>> replica.high.watermark.checkpoint.interval.ms = 5000 >>>>> log.cleaner.io.buffer.size = 524288 >>>>> sasl.kerberos.ticket.renew.window.factor = 0.8 >>>>> zookeeper.connection.timeout.ms = 25000 >>>>> controlled.shutdown.retry.backoff.ms = 5000 >>>>> log.roll.hours = 168 >>>>> log.cleanup.policy = delete >>>>> host.name = >>>>> log.roll.jitter.ms = null >>>>> max.connections.per.ip = 2147483647 >>>>> offsets.topic.segment.bytes = 104857600 >>>>> background.threads = 10 >>>>> quota.consumer.default = 9223372036854775807 >>>>> request.timeout.ms = 30000 >>>>> log.message.format.version = 0.10.0-IV1 >>>>> log.index.interval.bytes = 4096 >>>>> log.dir = /tmp/kafka-logs >>>>> log.segment.bytes = 1073741824 >>>>> log.cleaner.backoff.ms = 15000 >>>>> offset.metadata.max.bytes = 4096 >>>>> ssl.truststore.location = null >>>>> group.max.session.timeout.ms = 300000 >>>>> ssl.keystore.password = null >>>>> zookeeper.sync.time.ms = 2000 >>>>> port = 6667 >>>>> log.retention.minutes = null >>>>> log.segment.delete.delay.ms = 60000 >>>>> log.dirs = /data2/kafka-log >>>>> controlled.shutdown.enable = true >>>>> compression.type = producer >>>>> max.connections.per.ip.overrides = >>>>> log.message.timestamp.difference.max.ms = 9223372036854775807 >>>>> sasl.kerberos.kinit.cmd = /usr/bin/kinit >>>>> log.cleaner.io.max.bytes.per.second = 1.7976931348623157E308 >>>>> auto.leader.rebalance.enable = true >>>>> leader.imbalance.check.interval.seconds = 300 >>>>> log.cleaner.min.cleanable.ratio = 0.5 >>>>> replica.lag.time.max.ms = 10000 >>>>> num.network.threads = 3 >>>>> ssl.key.password = null >>>>> reserved.broker.max.id = 1000 >>>>> metrics.num.samples = 2 >>>>> socket.send.buffer.bytes = 102400 >>>>> ssl.protocol = TLS >>>>> socket.receive.buffer.bytes = 102400 >>>>> ssl.keystore.location = null >>>>> replica.fetch.min.bytes = 1 >>>>> broker.rack = null >>>>> unclean.leader.election.enable = true >>>>> sasl.enabled.mechanisms = [GSSAPI] >>>>> group.min.session.timeout.ms = 6000 >>>>> log.cleaner.io.buffer.load.factor = 0.9 >>>>> offsets.retention.check.interval.ms = 600000 >>>>> producer.purgatory.purge.interval.requests = 10000 >>>>> metrics.sample.window.ms = 30000 >>>>> broker.id = -1 >>>>> offsets.topic.compression.codec = 0 >>>>> log.retention.check.interval.ms = 300000 >>>>> advertised.listeners = null >>>>> leader.imbalance.per.broker.percentage = 10 >>>>> (kafka.server.KafkaConfig) >>>>> [2017-05-17 17:03:14,990] INFO starting (kafka.server.KafkaServer) >>>>> [2017-05-17 17:03:14,998] INFO Connecting to zookeeper on >>>>> ec2-34-223-200-113.us-west-2.compute.amazonaws.com:2181 >>>>> (kafka.server.KafkaServer) >>>>> [2017-05-17 17:03:40,740] FATAL Fatal error during KafkaServer startup. >>>>> Prepare to shutdown (kafka.server.KafkaServer) >>>>> org.I0Itec.zkclient.exception.ZkTimeoutException: Unable to connect to >>>>> zookeeper server within timeout: 25000 >>>>> at org.I0Itec.zkclient.ZkClient.connect(ZkClient.java:1232) >>>>> at org.I0Itec.zkclient.ZkClient.<init>(ZkClient.java:156) >>>>> at org.I0Itec.zkclient.ZkClient.<init>(ZkClient.java:130) >>>>> at kafka.utils.ZkUtils$.createZkClientAndConnection(ZkUtils. >>>>> scala:75) >>>>> at kafka.utils.ZkUtils$.apply(ZkUtils.scala:57) >>>>> at kafka.server.KafkaServer.initZk(KafkaServer.scala:294) >>>>> at kafka.server.KafkaServer.startup(KafkaServer.scala:180) >>>>> at kafka.server.KafkaServerStarta >>>>> ble.startup(KafkaServerStartab >>>>> le.scala:37) >>>>> at kafka.Kafka$.main(Kafka.scala:67) >>>>> at kafka.Kafka.main(Kafka.scala) >>>>> [2017-05-17 17:03:40,740] INFO shutting down (kafka.server.KafkaServer) >>>>> [2017-05-17 17:03:40,747] INFO shut down completed >>>>> (kafka.server.KafkaServer) >>>>> [2017-05-17 17:03:40,748] FATAL Fatal error during KafkaServerStartable >>>>> startup. Prepare to shutdown (kafka.server.KafkaServerStartable) >>>>> org.I0Itec.zkclient.exception.ZkTimeoutException: Unable to connect to >>>>> zookeeper server within timeout: 25000 >>>>> at org.I0Itec.zkclient.ZkClient.connect(ZkClient.java:1232) >>>>> at org.I0Itec.zkclient.ZkClient.<init>(ZkClient.java:156) >>>>> at org.I0Itec.zkclient.ZkClient.<init>(ZkClient.java:130) >>>>> at kafka.utils.ZkUtils$.createZkClientAndConnection(ZkUtils. >>>>> scala:75) >>>>> at kafka.utils.ZkUtils$.apply(ZkUtils.scala:57) >>>>> at kafka.server.KafkaServer.initZk(KafkaServer.scala:294) >>>>> at kafka.server.KafkaServer.startup(KafkaServer.scala:180) >>>>> at kafka.server.KafkaServerStarta >>>>> ble.startup(KafkaServerStartab >>>>> le.scala:37) >>>>> at kafka.Kafka$.main(Kafka.scala:67) >>>>> at kafka.Kafka.main(Kafka.scala) >>>>> [2017-05-17 17:03:40,764] INFO shutting down (kafka.server.KafkaServer) >>>>> >>>>> Manually starting Kafka via Ambari fixes this... However, Metron REST >>>>> still doesn't work: >>>>> >>>>> . ____ _ __ _ _ >>>>> /\\ / ___'_ __ _ _(_)_ __ __ _ \ \ \ \ >>>>> ( ( )\___ | '_ | '_| | '_ \/ _` | \ \ \ \ >>>>> \\/ ___)| |_)| | | | | || (_| | ) ) ) ) >>>>> ' |____| .__|_| |_|_| |_\__, | / / / / >>>>> =========|_|==============|___/=/_/_/_/ >>>>> :: Spring Boot :: (v1.4.1.RELEASE) >>>>> >>>>> 2017-05-17T17:03:52.406 ERROR [org.springframework.boot.Spri >>>>> ngApplication] >>>>> - Application startup failed >>>>> org.springframework.beans.factory.UnsatisfiedDependencyException: >>>>> Error >>>>> creating bean with name 'kafkaServiceImpl' defined in URL >>>>> [jar:file:/usr/metron/0.4.0/lib/metron-rest-0.4.0.jar!/BOOT- >>>>> INF/classes!/org/apache/metron/rest/service/impl/KafkaServic >>>>> eImpl.class]: >>>>> Unsatisfied dependency expressed through constructor parameter 0; >>>>> nested >>>>> exception is org.springframework.beans.fact >>>>> ory.UnsatisfiedDependencyException: >>>>> Error creating bean with name 'kafkaConfig': Unsatisfied dependency >>>>> expressed through field 'zkClient'; nested exception is >>>>> org.springframework.beans.factory.BeanCreationException: Error >>>>> creating >>>>> bean with name 'zkClient' defined in class path resource >>>>> [org/apache/metron/rest/config/ZookeeperConfig.class]: Bean >>>>> instantiation >>>>> via factory method failed; nested exception is >>>>> org.springframework.beans.BeanInstantiationException: Failed to >>>>> instantiate [org.I0Itec.zkclient.ZkClient]: Factory method 'zkClient' >>>>> threw exception; nested exception is org.I0Itec.zkclient.exception. >>>>> ZkTimeoutException: >>>>> Unable to connect to zookeeper server within timeout: 10000 >>>>> at org.springframework.beans.fact >>>>> ory.support.ConstructorResolve >>>>> r.createArgumentArray(ConstructorResolver.java:749) >>>>> at org.springframework.beans.fact >>>>> ory.support.ConstructorResolve >>>>> r.autowireConstructor(ConstructorResolver.java:189) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.autowireConstructor(AbstractAu >>>>> towireCapableBeanFactory.java:1148) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.createBeanInstance(AbstractAut >>>>> owireCapableBeanFactory.java:1051) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFac >>>>> tory.java:510) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.createBean(AbstractAutowireCapableBeanFacto >>>>> ry.java:482) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractBeanFactor >>>>> y$1.getObject(AbstractBeanFactory.java:306) >>>>> at org.springframework.beans.fact >>>>> ory.support.DefaultSingletonBe >>>>> anRegistry.getSingleton(DefaultSingletonBeanRegistry.java:230) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractBeanFactor >>>>> y.doGetBean(AbstractBeanFactory.java:302) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractBeanFactor >>>>> y.getBean(AbstractBeanFactory.java:197) >>>>> at org.springframework.beans.fact >>>>> ory.support.DefaultListableBea >>>>> nFactory.preInstantiateSingletons(DefaultListableBeanFactory.java:751) >>>>> at org.springframework.context.su >>>>> pport.AbstractApplicationConte >>>>> xt.finishBeanFactoryInitialization(AbstractApplicationContex >>>>> t.java:861) >>>>> at org.springframework.context.su >>>>> pport.AbstractApplicationConte >>>>> xt.refresh(AbstractApplicationContext.java:541) >>>>> at org.springframework.boot.conte >>>>> xt.embedded.EmbeddedWebApplica >>>>> tionContext.refresh(EmbeddedWebApplicationContext.java:122) >>>>> at org.springframework.boot.Sprin >>>>> gApplication.refresh(SpringApp >>>>> lication.java:761) >>>>> at org.springframework.boot.Sprin >>>>> gApplication.refreshContext(Sp >>>>> ringApplication.java:371) >>>>> at org.springframework.boot.Sprin >>>>> gApplication.run(SpringApplica >>>>> tion.java:315) >>>>> at org.springframework.boot.Sprin >>>>> gApplication.run(SpringApplica >>>>> tion.java:1186) >>>>> at org.springframework.boot.Sprin >>>>> gApplication.run(SpringApplica >>>>> tion.java:1175) >>>>> at org.apache.metron.rest.MetronR >>>>> estApplication.main(MetronRest >>>>> Application.java:27) >>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>> at sun.reflect.NativeMethodAccess >>>>> orImpl.invoke(NativeMethodAcce >>>>> ssorImpl.java:62) >>>>> at sun.reflect.DelegatingMethodAc >>>>> cessorImpl.invoke(DelegatingMe >>>>> thodAccessorImpl.java:43) >>>>> at java.lang.reflect.Method.invoke(Method.java:498) >>>>> at org.springframework.boot.loade >>>>> r.MainMethodRunner.run(MainMet >>>>> hodRunner.java:48) >>>>> at org.springframework.boot.loader.Launcher.launch(Launcher. >>>>> java:87) >>>>> at org.springframework.boot.loader.Launcher.launch(Launcher. >>>>> java:50) >>>>> at org.springframework.boot.loader.PropertiesLauncher.main( >>>>> PropertiesLauncher.java:525) >>>>> Caused by: org.springframework.beans.fact >>>>> ory.UnsatisfiedDependencyException: >>>>> Error creating bean with name 'kafkaConfig': Unsatisfied dependency >>>>> expressed through field 'zkClient'; nested exception is >>>>> org.springframework.beans.factory.BeanCreationException: Error >>>>> creating >>>>> bean with name 'zkClient' defined in class path resource >>>>> [org/apache/metron/rest/config/ZookeeperConfig.class]: Bean >>>>> instantiation >>>>> via factory method failed; nested exception is >>>>> org.springframework.beans.BeanInstantiationException: Failed to >>>>> instantiate [org.I0Itec.zkclient.ZkClient]: Factory method 'zkClient' >>>>> threw exception; nested exception is org.I0Itec.zkclient.exception. >>>>> ZkTimeoutException: >>>>> Unable to connect to zookeeper server within timeout: 10000 >>>>> at org.springframework.beans.fact >>>>> ory.annotation.AutowiredAnnota >>>>> tionBeanPostProcessor$AutowiredFieldElement.inject(A >>>>> utowiredAnnotationBeanPostProcessor.java:569) >>>>> at org.springframework.beans.fact >>>>> ory.annotation.InjectionMetada >>>>> ta.inject(InjectionMetadata.java:88) >>>>> at org.springframework.beans.fact >>>>> ory.annotation.AutowiredAnnota >>>>> tionBeanPostProcessor.postProcessPropertyValues(AutowiredAnn >>>>> otationBeanPostProcessor.java:349) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.populateBean(AbstractAutowireCapableBeanFac >>>>> tory.java:1219) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFac >>>>> tory.java:543) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.createBean(AbstractAutowireCapableBeanFacto >>>>> ry.java:482) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractBeanFactor >>>>> y$1.getObject(AbstractBeanFactory.java:306) >>>>> at org.springframework.beans.fact >>>>> ory.support.DefaultSingletonBe >>>>> anRegistry.getSingleton(DefaultSingletonBeanRegistry.java:230) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractBeanFactor >>>>> y.doGetBean(AbstractBeanFactory.java:302) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractBeanFactor >>>>> y.getBean(AbstractBeanFactory.java:197) >>>>> at org.springframework.beans.fact >>>>> ory.support.ConstructorResolve >>>>> r.instantiateUsingFactoryMethod(ConstructorResolver.java:372) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.instantiateUsingFactoryMethod(AbstractAutow >>>>> ireCapableBeanFactory.java:1128) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.createBeanInstance(AbstractAut >>>>> owireCapableBeanFactory.java:1023) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFac >>>>> tory.java:510) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.createBean(AbstractAutowireCapableBeanFacto >>>>> ry.java:482) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractBeanFactor >>>>> y$1.getObject(AbstractBeanFactory.java:306) >>>>> at org.springframework.beans.fact >>>>> ory.support.DefaultSingletonBe >>>>> anRegistry.getSingleton(DefaultSingletonBeanRegistry.java:230) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractBeanFactor >>>>> y.doGetBean(AbstractBeanFactory.java:302) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractBeanFactor >>>>> y.getBean(AbstractBeanFactory.java:202) >>>>> at org.springframework.beans.fact >>>>> ory.config.DependencyDescripto >>>>> r.resolveCandidate(DependencyDescriptor.java:207) >>>>> at org.springframework.beans.fact >>>>> ory.support.DefaultListableBea >>>>> nFactory.doResolveDependency(DefaultListableBeanFactory.java:1128) >>>>> at org.springframework.beans.fact >>>>> ory.support.DefaultListableBea >>>>> nFactory.resolveDependency(DefaultListableBeanFactory.java:1056) >>>>> at org.springframework.beans.fact >>>>> ory.support.ConstructorResolve >>>>> r.resolveAutowiredArgument(ConstructorResolver.java:835) >>>>> at org.springframework.beans.fact >>>>> ory.support.ConstructorResolve >>>>> r.createArgumentArray(ConstructorResolver.java:741) >>>>> ... 27 more >>>>> Caused by: org.springframework.beans.factory.BeanCreationException: >>>>> Error >>>>> creating bean with name 'zkClient' defined in class path resource >>>>> [org/apache/metron/rest/config/ZookeeperConfig.class]: Bean >>>>> instantiation >>>>> via factory method failed; nested exception is >>>>> org.springframework.beans.BeanInstantiationException: Failed to >>>>> instantiate [org.I0Itec.zkclient.ZkClient]: Factory method 'zkClient' >>>>> threw exception; nested exception is org.I0Itec.zkclient.exception. >>>>> ZkTimeoutException: >>>>> Unable to connect to zookeeper server within timeout: 10000 >>>>> at org.springframework.beans.fact >>>>> ory.support.ConstructorResolve >>>>> r.instantiateUsingFactoryMethod(ConstructorResolver.java:599) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.instantiateUsingFactoryMethod(AbstractAutow >>>>> ireCapableBeanFactory.java:1128) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.createBeanInstance(AbstractAut >>>>> owireCapableBeanFactory.java:1023) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFac >>>>> tory.java:510) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractAutowireCa >>>>> pableBeanFactory.createBean(AbstractAutowireCapableBeanFacto >>>>> ry.java:482) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractBeanFactor >>>>> y$1.getObject(AbstractBeanFactory.java:306) >>>>> at org.springframework.beans.fact >>>>> ory.support.DefaultSingletonBe >>>>> anRegistry.getSingleton(DefaultSingletonBeanRegistry.java:230) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractBeanFactor >>>>> y.doGetBean(AbstractBeanFactory.java:302) >>>>> at org.springframework.beans.fact >>>>> ory.support.AbstractBeanFactor >>>>> y.getBean(AbstractBeanFactory.java:202) >>>>> at org.springframework.beans.fact >>>>> ory.config.DependencyDescripto >>>>> r.resolveCandidate(DependencyDescriptor.java:207) >>>>> at org.springframework.beans.fact >>>>> ory.support.DefaultListableBea >>>>> nFactory.doResolveDependency(DefaultListableBeanFactory.java:1128) >>>>> at org.springframework.beans.fact >>>>> ory.support.DefaultListableBea >>>>> nFactory.resolveDependency(DefaultListableBeanFactory.java:1056) >>>>> at org.springframework.beans.fact >>>>> ory.annotation.AutowiredAnnota >>>>> tionBeanPostProcessor$AutowiredFieldElement.inject(A >>>>> utowiredAnnotationBeanPostProcessor.java:566) >>>>> ... 50 more >>>>> Caused by: org.springframework.beans.BeanInstantiationException: >>>>> Failed >>>>> to instantiate [org.I0Itec.zkclient.ZkClient]: Factory method >>>>> 'zkClient' >>>>> threw exception; nested exception is org.I0Itec.zkclient.exception. >>>>> ZkTimeoutException: >>>>> Unable to connect to zookeeper server within timeout: 10000 >>>>> at org.springframework.beans.fact >>>>> ory.support.SimpleInstantiatio >>>>> nStrategy.instantiate(SimpleInstantiationStrategy.java:189) >>>>> at org.springframework.beans.fact >>>>> ory.support.ConstructorResolve >>>>> r.instantiateUsingFactoryMethod(ConstructorResolver.java:588) >>>>> ... 62 more >>>>> Caused by: org.I0Itec.zkclient.exception.ZkTimeoutException: Unable to >>>>> connect to zookeeper server within timeout: 10000 >>>>> at org.I0Itec.zkclient.ZkClient.connect(ZkClient.java:1232) >>>>> at org.I0Itec.zkclient.ZkClient.<init>(ZkClient.java:156) >>>>> at org.I0Itec.zkclient.ZkClient.<init>(ZkClient.java:130) >>>>> at org.I0Itec.zkclient.ZkClient.<init>(ZkClient.java:97) >>>>> at org.apache.metron.rest.config. >>>>> ZookeeperConfig.zkClient(Zooke >>>>> eperConfig.java:52) >>>>> at org.apache.metron.rest.config. >>>>> ZookeeperConfig$$EnhancerBySpr >>>>> ingCGLIB$$6e0c6742.CGLIB$zkClient$1(<generated>) >>>>> at org.apache.metron.rest.config. >>>>> ZookeeperConfig$$EnhancerBySpr >>>>> ingCGLIB$$6e0c6742$$FastClassBySpringCGLIB$$3219d861.invoke( >>>>> <generated>) >>>>> at org.springframework.cglib.prox >>>>> y.MethodProxy.invokeSuper(Meth >>>>> odProxy.java:228) >>>>> at org.springframework.context.an >>>>> notation.ConfigurationClassEnh >>>>> ancer$BeanMethodInterceptor.intercept(ConfigurationClassEnha >>>>> ncer.java:356) >>>>> at org.apache.metron.rest.config. >>>>> ZookeeperConfig$$EnhancerBySpr >>>>> ingCGLIB$$6e0c6742.zkClient(<generated>) >>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>> at sun.reflect.NativeMethodAccess >>>>> orImpl.invoke(NativeMethodAcce >>>>> ssorImpl.java:62) >>>>> at sun.reflect.DelegatingMethodAc >>>>> cessorImpl.invoke(DelegatingMe >>>>> thodAccessorImpl.java:43) >>>>> at java.lang.reflect.Method.invoke(Method.java:498) >>>>> at org.springframework.beans.fact >>>>> ory.support.SimpleInstantiatio >>>>> nStrategy.instantiate(SimpleInstantiationStrategy.java:162) >>>>> ... 63 more >>>>> >>>>> Any ideas on what might be going or why the timeout triggers? >>>>> >>>>> >>>>> On 2017-05-17 08:31, Laurens Vets wrote: >>>>> >>>>> Hello list, >>>>>> >>>>>> I'm trying to spin up Metron in EC2 for testing purposes. I made the >>>>>> following changes in playbook.yml so that I'm left with 5 EC2 >>>>>> instances instead of 10: >>>>>> >>>>>> <-- SNIP --> >>>>>> tasks: >>>>>> - include: tasks/create-keypair.yml >>>>>> - include: tasks/create-vpc.yml >>>>>> - include: tasks/create-open-inbound-security-group.yml >>>>>> - include: tasks/create-open-outbound-security-group.yml >>>>>> - include: tasks/create-hosts.yml host_count=1 >>>>>> host_type=sensors,ambari_master,ec2,monit >>>>>> # - include: tasks/create-hosts.yml host_count=4 >>>>>> host_type=ambari_slave,ec2 >>>>>> - include: tasks/create-hosts.yml host_count=1 >>>>>> host_type=pcap_server,monit,ec2 >>>>>> - include: tasks/create-hosts.yml host_count=1 >>>>>> host_type=ambari_slave,enrichment,metron,ec2,zeppelin >>>>>> # - include: tasks/create-hosts.yml host_count=2 >>>>>> host_type=ambari_slave,search,ec2 >>>>>> - include: tasks/create-hosts.yml host_count=1 >>>>>> host_type=ambari_slave,search,ec2 >>>>>> - include: tasks/create-hosts.yml host_count=1 >>>>>> host_type=ambari_slave,web,ec2 >>>>>> <-- END SNIP --> >>>>>> >>>>>> I also had to change the line "ansible-playbook -vvv -i ec2.py >>>>>> playbook.yml \ >>>>>> --skip-tags="quick_dev,sensor-stubs" \ >>>>>> --extra-vars="env=$ENV" \ >>>>>> $EXTRA_ARGS" >>>>>> >>>>>> to "ansible-playbook -vvv -i ec2.py playbook.yml \ >>>>>> --skip-tags="quick_dev,sensor-stubs" \ >>>>>> --extra-vars="env=$ENV" \ >>>>>> $EXTRA_ARGS -c paramiko" >>>>>> in run.sh otherwise ssh would time out. >>>>>> >>>>>> However, Ansible errors out with: >>>>>> >>>>>> 2017-05-16 18:02:50,615 p=5362 u=laurens | TASK [ambari_config : >>>>>> Deploy cluster with Ambari; >>>>>> http://ec2-35-164-26-255.us-west-2.compute.amazonaws.com:8080] *** >>>>>> 2017-05-16 18:02:50,615 p=5362 u=laurens | task path: >>>>>> /home/laurens/SAPSource/metron/metron-deployment/roles/ >>>>>> ambari_config/tasks/main.yml:36 >>>>>> 2017-05-16 18:02:52,660 p=5362 u=laurens | fatal: >>>>>> [ec2-35-164-26-255.us-west-2.compute.amazonaws.com]: FAILED! => >>>>>> {"changed": false, "failed": true, "invocation": {"module_args": >>>>>> {"blueprint_name": "metron_blueprint", "blueprint_var": {"groups": >>>>>> [{"cardinality": 1, "components": [{"name": "NAMENODE"}, {"name": >>>>>> "SECONDARY_NAMENODE"}, {"name": "RESOURCEMANAGER"}, {"name": >>>>>> "HISTORYSERVER"}, {"name": "HDFS_CLIENT"}, {"name": "YARN_CLIENT"}, >>>>>> {"name": "MAPREDUCE2_CLIENT"}, {"name": "SPARK_CLIENT"}, {"name": >>>>>> "ZOOKEEPER_CLIENT"}, {"name": "HBASE_CLIENT"}, {"name": "ES_SLAVE"}], >>>>>> "configuration": [], "name": "master_1"}, {"cardinality": 1, >>>>>> "components": [{"name": "ZOOKEEPER_SERVER"}, {"name": "NIMBUS"}, >>>>>> {"name": "STORM_UI_SERVER"}, {"name": "DRPC_SERVER"}, {"name": >>>>>> "SPARK_JOBHISTORYSERVER"}, {"name": "HBASE_MASTER"}, {"name": >>>>>> "HDFS_CLIENT"}, {"name": "YARN_CLIENT"}, {"name": >>>>>> "MAPREDUCE2_CLIENT"}, {"name": "SPARK_CLIENT"}, {"name": >>>>>> "ZOOKEEPER_CLIENT"}, {"name": "HBASE_CLIENT"}, {"name": >>>>>> "APP_TIMELINE_SERVER"}, {"name": "ES_SLAVE"}], "configuration": [], >>>>>> "name": "master_2"}, {"cardinality": 1, "components": [{"name": >>>>>> "METRON_INDEXING"}, {"name": "METRON_ENRICHMENT_MASTER"}, {"name": >>>>>> "METRON_PARSERS"}, {"name": "METRON_REST"}, {"name": "DATANODE"}, >>>>>> {"name": "NODEMANAGER"}, {"name": "SUPERVISOR"}, {"name": >>>>>> "KAFKA_BROKER"}, {"name": "HBASE_REGIONSERVER"}, {"name": >>>>>> "HDFS_CLIENT"}, {"name": "YARN_CLIENT"}, {"name": >>>>>> "MAPREDUCE2_CLIENT"}, {"name": "SPARK_CLIENT"}, {"name": >>>>>> "ZOOKEEPER_CLIENT"}, {"name": "HBASE_CLIENT"}], "configuration": [], >>>>>> "name": "metron"}, {"cardinality": 1, "components": [{"name": >>>>>> "KIBANA_MASTER"}, {"name": "ES_MASTER"}], "configuration": [], "name": >>>>>> "web"}, {"cardinality": "1+", "components": [{"name": "DATANODE"}, >>>>>> {"name": "NODEMANAGER"}, {"name": "SUPERVISOR"}, {"name": >>>>>> "KAFKA_BROKER"}, {"name": "HBASE_REGIONSERVER"}, {"name": >>>>>> "HDFS_CLIENT"}, {"name": "YARN_CLIENT"}, {"name": >>>>>> "MAPREDUCE2_CLIENT"}, {"name": "SPARK_CLIENT"}, {"name": >>>>>> "ZOOKEEPER_CLIENT"}, {"name": "HBASE_CLIENT"}], "configuration": [], >>>>>> "name": "slaves"}], "required_configurations": [{"metron-env": >>>>>> {"es_hosts": >>>>>> "ec2-35-164-136-88.us-west-2.compute.amazonaws.com,ec2-34-20 >>>>>> 8-94-45.us-west-2.compute.amazonaws.com", >>>>>> "metron_jdbc_driver": "org.h2.Driver", "metron_jdbc_password": "root", >>>>>> "metron_jdbc_platform": "h2", "metron_jdbc_url": >>>>>> "jdbc:h2:file:~/metrondb", "metron_jdbc_username": "root", >>>>>> "storm_rest_addr": >>>>>> "ec2-34-208-94-45.us-west-2.compute.amazonaws.com:8744", >>>>>> "zeppelin_server_url": >>>>>> "ec2-34-210-151-29.us-west-2.compute.amazonaws.com"}}, {"kibana-env": >>>>>> {"kibana_default_application": "dashboard/Metron-Dashboard", >>>>>> "kibana_es_url": >>>>>> "http://ec2-35-164-136-88.us-west-2.compute.amazonaws.com:9200", >>>>>> "kibana_log_dir": "/var/log/kibana", "kibana_pid_dir": >>>>>> "/var/run/kibana", "kibana_server_port": 5000}}], "stack_name": "HDP", >>>>>> "stack_version": "2.5"}, "cluster_name": "metron", "cluster_state": >>>>>> "present", "configurations": [{"zoo.cfg": {"dataDir": >>>>>> "/data1/hadoop/zookeeper"}}, {"hadoop-env": {"dtnode_heapsize": 1024, >>>>>> "namenode_heapsize": 2048}}, {"hbase-env": {"hbase_master_heapsize": >>>>>> 1024, "hbase_regionserver_heapsize": 1024}}, {"hdfs-site": >>>>>> {"dfs.datanode.data.dir": >>>>>> "/data1/hadoop/hdfs/data,/data2/hadoop/hdfs/data", >>>>>> "dfs.journalnode.edits.dir": "/data1/hadoop/hdfs/journalnode", >>>>>> "dfs.namenode.checkpoint.dir": "/data1/hadoop/hdfs/namesecondary", >>>>>> "dfs.namenode.name.dir": "/data1/hadoop/hdfs/namenode"}}, >>>>>> {"mapred-site": {"mapreduce.jobhistory.recovery.store.leveldb.path": >>>>>> "/data1/hadoop/mapreduce/jhs", "mapreduce.map.memory.mb": 1229, >>>>>> "mapreduce.reduce.memory.mb": 1229}}, {"yarn-site": >>>>>> {"yarn.nodemanager.local-dirs": "/data1/hadoop/yarn/local", >>>>>> "yarn.nodemanager.log-dirs": "/data1/hadoop/yarn/log", >>>>>> "yarn.nodemanager.resource.memory-mb": 4096, >>>>>> "yarn.timeline-service.leveldb-state-store.path": >>>>>> "/data1/hadoop/yarn/timeline", >>>>>> "yarn.timeline-service.leveldb-timeline-store.path": >>>>>> "/data1/hadoop/yarn/timeline"}}, {"storm-site": {"storm.local.dir": >>>>>> "/data1/hadoop/storm", "supervisor.slots.ports": "[6700, 6701, 6702, >>>>>> 6703]", "topology.classpath": "/etc/hbase/conf:/etc/hadoop/conf"}}, >>>>>> {"kafka-broker": {"log.dirs": "/data2/kafka-log"}}, {"metron-env": >>>>>> {"parsers": "bro,snort,yaf"}}, {"elastic-site": >>>>>> {"gateway_recover_after_data_nodes": 1, "index_number_of_replicas": >>>>>> 1, >>>>>> "index_number_of_shards": 2, "network_host": "[ _local_, _eth0_ ]", >>>>>> "zen_discovery_ping_unicast_hosts": "[ >>>>>> ec2-35-164-136-88.us-west-2.compute.amazonaws.com, >>>>>> ec2-34-208-94-45.us-west-2.compute.amazonaws.com ]"}}], "host": >>>>>> "ec2-35-164-26-255.us-west-2.compute.amazonaws.com", "password": >>>>>> "admin", "port": 8080, "username": "admin", "wait_for_complete": >>>>>> true}, "module_name": "ambari_cluster_state"}, "msg": "Ambari client >>>>>> exception occurred: Could not create cluster: request code 400, >>>>>> request message {\n \"status\" : 400,\n \"message\" : >>>>>> \"Invalid Cluster Creation Template: >>>>>> org.apache.ambari.server.topology.InvalidTopologyTemplateException: >>>>>> Host group 'slaves' must contain at least one 'hosts/fqdn' or a >>>>>> 'host_count' value\"\n}"} >>>>>> >>>>>> PLAY RECAP ****************************** >>>>>> ****************************** >>>>>> ********* >>>>>> ec2-34-208-94-45.us-west-2.compute.amazonaws.com : ok=49 changed=27 >>>>>> unreachable=0 failed=0 >>>>>> ec2-34-209-224-190.us-west-2.compute.amazonaws.com : ok=23 >>>>>> changed=8 >>>>>> unreachable=0 failed=0 >>>>>> ec2-34-210-151-29.us-west-2.compute.amazonaws.com : ok=58 >>>>>> changed=31 >>>>>> unreachable=0 failed=0 >>>>>> ec2-35-164-136-88.us-west-2.compute.amazonaws.com : ok=49 >>>>>> changed=27 >>>>>> unreachable=0 failed=0 >>>>>> ec2-35-164-26-255.us-west-2.compute.amazonaws.com : ok=54 >>>>>> changed=28 >>>>>> unreachable=0 failed=1 >>>>>> localhost : ok=29 changed=12 unreachable=0 >>>>>> failed=0 >>>>>> >>>>>> Any idea what might be going on? >>>>>> >>>>>> (Don't worry about the URLs, the instances have been terminated) >>>>>> >>>>>> >>>>> >