Hi Adam,

Thanks a lot for replying.  I am using virtualbox for installing the VMs,
earlier I was using bridge adapter so it takes the network IP from dhcp, I
assign them as public and local_ip same.

As you mention for static IP I tried with with *NAT + Host-only Network (*NAT
as primary interface eth0*), but all node having same IP in eth0 in NAT as
10.0.2.15 (is it fine to have all node same IP because I am not using that
IP)* and i assign host-only ip as static as 192.168.56.110 etc.

1) can I use both local_ip and public_ip same as host only ip
(192.168.56.110 etc) or public_ip would be the ip of host machine on which
virtualbox is installed as in NAT VM use Host IP as public IP to contact
outer world.

2) Is public_ip necessary as I only want stress testing to run in same
network, I don't want to install the no. on client like zoiper and all.

3) Is port forwarding isnecessary in  *NAT + Host-only Network, because * nodes
are able to communicate each other and in host only network so I don't
think port forwarding is necessary.

4) I just want to run stress testing for handling 1 lack call/sec. so how
many sprout, vellum node is needed for this much calls.

Thanks,
Sunil


On Thu, Apr 19, 2018 at 2:42 PM, Adam Lindley <[email protected]>
wrote:

> Hi Sunil,
>
>
>
> I’m afraid the steps you’ve taken are not supported in Project Clearwater
> deployments. Both changing the ‘local_ip’ of a node, and removing nodes
> just by deleting the VMs.
>
>
>
> On the first point, you need to be able to give your VMs permanent static
> IP addresses.
>
> On the second, by deleting the VMs in your cluster, your underlying etcd
> cluster has lost quorum. I would suggest http://clearwater.readthedocs.
> io/en/stable/Handling_Multiple_Failed_Nodes.htm as a starting point for
> recovering information from it. However, as your single remaining node will
> likely also have problems due to the local IP changing, you may simply want
> to redeploy from scratch.
>
>
>
> More in general, you seem to have hit a substantial number of issues in
> deploying Project Clearwater, which is both not what we want, and not what
> the experience of many other users seems to be. I would suggest taking a
> wider look over our provided documentation, and making sure your
> environment matches our expectations, and that you’re clear on our
> processes. This should make your next deployment a lot smoother.
>
>
>
> Cheers, and good luck,
>
> Adam
>
>
>
> *From:* Clearwater [mailto:[email protected]]
> *On Behalf Of *Sunil Kumar
> *Sent:* 19 April 2018 07:16
> *To:* [email protected]
> *Subject:* Re: [Project Clearwater] Unable to contact the etcd cluster
>
>
>
> Hi,
>
> the node with ip 10.224.61.109, 10.224.61.112 etc is no more there, I have
> deleted the node directly. It looks like they are still in the etcd
> cluster. Can you please tell me how to remove them
>
>
>
> [IST Apr 19 19:32:45] error    : 'etcd_process' process is not running
>
> [IST Apr 19 19:32:45] info     : 'etcd_process' trying to restart
>
> [IST Apr 19 19:32:45] info     : 'etcd_process' restart: /bin/bash
>
> [IST Apr 19 19:33:15] error    : 'etcd_process' failed to restart (exit
> status -1) -- /bin/bash: Program timed out -- zmq_msg_recv: Resource
> temporarily unavailable
>
> cat: /var/run/clearwater-etcd/clearwater-etcd.pid: No such file or
> directory
>
> cat: /var/run/clearwater-etcd/clearwater-etcd.pid: No such file or
> directory
>
> context deadline excee
>
> [IST Apr 19 19:33:25] error    : 'etcd_process' process is not running
>
> [IST Apr 19 19:33:25] info     : 'etcd_process' trying to restart
>
> [IST Apr 19 19:33:25] info     : 'etcd_process' restart: /bin/bash
>
> [IST Apr 19 19:33:55] error    : 'etcd_process' failed to restart (exit
> status -1) -- /bin/bash: Program timed out -- zmq_msg_recv: Resource
> temporarily unavailable
>
> client: etcd cluster is unavailable or misconfigured; error #0: *dial tcp
> 10.224.61.109:4000 <http://10.224.61.109:4000>*: getsockopt: no route to
> host
>
> ; error #1: dial tcp 10.224.61.47:4000: getsockopt: co
>
> [IST Apr 19 19:34:05] error    : 'etcd_process' process is not running
>
> [IST Apr 19 19:34:05] info     : 'etcd_process' trying to restart
>
> [IST Apr 19 19:34:05] info     : 'etcd_process' restart: /bin/bash
>
> [IST Apr 19 19:34:36] error    : 'etcd_process' failed to restart (exit
> status 2) -- /bin/bash: zmq_msg_recv: Resource temporarily unavailable
>
> context deadline exceeded
>
>
>
>
>
> On Thu, Apr 19, 2018 at 11:03 AM, Sunil Kumar <[email protected]>
> wrote:
>
> Hi,
>
> Any body can help me on this. after ip lost, i update the ip in
> local_config and dns and restart the service. extra vm is deleted lik  i
> had 3 sprout node so 2 are deleted.
>
>
>
> [vellum]ubuntu@vellum:~$ cw-config upload shared_config
>
> Unable to contact the etcd cluster.
>
>
>
> thanks
>
> sunil
>
>
>
> _______________________________________________
> Clearwater mailing list
> [email protected]
> http://lists.projectclearwater.org/mailman/listinfo/
> clearwater_lists.projectclearwater.org
>
>
_______________________________________________
Clearwater mailing list
[email protected]
http://lists.projectclearwater.org/mailman/listinfo/clearwater_lists.projectclearwater.org

Reply via email to