the lost node is not the master node but IP of master nodes is changed and I update them.
thanks On Thu, Apr 19, 2018 at 3:23 PM, Sunil Kumar <skgola1...@gmail.com> wrote: > Hi Adam, > > Thanks a lot for replying. I am using virtualbox for installing the VMs, > earlier I was using bridge adapter so it takes the network IP from dhcp, I > assign them as public and local_ip same. > > As you mention for static IP I tried with with *NAT + Host-only Network (*NAT > as primary interface eth0*), but all node having same IP in eth0 in NAT > as 10.0.2.15 (is it fine to have all node same IP because I am not using > that IP)* and i assign host-only ip as static as 192.168.56.110 etc. > > 1) can I use both local_ip and public_ip same as host only ip > (192.168.56.110 etc) or public_ip would be the ip of host machine on which > virtualbox is installed as in NAT VM use Host IP as public IP to contact > outer world. > > 2) Is public_ip necessary as I only want stress testing to run in same > network, I don't want to install the no. on client like zoiper and all. > > 3) Is port forwarding isnecessary in *NAT + Host-only Network, because * > nodes > are able to communicate each other and in host only network so I don't > think port forwarding is necessary. > > 4) I just want to run stress testing for handling 1 lack call/sec. so how > many sprout, vellum node is needed for this much calls. > > Thanks, > Sunil > > > On Thu, Apr 19, 2018 at 2:42 PM, Adam Lindley <adam.lind...@metaswitch.com > > wrote: > >> Hi Sunil, >> >> >> >> I’m afraid the steps you’ve taken are not supported in Project Clearwater >> deployments. Both changing the ‘local_ip’ of a node, and removing nodes >> just by deleting the VMs. >> >> >> >> On the first point, you need to be able to give your VMs permanent static >> IP addresses. >> >> On the second, by deleting the VMs in your cluster, your underlying etcd >> cluster has lost quorum. I would suggest http://clearwater.readthedocs. >> io/en/stable/Handling_Multiple_Failed_Nodes.htm as a starting point for >> recovering information from it. However, as your single remaining node will >> likely also have problems due to the local IP changing, you may simply want >> to redeploy from scratch. >> >> >> >> More in general, you seem to have hit a substantial number of issues in >> deploying Project Clearwater, which is both not what we want, and not what >> the experience of many other users seems to be. I would suggest taking a >> wider look over our provided documentation, and making sure your >> environment matches our expectations, and that you’re clear on our >> processes. This should make your next deployment a lot smoother. >> >> >> >> Cheers, and good luck, >> >> Adam >> >> >> >> *From:* Clearwater [mailto:clearwater-boun...@lists.projectclearwater.org] >> *On Behalf Of *Sunil Kumar >> *Sent:* 19 April 2018 07:16 >> *To:* clearwater@lists.projectclearwater.org >> *Subject:* Re: [Project Clearwater] Unable to contact the etcd cluster >> >> >> >> Hi, >> >> the node with ip 10.224.61.109, 10.224.61.112 etc is no more there, I >> have deleted the node directly. It looks like they are still in the etcd >> cluster. Can you please tell me how to remove them >> >> >> >> [IST Apr 19 19:32:45] error : 'etcd_process' process is not running >> >> [IST Apr 19 19:32:45] info : 'etcd_process' trying to restart >> >> [IST Apr 19 19:32:45] info : 'etcd_process' restart: /bin/bash >> >> [IST Apr 19 19:33:15] error : 'etcd_process' failed to restart (exit >> status -1) -- /bin/bash: Program timed out -- zmq_msg_recv: Resource >> temporarily unavailable >> >> cat: /var/run/clearwater-etcd/clearwater-etcd.pid: No such file or >> directory >> >> cat: /var/run/clearwater-etcd/clearwater-etcd.pid: No such file or >> directory >> >> context deadline excee >> >> [IST Apr 19 19:33:25] error : 'etcd_process' process is not running >> >> [IST Apr 19 19:33:25] info : 'etcd_process' trying to restart >> >> [IST Apr 19 19:33:25] info : 'etcd_process' restart: /bin/bash >> >> [IST Apr 19 19:33:55] error : 'etcd_process' failed to restart (exit >> status -1) -- /bin/bash: Program timed out -- zmq_msg_recv: Resource >> temporarily unavailable >> >> client: etcd cluster is unavailable or misconfigured; error #0: *dial >> tcp 10.224.61.109:4000 <http://10.224.61.109:4000>*: getsockopt: no >> route to host >> >> ; error #1: dial tcp 10.224.61.47:4000: getsockopt: co >> >> [IST Apr 19 19:34:05] error : 'etcd_process' process is not running >> >> [IST Apr 19 19:34:05] info : 'etcd_process' trying to restart >> >> [IST Apr 19 19:34:05] info : 'etcd_process' restart: /bin/bash >> >> [IST Apr 19 19:34:36] error : 'etcd_process' failed to restart (exit >> status 2) -- /bin/bash: zmq_msg_recv: Resource temporarily unavailable >> >> context deadline exceeded >> >> >> >> >> >> On Thu, Apr 19, 2018 at 11:03 AM, Sunil Kumar <skgola1...@gmail.com> >> wrote: >> >> Hi, >> >> Any body can help me on this. after ip lost, i update the ip in >> local_config and dns and restart the service. extra vm is deleted lik i >> had 3 sprout node so 2 are deleted. >> >> >> >> [vellum]ubuntu@vellum:~$ cw-config upload shared_config >> >> Unable to contact the etcd cluster. >> >> >> >> thanks >> >> sunil >> >> >> >> _______________________________________________ >> Clearwater mailing list >> Clearwater@lists.projectclearwater.org >> http://lists.projectclearwater.org/mailman/listinfo/clearwat >> er_lists.projectclearwater.org >> >> >
_______________________________________________ Clearwater mailing list Clearwater@lists.projectclearwater.org http://lists.projectclearwater.org/mailman/listinfo/clearwater_lists.projectclearwater.org