Hi Pushpendra, Firstly I have CC’d this onto the clearwater mailing list and please send all replies via this as otherwise your emails may not be seen by my colleagues or me! In answer to your questions I don’t think what the ns value is matters so what you’ve got should be fine, and the second value using your bind IP looks good. It also doesn’t look like the DNS server/queries is the problem in your situation, as homestead is trying to contact the vellum node and failing there. If you have a look at the logs on the vellum node for Cassandra, Astaire and Rogers, these are at /var/log/Cassandra/, /var/log/Astaire and /var/log/rogers. This should give a clue as to what is happening, as based on the monit.log they don’t look like they’re working properly. Let us know how it goes, Ben
From: Pushpendra [mailto:pushpendra16mn...@gmail.com] Sent: 09 March 2018 11:06 To: Bennett Allen <bennett.al...@metaswitch.com> Subject: Failed to register to client (Zoiper5 and XLite) Hi Ben, I am able to login and create private identity on ellis, but when I am trying to register these identity on client app (zoiper5 an XLite), its not registered. Its in an endless loop. I am using <zone> as domain while registering to client. Please find the log detail below: Please let me what is wrong and how can I fix, thanks in advance ☺ [vellum]ubuntu@vellum:/var/log$ tail -40 monit.log [IST Mar 7 23:54:05] error : 'cassandra_process' process is not running [IST Mar 7 23:54:05] info : 'cassandra_process' trying to restart [IST Mar 7 23:54:05] info : 'cassandra_process' restart: /bin/bash [IST Mar 7 23:54:06] info : 'cassandra_process' process is running with pid 31746 [IST Mar 7 23:54:06] error : 'cassandra_uptime' '/usr/share/clearwater/infrastructure/monit_uptime/check-cassandra-uptime' failed with exit status (1) -- no output [IST Mar 7 23:54:06] error : 'astaire_uptime' '/usr/share/clearwater/infrastructure/monit_uptime/check-astaire-uptime' failed with exit status (1) -- no output [IST Mar 7 23:54:16] error : 'rogers_process' process is not running [IST Mar 7 23:54:16] info : 'rogers_process' trying to restart [IST Mar 7 23:54:16] info : 'rogers_process' restart: /bin/bash [IST Mar 7 23:54:16] info : 'rogers_process' process is running with pid 31827 [IST Mar 7 23:54:16] error : 'clearwater_cluster_manager_process' process is not running [IST Mar 7 23:54:16] info : 'clearwater_cluster_manager_process' trying to restart [IST Mar 7 23:54:16] info : 'clearwater_cluster_manager_process' restart: /bin/bash [IST Mar 7 23:54:17] error : 'cassandra_process' process is not running [IST Mar 7 23:54:17] info : 'cassandra_process' trying to restart [IST Mar 7 23:54:17] info : 'cassandra_process' restart: /bin/bash [IST Mar 7 23:54:17] info : 'cassandra_process' process is running with pid 32022 [IST Mar 7 23:54:17] error : 'cassandra_uptime' '/usr/share/clearwater/infrastructure/monit_uptime/check-cassandra-uptime' failed with exit status (1) -- no output [IST Mar 7 23:54:17] error : 'astaire_uptime' '/usr/share/clearwater/infrastructure/monit_uptime/check-astaire-uptime' failed with exit status (1) -- no output [IST Mar 7 23:54:27] error : 'rogers_uptime' '/usr/share/clearwater/infrastructure/monit_uptime/check-rogers-uptime' failed with exit status (1) -- no output [IST Mar 7 23:54:27] info : 'clearwater_cluster_manager_process' process is running with pid 31936 [IST Mar 7 23:54:27] error : 'cassandra_process' process is not running [IST Mar 7 23:54:27] info : 'cassandra_process' trying to restart [IST Mar 7 23:54:27] info : 'cassandra_process' restart: /bin/bash [IST Mar 7 23:54:28] info : 'cassandra_process' process is running with pid 32610 [IST Mar 7 23:54:28] error : 'cassandra_uptime' '/usr/share/clearwater/infrastructure/monit_uptime/check-cassandra-uptime' failed with exit status (1) -- no output [IST Mar 7 23:54:28] error : 'astaire_uptime' '/usr/share/clearwater/infrastructure/monit_uptime/check-astaire-uptime' failed with exit status (1) -- no output [IST Mar 7 23:54:38] error : 'rogers_uptime' '/usr/share/clearwater/infrastructure/monit_uptime/check-rogers-uptime' failed with exit status (1) -- no output [IST Mar 7 23:54:38] error : 'cassandra_uptime' '/usr/share/clearwater/infrastructure/monit_uptime/check-cassandra-uptime' failed with exit status (1) -- no output [IST Mar 7 23:54:39] info : 'astaire_uptime' status succeeded [status=0] -- zmq_msg_recv: Resource temporarily unavailable [IST Mar 7 23:54:46] info : Awakened by the SIGHUP signal Reinitializing Monit - Control file '/etc/monit/monitrc' [IST Mar 7 23:54:47] info : 'node-vellum' Monit reloaded [IST Mar 7 23:54:47] error : Cannot create socket to [localhost]:2812 -- Connection refused [IST Mar 7 23:54:57] info : Awakened by the SIGHUP signal Reinitializing Monit - Control file '/etc/monit/monitrc' [IST Mar 7 23:54:57] info : 'node-vellum' Monit reloaded [IST Mar 7 23:54:57] error : Cannot create socket to [localhost]:2812 -- Connection refused [IST Mar 7 23:55:07] error : 'cassandra_uptime' '/usr/share/clearwater/infrastructure/monit_uptime/check-cassandra-uptime' failed with exit status (1) -- no output [IST Mar 7 23:55:17] info : 'cassandra_uptime' status succeeded [status=0] -- zmq_msg_recv: Resource temporarily unavailable [dime]ubuntu@dime:/var/log/homestead$ tail -30 homestead_err.log Thrift: Wed Mar 7 23:54:03 2018 TSocket::open() error on socket (after THRIFT_POLL) <Host: 10.224.61.24 Port: 9160>Connection refused Thrift: Wed Mar 7 23:54:03 2018 TSocket::open() error on socket (after THRIFT_POLL) <Host: 10.224.61.24 Port: 9160>Connection refused [dime]ubuntu@dime:/var/log/ralf$ tail -40 ralf_current.txt 08-03-2018 14:00:02.739 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:00:07.740 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:00:12.740 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:00:17.744 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:00:22.745 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:00:27.745 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:00:31.591 UTC [7f0837fff700] Status alarm.cpp:244: Reraising all alarms with a known state 08-03-2018 14:00:32.745 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:00:37.746 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:00:42.746 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:00:47.747 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:00:52.747 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:00:57.747 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:01:01.591 UTC [7f0837fff700] Status alarm.cpp:244: Reraising all alarms with a known state 08-03-2018 14:01:02.748 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:01:07.748 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:01:12.749 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:01:17.749 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:01:22.749 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:01:27.750 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:01:31.591 UTC [7f0837fff700] Status alarm.cpp:244: Reraising all alarms with a known state 08-03-2018 14:03:57.762 UTC [7f07ee7bc700] Error diameterstack.cpp:853: No Diameter peers have been found 08-03-2018 14:04:01.599 UTC [7f0837fff700] Status alarm.cpp:244: Reraising all alarms with a known state 08-03-2018 14:04:02.764 UTC [7f07ee7bc700] Warning dnscachedresolver.cpp:836: Failed to retrieve record for _diameter._tcp.iind.intel.com<http://tcp.iind.intel.com>: Domain name not found 08-03-2018 14:04:02.764 UTC [7f07ee7bc700] Warning dnscachedresolver.cpp:836: Failed to retrieve record for _diameter._sctp.iind.intel.com<http://sctp.iind.intel.com>: Domain name not found [dime]ubuntu@dime:/var/log/homestead$ tail -30 homestead_current.txt 08-03-2018 16:00:25.983 UTC [7fd5977fe700] Status alarm.cpp:244: Reraising all alarms with a known state 08-03-2018 16:00:55.984 UTC [7fd5977fe700] Status alarm.cpp:244: Reraising all alarms with a known state 08-03-2018 16:01:25.984 UTC [7fd5977fe700] Status alarm.cpp:244: Reraising all alarms with a known state 08-03-2018 16:01:55.984 UTC [7fd5977fe700] Status alarm.cpp:244: Reraising all alarms with a known state 08-03-2018 16:02:25.989 UTC [7fd5977fe700] Status alarm.cpp:244: Reraising all alarms with a known state 08-03-2018 16:02:55.989 UTC [7fd5977fe700] Status alarm.cpp:244: Reraising all alarms with a known state 08-03-2018 16:03:25.989 UTC [7fd5977fe700] Status alarm.cpp:244: Reraising all alarms with a known state [sprout]ubuntu@sprout:/var/log/sprout$ tail -30 sprout_current.txt 08-03-2018 17:45:38.954 UTC [7fd0b37be700] Debug pjsip: sip_endpoint.c Distributing rdata to modules: Request msg OPTIONS/cseq=97208 (rdata0x7fd04c092988) 08-03-2018 17:45:38.954 UTC [7fd0b37be700] Debug uri_classifier.cpp:139: home domain: false, local_to_node: true, is_gruu: false, enforce_user_phone: false, prefer_sip: true, treat_number_as_phone: false 08-03-2018 17:45:38.954 UTC [7fd0b37be700] Debug uri_classifier.cpp:172: Classified URI as 3 08-03-2018 17:45:38.954 UTC [7fd0b37be700] Debug pjsip: endpoint Response msg 200/OPTIONS/cseq=97208 (tdta0x7fd0d40829d0) created 08-03-2018 17:45:38.954 UTC [7fd0b37be700] Verbose common_sip_processing.cpp:103: TX 277 bytes Response msg 200/OPTIONS/cseq=97208 (tdta0x7fd0d40829d0) to TCP 10.224.61.22:52668<http://10.224.61.22:52668>: --start msg-- SIP/2.0 200 OK Via: SIP/2.0/TCP 10.224.61.22;rport=52668;received=10.224.61.22;branch=z9hG4bK-97208 Call-ID: poll-sip-97208 From: "poll-sip" <sip:poll-sip@10.224.61.22>;tag=97208 To: <sip:poll-sip@10.224.61.22>;tag=z9hG4bK-97208 CSeq: 97208 OPTIONS Content-Length: 0 --end msg-- 08-03-2018 17:45:38.954 UTC [7fd0b37be700] Debug common_sip_processing.cpp:275: Skipping SAS logging for OPTIONS response 08-03-2018 17:45:38.954 UTC [7fd0b37be700] Debug pjsip: tdta0x7fd0d408 Destroying txdata Response msg 200/OPTIONS/cseq=97208 (tdta0x7fd0d40829d0) 08-03-2018 17:45:38.954 UTC [7fd0b37be700] Debug thread_dispatcher.cpp:270: Worker thread completed processing message 0x7fd04c092988 08-03-2018 17:45:38.954 UTC [7fd0b37be700] Debug thread_dispatcher.cpp:284: Request latency = 292us 08-03-2018 17:45:38.954 UTC [7fd0b37be700] Debug event_statistic_accumulator.cpp:32: Accumulate 292 for 0x187c788 08-03-2018 17:45:38.954 UTC [7fd0b37be700] Debug event_statistic_accumulator.cpp:32: Accumulate 292 for 0x187c7d0 08-03-2018 17:45:38.954 UTC [7fd0b37be700] Debug load_monitor.cpp:341: Not recalculating rate as we haven't processed 20 requests yet (only 10). 08-03-2018 17:45:38.954 UTC [7fd0b37be700] Debug utils.cpp:878: Removed IOHook 0x7fd0b37bde30 to stack. There are now 0 hooks 08-03-2018 17:45:38.954 UTC [7fd0b37be700] Debug thread_dispatcher.cpp:158: Attempting to process queue element 08-03-2018 17:45:39.463 UTC [7fd0d17fa700] Warning (Net-SNMP): Warning: Failed to connect to the agentx master agent ([NIL]): 08-03-2018 17:45:40.955 UTC [7fd053eff700] Verbose pjsip: tcps0x7fd04c09 TCP connection closed 08-03-2018 17:45:40.955 UTC [7fd053eff700] Debug connection_tracker.cpp:67: Connection 0x7fd04c0910f8 has been destroyed 08-03-2018 17:45:40.955 UTC [7fd053eff700] Verbose pjsip: tcps0x7fd04c09 TCP transport destroyed with reason 70016: End of file (PJ_EEOF) $TTL 5m ; Default TTL ; SOA, NS and A record for DNS server itself @ 3600 IN SOA ns admin ( 2014010800 ; Serial 3600 ; Refresh 3600 ; Retry 3600 ; Expire 300 ) ; Minimum TTL @ 3600 IN NS ns ns 3600 IN A 10.224.61.82 ; IPv4 address of BIND server ; bono ; ==== ; ; Per-node records - not required to have both IPv4 and IPv6 records bono-1 IN A 10.224.61.8 . . . . In zone file infront of ns (you can see above) I am using the IP address of machine where i installed the bind9, but in that machine i checked in /etc/resolv.conf file nameserver is 127.0.0.53. (what should I use in place of ns?) -> In clearwater node also i have write in /etc/dnsmasq.resolv.conf naeserver as 10.224.61.82 (that machine's IP where i installed bind9), what should i write? Thanks, Pushpendra
_______________________________________________ Clearwater mailing list Clearwater@lists.projectclearwater.org http://lists.projectclearwater.org/mailman/listinfo/clearwater_lists.projectclearwater.org