[
https://issues.apache.org/jira/browse/CLOUDSTACK-105?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13471641#comment-13471641
]
Caleb Call commented on CLOUDSTACK-105:
---------------------------------------
Our servers are all in bridge mode (per the setup instructions in the
Cloudstack setup guide).
[root@cloud-hv01 log]# cat /etc/xensource/network.conf
bridge
[root@cloud-hv01 log]#
I also noticed in /var/log/messages, the following are being logged over and
over. Looks like this is something related to openvswitch.
Oct 8 10:06:54 cloud-hv01 syslogd 1.4.1: restart.
Oct 8 10:06:57 cloud-hv01 ovs-vsctl:
1568403|stream_unix|ERR|/tmp/stream-unix.9679.784180: connection to
/var/run/openvswitch/db.sock failed: No such file or directory
Oct 8 10:06:57 cloud-hv01 ovs-vsctl:
1568404|reconnect|WARN|unix:/var/run/openvswitch/db.sock: connection attempt
failed (No such file or directory)
Oct 8 10:07:05 cloud-hv01 ovs-vsctl:
1568405|stream_unix|ERR|/tmp/stream-unix.9679.784181: connection to
/var/run/openvswitch/db.sock failed: No such file or directory
Oct 8 10:07:05 cloud-hv01 ovs-vsctl:
1568406|reconnect|WARN|unix:/var/run/openvswitch/db.sock: connection attempt
failed (No such file or directory)
Oct 8 10:07:13 cloud-hv01 ovs-vsctl:
1568407|stream_unix|ERR|/tmp/stream-unix.9679.784182: connection to
/var/run/openvswitch/db.sock failed: No such file or directory
Oct 8 10:07:13 cloud-hv01 ovs-vsctl:
1568408|reconnect|WARN|unix:/var/run/openvswitch/db.sock: connection attempt
failed (No such file or directory)
I've also noticed this same behavior of leaving these stale socket files behind
on a completely different Cloudstack cluster using Xenserver in bridge mode.
> /tmp/stream-unix.####.###### stale sockets causing inodes to run out on
> Xenserver
> ---------------------------------------------------------------------------------
>
> Key: CLOUDSTACK-105
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-105
> Project: CloudStack
> Issue Type: Bug
> Components: XenServer
> Affects Versions: pre-4.0.0
> Environment: Xenserver 6.0.2
> Cloudstack 3.0.2
> Reporter: Caleb Call
> Assignee: Devdeep Singh
> Fix For: 4.1.0
>
>
> We came across an interesting issue in one of our clusters. We ran out of
> inodes on all of our cluster members (since when does this happen in 2012?).
> When this happened, it in turn made the / filesystem a read-only filesystem
> which in turn made all the hosts go in to emergency maintenance mode and as a
> result get marked down by Cloudstack. We found that it was caused by
> hundreds of thousands of stale socket files in /tmp named
> "stream-unix.####.######". To resolve the issue, we had to delete those
> stale socket files (find /tmp -name "*stream*" -mtime +7 -exec rm -v {} \;),
> then kill and restart xapi, then correct the emergency maintenance mode.
> These hosts had only been up for 45 days before this issue occurred.
> In our scouring of the interwebs, the only other instance we've been able to
> find of this (or similar) happening is in the same setup we are currently
> running. Xenserver 6.0.2 with CS 3.0.2. Do these stream-unix sockets have
> anything to do with Cloudstack? I would think if this was a Xenserver issue
> (bug), there would be a lot more on the internet about this happening. For a
> temporary workaround, we've added a cronjob to cleanup these files but we'd
> really like to address the actual issue that's causing these sockets to
> become stale and not get cleaned-up.
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira