On 09/10/2010 11:59 AM, Ferenc Holzhauser wrote: > On 10 September 2010 09:48, Daniel Lezcano<daniel.lezc...@free.fr> wrote: > >> On 09/09/2010 11:22 PM, Papp Tamás wrote: >> >>> Daniel Lezcano wrote, On 2010. 09. 05. 22:30: >>> >>>>>> >>>>> Well, I can. Now again, right after I start a container I get the >>>>> kernel >>>>> panic. I see the console through a KVM, this is a screenshot: >>>>> >>>> Papp, I suppose you attached an image but I don't see it. It is >>>> possible to resend. >>>> I increased the message body size limit to 128KB to the mailing setting. >>>> >>> hi Daniel, >>> >>> Have you checked the screenshots? >>> >> Yes, thanks for sending the screenshots. >> This is a critical bug :s >> >> It seems you didn't face this problem before, right ? Did you change >> something in your configuration ? >> Do you have the same problem with a 2.6.32-23-server kernel ? Are you >> able to reproduce this bug with qemu ? >> >> Thanks >> -- Daniel >> >> >> >> ------------------------------------------------------------------------------ >> Automate Storage Tiering Simply >> Optimize IT performance and efficiency through flexible, powerful, >> automated storage tiering capabilities. View this brief to learn how >> you can reduce costs and improve performance. >> http://p.sf.net/sfu/dell-sfdev2dev >> _______________________________________________ >> Lxc-users mailing list >> Lxc-users@lists.sourceforge.net >> https://lists.sourceforge.net/lists/listinfo/lxc-users >> >> > Hi, > > I might be able to answer that question: > > From the screenshots this appears to be the same issue what I (and at > least another user) have experienced and reported earlier on the list. > I've created a qemu-kvm VM on the server where I had this issue. I can > not reproduce the problem anymore either in this qemu VM or an LXC > container INSIDE that VM. > Yep, sounds the same problem. Let's look at this closely ...
1) Papp is using a 2.6.32-24-server kernel => kernel crash Physical interfaces: =================== ?? Interfaces configuration: ======================== ?? Ifconfig: ======== ?? bridge info: =========== ?? Lxc configuration: ================= lxc.utsname = test lxc.tty = 4 lxc.network.type = veth lxc.network.flags = up lxc.network.link = br1 lxc.network.name = eth1 lxc.network.mtu = 1500 lxc.network.ipv4 = 10.1.1.219/16 lxc.network.hwaddr = AC:DD:22:63:22:22 lxc.network.veth.pair = veth118 lxc.rootfs = /data/lxc/test/rootfs lxc.cgroup.devices.deny = a lxc.cgroup.devices.allow = c 1:3 rwm lxc.cgroup.devices.allow = c 1:5 rwm lxc.cgroup.devices.allow = c 5:1 rwm lxc.cgroup.devices.allow = c 5:0 rwm lxc.cgroup.devices.allow = c 4:0 rwm lxc.cgroup.devices.allow = c 4:1 rwm lxc.cgroup.devices.allow = c 1:9 rwm lxc.cgroup.devices.allow = c 1:8 rwm lxc.cgroup.devices.allow = c 136:* rwm lxc.cgroup.devices.allow = c 5:2 rwm lxc.cgroup.devices.allow = c 254:0 rm 2) Daniel tried on a 2.6.32-23-server kernel => no problem Physical interfaces: =================== 05:00.0 Ethernet controller: Intel Corporation 80003ES2LAN Gigabit Ethernet Controller (Copper) (rev 01) 05:00.1 Ethernet controller: Intel Corporation 80003ES2LAN Gigabit Ethernet Controller (Copper) (rev 01) Interfaces configuration: ======================== auto lo iface lo inet loopback auto eth0 iface eth0 inet dhcp auto br0 iface br0 inet static address 172.20.0.1 netmask 255.255.0.0 bridge_stp off bridge_maxwait 5 pre-up /usr/sbin/brctl addbr br0 post-up /usr/sbin/brctl setfd br0 0 post-up /sbin/iptables -t nat -A POSTROUTING -o eth0 -j MASQUERADE post-up echo 1> /proc/sys/net/ipv4/ip_forward Ifconfig: ======== br0 Link encap:Ethernet HWaddr 56:1e:16:49:91:8a inet addr:172.20.0.1 Bcast:172.20.255.255 Mask:255.255.0.0 inet6 addr: fe80::541e:16ff:fe49:918a/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:59748 errors:0 dropped:0 overruns:0 frame:0 TX packets:117950 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:3722380 (3.7 MB) TX bytes:170211775 (170.2 MB) eth0 Link encap:Ethernet HWaddr 00:e0:81:74:bc:fa inet addr:192.168.254.103 Bcast:192.168.254.255 Mask:255.255.255.0 inet6 addr: fe80::2e0:81ff:fe74:bcfa/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:2700923 errors:0 dropped:0 overruns:0 frame:0 TX packets:1786015 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:100 RX bytes:3651075205 (3.6 GB) TX bytes:147340322 (147.3 MB) Memory:d3000000-d3020000 eth1 Link encap:Ethernet HWaddr 00:e0:81:74:bc:fb UP BROADCAST MULTICAST MTU:1500 Metric:1 RX packets:0 errors:0 dropped:0 overruns:0 frame:0 TX packets:0 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:0 (0.0 B) TX bytes:0 (0.0 B) Memory:d3020000-d3040000 vethiCcp1W Link encap:Ethernet HWaddr 56:1e:16:49:91:8a inet6 addr: fe80::541e:16ff:fe49:918a/64 Scope:Link UP BROADCAST RUNNING PROMISC MULTICAST MTU:1500 Metric:1 RX packets:9744 errors:0 dropped:0 overruns:0 frame:0 TX packets:19033 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:525357 (525.3 KB) TX bytes:28466570 (28.4 MB) bridge info: =========== bridge name bridge id STP enabled interfaces br0 8000.561e1649918a no vethiCcp1W Lxc configuration: ================= lxc.network.type=veth lxc.network.link=br0 lxc.network.flags=up lxc.utsname = ubuntu lxc.tty = 4 lxc.pts = 1024 lxc.rootfs = /var/lib/lxc/ubuntu/rootfs lxc.mount = /var/lib/lxc/ubuntu/fstab lxc.cgroup.devices.deny = a # /dev/null and zero lxc.cgroup.devices.allow = c 1:3 rwm lxc.cgroup.devices.allow = c 1:5 rwm # consoles lxc.cgroup.devices.allow = c 5:1 rwm lxc.cgroup.devices.allow = c 5:0 rwm lxc.cgroup.devices.allow = c 4:0 rwm lxc.cgroup.devices.allow = c 4:1 rwm # /dev/{,u}random lxc.cgroup.devices.allow = c 1:9 rwm lxc.cgroup.devices.allow = c 1:8 rwm lxc.cgroup.devices.allow = c 136:* rwm lxc.cgroup.devices.allow = c 5:2 rwm # rtc lxc.cgroup.devices.allow = c 254:0 rwm 3) Ferenc used a 2.6.32-22 server kernel => kernel crash Physical interfaces: =================== 01:00.0 Ethernet controller: Intel Corporation 82575EB Gigabit Network Connection (rev 02) 01:00.1 Ethernet controller: Intel Corporation 82575EB Gigabit Network Connection (rev 02) 07:00.0 Ethernet controller: Intel Corporation 82575EB Gigabit Network Connection (rev 02) 07:00.1 Ethernet controller: Intel Corporation 82575EB Gigabit Network Connection (rev 02) bridge info: =========== ?? Ifconfig: ======== ?? Interfaces configuration: ======================== # The loopback network interface auto lo iface lo inet loopback # The primary network interface auto br0 iface br0 inet static address 10.0.0.30 network 10.0.0.0 netmask 255.255.255.0 broadcast 10.0.0.255 gateway 10.0.0.1 bridge_ports eth0 bridge_fd 9 bridge_hello 2 bridge_maxage 12 bridge_stp off dns-nameservers 10.0.0.2 dns-search asd.internal_net.com auto br1 iface br1 inet static address 192.168.1.30 network 192.168.1.0 netmask 255.255.255.0 broadcast 192.168.1.255 bridge_ports eth1 bridge_fd 9 bridge_hello 2 bridge_maxage 12 bridge_stp off Lxc configuration: ================= lxc.utsname = backend1 lxc.tty = 4 lxc.network.type = veth lxc.network.flags = up lxc.network.link = br0 lxc.network.name = eth0 lxc.network.mtu = 1500 lxc.network.ipv4 = 10.0.0.45/24 lxc.network.type = veth lxc.network.flags = up lxc.network.link = br1 lxc.network.name = eth1 lxc.network.mtu = 1500 lxc.network.ipv4 = 192.168.1.45/24 lxc.rootfs = /lxc/backend1/root lxc.mount = /lxc/backend1/fstab lxc.cgroup.devices.deny = a # /dev/null and zero lxc.cgroup.devices.allow = c 1:3 rwm lxc.cgroup.devices.allow = c 1:5 rwm # consoles lxc.cgroup.devices.allow = c 5:1 rwm lxc.cgroup.devices.allow = c 5:0 rwm lxc.cgroup.devices.allow = c 4:0 rwm lxc.cgroup.devices.allow = c 4:1 rwm # /dev/{,u}random lxc.cgroup.devices.allow = c 1:9 rwm lxc.cgroup.devices.allow = c 1:8 rwm # /dev/pts/* - pts namespaces are "coming soon" lxc.cgroup.devices.allow = c 136:* rwm lxc.cgroup.devices.allow = c 5:2 rwm # rtc lxc.cgroup.devices.allow = c 254:0 rwm ------------------------------------------------------------------------------ Automate Storage Tiering Simply Optimize IT performance and efficiency through flexible, powerful, automated storage tiering capabilities. View this brief to learn how you can reduce costs and improve performance. http://p.sf.net/sfu/dell-sfdev2dev _______________________________________________ Lxc-users mailing list Lxc-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/lxc-users