On Sat, 2013-05-18 at 12:09 -0700, jjs - mainphrame wrote: 
> Interesting. I didn't realize how spoiled I am and how easy I have it with
> lxc on ubuntu!

Don't get too too comfortable.  I don't know if Ubuntu is transitioning
to systemd yet or not (or planning to or creating a new alternative) but
this was a talk tease from a talk delivered by my friend and co-guru
Mark Torres (formerly of Suse fame) to the Atlanta Unix Users Group a
couple of months ago...

-- 
  Marc will be discussing some of the myths, uses and advantages of
  this replacement of the Linux sysvinit package.  Despite the best
  efforts of many (and a never ending supply of mailing list
  flames, holy water, silver bullets, wooden stakes, poisons and
  evil looks) it is coming soon to a Linux Distro near you.
-- 

You may have to deal with the likes of systemd sooner or later (or some
other variant in similar concept).  This is not something peculiar to
Fedora (OpenSuse, and Arch Linux both have systemd as will RHEL, CentOS,
Scientific Linux, etc...).  By the time it gets to Ubuntu, though, I
suspect it will be much more stable.

Seems like Fedora adopted it early (very much like Ubuntu with Unity)
and Fedora has sort of been the proving ground for systemd (that's what
Fedora is for).  Personally, I'm in the "holy water and wooden stakes"
category (and I have at least one flame war on the systemd list to back
me up) with regards to systemd but it is what it is.  We've been dealing
with it as it evolves.

It's worth noting that I've decided that I will never get Fedora 15 or
Fedora 16 running in a container (they will work as a host, however) due
to incompatible versions of systemd.  But F15 and F16 are EOL and thus
not worth any effort or stress.  They (systemd) obviously found
solutions for the havoc they caused in those versions and we have
managed to get F17 and F18 working with the recent versions of systemd.
It's not Fedora, it's systemd that create these problems.

OTOH...  Unity caused me to totally abandon Ubuntu.  It totally broke
freenx and the NX support and try as I might, I could not get it to work
after trying for 2 or 3 releases of Ubuntu.  Even though the freenx
package was still in the repos, several of the libraries upon which it
depended were gone or deprecated.  I tried all the manual patches and
source compiles and nothing worked or created irreconcilable conflicts.
If I can't use NX to get graphic desktops in a container or on a remote
host system, Ubuntu became useless to me and I had to rip it out where I
had it.  Even the laptops for my grandkids (which were running Ubuntu)
are now running on Fedora for that reason (so grandpa can have a remote
desktop on their laptop and do remote graphical administration).

Maybe they have since fixed it but they already lost me on that one
point alone.  I do need to get some (at least one) containers running
Ubuntu running just so I can run Oregano (Electronics Schematic CAD
program with simulation capabilities) which I can not get to work on
Fedora 18 (Fedora 19 might work) due to some gtk3+ library issues.  So
maybe I'll revisit Ubuntu again soon.  I have successfully built an
Ubuntu container on a Fedora host.  That's one of the nice things, that
I can cross build Ubuntu on a Fedora host.  I think (but I haven't
tried) the reverse is true.  If not, I may have to fix that in the
Fedora template.  That's not true of many of the other templates.  I
can't build an Arch Linux container on a Fedora host.  That's something
the container owners have to address.

We've gotten it (systemd) working and functional (maybe not totally
"pretty") and now it's more a matter of getting the updated revs into
the downstream releases to make it just as easy.  That's the way it
works.  All the template maintainers (and I'm just a contributor) are
shooting at moving targets.  What's in the git repo right now for the
Fedora template is pretty sweet (if I do say so myself) and will build
Fedora containers out of the box.

Which reminds me.  The dudes that do the Fedora Remix for the Raspberry
Pi just renamed their distribution to Pidor (shades of Kubuntu, Edubuntu
and the other remixes of Ubuntu).  That's going to break the lxc-fedora
template on the Raspberry Pi (yet again - pounding head on desk).  Right
now, AFAICT, the ARM support for Ubuntu is only for ARM7 and up which
leaves out the RPi (ARM6).  Oh well.  It never ends.

> Joe

Regards,
Mike 

> On Sat, May 18, 2013 at 11:19 AM, Michael H. Warfield 
> <m...@wittsend.com>wrote:
> 
> > On Sat, 2013-05-18 at 19:41 +0530, Ajith Adapa wrote:
> > > Hmm sounds one more road block for using lxc in fedora 17 because of
> > > systemd.
> >
> > It's not a roadblock.  More like a mile long stretch of stingers (stop
> > spike strips / tire deflators).  We're getting there.  It's just one
> > more unnecessary puzzle to solve.  Sigh...
> >
> > > Currently there is no place where there is a guide for starting up
> > > with LXC for latest fedora versions. I think a page in fedoraproject
> > > would be of great help with the known issues and steps using lxc under
> > > various fedora versions.
> >
> > First we get it working but, yeah, that would be incredibly nice and
> > then also add it to this project as well.
> >
> > > I am really thinking to start using LXC containers in fedora 14. Build
> > > and Boot it up with latest stable kernel version (Might be 3.4) and
> > > LXC version (>0.9) and try out using LXC- containers :)
> > >
> > >
> > >
> > >
> > > On Sat, May 18, 2013 at 7:28 PM, Michael H. Warfield
> > > <m...@wittsend.com> wrote:
> > >         On Sat, 2013-05-18 at 19:02 +0530, Ajith Adapa wrote:
> > >         > Sorry for the confusion.
> > >
> > >         > In case of issue 3, I felt host kernel crashed because of
> > >         the soft
> > >         > lock issue mentioned in Issue 2.That's the reason I was
> > >         saying "as a
> > >         > result of ..". Ideally speaking I haven't done anything
> > >         other than
> > >         > creating the lxc-container at the time. Once I restarted the
> > >         host
> > >         > machine after crash I havent observed any issues.
> > >
> > >         > Then I have started the container using below command and
> > >         tried to
> > >         > connect to its shell using lxc-console command but I ended
> > >         up with
> > >         > below message. Ideally I should see a prompt but its just
> > >         hangs down
> > >         > there. <Ctl+a q> works and nothing else.
> > >
> > >         > [root@ipiblr ~]# lxc-start -n TEST -d
> > >         > [root@ipiblr ~]# lxc-console -n TEST
> > >
> > >         > Type <Ctrl+a q> to exit the console, <Ctrl+a Ctrl+a> to
> > >         enter Ctrl+a
> > >         > itself
> > >
> > >
> > >         Oh, crap...  I keep forgetting about that (because I don't use
> > >         it).
> > >         That needs to be noted somewhere in the documentation.
> > >
> > >         That's yet another BAD decision on the part of the systemd
> > >         crowd,
> > >         lxc-console is probably not going to work, at least for the
> > >         time being.
> > >         They (systemd) intentionally, with documented malice a
> > >         forethought,
> > >         disable gettys on the vtys in the container if systemd detects
> > >         that it's
> > >         in a container.  However, /dev/console in the container is
> > >         still active
> > >         and is connected to lxc-start and I'm able to log in there but
> > >         I have
> > >         never gotten lxc-console to work with a systemd container and
> > >         I don't
> > >         know of anything I can do about it.  You would need some way
> > >         to force
> > >         the container to start gettys on the vtys.
> > >
> > >         Maybe, if I (or someone else) can figure out a way to do that
> > >         (force the
> > >         gettys to start on the vtys), it could be integrated into the
> > >         Fedora
> > >         template.  My patches for the autodev stuff (plus other stuff)
> > >         have now
> > >         been accepted and applied by Serge, so that's done.  Maybe I
> > >         can look
> > >         deeper into this morass now.
> > >
> > >         Regards,
> > >         Mike
> > >
> > >         > Regards,
> > >         > Ajith
> > >         >
> > >         >
> > >         >
> > >         >
> > >         > On Sat, May 18, 2013 at 5:55 PM, Michael H. Warfield
> > >         > <m...@wittsend.com> wrote:
> > >         >         Hello,
> > >         >
> > >         >         On Sat, 2013-05-18 at 12:35 +0530, Ajith Adapa
> > >         wrote:
> > >         >         > Hi,
> > >         >
> > >         >         > I have installed all the rpms created by @thomas
> > >         and
> > >         >         followed @michael
> > >         >         > steps to start a lxc container.
> > >         >
> > >         >         > I have a doubt.
> > >         >
> > >         >         > 1. When I give lxc-create command I came across
> > >         huge
> > >         >         download of various
> > >         >         > files.
> > >         >         > As per my understanding rootfs is created for new
> > >         container
> > >         >         (where can i
> > >         >         > get the steps for it ? ).
> > >         >
> > >         >
> > >         >         Steps for what?  It's
> > >         in /var/lib/lxc/{Container}/rootfs/
> > >         >
> > >         >         > But I see below log. Is there any issue ?
> > >         >
> > >         >         > Copy /var/cache/lxc/fedora/i686/17/rootfs
> > >         >         to /var/lib/lxc/TEST/TEST/rootfs
> > >         >         > ...
> > >         >         > Copying rootfs
> > >         to /var/lib/lxc/TEST/TEST/rootfs ...setting
> > >         >         root passwd to
> > >         >         > root
> > >         >         > installing fedora-release package
> > >         >         > warning: Failed to read auxiliary vector, /proc
> > >         not mounted?
> > >         >         > warning: Failed to read auxiliary vector, /proc
> > >         not mounted?
> > >         >         > warning: Failed to read auxiliary vector, /proc
> > >         not mounted?
> > >         >         > warning: Failed to read auxiliary vector, /proc
> > >         not mounted?
> > >         >         > warning: Failed to read auxiliary vector, /proc
> > >         not mounted?
> > >         >         > warning: Failed to read auxiliary vector, /proc
> > >         not mounted?
> > >         >         > warning: Failed to read auxiliary vector, /proc
> > >         not mounted?
> > >         >         > warning: Failed to read auxiliary vector, /proc
> > >         not mounted?
> > >         >
> > >         >
> > >         >         The warnings are perfectly normal and harmless.  I
> > >         ran into
> > >         >         this with
> > >         >         recent versions of yum and researched it.  It's
> > >         because /proc
> > >         >         is not
> > >         >         mounted in the container itself when the container
> > >         is being
> > >         >         created.
> > >         >         You can ignore them.
> > >         >
> > >         >         > Package fedora-release-17-2.noarch already
> > >         installed and
> > >         >         latest version
> > >         >         > Nothing to do
> > >         >
> > >         >
> > >         >         Again, normal.
> > >         >
> > >         >         > container rootfs and config created
> > >         >         > 'fedora' template installed
> > >         >         > 'TEST' created
> > >         >
> > >         >
> > >         >         Looks like your container was created.  I don't see
> > >         a problem.
> > >         >
> > >         >         > 2.I see a SOFT LOCK issue with latest version
> > >         kernel shown
> > >         >         below.
> > >         >
> > >         >         > # uname -a
> > >         >         > Linux blr 3.8.8-100.fc17.i686 #1 SMP Wed Apr 17
> > >         17:26:59 UTC
> > >         >         2013 i686 i686
> > >         >         > i386 GNU/Linux
> > >         >         >
> > >         >         >
> > >         >         > [1098069.351017] SELinux: initialized (dev
> > >         binfmt_misc, type
> > >         >         binfmt_misc),
> > >         >         > uses genfs_contexts
> > >         >         > [1281973.370052] BUG: soft lockup - CPU#0 stuck
> > >         for 23s!
> > >         >         [kworker/0:1:2201]
> > >         >
> > >         >
> > >         >         I've seen that on my Dell 610's but they haven't
> > >         caused any
> > >         >         real
> > >         >         failures.  Not quite sure what that is.
> > >         >
> > >         >         > [1281973.370052] Modules linked in: binfmt_misc
> > >         lockd sunrpc
> > >         >         snd_intel8x0
> > >         >         > snd_ac97_codec ac97_bus snd_seq snd_seq_device
> > >         snd_pcm
> > >         >         i2c_piix4 i2c_core
> > >         >         > microcode virtio_balloon snd_page_alloc snd_timer
> > >         snd
> > >         >         soundcore virtio_net
> > >         >         > uinput virtio_blk
> > >         >         > [1281973.370052] Pid: 2201, comm: kworker/0:1 Not
> > >         tainted
> > >         >         > 3.8.8-100.fc17.i686 #1 Bochs Bochs
> > >         >         > [1281973.370052] EIP: 0060:[<c068b17a>] EFLAGS:
> > >         00000206
> > >         >         CPU: 0
> > >         >         > [1281973.370052] EIP is at iowrite16+0x1a/0x40
> > >         >         > [1281973.370052] EAX: 00000001 EBX: f69b3000 ECX:
> > >         0001c050
> > >         >         EDX: 0000c050
> > >         >         > [1281973.370052] ESI: e9d9b600 EDI: 00000000 EBP:
> > >         f5009b90
> > >         >         ESP: f5009b8c
> > >         >         > [1281973.370052]  DS: 007b ES: 007b FS: 00d8 GS:
> > >         00e0 SS:
> > >         >         0068
> > >         >         > [1281973.370052] CR0: 8005003b CR2: 09cae530 CR3:
> > >         345e0000
> > >         >         CR4: 000006d0
> > >         >         > [1281973.370052] DR0: 00000000 DR1: 00000000 DR2:
> > >         00000000
> > >         >         DR3: 00000000
> > >         >         > [1281973.370052] DR6: ffff0ff0 DR7: 00000400
> > >         >         > [1281973.370052] Process kworker/0:1 (pid: 2201,
> > >         ti=f5008000
> > >         >         task=f6830cb0
> > >         >         > task.ti=f4bb2000)
> > >         >         > [1281973.370052] Stack:
> > >         >         > [1281973.370052]  c07107cd f5009b9c c070ffb9
> > >         f4a17a00
> > >         >         f5009bcc f7c36f2b
> > >         >         > 00000000 e9d9b600
> > >         >         > [1281973.370052]  00000020 00000000 e9d9b600
> > >         00000000
> > >         >         f69b2000 00000000
> > >         >         > f4b5a740 00000036
> > >         >         > [1281973.370052]  f5009c00 c088ea5e e9d9b600
> > >         00000000
> > >         >         f7c384c0 f6822600
> > >         >         > f69b2000 00000000
> > >         >         > [1281973.370052] Call Trace:
> > >         >         > [1281973.370052]  [<c07107cd>] ? vp_notify
> > >         +0x1d/0x20
> > >         >         > [1281973.370052]  [<c070ffb9>] virtqueue_kick
> > >         +0x19/0x20
> > >         >         > [1281973.370052]  [<f7c36f2b>] start_xmit
> > >         +0x14b/0x370
> > >         >         [virtio_net]
> > >         >         > [1281973.370052]  [<c088ea5e>] dev_hard_start_xmit
> > >         >         +0x24e/0x4c0
> > >         >         > [1281973.370052]  [<c08a793f>] sch_direct_xmit
> > >         +0xaf/0x180
> > >         >         > [1281973.370052]  [<c088f01e>] dev_queue_xmit
> > >         +0x12e/0x370
> > >         >         > [1281973.370052]  [<c08bf670>] ? ip_fragment
> > >         +0x870/0x870
> > >         >         > [1281973.370052]  [<c08bf88e>] ip_finish_output
> > >         +0x21e/0x3b0
> > >         >         > [1281973.370052]  [<c08bf670>] ? ip_fragment
> > >         +0x870/0x870
> > >         >         > [1281973.370052]  [<c08c0354>] ip_output+0x84/0xd0
> > >         >         > [1281973.370052]  [<c08bf670>] ? ip_fragment
> > >         +0x870/0x870
> > >         >         > [1281973.370052]  [<c08bfb00>] ip_local_out
> > >         +0x20/0x30
> > >         >         > [1281973.370052]  [<c08bfc3f>] ip_queue_xmit
> > >         +0x12f/0x3b0
> > >         >         > [1281973.370052]  [<c08d62fb>] tcp_transmit_skb
> > >         +0x3cb/0x850
> > >         >         > [1281973.370052]  [<c097a440>] ?
> > >         apic_timer_interrupt
> > >         >         +0x34/0x3c
> > >         >         > [1281973.370052]  [<c08d8b50>] tcp_send_ack
> > >         +0xd0/0x120
> > >         >         > [1281973.370052]  [<c08cc096>] __tcp_ack_snd_check
> > >         +0x56/0x90
> > >         >         > [1281973.370052]  [<c08d3038>] tcp_rcv_established
> > >         >         +0x1c8/0x890
> > >         >         > [1281973.370052]  [<c08dc8f3>] tcp_v4_do_rcv
> > >         +0x223/0x3e0
> > >         >         > [1281973.370052]  [<c06233f4>] ?
> > >         security_sock_rcv_skb
> > >         >         +0x14/0x20
> > >         >         > [1281973.370052]  [<c08de39c>] tcp_v4_rcv
> > >         +0x53c/0x770
> > >         >         > [1281973.370052]  [<c08bb110>] ? ip_rcv_finish
> > >         +0x320/0x320
> > >         >         > [1281973.370052]  [<c08bb1c2>]
> > >         ip_local_deliver_finish
> > >         >         +0xb2/0x260
> > >         >         > [1281973.370052]  [<c08bb4ac>] ip_local_deliver
> > >         +0x3c/0x80
> > >         >         > [1281973.370052]  [<c08bb110>] ? ip_rcv_finish
> > >         +0x320/0x320
> > >         >         > [1281973.370052]  [<c08bae50>] ip_rcv_finish
> > >         +0x60/0x320
> > >         >         > [1281973.370052]  [<c043009c>] ?
> > >         pvclock_clocksource_read
> > >         >         +0x9c/0x130
> > >         >         > [1281973.370052]  [<c08bb73c>] ip_rcv+0x24c/0x370
> > >         >         > [1281973.370052]  [<c088d5db>] __netif_receive_skb
> > >         >         +0x5bb/0x740
> > >         >         > [1281973.370052]  [<c088d8ce>] netif_receive_skb
> > >         +0x2e/0x90
> > >         >         > [1281973.370052]  [<f7c36a49>] virtnet_poll
> > >         +0x449/0x6a0
> > >         >         [virtio_net]
> > >         >         > [1281973.370052]  [<c044d6aa>] ? run_timer_softirq
> > >         >         +0x1a/0x210
> > >         >         > [1281973.370052]  [<c088decd>] net_rx_action
> > >         +0x11d/0x1f0
> > >         >         > [1281973.370052]  [<c044695b>] __do_softirq
> > >         +0xab/0x1c0
> > >         >         > [1281973.370052]  [<c04468b0>] ?
> > >         local_bh_enable_ip
> > >         >         +0x90/0x90
> > >         >         > [1281973.370052]  <IRQ>
> > >         >         > [1281973.370052]  [<c0446bdd>] ? irq_exit
> > >         +0x9d/0xb0
> > >         >         > [1281973.370052]  [<c04258ee>] ?
> > >         smp_apic_timer_interrupt
> > >         >         +0x5e/0x90
> > >         >         > [1281973.370052]  [<c097a440>] ?
> > >         apic_timer_interrupt
> > >         >         +0x34/0x3c
> > >         >         > [1281973.370052]  [<c044007b>] ? console_start
> > >         +0xb/0x20
> > >         >         > [1281973.370052]  [<c0979bbf>] ?
> > >         _raw_spin_unlock_irqrestore
> > >         >         +0xf/0x20
> > >         >         > [1281973.370052]  [<c07918d6>] ? ata_scsi_queuecmd
> > >         >         +0x96/0x250
> > >         >         > [1281973.370052]  [<c076ad18>] ? scsi_dispatch_cmd
> > >         >         +0xb8/0x260
> > >         >         > [1281973.370052]  [<c066007b>] ?
> > >         queue_store_random
> > >         >         +0x4b/0x70
> > >         >         > [1281973.370052]  [<c07711b3>] ? scsi_request_fn
> > >         +0x2c3/0x4b0
> > >         >         > [1281973.370052]  [<c042f2b7>] ? kvm_clock_read
> > >         +0x17/0x20
> > >         >         > [1281973.370052]  [<c0409448>] ? sched_clock
> > >         +0x8/0x10
> > >         >         > [1281973.370052]  [<c065cace>] ? __blk_run_queue
> > >         +0x2e/0x40
> > >         >         > [1281973.370052]  [<c066214a>] ?
> > >         blk_execute_rq_nowait
> > >         >         +0x6a/0xd0
> > >         >         > [1281973.370052]  [<c066221d>] ? blk_execute_rq
> > >         +0x6d/0xe0
> > >         >         > [1281973.370052]  [<c06620b0>] ?
> > >         __raw_spin_unlock_irq
> > >         >         +0x10/0x10
> > >         >         > [1281973.370052]  [<c0446ba7>] ? irq_exit
> > >         +0x67/0xb0
> > >         >         > [1281973.370052]  [<c04258ee>] ?
> > >         smp_apic_timer_interrupt
> > >         >         +0x5e/0x90
> > >         >         > [1281973.370052]  [<c097a440>] ?
> > >         apic_timer_interrupt
> > >         >         +0x34/0x3c
> > >         >         > [1281973.370052]  [<c076ffa0>] ? scsi_execute
> > >         +0xb0/0x140
> > >         >         > [1281973.370052]  [<c0771429>] ? scsi_execute_req
> > >         +0x89/0x100
> > >         >         > [1281973.370052]  [<c077f3d5>] ? sr_check_events
> > >         +0xb5/0x2e0
> > >         >         > [1281973.370052]  [<c07a64cd>] ?
> > >         cdrom_check_events
> > >         >         +0x1d/0x40
> > >         >         > [1281973.370052]  [<c077f856>] ?
> > >         sr_block_check_events
> > >         >         +0x16/0x20
> > >         >         > [1281973.370052]  [<c06663c5>] ? disk_check_events
> > >         +0x45/0xf0
> > >         >         > [1281973.370052]  [<c0666485>] ?
> > >         disk_events_workfn
> > >         >         +0x15/0x20
> > >         >         > [1281973.370052]  [<c045788e>] ? process_one_work
> > >         >         +0x12e/0x3d0
> > >         >         > [1281973.370052]  [<c097a440>] ?
> > >         apic_timer_interrupt
> > >         >         +0x34/0x3c
> > >         >         > [1281973.370052]  [<c0459939>] ? worker_thread
> > >         +0x119/0x3b0
> > >         >         > [1281973.370052]  [<c0459820>] ?
> > >         flush_delayed_work
> > >         >         +0x50/0x50
> > >         >         > [1281973.370052]  [<c045e2a4>] ? kthread+0x94/0xa0
> > >         >         > [1281973.370052]  [<c0980ef7>] ?
> > >         ret_from_kernel_thread
> > >         >         +0x1b/0x28
> > >         >         > [1281973.370052]  [<c045e210>] ?
> > >         kthread_create_on_node
> > >         >         +0xc0/0xc0
> > >         >         > [1281973.370052] Code: 5d c3 8d b4 26 00 00 00 00
> > >         89 02 c3
> > >         >         90 8d 74 26 00
> > >         >         > 81 fa ff ff 03 00 89 d1 77 2e 81 fa 00 00 01 00 76
> > >         0e 81 e2
> > >         >         ff ff 00 00 66
> > >         >         > ef <c3> 90 8d 74 26 00 55 ba 2c 5a b2 c0 89 e5 89
> > >         c8 e8 01
> > >         >         ff ff ff
> > >         >         > [1281991.139165] ata2: lost interrupt (Status
> > >         0x58)
> > >         >         > [1281991.148055] ata2: drained 12 bytes to clear
> > >         DRQ
> > >         >         > [1281991.165039] ata2.00: exception Emask 0x0 SAct
> > >         0x0 SErr
> > >         >         0x0 action 0x6
> > >         >         > frozen
> > >         >         > [1281991.172924] sr 1:0:0:0: CDB:
> > >         >         > [1281991.172932] Get event status notification: 4a
> > >         01 00 00
> > >         >         10 00 00 00 08
> > >         >         > 00
> > >         >         > [1281991.497342] ata2.00: cmd
> > >         >         a0/00:00:00:08:00/00:00:00:00:00/a0 tag 0 pio
> > >         >         > 16392 in
> > >         >         > [1281991.497342]          res
> > >         >         40/00:02:00:04:00/00:00:00:00:00/a0 Emask 0x4
> > >         >         > (timeout)
> > >         >         > [1281991.523767] ata2.00: status: { DRDY }
> > >         >         > [1281991.616161] ata2: soft resetting link
> > >         >         > [1281998.232648] ata2.01: qc timeout (cmd 0xec)
> > >         >         > [1281998.238559] ata2.01: failed to IDENTIFY (I/O
> > >         error,
> > >         >         err_mask=0x4)
> > >         >         > [1281998.247432] ata2: soft resetting link
> > >         >         > [1281998.575468] ata2.01: NODEV after polling
> > >         detection
> > >         >         > [1281998.698009] ata2.00: configured for MWDMA2
> > >         >         > [1281998.714460] ata2: EH complete
> > >         >
> > >         >
> > >         >         Not sure what the deal is with that ATA error.
> > >          That's a hard
> > >         >         drive lost
> > >         >         interrupt problem.  Looks to be on your CD Rom
> > >         drive?  Looks
> > >         >         like it
> > >         >         recovered.
> > >         >
> > >         >         > 3. Last but not least after sometime my host
> > >         kernel crashed
> > >         >         as a result
> > >         >         > need to restart the VPC.
> > >         >
> > >         >
> > >         >         I don't understand what you are saying here.  You're
> > >         saying
> > >         >         your kernel
> > >         >         crashed but I don't understand the "as a result
> > >         of..."  What
> > >         >         did you do,
> > >         >         why did you do it, and what happened?
> > >         >
> > >         >         > Regards,
> > >         >         > Ajith
> > >         >
> > >         >         Regards,
> > >         >         Mike
> > >         >
> > >         >         > On Thu, May 16, 2013 at 8:09 PM, Ajith Adapa
> > >         >         <ajith.ad...@gmail.com> wrote:
> > >         >         >
> > >         >         > > Thanks @thomas and @michael.
> > >         >         > >
> > >         >         > > I will try the RPMs and steps provided to start
> > >         a
> > >         >         container.
> > >         >         > >
> > >         >         > > Regards,
> > >         >         > > Ajith
> > >         >         > >
> > >         >         > >
> > >         >         > > On Wed, May 15, 2013 at 2:01 PM, Thomas Moschny
> > >         >         <thomas.mosc...@gmail.com>wrote:
> > >         >         > >
> > >         >         > >> 2013/5/14 Michael H. Warfield
> > >         <m...@wittsend.com>:
> > >         >         > >> > What I would recommend as steps on Fedora
> > >         17...
> > >         >         > >> >
> > >         >         > >> > Download lxc-0.9.0 here:
> > >         >         > >> >
> > >         >         > >> >
> > >         >
> > >         http://lxc.sourceforge.net/download/lxc/lxc-0.9.0.tar.gz
> > >         >         > >> >
> > >         >         > >> > You should have rpm-build and friends
> > >         installed via yum
> > >         >         on your system.
> > >         >         > >> > Build the lxc rpms by running rpmbuild (as
> > >         any user) as
> > >         >         follows:
> > >         >         > >>
> > >         >         > >> You could also try using the pre-built packages
> > >         I put
> > >         >         here:
> > >         >         > >> http://thm.fedorapeople.org/lxc/ .
> > >         >         > >>
> > >         >         > >> Regards,
> > >         >         > >> Thomas
> > >         >         > >>
> > >         >         > >>
> > >         >         > >>
> > >         >
> > >
> > ------------------------------------------------------------------------------
> > >         >         > >> AlienVault Unified Security Management (USM)
> > >         platform
> > >         >         delivers complete
> > >         >         > >> security visibility with the essential security
> > >         >         capabilities. Easily and
> > >         >         > >> efficiently configure, manage, and operate all
> > >         of your
> > >         >         security controls
> > >         >         > >> from a single console and one unified
> > >         framework. Download
> > >         >         a free trial.
> > >         >         > >> http://p.sf.net/sfu/alienvault_d2d
> > >         >         > >> _______________________________________________
> > >         >         > >> Lxc-users mailing list
> > >         >         > >> Lxc-users@lists.sourceforge.net
> > >         >         > >>
> > >         https://lists.sourceforge.net/lists/listinfo/lxc-users
> > >         >         > >>
> > >         >         > >
> > >         >         > >
> > >         >         >
> > >         >         >
> > >         >
> > >
> > ------------------------------------------------------------------------------
> > >         >         > AlienVault Unified Security Management (USM)
> > >         platform
> > >         >         delivers complete
> > >         >         > security visibility with the essential security
> > >         >         capabilities. Easily and
> > >         >         > efficiently configure, manage, and operate all of
> > >         your
> > >         >         security controls
> > >         >         > from a single console and one unified framework.
> > >         Download a
> > >         >         free trial.
> > >         >         > http://p.sf.net/sfu/alienvault_d2d
> > >         >         > _______________________________________________
> > >         Lxc-users
> > >         >         mailing list Lxc-users@lists.sourceforge.net
> > >         >
> > >         https://lists.sourceforge.net/lists/listinfo/lxc-users
> > >         >
> > >         >
> > >         >         --
> > >         >         Michael H. Warfield (AI4NB) | (770) 985-6132 |
> > >         >          m...@wittsend.com
> > >         >            /\/\|=mhw=|\/\/          | (678) 463-0932 |
> > >         >          http://www.wittsend.com/mhw/
> > >         >            NIC whois: MHW9          | An optimist believes
> > >         we live in
> > >         >         the best of all
> > >         >          PGP Key: 0x674627FF        | possible worlds.  A
> > >         pessimist is
> > >         >         sure of it!
> > >         >
> > >         >
> > >         >
> > >         >
> > >
> > >         > --
> > >         > This message has been scanned for viruses and
> > >         > dangerous content by MailScanner, and is
> > >         > believed to be clean.
> > >
> > >         --
> > >         Michael H. Warfield (AI4NB) | (770) 985-6132 |
> > >          m...@wittsend.com
> > >            /\/\|=mhw=|\/\/          | (678) 463-0932 |
> > >          http://www.wittsend.com/mhw/
> > >            NIC whois: MHW9          | An optimist believes we live in
> > >         the best of all
> > >          PGP Key: 0x674627FF        | possible worlds.  A pessimist is
> > >         sure of it!
> > >
> > >
> > >
> > >
> > > --
> > > This message has been scanned for viruses and
> > > dangerous content by MailScanner, and is
> > > believed to be clean.
> >
> > --
> > Michael H. Warfield (AI4NB) | (770) 985-6132 |  m...@wittsend.com
> >    /\/\|=mhw=|\/\/          | (678) 463-0932 |
> > http://www.wittsend.com/mhw/
> >    NIC whois: MHW9          | An optimist believes we live in the best of
> > all
> >  PGP Key: 0x674627FF        | possible worlds.  A pessimist is sure of it!
> >
> >
> > ------------------------------------------------------------------------------
> > AlienVault Unified Security Management (USM) platform delivers complete
> > security visibility with the essential security capabilities. Easily and
> > efficiently configure, manage, and operate all of your security controls
> > from a single console and one unified framework. Download a free trial.
> > http://p.sf.net/sfu/alienvault_d2d
> > _______________________________________________
> > Lxc-users mailing list
> > Lxc-users@lists.sourceforge.net
> > https://lists.sourceforge.net/lists/listinfo/lxc-users
> >
> >
> 
> ------------------------------------------------------------------------------
> Try New Relic Now & We'll Send You this Cool Shirt
> New Relic is the only SaaS-based application performance monitoring service 
> that delivers powerful full stack analytics. Optimize and monitor your
> browser, app, & servers with just a few lines of code. Try New Relic
> and get this awesome Nerd Life shirt! http://p.sf.net/sfu/newrelic_d2d_may
> _______________________________________________
> Lxc-users mailing list
> Lxc-users@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/lxc-users

-- 
Michael H. Warfield (AI4NB) | (770) 985-6132 |  m...@wittsend.com
   /\/\|=mhw=|\/\/          | (678) 463-0932 |  http://www.wittsend.com/mhw/
   NIC whois: MHW9          | An optimist believes we live in the best of all
 PGP Key: 0x674627FF        | possible worlds.  A pessimist is sure of it!


------------------------------------------------------------------------------
Try New Relic Now & We'll Send You this Cool Shirt
New Relic is the only SaaS-based application performance monitoring service 
that delivers powerful full stack analytics. Optimize and monitor your
browser, app, & servers with just a few lines of code. Try New Relic
and get this awesome Nerd Life shirt! http://p.sf.net/sfu/newrelic_d2d_may
_______________________________________________
Lxc-users mailing list
Lxc-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/lxc-users

Reply via email to