I am having a bit of difficulty with the user ID namespacing/mapping
used by lxd.
Firstly, let me see if I understand this properly. Reading
https://www.stgraber.org/2014/01/17/lxc-1-0-unprivileged-containers/ it
seems each host user is allocated one mapped range of uids. For example:
$ cat /etc/subuid
lxd:100000:65536
root:100000:65536
Here, both these users get the same mapped range. Furthermore,
presumably every container launched by those users has the *same* mapped
range; that is, root in container foo and root in container bar both map
to 100000 on the host, correct? That sounds like it doesn't give proper
isolation from one container to another, but I'll assume that's not a
problem in practice.
Here's the actual issue I'm dealing with.
I am trying to deploy FreeIPA as a way to do centralised authentication
for servers - and of course, with lots of lxd containers, central
authentication would be extremely helpful!
The trouble is that FreeIPA allocates very high UIDs and GIDs by
default. Example from a fresh install of FreeIPA inside a centos/7/amd64
container:
[root@test ~]# id admin
uid=1134400000(admin) gid=1134400000(admins) groups=1134400000(admins)
[root@test ~]# getent passwd admin
admin:*:1134400000:1134400000:Administrator:/home/admin:/bin/bash
But this id isn't usable within the container, as I find if I try to ssh
to it:
Oct 14 11:01:23 test sshd[3896]: Authorized to admin, krb5 principal
ad...@ipa.example.com (ssh_gssapi_krb5_cmdok)
Oct 14 11:01:23 test sshd[3896]: Accepted gssapi-with-mic for admin from
10.15.6.253 port 49800 ssh2
*Oct 14 11:01:23 test sshd[3896]: fatal: initgroups: admin: Invalid
argument*
Or even just use su within the container:
[root@test ~]# su - admin
*su: cannot set groups: Invalid argument*
So I'm wondering about the best way to deal with this.
(1) I can try to configure FreeIPA to allocate uids in the "low" range,
say 2000+.
(2) I can try to do get lxd to map a larger range, for example map
1134400000 upwards to 2134400000 on the host. But are there problems in
that approach? Is there a reason why the default lxd config only maps
64K worth of uids?
Plus, I imagine this means all existing containers will need their
filesystems changed because of the new mapped IDs?
# ls -l /var/lib/lxd/containers/ldap-1/rootfs/sbin/suexec
-r-x--x--- 1 100000 100048 15352 Jul 18 15:31
/var/lib/lxd/containers/ldap-1/rootfs/sbin/suexec
Thanks,
Brian.
_______________________________________________
lxc-users mailing list
lxc-users@lists.linuxcontainers.org
http://lists.linuxcontainers.org/listinfo/lxc-users