Thanks Abhishek!!! I now have a working xcpu (sxcpu) cluster. Now on to making it useful... :-)
Daniel On Fri, Aug 29, 2008 at 9:18 PM, Abhishek Kulkarni <[EMAIL PROTECTED]> wrote: > Gotcha! > You have xcpu2 installed on your master, whereas the nodes are using the > sxcpu xcpufs. > I would recommend starting over on the master with a complete make uninstall > (wait, i am not sure if the xcpu2 Makefile has a uninstall target or not) In > any case, get the latest sxcpu from the trunk and build, install it. > Alternatively, the simplest would be to get the xcpu rpms from NSA repo at > http://mirrors.versaweb.com/caos/Caos-NSA-1.0/packages/i386/ > > Thanks, > -- Abhishek > > On Fri, Aug 29, 2008 at 6:41 PM, Daniel Gruner <[EMAIL PROTECTED]> wrote: >> >> And here is the output from using the '-d' switch to xrx (it also hangs): >> >> [EMAIL PROTECTED] ~]# xrx -d n0000\!6667 /bin/ls >> <<<-- (0x627000) Tversion tag 65535 msize 32792 version '9P2000.u' >> -->>> (0x627000) Rversion tag 65535 msize 32792 version '9P2000.u' >> <<<-- (0x627000) Tauth tag 0 afid 0 uname root aname nuname 0 >> -->>> (0x627000) Rauth tag 0 qid (0000000000000000 0 'A') >> <<<-- (0x627000) Tread tag 0 fid 0 offset 0 count 4096 >> -->>> (0x627000) Rread tag 0 count 16 data dd342060 80f55a6d 4f1b773b >> 4a944c66 >> >> <<<-- (0x627000) Tread tag 0 fid 0 offset 16 count 4080 >> -->>> (0x627000) Rread tag 0 count 0 data >> >> <<<-- (0x627000) Twrite tag 0 fid 0 offset 0 count 256 data 239736bc >> 9d0ba55d ea0f8f7b 51180217 55eacab4 b4e58ec6 fcbdbf80 50327775 >> 64143acf f596be07 9729cddb a681c51c fac2ede0 5f28de13 7b23042a acf32bcd >> >> >> -->>> (0x627000) Rwrite tag 0 count 256 >> <<<-- (0x627000) Tattach tag 0 fid 1 afid 0 uname root aname nuname 0 >> -->>> (0x627000) Rattach tag 0 qid (0000000000000001 0 'd') >> connection 0x627000 to 10.10.0.10 opened >> <<<-- (0x627000) Twalk tag 0 fid 1 newfid 2 nwname 1 'arch' >> -->>> (0x627000) Rwalk tag 0 nwqid 1 (0000000000000004 0 '') >> <<<-- (0x627000) Topen tag 0 fid 2 mode 0 >> -->>> (0x627000) Ropen tag 0 (0000000000000004 0 '') iounit 0 >> <<<-- (0x627000) Tread tag 0 fid 2 offset 0 count 63 >> -->>> (0x627000) Rread tag 0 count 13 data 2f4c696e 75782f78 38365f36 34 >> <<<-- (0x627000) Tread tag 0 fid 2 offset 13 count 50 >> -->>> (0x627000) Rread tag 0 count 0 data >> >> <<<-- (0x627000) Tclunk tag 0 fid 2 >> -->>> (0x627000) Rclunk tag 0 >> <<<-- (0x627000) Twalk tag 0 fid 1 newfid 2 nwname 1 'clone' >> -->>> (0x627000) Rwalk tag 0 nwqid 1 (0000000000000002 0 '') >> <<<-- (0x627000) Topen tag 0 fid 2 mode 0 >> -->>> (0x627000) Ropen tag 0 (0000000000000002 0 '') iounit 0 >> <<<-- (0x627000) Tread tag 0 fid 2 offset 0 count 63 >> -->>> (0x627000) Rread tag 0 count 1 data 38 >> >> <<<-- (0x627000) Tread tag 0 fid 2 offset 1 count 62 >> -->>> (0x627000) Rread tag 0 count 0 data >> >> <<<-- (0x627000) Twalk tag 0 fid 1 newfid 3 nwname 2 '8' 'ctl' >> -->>> (0x627000) Rwalk tag 0 nwqid 2 (0000000009000000 0 'd') >> (0000000009000001 0 '') >> <<<-- (0x627000) Topen tag 0 fid 3 mode 1 >> -->>> (0x627000) Ropen tag 0 (0000000009000001 0 '') iounit 0 >> <<<-- (0x627000) Twalk tag 0 fid 1 newfid 4 nwname 2 '8' 'wait' >> -->>> (0x627000) Rwalk tag 0 nwqid 2 (0000000009000000 0 'd') >> (0000000009000009 0 '') >> <<<-- (0x627000) Topen tag 0 fid 4 mode 0 >> -->>> (0x627000) Ropen tag 0 (0000000009000009 0 '') iounit 0 >> <<<-- (0x627000) Twalk tag 0 fid 1 newfid 5 nwname 2 '8' 'stdin' >> -->>> (0x627000) Rwalk tag 0 nwqid 2 (0000000009000000 0 'd') >> (0000000009000005 0 '') >> <<<-- (0x627000) Topen tag 0 fid 5 mode 1 >> -->>> (0x627000) Ropen tag 0 (0000000009000005 0 '') iounit 0 >> <<<-- (0x627000) Twalk tag 0 fid 1 newfid 6 nwname 2 '8' 'stdout' >> -->>> (0x627000) Rwalk tag 0 nwqid 2 (0000000009000000 0 'd') >> (0000000009000006 0 '') >> <<<-- (0x627000) Topen tag 0 fid 6 mode 0 >> -->>> (0x627000) Ropen tag 0 (0000000009000006 0 '') iounit 0 >> <<<-- (0x627000) Twalk tag 0 fid 1 newfid 7 nwname 2 '8' 'stderr' >> -->>> (0x627000) Rwalk tag 0 nwqid 2 (0000000009000000 0 'd') >> (0000000009000007 0 '') >> <<<-- (0x627000) Topen tag 0 fid 7 mode 0 >> -->>> (0x627000) Ropen tag 0 (0000000009000007 0 '') iounit 0 >> <<<-- (0x627000) Tclunk tag 0 fid 2 >> -->>> (0x627000) Rclunk tag 0 >> <<<-- (0x627000) Twalk tag 0 fid 1 newfid 2 nwname 2 '8' 'env' >> -->>> (0x627000) Rwalk tag 0 nwqid 2 (0000000009000000 0 'd') >> (0000000009000004 0 '') >> <<<-- (0x627000) Topen tag 0 fid 2 mode 129 >> -->>> (0x627000) Ropen tag 0 (0000000009000004 0 '') iounit 0 >> <<<-- (0x627000) Twalk tag 0 fid 1 newfid 8 nwname 2 '8' 'argv' >> -->>> (0x627000) Rwalk tag 0 nwqid 2 (0000000009000000 0 'd') >> (0000000009000003 0 '') >> <<<-- (0x627000) Topen tag 0 fid 8 mode 17 >> -->>> (0x627000) Ropen tag 0 (0000000009000003 0 '') iounit 0 >> <<<-- (0x627000) Twalk tag 0 fid 1 newfid 9 nwname 2 '8' 'ns' >> -->>> (0x627000) Rwalk tag 0 nwqid 1 (0000000009000000 0 'd') >> <<<-- (0x627000) Tclunk tag 0 fid 9 >> -->>> (0x627000) Rclunk tag 0 >> >> >> On Fri, Aug 29, 2008 at 8:29 PM, Daniel Gruner <[EMAIL PROTECTED]> wrote: >> > Yes, I used the xrx command as normal. The back-slash is to escape >> > the '!', which the shell >> > (bash) interprets. That is not the problem. I will try to do the >> > debug bit and/or strace, and post >> > it for you guys to decipher... >> > >> > Daniel >> > >> > On Fri, Aug 29, 2008 at 6:11 PM, Abhishek Kulkarni <[EMAIL PROTECTED]> >> > wrote: >> >> >> >> >> >> On Fri, Aug 29, 2008 at 12:45 PM, Daniel Gruner <[EMAIL PROTECTED]> >> >> wrote: >> >>> >> >>> On 8/29/08, Abhishek Kulkarni <[EMAIL PROTECTED]> wrote: >> >>> > >> >>> > >> >>> > >> >>> > On Fri, Aug 29, 2008 at 10:13 AM, Daniel Gruner <[EMAIL PROTECTED]> >> >>> > wrote: >> >>> > > >> >>> > > Hi Ab >> >>> > > >> >>> > > >> >>> > > On 8/29/08, Abhishek Kulkarni <[EMAIL PROTECTED]> wrote: >> >>> > > > Hi Daniel, >> >>> > > > >> >>> > > > Understand the way in which XCPU is supposed to integrate with >> >>> > > > oneSIS >> >>> > and/or >> >>> > > > Perceus. It uses these as a "launch vehicle" to build minimal >> >>> > > > images >> >>> > with >> >>> > > > xcpufs running on them, and provision the nodes with these >> >>> > > > images. >> >>> > > > In >> >>> > the >> >>> > > > best case, that's all that you need to be running on the compute >> >>> > > > nodes. >> >>> > > >> >>> > > I understand. >> >>> > > >> >>> > > >> >>> > > > >> >>> > > > On Fri, Aug 29, 2008 at 8:46 AM, Daniel Gruner >> >>> > > > <[EMAIL PROTECTED]> >> >>> > wrote: >> >>> > > > > >> >>> > > > > Hi Greg, >> >>> > > > > >> >>> > > > > I definitely have additional questions! :-) >> >>> > > > > >> >>> > > > > Ok, here we go: >> >>> > > > > >> >>> > > > > - assume I am totally new to this - what would one do in order >> >>> > > > > to >> >>> > > > > set >> >>> > > > > up a perceus/xcpu cluster? >> >>> > > > >> >>> > > > As Greg said, you have two ways to go about it. You could choose >> >>> > > > either >> >>> > of >> >>> > > > them or try both to see what works for ya. It's just a matter of >> >>> > > > playing >> >>> > > > with different configurations and rebooting your nodes to try >> >>> > > > them. >> >>> > > > >> >>> > > > > >> >>> > > > > >> >>> > > > > - now, I am not totally new to this game, and my background is >> >>> > > > > with >> >>> > > > > bproc clusters, so I would like to have a replacement for >> >>> > > > > these, >> >>> > > > > but >> >>> > > > > with the same basic principle of having a minimal node >> >>> > > > > installation, >> >>> > > > > and basically no management of nodes needed. I definitely do >> >>> > > > > not >> >>> > > > > want >> >>> > > > > to go to a model where the nodes have password files, and you >> >>> > > > > ssh >> >>> > > > > into >> >>> > > > > them in order to run your codes. >> >>> > > > > >> >>> > > > > - in the caos-NSA installation, the warewulfd is started by >> >>> > > > > default. >> >>> > > > > I assume it needs to be stopped and perceus started, correct? >> >>> > > > >> >>> > > > You can enable Perceus from "sidekick" in NSA. Warewulf focuses >> >>> > > > on >> >>> > cluster >> >>> > > > monitoring starting with 3.0. >> >>> > > >> >>> > > Ok, I am concentrating on my RHEL5 machine for now. It seems to >> >>> > > be >> >>> > > working, at least insofar as the nodes boot. I haven't been able >> >>> > > to >> >>> > > contact them to try to do anything, other than running xstat with >> >>> > > a >> >>> > > positive response: >> >>> > > >> >>> > > n0000 tcp!10.10.0.170!6667 /Linux/x86_64 up 0 >> >>> > > n0001 tcp!10.10.0.185!6667 /Linux/x86_64 up 0 >> >>> > > >> >>> > > I'd like the nodes to get sequential IP addresses, for ease of >> >>> > > identification and management, and I have yet to find out how you >> >>> > > do >> >>> > > that in perceus. >> >>> > >> >>> > Take a look at the ipaddr module in Perceus. >> >>> >> >>> Thanks for the pointer. I am looking at it, but the manual leaves a >> >>> lot to be desired in terms of describing what each module does and how >> >>> to configure them. I'll try the perceus list if I keep getting stuck >> >>> on this. >> >>> >> >>> > >> >>> > > >> >>> > > >> >>> > > Now, when I try to do anything on the nodes I get, for example: >> >>> > > >> >>> > > xgroupset 10.10.0.170 root 0 >> >>> > > xgroupset: Error: Connection refused:10.10.0.170 >> >>> > >> >>> > Whoops! What about telnet 10.10.0.170 6667? >> >>> > Perceus might possibly be running xcpufs on some non-standard port. >> >>> > I'm >> >>> > not >> >>> > sure about that but I remember seeing something like that a while >> >>> > back. >> >>> >> >>> You seem to have hit it! I can in fact telnet using port 6667 >> >>> explicitly (can't do anything while in there...:-). I thought that >> >>> was the default port anyway, correct? >> >>> >> >>> > >> >>> > > >> >>> > > >> >>> > > similarly with xrx. >> >>> > > >> >>> > > xrx 10.10.0.170 /bin/date >> >>> > > Error: Connection refused:10.10.0.170 >> >>> > >> >>> > Ditto with this, if it's running on a different port you would want >> >>> > to >> >>> > do >> >>> > xrx 10.10.0.170!port /bin/date >> >>> > >> >>> > Alternatively you could specify the "-a" flag to retrieve the nodes >> >>> > from >> >>> > the >> >>> > statfs. >> >>> >> >>> I can get xgroupset and xuserset to work with the -a flag, with no >> >>> complaints. However, when I try to run anything on the nodes using >> >>> xrx, whether I use the -a flag or explicitly set the port (xrx >> >>> 10.10.0.170\!6667 /bin/date) the command just hangs. I realize this >> >>> is some progress, but no cigar yet. >> >> >> >> Weird. Did you use the above xrx command as-is? If yes, notice the >> >> back-slash in your command. >> >> I have not ever seen xrx hang that way. Try passing the "-d" switch >> >> alongwith -a, to enable the debug mode. >> >> If not, do the strace and post the output. It's not that tough. Really. >> >> It >> >> should just work. >> >> >> >>> >> >>> > >> >>> > > >> >>> > > >> >>> > > I also don't get name resolution for the nXXXX names assigned to >> >>> > > the >> >>> > > nodes by perceus. >> >>> > > >> >>> > > >> >>> > >> >>> > Check your /etc/resolv.conf. >> >>> > Probably try adding the following to it. >> >>> > nameserver 127.0.0.1 >> >>> > >> >>> > If that doesn't work, the right place to ask this would be the >> >>> > Perceus >> >>> > ML. >> >>> >> >>> Doesn't work. I'll try the perceus gurus... >> >>> >> >>> > >> >>> > > >> >>> > > >> >>> > > >> >>> > > > >> >>> > > > > >> >>> > > > > >> >>> > > > > - what initialization of perceus needs to be done (the first >> >>> > > > > time >> >>> > > > > it >> >>> > > > > runs)? I know about the network interface specification, and >> >>> > > > > that >> >>> > > > > I >> >>> > > > > want it to use xget (the default), but is running the "perceus >> >>> > > > > module >> >>> > > > > activate xcpu" enough to get the nodes booting into xcpu? >> >>> > > > >> >>> > > > Yes, it is enough to get xcpufs running on the compute nodes. >> >>> > > > >> >>> > > > > >> >>> > > > > >> >>> > > > > - what about configuring the resource manager (e.g. slurm) for >> >>> > > > > use >> >>> > > > > in >> >>> > > > > the perceus/xcpu environment? >> >>> > > > >> >>> > > > XCPU only supports Moab Torque for now. >> >>> > > >> >>> > > Is this the open source torque, or just the commercial product? >> >>> >> >>> >> >>> Who would know about which version of Torque this is? I can't afford >> >>> the commercial Moab right now... >> >>> >> >>> >> >>> > > >> >>> > > >> >>> > > >> >>> > > >> >>> > > > >> >>> > > > > >> >>> > > > > >> >>> > > > > - I don't see the xcpufs and statfs daemons running on the >> >>> > > > > master >> >>> > > > > after starting perceus even though I told it to activate xcpu. >> >>> > > > > I >> >>> > > > > haven't tried to boot nodes yet, but I'd like to understand >> >>> > > > > what I >> >>> > > > > am >> >>> > > > > doing first (I hate black boxes...). >> >>> > > > > >> >>> > > > >> >>> > > > You shouldn't need to run xcpufs on the master. As for statfs, >> >>> > > > you >> >>> > > > can >> >>> > start >> >>> > > > it manually if it is not running already. >> >>> > > > >> >>> > > > Again, considering that you have fully configured the master and >> >>> > > > have >> >>> > the >> >>> > > > nodes provisioned to the init state, this is what I would do to >> >>> > > > generate >> >>> > my >> >>> > > > statfs.conf -- >> >>> > > > >> >>> > > > perceus node status | awk 'NR > 2 {print $1 "=tcp!" $3 "!6667"}' >> >>> > > > > >> >>> > > > /etc/xcpu/statfs.conf >> >>> > > >> >>> > > I had to replace the part "NR>2" with "NR>0" for the above >> >>> > > incantation >> >>> > > to work (??). >> >>> > >> >>> > Strange, I might probably be running a different version of Perceus. >> >>> >> >>> Actually, what happens is that the first two lines of output from the >> >>> "perceus node status" command are output to stderr, and the rest to >> >>> stdout. If your '|' redirection included stderr then the command as >> >>> you wrote it would work. What shell are you using? >> >> >> >> Ahh, maybe. I'm using bash. I didn't realize that. >> >> >> >> Thanks, >> >> -- Abhishek >> >> >> >>> >> >>> Thanks, >> >>> >> >>> Daniel >> >>> >> >>> >> >>> >> >>> > >> >>> > > >> >>> > > >> >>> > > >> >>> > > > >> >>> > > > And then, >> >>> > > > >> >>> > > > statfs -c /etc/xcpu/statfs >> >>> > > >> >>> > > statfs seems to work. Here is the output from xstat: >> >>> > > >> >>> > > n0000 tcp!10.10.0.170!6667 /Linux/x86_64 up 0 >> >>> > > n0001 tcp!10.10.0.185!6667 /Linux/x86_64 up 0 >> >>> > > >> >>> > > In any case, there is some progress, but it is not quite there >> >>> > > yet... >> >>> > >> >>> > I'm glad you are almost there. >> >>> > >> >>> > Thanks, >> >>> > -- Abhishek >> >>> > >> >>> > >> >>> > > >> >>> > > >> >>> > > Thanks, >> >>> > > Daniel >> >>> > > >> >>> > > >> >>> > > >> >>> > > >> >>> > > >> >>> > > >> >>> > > >> >>> > > >> >>> > > >> >>> > > > >> >>> > > > >> >>> > > > > >> >>> > > > > etc. >> >>> > > > > >> >>> > > > > I guess the main problem I have is not with perceus itself (I >> >>> > > > > have >> >>> > > > > read the manual), but rather with its integration and >> >>> > > > > provisioning >> >>> > > > > for >> >>> > > > > xcpu, and for the subsequent configuration of those pieces >> >>> > > > > that >> >>> > > > > make >> >>> > > > > the cluster useable in a production environment. >> >>> > > > > >> >>> > > > > >> >>> > > > > Thanks for your help, >> >>> > > > > Daniel >> >>> > > > >> >>> > > > Thanks >> >>> > > > -- Abhishek >> >>> > > > >> >>> > > > > >> >>> > > > > >> >>> > > > > >> >>> > > > > >> >>> > > > > >> >>> > > > > >> >>> > > > > On 8/29/08, Greg Kurtzer <[EMAIL PROTECTED]> wrote: >> >>> > > > > > >> >>> > > > > > You have multiple choices on how to move forward. >> >>> > > > > > >> >>> > > > > > First you can run the xcpu Perceus module like: >> >>> > > > > > >> >>> > > > > > # perceus module activate xcpu >> >>> > > > > > >> >>> > > > > > That will interrupt the node provisioning process and >> >>> > > > > > instead >> >>> > > > > > of >> >>> > > > > > copying the VNFS to the node it will just start up xcpu and >> >>> > > > > > start >> >>> > > > > > accepting connections. >> >>> > > > > > >> >>> > > > > > The second option would be to run xcpu from within the VNFS >> >>> > > > > > of >> >>> > > > > > your >> >>> > > > > > choice. That mechanism basically involves installing xcpu >> >>> > > > > > into >> >>> > > > > > the >> >>> > > > > > mounted VNFS image and then provision your nodes with that. >> >>> > > > > > >> >>> > > > > > Let me know if that helps or if you have additional >> >>> > > > > > questions. >> >>> > > > > > :) >> >>> > > > > > >> >>> > > > > > >> >>> > > > > > Greg >> >>> > > > > > >> >>> > > > > > >> >>> > > > > > >> >>> > > > > > >> >>> > > > > > On Fri, Aug 29, 2008 at 6:45 AM, Daniel Gruner >> >>> > > > > > <[EMAIL PROTECTED]> >> >>> > > > wrote: >> >>> > > > > > > >> >>> > > > > > > Hi Kevin, >> >>> > > > > > > >> >>> > > > > > > Well, I've just completed installing xcpu2 and perceus >> >>> > > > > > into >> >>> > > > > > my >> >>> > RHEL5 >> >>> > > > > > > machine, but now I am stumped with the configuration. >> >>> > > > > > How do >> >>> > > > > > you >> >>> > > > tell >> >>> > > > > > > perceus that you want your cluster to run xcpu? I sure >> >>> > > > > > don't >> >>> > > > > > > understand where this is configured (I assume somewhere >> >>> > > > > > in >> >>> > > > > > the >> >>> > > > > > > /etc/perceus .conf files), and there is no mention of >> >>> > > > > > that in >> >>> > > > > > the >> >>> > > > > > > manual other than saying that xcpu works. >> >>> > > > > > > >> >>> > > > > > > If you install xcpu2 you surely would need 9p, right? >> >>> > > > > > > >> >>> > > > > > > Also, how does slurm integrate into the perceus/xcpu >> >>> > > > > > world? >> >>> > > > > > > >> >>> > > > > > > I have also installed this on a caos-NSA test machine, >> >>> > > > > > but >> >>> > > > > > again >> >>> > I >> >>> > > > > > > don't know how to configure the provisioning. >> >>> > > > > > > >> >>> > > > > > > Any help with this would be much appreciated... >> >>> > > > > > > >> >>> > > > > > > Daniel >> >>> > > > > > > >> >>> > > > > > > >> >>> > > > > > > On 8/28/08, Kevin Tegtmeier <[EMAIL PROTECTED]> wrote: >> >>> > > > > > >> We used RHEL5 + perceus successfully. I had to modify >> >>> > > > > > the >> >>> > perceus >> >>> > > > boot >> >>> > > > > > >> image for x86_64, but it may have been a kexec/hardware >> >>> > > > > > specific >> >>> > > > issue I ran >> >>> > > > > > >> into. If you run into an issue with it I can help you >> >>> > > > > > along. >> >>> > > > > > >> >> >>> > > > > > >> I don't think the 9P module was built in, but I don't >> >>> > > > > > think >> >>> > > > > > you >> >>> > > > would use >> >>> > > > > > >> it. >> >>> > > > > > >> >> >>> > > > > > >> >> >>> > > > > > >> On Thu, Aug 28, 2008 at 11:31 AM, Daniel Gruner >> >>> > <[EMAIL PROTECTED]> >> >>> > > > wrote: >> >>> > > > > > >> >> >>> > > > > > >> > >> >>> > > > > > >> > Thanks, Abhishek. >> >>> > > > > > >> > >> >>> > > > > > >> > I will try it and report on my success/lack thereof. >> >>> > > > > > >> > >> >>> > > > > > >> > Just for info, I am using a RHEL5 distribution, but >> >>> > > > > > with >> >>> > > > > > the >> >>> > > > 2.6.26 >> >>> > > > > > >> > kernel so that it supports 9p. Has anybody been >> >>> > > > > > successful >> >>> > with >> >>> > > > this >> >>> > > > > > >> > distribution? Otherwise, is there a preferred one? >> >>> > > > > > >> > >> >>> > > > > > >> > Daniel >> >>> > > > > > >> > >> >>> > > > > > >> > >> >>> > > > > > >> > >> >>> > > > > > >> > >> >>> > > > > > >> > On 8/28/08, Abhishek Kulkarni <[EMAIL PROTECTED]> >> >>> > > > > > wrote: >> >>> > > > > > >> > > >> >>> > > > > > >> > > Daniel, >> >>> > > > > > >> > > >> >>> > > > > > >> > > It is _not_ necessary to install cAos Linux to use >> >>> > > > > > Perceus. >> >>> > > > Perceus >> >>> > > > > > >> > > supports most, if not all, distributions. >> >>> > > > > > >> > > >> >>> > > > > > >> > > XCPU is bundled up as a module within Perceus. The >> >>> > > > documentation at >> >>> > > > > > >> > > >> >>> > > > > > >> >> >>> > > > http://www.perceus.org/docs/perceus-userguide-1.4.0.pdf >> >>> > is >> >>> > > > > > >> quite >> >>> > > > > > >> > > extensive at that and has details on importing and >> >>> > activating >> >>> > > > modules. >> >>> > > > > > >> > > It's quite simple even if you find yourself wanting >> >>> > > > > > to >> >>> > tinker >> >>> > > > with the >> >>> > > > > > >> > > XCPU Perceus module (it's just a shell script that >> >>> > > > > > runs >> >>> > > > > > at >> >>> > a >> >>> > > > specified >> >>> > > > > > >> > > provisioning state/level) >> >>> > > > > > >> > > >> >>> > > > > > >> > > >> >>> > > > > > >> > > -- Abhishek >> >>> > > > > > >> > > >> >>> > > > > > >> > > >> >>> > > > > > >> > > On Thu, 2008-08-28 at 14:17 -0400, Daniel Gruner >> >>> > > > > > wrote: >> >>> > > > > > >> > > > Yes, that is a possibility. Instructions on >> >>> > > > > > that, >> >>> > please? >> >>> > > > > > >> > > > I tried installing caos linux, but it doesn't >> >>> > > > > > quite >> >>> > finish >> >>> > > > doing the >> >>> > > > > > >> install. >> >>> > > > > > >> > > > >> >>> > > > > > >> > > > Daniel >> >>> > > > > > >> > > > >> >>> > > > > > >> > > > On 8/28/08, ron minnich <[EMAIL PROTECTED]> >> >>> > > > > > wrote: >> >>> > > > > > >> > > > > >> >>> > > > > > >> > > > > Use perceus. >> >>> > > > > > >> > > > > >> >>> > > > > > >> > > > > Ron >> >>> > > > > > >> > > > > >> >>> > > > > > >> > > > > >> >>> > > > > > >> > > > > On 8/28/08, Daniel Gruner <[EMAIL PROTECTED]> >> >>> > > > > > wrote: >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > Hi All, >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > The list has been very quiet lately... :-) >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > I've been trying, yet again, to install the >> >>> > > > > > latest >> >>> > xcpu2 >> >>> > > > in a >> >>> > > > > > >> test >> >>> > > > > > >> > > > > > cluster. Ron's instructions on the xcpu.org >> >>> > > > > > site >> >>> > seem >> >>> > > > to be >> >>> > > > > > >> outdated, >> >>> > > > > > >> > > > > > and partly buggy too. For instance, here >> >>> > > > > > are a >> >>> > couple >> >>> > > > of >> >>> > > > > > >> points: >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > - After doing: >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > make xcpu-tarball >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > make ramfs-tarball >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > make install >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > I don't know whether xcpu2 has actually been >> >>> > > > > > built >> >>> > (I >> >>> > > > suspect >> >>> > > > > > >> not), >> >>> > > > > > >> > > > > > and it certainly has not been installed >> >>> > > > > > (e.g. no >> >>> > xrx, or >> >>> > > > xcpufs, >> >>> > > > > > >> or >> >>> > > > > > >> > > > > > any of that stuff has been installed). >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > - The command >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > export u=`uname -r` >> >>> > > > > > >> > > > > > ./mk-initramfs-oneSIS -f initrd-$u.img $u >> >>> > > > > > -nn >> >>> > > > > > -rr \ >> >>> > > > > > >> > > > > > -o ../overlays/xcpu-64 \ >> >>> > > > > > >> > > > > > -w e1000 \ >> >>> > > > > > >> > > > > > -w forcedeth \ >> >>> > > > > > >> > > > > > -w ext3 >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > should really be >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > ./mk-xcpu-oneSIS .... >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > in order that the 9p and 9pnet modules get >> >>> > > > > > loaded >> >>> > into >> >>> > > > the >> >>> > > > > > >> initrd. >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > Can someone please take a look and revise >> >>> > > > > > the >> >>> > > > instructions (and >> >>> > > > > > >> let us >> >>> > > > > > >> > > > > > mere mortals know what to do)? >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > Furthermore, is xcpu2 actualy useable for >> >>> > > > > > production >> >>> > > > work? What >> >>> > > > > > >> about >> >>> > > > > > >> > > > > > its integration with a scheduler/resource >> >>> > > > > > manager? >> >>> > What >> >>> > > > about >> >>> > > > > > >> MPI? >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > > Regards, >> >>> > > > > > >> > > > > > Daniel >> >>> > > > > > >> > > > > > >> >>> > > > > > >> > > > > >> >>> > > > > > >> > > > > >> >>> > > > > > >> > > > > -- >> >>> > > > > > >> > > > > Sent from Gmail for mobile | mobile.google.com >> >>> > > > > > >> > > > > >> >>> > > > > > >> > > >> >>> > > > > > >> > > >> >>> > > > > > >> > >> >>> > > > > > >> >> >>> > > > > > >> >> >>> > > > > > > >> >>> > > > > > >> >>> > > > > > >> >>> > > > > > >> >>> > > > > > >> >>> > > > > > -- >> >>> > > > > > Greg Kurtzer >> >>> > > > > > http://www.infiscale.com/ >> >>> > > > > > http://www.runlevelzero.net/ >> >>> > > > > > http://www.perceus.org/ >> >>> > > > > > http://www.caoslinux.org/ >> >>> > > > > > >> >>> > > > > >> >>> > > > >> >>> > > > >> >>> > > >> >>> > >> >>> > >> >> >> >> >> > > >
