Hi Zhu Could you please try out version 1.3.1 and let us know if it works or not? ftp://ftp.parl.clemson.edu/pub/pvfs2/pvfs2-1.3.1.tar.gz Thanks, Murali
On Thu, 1 Dec 2005, zhu min wrote: > hello > > i'm already install the pvfs2-1.3.0 and it's run.but not over 10 > minutes,they all down. > > PS: > > cd /home/tools > > wget ftp://ftp.parl.clemson.edu/pub/pvfs2/pvfs2-1.3.0.tar.gz > > tar zxvf pvfs2-1.3.0.tar.gz > > cd pvfs2-1.3.0 > > ./configure --prefix=/home/pvfs2 --with-kernel24=/usr/src/linux-2.4 > --enable-redhat24 --enable-nptl-workaround > > make > > make install > > make kmod24_install > > cd /home/pvfs2/bin > > ./pvfs2-genconfig /etc/pvfs2-fs.conf /etc/pvfs2-server.conf > > ********************************************************************** > Welcome to the PVFS2 Configuration Generator: > > This interactive script will generate configuration files suitable > for use with a new PVFS2 file system. Please see the PVFS2 quickstart > guide for details. > > ********************************************************************** > > You must first select the network protocol that your file system will use. > The only currently supported options are "tcp" and "gm". > > * Enter protocol type [Default is tcp]: > > Choose a TCP/IP port for the servers to listen on. Note that this > script assumes that all servers will use the same port number. > > * Enter port number [Default is 3334]: > > Next you must list the hostnames of the machines that will act as > I/O servers. Acceptable syntax is "node1, node2, ..." or "node{#-#,#,#}". > > * Enter hostnames [Default is localhost]: cluster{1-5} > > Now list the hostnames of the machines that will act as Metadata > servers. This list may or may not overlap with the I/O server list. > > * Enter hostnames [Default is localhost]: cluster1 > > Configured a total of 8 servers: > 5 of them are I/O servers. > 1 of them are Metadata servers. > > * Would you like to verify server list (y/n) [Default is n]? y > > ****** I/O servers: > tcp://cluster1:3334 > tcp://cluster2:3334 > tcp://cluster3:3334 > tcp://cluster4:3334 > tcp://cluster5:3334 > > ****** Metadata servers: > tcp://cluster1:3334 > > * Does this look ok (y/n) [Default is y]? y > > Choose a file for each server to write log messages to. > > * Enter log file location [Default is /tmp/pvfs2-server.log]: > > Choose a directory for each server to store data in. > > * Enter directory name: [Default is /pvfs2-storage-space]: > > Writing fs config file... Done. > Writing 5 server config file(s)... Done. > > ls /etc > > pvfs2-fs.conf > pvfs2-server.conf-cluster1 > pvfs2-server.conf-cluster2 > pvfs2-server.conf-cluster3 > pvfs2-server.conf-cluster4 > pvfs2-server.conf-cluster5 > > cp /home/tools/pvfs2-1.3.0/examples/pvfs2-server.rc > /etc/rc.d/init.d/pvfs2-server > > chmod a+x /etc/rc.d/init.d/pvfs2-server > > chkconfig pvfs2-server on > > ls -al /etc/rc.d/rc3.d/S35pvfs2-server > > /home/pvfs2/sbin/pvfs2-server /etc/pvfs2-fs.conf > /etc/pvfs2-server.conf-cluster1 -f > > /etc/rc.d/init.d/pvfs2-server start > > mkdir /mnt/pvfs2 > touch /etc/pvfs2tab > chmod a+r /etc/pvfs2tab > vi /etc/pvfs2tab > tcp://cluster1:3334/pvfs2-fs /mnt/pvfs2 pvfs2 default,noauto 0 0 > > #other 4 server's /etc/pvfs2tab is tcp://cluster2:3334/pvfs2-fs /mnt/pvfs2 > pvfs2 default,noauto 0 0;tcp://cluster3:3334/pvfs2-fs /mnt/pvfs2 pvfs2 > default,noauto 0 0; > #tcp://cluster4:3334/pvfs2-fs /mnt/pvfs2 pvfs2 default,noauto 0 > 0;tcp://cluster5:3334/pvfs2-fs /mnt/pvfs2 pvfs2 default,noauto 0 0 > > /home/pvfs2/bin/pvfs2-ping -m /mnt/pvfs2 > > > (1) Parsing tab file... > > (2) Initializing system interface... > > (3) Initializing each file system found in tab file: /etc/pvfs2tab... > > /mnt/pvfs2: Ok > > (4) Searching for /mnt/pvfs2 in pvfstab... > > PVFS2 servers: tcp://cluster1:3334 > Storage name: pvfs2-fs > Local mount point: /mnt/pvfs2 > > meta servers: > tcp://cluster1:3334 > > data servers: > tcp://cluster1:3334 > tcp://cluster2:3334 > tcp://cluster3:3334 > tcp://cluster4:3334 > tcp://cluster5:3334 > > (5) Verifying that all servers are responding... > > meta servers: > tcp://cluster1:3334 Ok > > data servers: > tcp://cluster1:3334 Ok > tcp://cluster2:3334 Ok > tcp://cluster3:3334 Ok > tcp://cluster4:3334 Ok > tcp://cluster5:3334 Ok > > (6) Verifying that fsid 151391925 is acceptable to all servers... > > Ok; all servers understand fs_id 151391925 > > (7) Verifying that root handle is owned by one server... > > Root handle: 1048576 > Ok; root handle is owned by exactly one server. > > ============================================================= > > The PVFS filesystem at /mnt/pvfs2 appears to be correctly configured. > > cd /home/pvfs2/sbin > > ./pvfs2-client -p ./pvfs2-client-core > > insmod /home/tools/pvfs2-1.3.0/src/kernel/linux-2.4/pvfs2.o > > mount -t pvfs2 pvfs2 /mnt/pvfs2 -o tcp://cluster1:3334/pvfs2-fs > > ############################################################################ > > other 4 server's > { > cluster2: > > cd /home/pvfs2/sbin > > ./pvfs2-client -p ./pvfs2-client-core > > insmod /home/tools/pvfs2-1.3.0/src/kernel/linux-2.4/pvfs2.o > > mount -t pvfs2 pvfs2 /mnt/pvfs2 -o tcp://cluster2:3334/pvfs2-fs > > cluster3: > > cd /home/pvfs2/sbin > > ./pvfs2-client -p ./pvfs2-client-core > > insmod /home/tools/pvfs2-1.3.0/src/kernel/linux-2.4/pvfs2.o > > mount -t pvfs2 pvfs2 /mnt/pvfs2 -o tcp://cluster3:3334/pvfs2-fs > > cluster4: > > cd /home/pvfs2/sbin > > ./pvfs2-client -p ./pvfs2-client-core > > insmod /home/tools/pvfs2-1.3.0/src/kernel/linux-2.4/pvfs2.o > > mount -t pvfs2 pvfs2 /mnt/pvfs2 -o tcp://cluster4:3334/pvfs2-fs > > cluster5: > > cd /home/pvfs2/sbin > > ./pvfs2-client -p ./pvfs2-client-core > > insmod /home/tools/pvfs2-1.3.0/src/kernel/linux-2.4/pvfs2.o > > mount -t pvfs2 pvfs2 /mnt/pvfs2 -o tcp://cluster5:3334/pvfs2-fs > } > ############################################################################### > > > cp test.tar.gz /mnt/pvfs2 > > ls /mnt/pvfs2 > > can see the file test.tar.gz,in other 4 host run ls /mnt/pvfs2 also can > see the file. > > next,i run > > tar zxvf test.tar.gz > > it's ok,but A little minutes after,they all down,tar stop,nothing error > messages! > > _________________________________________________________________ > 与联机的朋友进行交流,请使用 MSN Messenger: http://messenger.msn.com/cn > > _______________________________________________ > PVFS2-developers mailing list > [email protected] > http://www.beowulf-underground.org/mailman/listinfo/pvfs2-developers > > _______________________________________________ PVFS2-users mailing list [email protected] http://www.beowulf-underground.org/mailman/listinfo/pvfs2-users
