Hello Phil, Thanks for you answer. Yes I delete the storage dir every time I make a new configuration and I run the pvfs2-server -f command before starting the daemons.
The only thing that I get from the servers is the batch_create, starting server, and the "PVFS2 server got signal 15 (server_status_flag: 507903" error message. Do you want me to try on an other log level? Also, this is how the server is configured: ***** Displaying PVFS Configuration Information ***** ------------------------------------------------------ PVFS2 configured to build karma gui : no PVFS2 configured to perform coverage analysis : no PVFS2 configured for aio threaded callbacks : yes PVFS2 configured to use FUSE : no PVFS2 configured for the 2.6.x kernel module : no PVFS2 configured for the 2.4.x kernel module : no PVFS2 configured for using the mmap-ra-cache : no PVFS2 will use workaround for redhat 2.4 kernels : no PVFS2 will use workaround for buggy NPTL : no PVFS2 server will be built : yes PVFS2 version string: 2.8.1 thanks again, On Mon, Apr 6, 2009 at 5:21 PM, Phil Carns <[email protected]> wrote: > Hello, > > I'm not sure what would cause that "Invalid argument" error. > > Could you try the following steps: > > - kill both servers > - modify your configuration files to set "EventLogging" to "none" > - delete your old log files (or move them to another directory) > - start the servers > > You can then send us the complete contents of both log files and we can go > from there. The "all" level is a little hard to interpret because it > generates a lot of information even when servers are idle. > > Also, when you went from one server to two, did you delete your old storage > space (/pvfs) and start over, or are you trying to keep that data and add > servers to it? > > thanks! > -Phil > > Asterios Katsifodimos wrote: > >> Hello all, >> >> I have been trying to install PVFS 2.8.1 on Ubuntu server, Centos4 and >> Scientific Linux 4. I compile it and can run it on a "single host" >> configuration >> without any problems. >> >> However, when I add more nodes to the configuration(always using the >> pvfs2-genconfig defaults ) I have the following problem: >> >> *On the metadata node I get these messages:* >> [E 04/02 20:16] batch_create request got: Invalid argument >> [E 04/02 20:16] batch_create request got: Invalid argument >> [E 04/02 20:16] batch_create request got: Invalid argument >> [E 04/02 20:16] batch_create request got: Invalid argument >> >> >> *In the IO nodes I get:* >> [r...@wn140 ~]# tail -50 /tmp/pvfs2-server.log >> [D 04/02 23:53] BMI_testcontext completing: 18446744072456767880 >> [D 04/02 23:53] [SM Entering]: (0x88f8b00) msgpairarray_sm:complete >> (status: 1) >> [D 04/02 23:53] [SM frame get]: (0x88f8b00) op-id: 37 index: 0 base-frm: 1 >> [D 04/02 23:53] msgpairarray_complete: sm 0x88f8b00 status_user_tag 1 >> msgarray_count 1 >> [D 04/02 23:53] msgpairarray: 1 operations remain >> [D 04/02 23:53] [SM Exiting]: (0x88f8b00) msgpairarray_sm:complete (error >> code: -1073742006), (action: DEFERRED) >> [D 04/02 23:53] [SM Entering]: (0x88f8b00) msgpairarray_sm:complete >> (status: 0) >> [D 04/02 23:53] [SM frame get]: (0x88f8b00) op-id: 37 index: 0 base-frm: 1 >> [D 04/02 23:53] msgpairarray_complete: sm 0x88f8b00 status_user_tag 0 >> msgarray_count 1 >> [D 04/02 23:53] msgpairarray: all operations complete >> [D 04/02 23:53] [SM Exiting]: (0x88f8b00) msgpairarray_sm:complete (error >> code: 190), (action: COMPLETE) >> [D 04/02 23:53] [SM Entering]: (0x88f8b00) msgpairarray_sm:completion_fn >> (status: 0) >> [D 04/02 23:53] [SM frame get]: (0x88f8b00) op-id: 37 index: 0 base-frm: 1 >> [D 04/02 23:53] (0x88f8b00) msgpairarray state: completion_fn >> [E 04/02 23:53] Warning: msgpair failed to tcp://wn141:3334, will retry: >> Connection refused >> [D 04/02 23:53] *** msgpairarray_completion_fn: msgpair 0 failed, retry 1 >> [D 04/02 23:53] *** msgpairarray_completion_fn: msgpair retrying after >> delay. >> [D 04/02 23:53] [SM Exiting]: (0x88f8b00) msgpairarray_sm:completion_fn >> (error code: 191), (action: COMPLETE) >> [D 04/02 23:53] [SM Entering]: (0x88f8b00) msgpairarray_sm:post_retry >> (status: 0) >> [D 04/02 23:53] [SM frame get]: (0x88f8b00) op-id: 37 index: 0 base-frm: 1 >> [D 04/02 23:53] msgpairarray_post_retry: sm 0x88f8b00, wait 2000 ms >> [D 04/02 23:53] [SM Exiting]: (0x88f8b00) msgpairarray_sm:post_retry >> (error code: 0), (action: DEFERRED) >> [D 04/02 23:53] [SM Entering]: (0x89476c0) perf_update_sm:do_work (status: >> 0) >> [P 04/02 23:53] Start times (hr:min:sec): 23:53:11.330 23:53:10.310 >> 23:53:09.287 23:53:08.268 23:53:07.245 23:53:06.225 >> [P 04/02 23:53] Intervals (hr:min:sec) : 00:00:01.026 00:00:01.020 >> 00:00:01.023 00:00:01.019 00:00:01.023 00:00:01.020 >> [P 04/02 23:53] >> ------------------------------------------------------------------------------------------------------------- >> [P 04/02 23:53] bytes read : 0 0 >> 0 0 0 0 >> [P 04/02 23:53] bytes written : 0 0 >> 0 0 0 0 >> [P 04/02 23:53] metadata reads : 0 0 >> 0 0 0 0 >> [P 04/02 23:53] metadata writes : 0 0 >> 0 0 0 0 >> [P 04/02 23:53] metadata dspace ops : 0 0 >> 0 0 0 0 >> [P 04/02 23:53] metadata keyval ops : 1 1 >> 1 1 1 1 >> [P 04/02 23:53] request scheduler : 0 0 >> 0 0 0 0 >> [D 04/02 23:53] [SM Exiting]: (0x89476c0) perf_update_sm:do_work (error >> code: 0), (action: DEFERRED) >> [D 04/02 23:53] [SM Entering]: (0x8948810) job_timer_sm:do_work (status: >> 0) >> [D 04/02 23:53] [SM Exiting]: (0x8948810) job_timer_sm:do_work (error >> code: 0), (action: DEFERRED) >> [D 04/02 23:53] [SM Entering]: (0x89476c0) perf_update_sm:do_work (status: >> 0) >> [P 04/02 23:53] Start times (hr:min:sec): 23:53:12.356 23:53:11.330 >> 23:53:10.310 23:53:09.287 23:53:08.268 23:53:07.245 >> [P 04/02 23:53] Intervals (hr:min:sec) : 00:00:01.020 00:00:01.026 >> 00:00:01.020 00:00:01.023 00:00:01.019 00:00:01.023 >> [P 04/02 23:53] >> ------------------------------------------------------------------------------------------------------------- >> [P 04/02 23:53] bytes read : 0 0 >> 0 0 0 0 >> [P 04/02 23:53] bytes written : 0 0 >> 0 0 0 0 >> [P 04/02 23:53] metadata reads : 0 0 >> 0 0 0 0 >> [P 04/02 23:53] metadata writes : 0 0 >> 0 0 0 0 >> [P 04/02 23:53] metadata dspace ops : 0 0 >> 0 0 0 0 >> [P 04/02 23:53] metadata keyval ops : 1 1 >> 1 1 1 1 >> [P 04/02 23:53] request scheduler : 0 0 >> 0 0 0 0 >> [D 04/02 23:53] [SM Exiting]: (0x89476c0) perf_update_sm:do_work (error >> code: 0), (action: DEFERRED) >> [D 04/02 23:53] [SM Entering]: (0x8948810) job_timer_sm:do_work (status: >> 0) >> [D 04/02 23:53] [SM Exiting]: (0x8948810) job_timer_sm:do_work (error >> code: 0), (action: DEFERRED) >> >> >> The metadata node keeps asking for something that the IO nodes cannot give >> the right way. So it complains. This makes the nodes and the metadata node >> not to work. >> >> I have installed those services many times. I have tested this using >> berkeley >> db 4.2 and 4.3 on Redhat systems(centos, scientific linnux) and on Ubuntu >> server. >> >> I have also tried the PVFS version 2.6.3 and I get the same problem. >> >> *My config files look like:* >> [r...@wn140 ~]# more /etc/pvfs2-fs.conf >> <Defaults> >> UnexpectedRequests 50 >> EventLogging all >> EnableTracing no >> LogStamp datetime >> BMIModules bmi_tcp >> FlowModules flowproto_multiqueue >> PerfUpdateInterval 1000 >> ServerJobBMITimeoutSecs 30 >> ServerJobFlowTimeoutSecs 30 >> ClientJobBMITimeoutSecs 300 >> ClientJobFlowTimeoutSecs 300 >> ClientRetryLimit 5 >> ClientRetryDelayMilliSecs 2000 >> PrecreateBatchSize 512 >> PrecreateLowThreshold 256 >> >> StorageSpace /pvfs >> LogFile /tmp/pvfs2-server.log >> </Defaults> >> >> <Aliases> >> Alias wn140 tcp://wn140:3334 >> Alias wn141 tcp://wn141:3334 >> </Aliases> >> >> <Filesystem> >> Name pvfs2-fs >> ID 320870944 >> RootHandle 1048576 >> FileStuffing yes >> <MetaHandleRanges> >> Range wn140 3-2305843009213693953 >> Range wn141 2305843009213693954-4611686018427387904 >> </MetaHandleRanges> >> <DataHandleRanges> >> Range wn140 4611686018427387905-6917529027641081855 >> Range wn141 6917529027641081856-9223372036854775806 >> </DataHandleRanges> >> <StorageHints> >> TroveSyncMeta yes >> TroveSyncData no >> TroveMethod alt-aio >> </StorageHints> >> </Filesystem> >> >> >> My setup is made from two nodes that are both IO and Metadata nodes. I >> have also tried >> a 4 node setup with 2I/O - 2 MD nodes resulting in the same thing. >> >> Any suggestions? >> >> thank you in advance, >> -- >> Asterios Katsifodimos >> High Performance Computing systems Lab >> Department of Computer Science, University of Cyprus >> http://www.asteriosk.gr <http://www.asteriosk.gr/> >> >> >> ------------------------------------------------------------------------ >> >> _______________________________________________ >> Pvfs2-users mailing list >> [email protected] >> http://www.beowulf-underground.org/mailman/listinfo/pvfs2-users >> > >
_______________________________________________ Pvfs2-users mailing list [email protected] http://www.beowulf-underground.org/mailman/listinfo/pvfs2-users
