On 2020/5/26 15:44, Chao Yu wrote:
> On 2020/5/26 10:26, Jaegeuk Kim wrote:
>> On 05/26, Chao Yu wrote:
>>> Hi Jaegeuk,
>>>
>>> On 2020/5/26 9:59, Jaegeuk Kim wrote:
>>>> Hi Chao,
>>>>
>>>> I'm hitting segment.c:1065 when running longer fsstress (1000s) with error
>>>
>>> (1000s) do you mean time in single round or total time of multi rounds?
>>>
>>>> injection. Do you have any issue from your side?
>>>
>>> I haven't hit that before, in my test, in single round, fsstress won't last 
>>> long
>>> time (normally about 10s+ for each round).
>>>
>>> Below is por_fsstress() implementation in my code base:
>>>
>>> por_fsstress()
>>> {
>>>         _fs_opts
>>>
>>>         while true; do
>>>                 ltp/fsstress -x "echo 3 > /proc/sys/vm/drop_caches" -X 10 
>>> -r -f fsync=8 -f sync=0 -f write=4 -f dwrite=2 -f truncate=6 -f allocsp=0 
>>> -f bulkstat=0 -f bulkstat1=0 -f freesp=0 -f zero=1 -f collapse=1 -f 
>>> insert=1 -f resvsp=0 -f unresvsp=0 -S t -p 20 -n 200000 -d $TESTDIR/test &
>>>                 sleep 10
>>>                 src/godown $TESTDIR
>>>                 killall fsstress
>>>                 sleep 5
>>>                 umount $TESTDIR
>>>                 if [ $? -ne 0 ]; then
>>>                         for i in `seq 1 50`
>>>                         do
>>>                                 umount $TESTDIR
>>>                                 if [ $? -eq 0]; then
>>>                                         break
>>>                                 fi
>>>                                 sleep 5
>>>                         done
>>>                 fi
>>>                 echo 3 > /proc/sys/vm/drop_caches
>>>                 _fsck
>>>                 _mount f2fs
>>>                 rm $TESTDIR/testfile
>>>                 touch $TESTDIR/testfile
>>>                 umount $TESTDIR
>>>                 _fsck
>>>                 _mount f2fs
>>>                 _rm_50
>>>         done
>>> }
>>>
>>> Did you update this code?
>>>
>>> Could you share more test configuration, like mkfs option, device size, 
>>> mount option,
>>> new por_fsstress() implementation if it exists? I can try to reproduce this 
>>> issue
>>> in my env.
>>
>> I just changed, in __run_godown_fsstress(), sleep 1000 instead of 10.
>>
>> https://github.com/jaegeuk/xfstests-f2fs/blob/f2fs/run.sh#L249
>>
>> ./run.sh por_fsstress
> 
> Reproducing...

After one night reproducing, the issue still not occur..

BTW, I enabled below features in image:

extra_attr project_quota inode_checksum flexible_inline_xattr inode_crtime 
compression

and tagged compression flag on root inode.

> 
> Thanks,
> 
>>
>>>
>>> Thanks,
>>>
>>>>
>>>> Thanks,
>>>> .
>>>>
>> .
>>
> 
> 
> _______________________________________________
> Linux-f2fs-devel mailing list
> Linux-f2fs-devel@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/linux-f2fs-devel
> .
> 


_______________________________________________
Linux-f2fs-devel mailing list
Linux-f2fs-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/linux-f2fs-devel

Reply via email to