Re: ZFS crashing while zfs recv in progress

2013-06-07 Thread Pascal Braun, Continum
first I'd like to thank you for your time and effort.

 - Disk da3 has a different drive firmware (A580) than the A800
 drives.

Somehow I did miss that. I can replace this disk with a A800 one, although I 
don't think this will change much.

 - I have not verified if any of these disks use 4KByte sectors (dmesg
 is
   not going to tell you the entire truth).  I would appreciate seeing
   smartctl -x output from {da0,da1,da3} so I could get an idea.
Your
   pools use gpt labelling so I am left with the hope that your labels
   refer to the partition with proper 4KB alignment regardless.

The 'tank' disks are real 512bytes disks. The zpool currently in use is 
ashift=9. I've also tried ashift=12 in the past, but it didn't help. You'll 
find the output of smartctl in the attachment.

 Can you tell me what exact disk (e.g. daXX) in the above list you
 used
 for swap, and what kind of both system and disk load were going on at
 the time you saw the swap message?

 I'm looking for a capture of gstat -I500ms output (you will need a
 VERY long/big terminal window to capture this given how many disks
 you
 have) while I/O is happening, as well as top -s 1 in another
 window.
 I would also like to see zpool iostat -v 1 output while things are
 going on, to help possibly narrow down if there is a single disk
 causing
 the entire I/O subsystem for that controller to choke.

The swap disk in use is da28. 
The last output of top -s 1 that could be writen to disk was:

---
last pid:  3653;  load averages:  0.03,  0.19,  0.30up 
0+15:55:50  03:04:33
43 processes:  1 running, 41 sleeping, 1 zombie
CPU:  0.3% user,  0.0% nice,  0.6% system,  0.1% interrupt, 99.0% idle
Mem: 7456K Active, 27M Inact, 6767M Wired, 3404K Cache, 9053M Free
Swap: 256G Total, 5784K Used, 256G Free

  PID USERNAME  THR PRI NICE   SIZERES STATE   C   TIME   WCPU COMMAND
 1917 root1  220 33420K  2356K piperd  2  41:24  3.96% zfs
 1913 root1  210 71980K  5248K select  4 288:50  3.27% sshd
 1853 root1  200 29484K  2788K nanslp  0   3:13  0.00% gstat
 1803 root1  200 35476K  2128K nanslp  1   2:44  0.00% zpool
 1798 root1  200 16560K  2240K CPU07   1:07  0.00% top
 1780 root1  200 67884K  1792K select  2   0:23  0.00% sshd
 1800 root1  200 12052K  1484K select  6   0:17  0.00% script
 1747 root1  200 71980K  1868K select  1   0:13  0.00% sshd
 3148 root1  20  -20 21140K  8956K pause   7   0:11  0.00% atop
 1850 root1  200 12052K  1412K select  4   0:06  0.00% script
 1784 root1  200 67884K  1772K select  7   0:05  0.00% sshd
 1652 nagios  1  200 12012K  1044K select  7   0:02  0.00% nrpe2
 1795 root1  200 12052K  1408K select  1   0:02  0.00% script
 1538 root1  200 11996K   960K nanslp  1   0:01  0.00% ipmon
 1670 root1  200 20272K  1876K select  1   0:01  0.00% sendmail
 1677 root1  200 14128K  1548K nanslp  2   0:00  0.00% cron
 1547 root1  200 12052K  1172K select  5   0:00  0.00% syslogd
---

The last output of zpool iostat -v 1

   capacity operationsbandwidth
poolalloc   free   read  write   read  write
--  -  -  -  -  -  -
tank1.19T  63.8T 95  0   360K  0
  raidz2 305G  16.0T 25  0  92.2K  0
gpt/disk3   -  - 16  0  8.47K  0
gpt/disk9   -  - 17  0  18.9K  0
gpt/disk15  -  - 12  0  6.98K  0
gpt/disk19  -  - 12  0  6.48K  0
gpt/disk23  -  - 21  0  14.0K  0
gpt/disk27  -  - 18  0  10.5K  0
gpt/disk31  -  - 18  0  9.47K  0
gpt/disk36  -  - 16  0  18.4K  0
gpt/disk33  -  - 12  0  15.5K  0
  raidz2 305G  16.0T 25  0   103K  0
gpt/disk1   -  - 16  0  8.47K  0
gpt/disk4   -  - 24  0  16.0K  0
gpt/disk7   -  - 17  0  10.5K  0
gpt/disk10  -  - 17  0  8.97K  0
gpt/disk13  -  - 25  0  15.5K  0
gpt/disk16  -  - 15  0  8.97K  0
gpt/disk24  -  - 15  0  7.98K  0
gpt/disk32  -  - 25  0  16.9K  0
gpt/disk37  -  - 16  0  9.47K  0
  raidz2 305G  16.0T 20  0  81.3K  0
gpt/disk2   -  -  9  0  4.98K  0
gpt/disk5   -  - 20  0  14.0K  0
gpt/disk8   -  - 18  0  10.5K  0
gpt/disk11  -  - 18  0  9.47K  0
gpt/disk17  -  - 20  0  11.5K  0
gpt/disk21  -  - 12  0  6.48K  0
gpt/disk25  -  - 12  0  6.48K  0

Re: ZFS crashing while zfs recv in progress

2013-06-07 Thread Dmitry Morozovsky
On Fri, 7 Jun 2013, Pascal Braun, Continum wrote:

[snip]

  If you could put a swap disk on a dedicated controller (and no other
  disks on it), that would be ideal.  Please do not use USB for this
  task
  (the USB stack may introduce its own set of complexities pertaining
  to
  interrupt usage).
 
 I can't easily do this in the current setup. I would have to recreate the 
 primary pool differently.

Don't you have a place for (possibly, 2.5) disk inside a case, so you can 
connect it directly to mobo AHCI?

-- 
Sincerely,
D.Marck [DM5020, MCK-RIPE, DM3-RIPN]
[ FreeBSD committer: ma...@freebsd.org ]

*** Dmitry Morozovsky --- D.Marck --- Wild Woozle --- ma...@rinet.ru ***

___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org


Re: ZFS crashing while zfs recv in progress

2013-06-04 Thread Pascal Braun, Continum

I've put swap on a seperate disk this time and re-run the zfs send / recv, but 
i'm still getting the problem. Whats interesting is, i'm still getting the same 
output about swap space on the console:
swap_pager: indefinite wait buffer: bufobj: 0, blkno: 36, size: 24576

The System is running from zroot and i'm trying to transfer to tank. zpool 
Status output:
---
  pool: tank
 state: ONLINE
  scan: none requested
config:

NAMESTATE READ WRITE CKSUM
tankONLINE   0 0 0
  raidz2-0  ONLINE   0 0 0
gpt/disk3   ONLINE   0 0 0
gpt/disk9   ONLINE   0 0 0
gpt/disk15  ONLINE   0 0 0
gpt/disk19  ONLINE   0 0 0
gpt/disk23  ONLINE   0 0 0
gpt/disk27  ONLINE   0 0 0
gpt/disk31  ONLINE   0 0 0
gpt/disk36  ONLINE   0 0 0
gpt/disk33  ONLINE   0 0 0
  raidz2-1  ONLINE   0 0 0
gpt/disk1   ONLINE   0 0 0
gpt/disk4   ONLINE   0 0 0
gpt/disk7   ONLINE   0 0 0
gpt/disk10  ONLINE   0 0 0
gpt/disk13  ONLINE   0 0 0
gpt/disk16  ONLINE   0 0 0
gpt/disk24  ONLINE   0 0 0
gpt/disk32  ONLINE   0 0 0
gpt/disk37  ONLINE   0 0 0
  raidz2-2  ONLINE   0 0 0
gpt/disk2   ONLINE   0 0 0
gpt/disk5   ONLINE   0 0 0
gpt/disk8   ONLINE   0 0 0
gpt/disk11  ONLINE   0 0 0
gpt/disk17  ONLINE   0 0 0
gpt/disk21  ONLINE   0 0 0
gpt/disk25  ONLINE   0 0 0
gpt/disk29  ONLINE   0 0 0
gpt/disk38  ONLINE   0 0 0
  raidz2-3  ONLINE   0 0 0
gpt/disk12  ONLINE   0 0 0
gpt/disk14  ONLINE   0 0 0
gpt/disk18  ONLINE   0 0 0
gpt/disk22  ONLINE   0 0 0
gpt/disk26  ONLINE   0 0 0
gpt/disk30  ONLINE   0 0 0
gpt/disk34  ONLINE   0 0 0
gpt/disk35  ONLINE   0 0 0
gpt/disk39  ONLINE   0 0 0
spares
  gpt/disk20AVAIL   

errors: No known data errors

  pool: zroot
 state: ONLINE
  scan: none requested
config:

NAME   STATE READ WRITE CKSUM
zroot  ONLINE   0 0 0
  mirror-0 ONLINE   0 0 0
gpt/disk0  ONLINE   0 0 0
gpt/disk6  ONLINE   0 0 0

errors: No known data errors
---


Attatched you'll also find the dmesg.boot.

thanks,
Pascal

- Ursprüngliche Mail -
 On Wed, 29 May 2013 11:04:00 +0200, Pascal Braun, Continum
 
 Please send more information about the new server. Sometimes there
 are
 bugs found in drivers with large disks, etc. Or firmware of hardware.
 The contents of /var/run/dmesg.boot is interesting to a lot of
 people.
 As is the output of zpool status.
 
 As you are having trouble with swap on zfs. Is it possible to put
 that on
 a separate disk for the test?
 
 Ronald.
 ___
 freebsd-stable@freebsd.org mailing list
 http://lists.freebsd.org/mailman/listinfo/freebsd-stable
 To unsubscribe, send any mail to
 freebsd-stable-unsubscr...@freebsd.org
 


dmesg.boot
Description: Binary data
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org

Re: ZFS crashing while zfs recv in progress

2013-06-04 Thread Steven Hartland

If your seeing swap_pager warnings it indicates two things:-
1. Your seeing memory pressure
2. Something is causing significant delay on transations to disk.

Additional information which maybe useful:-
* mptutil show adapter
* iostat 1
* Capture from top when the system is hung

   Regards
   Steve
- Original Message - 
From: Pascal Braun, Continum pascal.br...@continum.net

To: Ronald Klop ronald-freeb...@klop.yi.org
Cc: freebsd-stable@freebsd.org
Sent: Tuesday, June 04, 2013 9:54 AM
Subject: Re: ZFS crashing while zfs recv in progress



I've put swap on a seperate disk this time and re-run the zfs send / recv, but i'm still getting the problem. Whats interesting 
is, i'm still getting the same output about swap space on the console:

swap_pager: indefinite wait buffer: bufobj: 0, blkno: 36, size: 24576

The System is running from zroot and i'm trying to transfer to tank. zpool 
Status output:
---
 pool: tank
state: ONLINE
 scan: none requested
config:

   NAMESTATE READ WRITE CKSUM
   tankONLINE   0 0 0
 raidz2-0  ONLINE   0 0 0
   gpt/disk3   ONLINE   0 0 0
   gpt/disk9   ONLINE   0 0 0
   gpt/disk15  ONLINE   0 0 0
   gpt/disk19  ONLINE   0 0 0
   gpt/disk23  ONLINE   0 0 0
   gpt/disk27  ONLINE   0 0 0
   gpt/disk31  ONLINE   0 0 0
   gpt/disk36  ONLINE   0 0 0
   gpt/disk33  ONLINE   0 0 0
 raidz2-1  ONLINE   0 0 0
   gpt/disk1   ONLINE   0 0 0
   gpt/disk4   ONLINE   0 0 0
   gpt/disk7   ONLINE   0 0 0
   gpt/disk10  ONLINE   0 0 0
   gpt/disk13  ONLINE   0 0 0
   gpt/disk16  ONLINE   0 0 0
   gpt/disk24  ONLINE   0 0 0
   gpt/disk32  ONLINE   0 0 0
   gpt/disk37  ONLINE   0 0 0
 raidz2-2  ONLINE   0 0 0
   gpt/disk2   ONLINE   0 0 0
   gpt/disk5   ONLINE   0 0 0
   gpt/disk8   ONLINE   0 0 0
   gpt/disk11  ONLINE   0 0 0
   gpt/disk17  ONLINE   0 0 0
   gpt/disk21  ONLINE   0 0 0
   gpt/disk25  ONLINE   0 0 0
   gpt/disk29  ONLINE   0 0 0
   gpt/disk38  ONLINE   0 0 0
 raidz2-3  ONLINE   0 0 0
   gpt/disk12  ONLINE   0 0 0
   gpt/disk14  ONLINE   0 0 0
   gpt/disk18  ONLINE   0 0 0
   gpt/disk22  ONLINE   0 0 0
   gpt/disk26  ONLINE   0 0 0
   gpt/disk30  ONLINE   0 0 0
   gpt/disk34  ONLINE   0 0 0
   gpt/disk35  ONLINE   0 0 0
   gpt/disk39  ONLINE   0 0 0
   spares
 gpt/disk20AVAIL

errors: No known data errors

 pool: zroot
state: ONLINE
 scan: none requested
config:

   NAME   STATE READ WRITE CKSUM
   zroot  ONLINE   0 0 0
 mirror-0 ONLINE   0 0 0
   gpt/disk0  ONLINE   0 0 0
   gpt/disk6  ONLINE   0 0 0

errors: No known data errors
---


Attatched you'll also find the dmesg.boot.

thanks,
Pascal

- Ursprüngliche Mail -

On Wed, 29 May 2013 11:04:00 +0200, Pascal Braun, Continum

Please send more information about the new server. Sometimes there
are
bugs found in drivers with large disks, etc. Or firmware of hardware.
The contents of /var/run/dmesg.boot is interesting to a lot of
people.
As is the output of zpool status.

As you are having trouble with swap on zfs. Is it possible to put
that on
a separate disk for the test?

Ronald.
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to
freebsd-stable-unsubscr...@freebsd.org









___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org 




This e.mail is private and confidential between Multiplay (UK) Ltd. and the person or entity to whom it is addressed. In the event of misdirection, the recipient is prohibited from using, copying, printing or otherwise disseminating it or any information contained in it. 


In the event of misdirection, illegible or incomplete transmission please 
telephone +44 845 868 1337
or return the E.mail to postmas

Re: ZFS crashing while zfs recv in progress

2013-06-04 Thread Jeremy Chadwick
On Tue, Jun 04, 2013 at 10:54:30AM +0200, Pascal Braun, Continum wrote:
 I've put swap on a seperate disk this time and re-run the zfs send / recv, 
 but i'm still getting the problem. Whats interesting is, i'm still getting 
 the same output about swap space on the console:
 swap_pager: indefinite wait buffer: bufobj: 0, blkno: 36, size: 24576

What this means is covered here:

http://www.freebsd.org/doc/en/books/faq/troubleshoot.html#idp75389104

Google swap_pager: indefinite wait buffer and you will see lots of
people talking about this.  The commonality in most situations is that
the I/O subsystem is stalled or too busy to swap in/out a page of
memory, and the kernel throws a nastygram about it.

Basically, your I/O subsystem is taking too long to accomplish a task
(swap in/out a page of memory to/from swap).  It could be the disk
taking too long, it could be the controller taking too long, it could be
the disk being in bad shape, it could be the overall PCI/PCI-X/PCIe bus
being overwhelmed, or could be something as simple as excessive CPU load
(your pool/vdev setup I imagine is very CPU intensive).

Your system is utterly massive.  Warning: this is the first time I have
taken a stab at such a huge system.  The topology, for those wondering:

* mpt0 (LSI SATA/SAS; rev 1.5.20.0), irq 17
  |-- 6 disks attached
  |- da0  = Hitachi HDS72101 A3MA, 1953525168 sectors
  |- da1  = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da2  = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da3  = Hitachi HDS5C302 A580, 3907029168 sectors
  |- da4  = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da5  = Hitachi HDS5C302 A800, 3907029168 sectors
* mpt1 (LSI SATA/SAS; rev 1.5.20.0), irq 18
  |-- 7 disks attached
  |- da6  = Hitachi HDS72101 A3MA, 1953525168 sectors
  |- da7  = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da8  = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da9  = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da10 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da11 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da12 = Hitachi HDS5C302 A800, 3907029168 sectors
* mpt2 (LSI SATA/SAS; rev 1.5.20.0), irq 18
  |-- 6 disks attached
  |- da13 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da14 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da15 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da16 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da17 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da18 = Hitachi HDS5C302 A800, 3907029168 sectors
* mpt3 (LSI SATA/SAS; rev 1.5.20.0), irq 40
  |-- 8 disks attached
  |- da19 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da20 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da21 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da22 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da23 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da24 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da25 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da26 = Hitachi HDS5C302 A800, 3907029168 sectors
* mpt4 (LSI SATA/SAS; rev 1.5.20.0), irq 41
  |-- 7 disks attached
  |- da27 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da28 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da29 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da30 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da31 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da32 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da33 = Hitachi HDS5C302 A800, 3907029168 sectors
* mpt5 (LSI SATA/SAS; rev 1.5.20.0), irq 41 (shares IRQ with mpt4)
  |-- 5 disks attached
  |- da34 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da35 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da36 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da37 = Hitachi HDS5C302 A800, 3907029168 sectors
  |- da38 = Hitachi HDS5C302 A800, 3907029168 sectors

Things to note / things that caught my eye:

- mpt2 shares an IRQ with mpt1 (irq 18)

- mpt5 shares an IRQ with mpt4 (irq 41)

- Disk da3 has a different drive firmware (A580) than the A800 drives.

- (For readers only): Disks {da0,da6} are a completely different model
  and capacity than all the other drives in the array.  These drives
  are used (in a mirror) for the ZFS root pool (zroot), not the big
  fat gigantic pool (tank).

- I have not verified if any of these disks use 4KByte sectors (dmesg is
  not going to tell you the entire truth).  I would appreciate seeing
  smartctl -x output from {da0,da1,da3} so I could get an idea.  Your
  pools use gpt labelling so I am left with the hope that your labels
  refer to the partition with proper 4KB alignment regardless.

- The system has only 16GB RAM.  That's a bit shocking for something of
  this size.

Moving on.

Can you tell me what exact disk (e.g. daXX) in the above list you used
for swap, and what kind of both system and disk load were going on at
the time you saw the 

Re: ZFS crashing while zfs recv in progress

2013-06-04 Thread Scot Hetzel
On Wed, May 29, 2013 at 4:04 AM, Pascal Braun, Continum
pascal.br...@continum.net wrote:

 Hi,


 I'm trying to send a zfs pool from an old freebsd 9.0 installation to a new 
 machine with freebsd 9.1. The pool is quite heavy (about 16TB, lots of 
 snapshots) and the receiving side keeps crashing on me. The command used to 
 transfer (run on the old 9.0 installation):
 zfs send -R tank@snapshot | ssh 10.10.xx.xx zfs recv -F -d -v tank


 After a few hours the system stops all writing and I can't start any new 
 processes. Processes still running like 'zpool iostat' are still working, or 
 at least it is still reporting something. To me it looks like the filesystem 
 just disappeared. Unfortunately I'm running root on zfs so I don't have any 
 logs about this.
 The only message I sometimes find on the console are about not being able to 
 write to swap, which is also on zfs.


This could be where your problem is happening.  While you can create a
swap vol on ZFS, that swap vol also requires available memory from the
system to perform the swap.  As was suggested, try using a dedicated
disk / partition as your swap volume.


 Do you have any ideas? I don't even know where to start.


 regards, Pascal

 ___
 freebsd-stable@freebsd.org mailing list
 http://lists.freebsd.org/mailman/listinfo/freebsd-stable
 To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org

-- 
DISCLAIMER:

No electrons were maimed while sending this message. Only slightly bruised.
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org


ZFS crashing while zfs recv in progress

2013-05-29 Thread Pascal Braun, Continum

Hi, 


I'm trying to send a zfs pool from an old freebsd 9.0 installation to a new 
machine with freebsd 9.1. The pool is quite heavy (about 16TB, lots of 
snapshots) and the receiving side keeps crashing on me. The command used to 
transfer (run on the old 9.0 installation): 
zfs send -R tank@snapshot | ssh 10.10.xx.xx zfs recv -F -d -v tank 


After a few hours the system stops all writing and I can't start any new 
processes. Processes still running like 'zpool iostat' are still working, or at 
least it is still reporting something. To me it looks like the filesystem just 
disappeared. Unfortunately I'm running root on zfs so I don't have any logs 
about this. 
The only message I sometimes find on the console are about not being able to 
write to swap, which is also on zfs. 


Do you have any ideas? I don't even know where to start. 


regards, Pascal 

___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org


Re: ZFS crashing while zfs recv in progress

2013-05-29 Thread Steven Hartland

Silly question but I assume your new 9.1 system isnt running from tank at the 
time?
- Original Message - 
From: Pascal Braun, Continum pascal.br...@continum.net


I'm trying to send a zfs pool from an old freebsd 9.0 installation to a new machine with freebsd 9.1. The pool is quite heavy 
(about 16TB, lots of snapshots) and the receiving side keeps crashing on me. The command used to transfer (run on the old 9.0 
installation):

zfs send -R tank@snapshot | ssh 10.10.xx.xx zfs recv -F -d -v tank


After a few hours the system stops all writing and I can't start any new processes. Processes still running like 'zpool iostat' 
are still working, or at least it is still reporting something. To me it looks like the filesystem just disappeared. 
Unfortunately I'm running root on zfs so I don't have any logs about this.

The only message I sometimes find on the console are about not being able to 
write to swap, which is also on zfs.


Do you have any ideas? I don't even know where to start.


regards, Pascal

___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org





This e.mail is private and confidential between Multiplay (UK) Ltd. and the person or entity to whom it is addressed. In the event of misdirection, the recipient is prohibited from using, copying, printing or otherwise disseminating it or any information contained in it. 


In the event of misdirection, illegible or incomplete transmission please 
telephone +44 845 868 1337
or return the E.mail to postmas...@multiplay.co.uk.

___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org


Re: ZFS crashing while zfs recv in progress

2013-05-29 Thread Ronald Klop
On Wed, 29 May 2013 11:04:00 +0200, Pascal Braun, Continum  
pascal.br...@continum.net wrote:




Hi,


I'm trying to send a zfs pool from an old freebsd 9.0 installation to a  
new machine with freebsd 9.1. The pool is quite heavy (about 16TB, lots  
of snapshots) and the receiving side keeps crashing on me. The command  
used to transfer (run on the old 9.0 installation):

zfs send -R tank@snapshot | ssh 10.10.xx.xx zfs recv -F -d -v tank


After a few hours the system stops all writing and I can't start any new  
processes. Processes still running like 'zpool iostat' are still  
working, or at least it is still reporting something. To me it looks  
like the filesystem just disappeared. Unfortunately I'm running root on  
zfs so I don't have any logs about this.
The only message I sometimes find on the console are about not being  
able to write to swap, which is also on zfs.



Do you have any ideas? I don't even know where to start.



Please send more information about the new server. Sometimes there are  
bugs found in drivers with large disks, etc. Or firmware of hardware.

The contents of /var/run/dmesg.boot is interesting to a lot of people.
As is the output of zpool status.

As you are having trouble with swap on zfs. Is it possible to put that on  
a separate disk for the test?


Ronald.
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org