[zfs-discuss] help with a BIG problem, can't import my zpool anymore

2008-05-23 Thread Hernan Freschi
Hello, I'm having a big problem here, disastrous maybe. I have a zpool consisting of 4x500GB SATA drives, this pool was born on S10U4 and was recently upgraded to snv85 because of iSCSI issues with some initiator. Last night I was doing housekeeping, deleting old snapshots. One snapshot

Re: [zfs-discuss] help with a BIG problem, can't import my zpool anymore

2008-05-23 Thread Hernan Freschi
I got more info. I can run zpool history and this is what I get: 2008-05-23.00:29:40 zfs destroy tera/[EMAIL PROTECTED] 2008-05-23.00:29:47 [internal destroy_begin_sync txg:3890809] dataset = 152 2008-05-23.01:28:38 [internal destroy_begin_sync txg:3891101] dataset = 152 2008-05-23.07:01:36 zpool

Re: [zfs-discuss] help with a BIG problem, can't import my zpool anymore

2008-05-23 Thread Hernan Freschi
I let it run for about 4 hours. when I returned, still the same: I can ping the machine but I can't SSH to it, or use the console. Please, I need urgent help with this issue! This message posted from opensolaris.org ___ zfs-discuss mailing list

Re: [zfs-discuss] help with a BIG problem, can't import my zpool anymore

2008-05-23 Thread Hernan Freschi
I let it run while watching TOP, and this is what I got just before it hung. Look at free mem. Is this memory allocated to the kernel? can I allow the kernel to swap? last pid: 7126; load avg: 3.36, 1.78, 1.11; up 0+01:01:11 21:16:49 88

Re: [zfs-discuss] help with a BIG problem,

2008-05-23 Thread Hernan Freschi
I forgot to post arcstat.pl's output: Time read miss miss% dmis dm% pmis pm% mmis mm% arcsz c 22:32:37 556K 525K 94 515K 949K 98 515K 97 1G1G 22:32:38636310063 100 0063 100 1G1G 22:32:39747410074

Re: [zfs-discuss] help with a BIG problem,

2008-05-23 Thread Hernan Freschi
oops. replied too fast. Ran without -n, and space was added successfully... but it didn't work. It died out of memory again. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org

Re: [zfs-discuss] help with a BIG problem,

2008-05-24 Thread Hernan Freschi
No, this is a 64-bit system (athlon64) with 64-bit kernel of course. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Re: [zfs-discuss] help with a BIG problem,

2008-05-24 Thread Hernan Freschi
So, I think I've narrowed it down to two things: * ZFS tries to destroy the dataset every time it's called because the last time it didn't finish destroying * In this process, ZFS makes the kernel run out of memory and die So I thought of two options, but I'm not sure if I'm right: Option 1:

Re: [zfs-discuss] help with a BIG problem,

2008-05-25 Thread Hernan Freschi
Hello, thanks for your suggestion. I tried settin zfs_arc_max to 0x3000 (768MB, out of 3GB). The system ran for almost 45 minutes before it froze. Here's an interesting piece of arcstat.pl, which I noticed just as it was pasing by: Time read miss miss% dmis dm% pmis pm% mmis

[zfs-discuss] Dtracing ZFS/ZIL

2008-05-30 Thread Hernan Freschi
Hello. I'm still having problems with my array. It's been replaying the ZIL (I think) for a week now and it hasn't finished. Now I don't know if it will ever finish: is it starting from scratch every time? I'm dtracing the ZIL and this is what I get: 0 46882dsl_pool_zil_clean:return

[zfs-discuss] can anyone help me?

2008-05-31 Thread Hernan Freschi
Seriously, can anyone help me? I've been asking for a week. No relevant answers, just a couple of answers but none solved my problem or even pointed me in the right way, and my posts were bumped down into oblivion. I don't know how to ask. My home server has been offline for over a week now

Re: [zfs-discuss] can anyone help me?

2008-05-31 Thread Hernan Freschi
fwiw, here are my previous posts: http://www.opensolaris.org/jive/thread.jspa?threadID=61301tstart=30 http://www.opensolaris.org/jive/thread.jspa?threadID=62120tstart=0 This message posted from opensolaris.org ___ zfs-discuss mailing list

Re: [zfs-discuss] can anyone help me?

2008-06-01 Thread Hernan Freschi
I'll provide you with the results of these commands soon. But for the record, solaris does hang (dies out of memory, can't type anything on the console, etc). What I can do is boot with -k and get to kmdb when it's hung (BREAK over serial line). I have a crashdump I can upload. I checked the

Re: [zfs-discuss] can anyone help me?

2008-06-01 Thread Hernan Freschi
Here's the output. Numbers may be a little off because I'm doing a nightly build and compressing a crashdump with bzip2 at the same time. extended device statistics r/sw/s Mr/s Mw/s wait actv wsvc_t asvc_t %w %b device 3.7 19.40.10.3 3.3 0.0

Re: [zfs-discuss] can anyone help me?

2008-06-02 Thread Hernan Freschi
Thanks for your answer, after looking at your posts my suggestion would be to try the OpenSolaris 2008.05 Live CD and to import your pool using the CD. That CD is nv86 + some extra fixes. I upgraded the snv85 to snv89 to see if it helped, but it didn't. I'll try to download the 2008.05 CD

Re: [zfs-discuss] can anyone help me? [SOLVED]

2008-06-02 Thread Hernan Freschi
Well, finally managed to solve my issue, thanks to the invaluable help of Victor Latushkin, who I can't thank enough. I'll post a more detailed step-by-step record of what he and I did (well, all credit to him actually) to solve this. Actually, the problem is still there (destroying a huge

Re: [zfs-discuss] can anyone help me?

2008-06-03 Thread Hernan Freschi
no, weird situation. I unplugged the disks from the controller (I have them labeled) before upgrading to snv89. after the upgrade, the controller names changed. This message posted from opensolaris.org ___ zfs-discuss mailing list

[zfs-discuss] problems with ludelete

2008-11-04 Thread Hernan Freschi
Hi, I'm not sure if this is the right place to ask. I'm having a little trouble deleting old solaris installs: [EMAIL PROTECTED]:~]# lustatus Boot Environment Is Active ActiveCanCopy Name Complete NowOn Reboot Delete Status

Re: [zfs-discuss] Deleting large amounts of files

2010-07-19 Thread Hernan Freschi
Hi, thanks for answering, How large is your ARC / your main memory?   Probably too small to hold all metadata (1/1000 of the data amount).   = metadata has to be read again and again Main memory is 8GB. ARC (according to arcstat.pl) usually stays at 5-7GB A recordsize smaller than 128k

Re: [zfs-discuss] Deleting large amounts of files

2010-07-21 Thread Hernan Freschi
On Tue, Jul 20, 2010 at 1:40 PM, Ulrich Graef ulrich.gr...@oracle.com wrote: When you are writing to a file and currently dedup is enabled, then the Data is entered into the dedup table of the pool. (There is one dedup table per pool not per zfs). Switching off the dedup does not change this