Hello there,
I think you should share it with the list, if you can, seems like an
interesting work. ZFS has some issues with snapshots and spa_sync performance
for snapshots deletion.
Thanks
Leal
[ http://www.eall.com.br/blog ]
--
This message posted from opensolaris.org
Thanks Adam,
So, if i understand well, the MLC SSD more appropriate for read cache is more
theory than pratice right now. Right? I mean, SUN is just using SLC SSD's?
That would explain the only support for SLC on SUN hardware (x42xx) series.
Thanks again.
Leal
[ http://www.eall.com.br/blog
Ok Bob, but i think that is the problem about picket fencing... and so we are
talking about commit the sync operations to disk. What i'm seeing is no read
activity from disks when the slog is beeing written. The disks are zero (no
read, no write).
Thanks a lot for your reply.
Leal
[
Hello,
Well, i'm trying to understand this workload, but what i'm seeing to reproduce
this is just flood the SSD with writes, and the disks show no activity. I'm
testing with aggr (two links), and for one or two seconds there is no read
activity (output from server).
Right now i'm suspecting
That's only one element of it Bob. ZFS also needs
devices to fail quickly and in a predictable manner.
A consumer grade hard disk could lock up your entire
pool as it fails. The kit Sun supply is more likely
to fail in a manner ZFS can cope with.
I agree 100%.
Hardware, firmware,
Hello all,
Somebody using iSCSI cache enable on 7000 series? I'm talking about
OpenSolaris (ZFS) as an iSCSI initiator, because i don't know another
filesystem that handles disk caches.
So, that option was created for ZFS ;-)?
Any suggestions on this?
Thanks
Leal
[
Hello all...
I'm seeing this behaviour in an old build (89), and i just want to hear from
you if there is some known bug about it. I'm aware of the picket fencing
problem, and that ZFS is not choosing right if write to slog is better or not
(thinking if we have a better throughput from disks).
Note that this issue does not apply at all to NFS
service, database
service, or any other usage which does synchronous
writes.
Bob
Hello Bob,
There is impact for all workloads.
The fact that the write is sync or not, is just a question to write on slog
(SSD) or not.
But the txg
Hello all,
I'm trying to understand the ZFS IO scheduler (
http://www.eall.com.br/blog/?p=1170 ), and why sometimes the system seems to be
stalled for some seconds, and every application that needs some IO (most
read, i think), have serious problems. What can be a big problem in iSCSI or
NFS
Hello all,
There is some project here to integrate amanda on opensolaris, or some howto
for integration with ZFS? Some use case (using the opensource version)?
The amanda site there is a few instructions, but i think here we can create
something more specific to OS.
Thanks.
--
This message
Hello all...
We are getting this error: E2BIG - Arg list too long, when trying to send
incremental backups (b89 - b101). Do you know about any bugs related to that?
I did a look on the archives, and google but could not find anything.
What i did find was something related with wrong
FYI (version 0.3):
http://www.eall.com.br/blog/?p=970
Leal
[ http://www.eall.com.br/blog ]
Hello all..
I did some tests to understand the behaviour of ZFS
and slog (SSD), and for understand the workload i
did implement a simple software to visualize the
data blocks (read/write).
I'm
Hello all..
I did some tests to understand the behaviour of ZFS and slog (SSD), and for
understand the workload i did implement a simple software to visualize the data
blocks (read/write).
I'm posting here the link in the case somebody wants to try it.
http://www.eall.com.br/blog/?p=906
Hello Bernd,
Now i see your point... ;-)
Well, following a very simple math:
- One txg each 5 seconds = 17280/day;
- Each txg writing 1MB (L0-L3) = 17GB/day
In the paper the math was 10 years = ( 2.7 * the size of the USB drive) writes
per day, right?
So, in a 4GB drive, would be
... it seems not related with txg,
sync of disks, labels, nothing... ;-)
Thanks a lot Bernd.
Leal
[http://www.eall.com.br/blog]
Regards,
Bernd
Marcelo Leal wrote:
Hello Bernd,
Now i see your point... ;-)
Well, following a very simple math:
- One txg each 5 seconds = 17280/day
Hi,
Hello Bernd,
After I published a blog entry about installing
OpenSolaris 2008.11 on a
USB stick, I read a comment about a possible issue
with wearing out
blocks on the USB stick after some time because ZFS
overwrites its
uberblocks in place.
I did not understand well what you
Hello,
- One way is virtualization, if you use a virtualization technology that uses
NFS for example, you could add your virtual images on a ZFS filesystem. NFS
can be used without virtualization too, but as you said the machines are
windows, i don't think the NFS client for windows is
Marcelo Leal writes:
Hello all,
Somedays ago i was looking at the code and did see
some variable that
seems to make a correlation between the size of
the data, and if the
data is written to the slog or directly to the
pool. But i did not
find it anymore, and i think is way more
Thanks a lot Sanjeev!
If you look my first message you will see that discrepancy in zdb...
Leal.
[http://www.eall.com.br/blog]
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
Hello all,
# zpool status
pool: mypool
state: ONLINE
scrub: scrub completed after 0h2m with 0 errors on Fri Dec 19 09:32:42 2008
config:
NAME STATE READ WRITE CKSUM
storage ONLINE 0 0 0
mirror ONLINE 0 0 0
execve(/usr/bin/rm, 0x08047DBC, 0x08047DC8) argc = 2
mmap(0x, 4096, PROT_READ|PROT_WRITE|PROT_EXEC, MAP_PRIVATE|MAP_ANON,
-1, 0) = 0xFEFF
resolvepath(/usr/lib/ld.so.1, /lib/ld.so.1, 1023) = 12
resolvepath(/usr/bin/rm, /usr/bin/rm, 1023) = 11
sysconfig(_CONFIG_PAGESIZE)
execve(/usr/bin/ls, 0x08047DA8, 0x08047DB4) argc = 2
mmap(0x, 4096, PROT_READ|PROT_WRITE|PROT_EXEC, MAP_PRIVATE|MAP_ANON,
-1, 0) = 0xFEFF
resolvepath(/usr/lib/ld.so.1, /lib/ld.so.1, 1023) = 12
resolvepath(/usr/bin/ls, /usr/bin/ls, 1023) = 11
xstat(2, /usr/bin/ls, 0x08047A58)
Hello all...
Can that be caused by some cache on the LSI controller?
Some flush that the controller or disk did not honour?
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
Hello all,
Somedays ago i was looking at the code and did see some variable that seems to
make a correlation between the size of the data, and if the data is written to
the slog or directly to the pool. But i did not find it anymore, and i think is
way more complex than that.
For example, if
Hello all,
I'm getting many OpenSolaris kernel panic while send/receiving data. I did try
to create another pool and another host to test, and the same error. And the
send side can be any server (i did test with four different servers, all build
89).
The panic message:
--- cut here
A percentage of the total space is reserved for pool
overhead and is not
allocatable, but shows up as available in zpool
list.
Something to change/show in the future?
--
Leal
[http://www.posix.brte.com.br/blog]
--
This message posted from opensolaris.org
I think the better solution is to have two pools, and write a script to change
the recording destination time to time, or move the files after it. Like the
prototype by Hartz.
ps.: Is this a reality show? ;-)
--
This message posted from opensolaris.org
Hello,
In ZFS the DNLC concept is gone, or is in ARC too? I mean, all the cache in
ZFS is ARC right?
I was thinking if we can tune the DNLC in ZFS like in UFS.. if we have too
*many* files and directories, i guess we can have a better performance having
all the metadata cached, and that is
of tuning would be usefull for just a few workloads, but could
be a *huge* enhancement for that workloads.
Leal
-- posix rules --
[http://www.posix.brte.com.br/blog]
On 10/30/08 04:50, Marcelo Leal wrote:
Hello,
In ZFS the DNLC concept is gone, or is in ARC too?
I mean
Hello,
In the situation you have described, if i understood well, you would not have
any space. When you take a snapshot, your snapshot is referencing the blocks
older than it...
Ex.:
You have a 500gb disk, and create a 5gb file, you got 495gb free space.
So you delete the file, you have
Because of one change to just one file, the MOS is a brend new one
Yes, all writes in ZFS are done in transaction groups.. so, evertime there
is a commit, something is really write to disk, there is a new txg and all the
blocks written are related to that txg (even the ubberblock).
I don´t
I agree with you Constantin that the sync is a performance problem, in the same
way i think in a NFS environment it is just *required*. If the sync can be
relaxed in a specific NFS environment, my first opinion is that the NFS is
not necessary on that environment in first place.
IMHO a
Bah, I've done it again. I meant use it as a slog
device, not as the ZIL...
But the slog is the ZIL. formaly a *separate* intent log. What´s the matter? I
think everyone did understand. I think you did make a confusion some threads
before about ZIL and L2ARC. That is a different thing.. ;-)
Hello there,
It´s not a wiki, but has many considerations about your question:
http://www.opensolaris.org/jive/thread.jspa?threadID=78841tstart=60
Leal.
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
But the slog is the ZIL. formaly a *separate*
intent log.
No the slog is not the ZIL!
Ok, when you did write this:
I've been slogging for a while on support for separate intent logs (slogs)
for ZFS. Without slogs, the ZIL is allocated dynamically from the main pool.
You were talking
On 10/22/08 13:56, Marcelo Leal wrote:
But the slog is the ZIL. formaly a *separate*
intent log.
No the slog is not the ZIL!
Ok, when you did write this:
I've been slogging for a while on support for
separate intent logs (slogs) for ZFS.
Without slogs, the ZIL is allocated
Hello Roch!
Leave the default recordsize. With 128K recordsize,
files smaller than
128K are stored as single record
tightly fitted to the smallest possible # of disk
sectors. Reads and
writes are then managed with fewer ops.
In the write ZFS is dynamic, but in the read?
If i have
Hello all,
Did you make a install on the USB stick, or did you use the Distribution
Constructor (DC)?
Leal.
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
Hello there...
I did see that already, talk with some guys without answer too... ;-)
Actually, this week i did not see discrepancy between tools, but the pool
information was wrong (space used). Exporting/importing, scrub, and etc, did
not solve. I know that zfs is async in the status report
Hello all,
I think he got some point here... maybe that would be an interesting feature
for that kind of workload. Caching all the metadata, would make the rsync task
more fast (for many files). Try to cache the data is really waste of time,
because the data will not be read again, and will
Hello all,
I think in SS 11 should be -xarch=amd64.
Leal.
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
So, there is no raid10 in a solaris/zfs setup?
I´m talking about no redundancy...
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Hello all,
I think the problem here is the ZFS´ capacity for recovery from a failure.
Forgive me, but thinking about creating a code without failures, maybe the
hackers did forget that other people can make mistakes (if they can´t).
- ZFS does not need fsck.
Ok, that´s a great statement,
On Fri, Oct 10, 2008 at 06:15:16AM -0700, Marcelo
Leal wrote:
- ZFS does not need fsck.
Ok, that?s a great statement, but i think ZFS
needs one. Really does.
And in my opinion a enhanced zdb would be the
solution. Flexibility.
Options.
About 99% of the problems reported as I
ZFS has not limit for snapshots and filesystems too, but try to create a lot
snapshots and filesytems and you will have to wait a lot for your pool to
import too... ;-)
I think you should not think about the limits, but performance. Any
filesytem with *too many entries by directory will
Just curiosity, why don´t use SC?
Leal.
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Hello Adrian,
Thanks, i was using send/receive (that´s why i did put it on subject ;), and i
did like to know if ZFS could have some solution for that as i said before. The
send/receive is not an exact copy of the filesystem (creation time, fsid,
etc) are different. So, the FH using that for
Hello milek,
That information remains true?
ZFS algorithm for selecting block sizes
The initial block size is the smallest support block size larger than the first
write to the file.
Grow to the next largest block size for the entire file when the total file
length increases beyond the
On Fri, 5 Sep 2008, Marcelo Leal wrote:
4 - The last one... ;-) For the FSB allocation,
how the zfs knows
the file size, for know if the file is smaller than
the FSB?
Something related to the txg? When the write goes
to the disk, the
zfs knows (some way) if that write is a whole
Hello all,
Some way to workaround the filehandle issue with a send/receive ZFS procedure?
In the ZFS begining, i did a conversation with some of the devel guys, and did
ask about how ZFS would treat the NFS filehandle.. IIRC, the answere was: No
problem, the NFS filehandle will not depend on
You are right! Seeing the numbers i could not think very well ;-)
What matters is the used size, and not the storage capacity! My fault...
Thanks a lot for the answers.
Leal.
--
This message posted from opensolaris.org
___
zfs-discuss mailing
Thanks a lot for the answers!
Relling did say something about checksum, i did ask to him about a more
detailed explanation about it. Because i did not understand what checksum
the receive part has to check, as the send can be redirected to a file on a
disc or tape...
In the end, i think if
Hello!
Assuming the default recordsize (FSB) in zfs is 128k, so:
1 - If i have a file with 10k, the zfs will allocate a FSD of 10k. Right? As
zfs is not static like the other filesystems, i don´t have that old internal
fragmentation...
2 - If the above is right, i don´t need to adjust the
Hello all,
Any plans (or already have), a send/receive way to get the transfer backup
statistics? I mean, the how much was transfered, time and/or bytes/sec?
And the last question... i did see in many threads the question about the
consistency between the send/receive through ssh... but no
Hello all,
I was used to use mirrors and solaris 10, in which the scrub process for 500gb
took about two hours... and with solaris express (snv_79a) tests, terabytes in
minutes. I did search for release changes in the scrub process, and could not
find anything about enhancements in this
Hello all,
i would like to continue with this topic, and after doing some research
about the topic, i have some (many) doubts, and maybe we could use this thread
to give some responses to me and other users that can have the same questions...
First, sorry to CC to many forums, but i think is a
Hello all,
In a traditional filesystem, we have a few filesystems, but with ZFS, we can
have thousands..
The question is: There is a command or procedure to remake the filesystems,
in a recovery from backup scenario?
I mean, imagine that i have a ZFS pool with 1,000 filesystems, and for some
Hello...
If i have understood well, you will have a host with EMC RAID5 discs. Is that
right?
You pay a lot of money to have EMC discs, and i think is not a good idea have
another layer of *any* RAID on top of it. If you have EMC RAID5 (eg.
symmetrix), you don't need to have a software
Hello all,
I'm having the same problem here, any news?
I need to use ACL's on the GNU/Linux clients. I'm using nfsv3, and on the
GNU/Linux servers that feature was working, i think we need a solution for
solaris/opensolaris. Now, with the dmm project, how we can start a migration
process, if
No answer... well, do you not have this problem or there is another option to
delegate such administration? I was thinking if we can delegate a single
filesystem administration to some user through ZFS administration web console
(67889).
Can i create a user and give him administration rights
Hello,
If you believe that the problem can be related to ZIL code, you can try to
disable it to debug (isolate) the problem. If it is not a fileserver (NFS),
disabling the zil should not impact consistency.
Leal.
This message posted from opensolaris.org
Hello all,
Some time ago i did write a simple script to handle on the fly
filesystem(zfs) creation for linux clients
(http://www.posix.brte.com.br/blog/?p=102). I was thinking in improve that
script to handle more generic remote actions... but i think we could start a
project on this:
A
62 matches
Mail list logo