so 40ms for how many messages and what kind of payload? And any idea how much data is blocked? (msgs/payload)
Even though 40ms doesn't seem like much, it is def. something that can creep up in a high load environment, and something you can't really monitor unless you have some sort of metrics built into the system. Maybe have this built in: http://metrics.codahale.com/ On Fri, May 25, 2012 at 1:22 PM, Jay Kreps <jay.kr...@gmail.com> wrote: > It depends a great deal on the hw and the flush interval. I think for our > older generation hw we saw an avg flush time of 40ms, for newer stuff we > just got it is much less but I think that might be because the disks > themselves have some kind of nvram or something. > > -Jay > > On Fri, May 25, 2012 at 7:09 AM, S Ahmed <sahmed1...@gmail.com> wrote: > > > In practise (at linkedin), how long do you see the calls blocked for > during > > fsycs? > > > > On Thu, May 24, 2012 at 1:40 PM, Jay Kreps <jay.kr...@gmail.com> wrote: > > > > > One issue with using the filesystem for persistence is that the > > > synchronization in the filesystem is not great. In particular the fsync > > and > > > fsyncdata system calls block appends to the file, apparently for the > > entire > > > duration of the fsync (which can be quite long). This is documented in > > some > > > detail here: > > > http://antirez.com/post/fsync-different-thread-useless.html > > > > > > This is a problem in 0.7 because our definition of a committed message > is > > > one written prior to calling fsync(). This is the only way to guarantee > > the > > > message is on disk. We do not hand out any messages to consumers until > an > > > fsync call occurs. The problem is that regardless of whether the fsync > is > > > in a background thread or not it will block any produce requests to the > > > file. This is buffered a bit in the client since our produce request is > > > effectively async in 0.7, but it can lead to weird latency spikes > > > nontheless as this buffering gets filled. > > > > > > In 0.8 with replication the definition of a committed message changes > to > > > one that is replicated to multiple machines, not necessarily committed > to > > > disk. This is a different kind of guarantee with different strengths > and > > > weaknesses (pro: data can survive destruction of the file system on one > > > machine, con: you will lose a few messages if you haven't sync'd and > the > > > power goes out). We will likely retain the flush interval and time > > settings > > > for those who want fine grained control over flushing, but it is less > > > relevant. > > > > > > Unfortunately *any* call to fsync will block appends even in a > background > > > thread so how can we give control over physical disk persistence > without > > > introducing high latency for the producer? The answer is that the linux > > > pdflush daemon actually does a very similar thing to our flush > > parameters. > > > pdflush is a daemon running on every linux machine that controls the > > > writing of buffered/cached data back to disk. It allows you to control > > the > > > percentage of memory filled with dirty pages by giving it either a > > > percentage of memory, a time out for any dirty page to be written, or a > > > fixed number of dirty bytes. > > > > > > The question is, does pdflush block appends? The answer seems to be > > mostly > > > no. It locks the page being flushed but not the whole file. The time to > > > flush one page is actually usually pretty quick (plus I think it may > not > > be > > > flushing just written pages anyway). I wrote some test code for this > and > > > here are the results: > > > > > > I modified the code from the link above. Here are the results from my > > > desktop (Centos Linux 2.6.32). > > > > > > We run the test writing 1024 bytes every 100 us and flushing every 500 > > us: > > > > > > $ ./pdflush-test 1024 100 500 > > > 21 > > > 4 > > > 3 > > > 3 > > > 9 > > > 6 > > > Sync in 20277 us (0), sleeping for 500 us > > > 19819 > > > 7 > > > 7 > > > 8 > > > 38 > > > Sync in 19470 us (0), sleeping for 500 us > > > 19048 > > > 7 > > > 4 > > > 3 > > > 8 > > > 4 > > > Sync in 19405 us (0), sleeping for 500 us > > > 19017 > > > 6 > > > 6 > > > 10 > > > 6 > > > Sync in 19410 us (0), sleeping for 500 us > > > 19025 > > > 7 > > > 7 > > > 11 > > > 6 > > > > > > $ cat /proc/sys/vm/dirty_writeback_centisecs > > > 100 > > > $ cat /proc/sys/vm/dirty_expire_centisecs > > > 500 > > > > > > Now run the test with the background flush disabled (rarely running): > > > $ ./pdflush-test 1024 100 5000000000000 > times.txt > > > > > > I ran this for 298,028 writes. The 99.9th percentile for this test is > 17 > > us > > > and the max time was 2043 us (2ms). > > > > > > Here is the test code: > > > > > > #include <stdio.h> > > > #include <unistd.h> > > > #include <string.h> > > > #include <sys/types.h> > > > #include <pthread.h> > > > #include <sys/stat.h> > > > #include <fcntl.h> > > > #include <sys/time.h> > > > #include <stdlib.h> > > > > > > static long long microseconds(void) { > > > struct timeval tv; > > > long long mst; > > > > > > gettimeofday(&tv, NULL); > > > mst = ((long long)tv.tv_sec)*1000000; > > > mst += tv.tv_usec; > > > return mst; > > > } > > > > > > void *IOThreadEntryPoint(void *arg) { > > > int fd, retval; > > > long long start; > > > long sleep = (long) arg; > > > > > > while(1) { > > > usleep(sleep); > > > start = microseconds(); > > > fd = open("/tmp/foo.txt",O_RDONLY); > > > retval = fsync(fd); > > > close(fd); > > > printf("Sync in %lld us (%d), sleeping for %ld us\n", > > > microseconds()-start, retval, sleep); > > > } > > > return NULL; > > > } > > > > > > int main(int argc, char* argv[]) { > > > if(argc != 4) { > > > printf("USAGE: %s size write_sleep fsync_sleep\n", argv[0]); > > > exit(1); > > > } > > > > > > pthread_t thread; > > > int fd = open("/tmp/foo.txt",O_WRONLY|O_CREAT,0644); > > > long long start; > > > long long ellapsed; > > > int size = atoi(argv[1]); > > > long write_sleep = atol(argv[2]); > > > long fsync_sleep = atol(argv[3]); > > > char buff[size]; > > > > > > pthread_create(&thread,NULL,IOThreadEntryPoint, (void*) > fsync_sleep); > > > > > > while(1) { > > > start = microseconds(); > > > if (write(fd,buff,size) == -1) { > > > perror("write"); > > > exit(1); > > > } > > > ellapsed = microseconds()-start; > > > printf("%lld\n", ellapsed); > > > usleep(write_sleep); > > > } > > > close(fd); > > > exit(0); > > > } > > > > > > Cheers, > > > > > > -Jay > > > > > >