The following changes since commit d8b64af2a1688c3c3d218cba0faad57a3ba050e8:

  rbd: various fixes and cleanups (2014-10-27 10:53:52 -0600)

are available in the git repository at:

  git://git.kernel.dk/fio.git master

for you to fetch changes up to 8859391bb7d77680ede51b46e770237abae566dc:

  engines/glusterfs_async: cleanups (2014-10-28 20:48:33 -0600)

----------------------------------------------------------------
Jens Axboe (6):
      engines/rbd: add support for rbd_invalidate_cache()
      engines/rbd: use rbd_aio_is_complete() and add TRIM support
      engines/rbd: remove unused rbd_options->td
      engines/rbd: add option to busy poll on event completion
      engines/glfs_async: add trim support
      engines/glusterfs_async: cleanups

 configure                 |   47 +++++++++++++++++-
 engines/glusterfs_async.c |   71 ++++++++++++---------------
 engines/rbd.c             |  119 +++++++++++++++++++++++----------------------
 3 files changed, 137 insertions(+), 100 deletions(-)

---

Diff of recent changes:

diff --git a/configure b/configure
index 58f02fa..de0fc89 100755
--- a/configure
+++ b/configure
@@ -1203,6 +1203,26 @@ fi
 echo "Rados Block Device engine     $rbd"
 
 ##########################################
+# check for rbd_invaidate_cache()
+rbd_inval="no"
+if test "$rbd" = "yes"; then
+cat > $TMPC << EOF
+#include <rbd/librbd.h>
+
+int main(int argc, char **argv)
+{
+  rbd_image_t image;
+
+  return rbd_invalidate_cache(image);
+}
+EOF
+if compile_prog "" "-lrbd -lrados" "rbd"; then
+  rbd_inval="yes"
+fi
+echo "rbd_invalidate_cache          $rbd_inval"
+fi
+
+##########################################
 # Check whether we have setvbuf
 setvbuf="no"
 cat > $TMPC << EOF
@@ -1241,6 +1261,7 @@ fi
 
 ##########################################
 # check for gfapi fadvise support
+if test "$gfapi" = "yes" ; then
 gf_fadvise="no"
 cat > $TMPC << EOF
 #include <glusterfs/api/glfs.h>
@@ -1253,11 +1274,29 @@ int main(int argc, char **argv)
   return 0;
 }
 EOF
-
 if compile_prog "" "-lgfapi -lglusterfs" "gfapi"; then
   gf_fadvise="yes"
 fi
 echo "Gluster API use fadvise       $gf_fadvise"
+fi
+
+##########################################
+# check for gfapi trim support
+gf_trim="no"
+if test "$gfapi" = "yes" ; then
+cat > $TMPC << EOF
+#include <glusterfs/api/glfs.h>
+
+int main(int argc, char **argv)
+{
+  return glfs_discard_async(NULL, 0, 0);
+}
+EOF
+if compile_prog "" "-lgfapi -lglusterfs" "gf trim"; then
+  gf_trim="yes"
+fi
+echo "Gluster API trim support      $gf_trim"
+fi
 
 ##########################################
 # Check if we support stckf on s390
@@ -1489,6 +1528,9 @@ fi
 if test "$rbd" = "yes" ; then
   output_sym "CONFIG_RBD"
 fi
+if test "$rbd_inval" = "yes" ; then
+  output_sym "CONFIG_RBD_INVAL"
+fi
 if test "$setvbuf" = "yes" ; then
   output_sym "CONFIG_SETVBUF"
 fi
@@ -1502,6 +1544,9 @@ fi
 if test "$gf_fadvise" = "yes" ; then
   output_sym "CONFIG_GF_FADVISE"
 fi
+if test "$gf_trim" = "yes" ; then
+  output_sym "CONFIG_GF_TRIM"
+fi
 if test "$libhdfs" = "yes" ; then
   output_sym "CONFIG_LIBHDFS"
 fi
diff --git a/engines/glusterfs_async.c b/engines/glusterfs_async.c
index cf6ea70..d9d6693 100644
--- a/engines/glusterfs_async.c
+++ b/engines/glusterfs_async.c
@@ -10,11 +10,11 @@ struct fio_gf_iou {
        struct io_u *io_u;
        int io_complete;
 };
-static ulong cb_count = 0, issued = 0;
 
 static struct io_u *fio_gf_event(struct thread_data *td, int event)
 {
        struct gf_data *gf_data = td->io_ops->data;
+
        dprint(FD_IO, "%s\n", __FUNCTION__);
        return gf_data->aio_events[event];
 }
@@ -25,18 +25,18 @@ static int fio_gf_getevents(struct thread_data *td, 
unsigned int min,
        struct gf_data *g = td->io_ops->data;
        unsigned int events = 0;
        struct io_u *io_u;
-       int i = 0;
-       struct fio_gf_iou *io = NULL;
+       int i;
 
        dprint(FD_IO, "%s\n", __FUNCTION__);
        do {
                io_u_qiter(&td->io_u_all, io_u, i) {
+                       struct fio_gf_iou *io;
+
                        if (!(io_u->flags & IO_U_F_FLIGHT))
                                continue;
 
-                       io = (struct fio_gf_iou *)io_u->engine_data;
-
-                       if (io && io->io_complete) {
+                       io = io_u->engine_data;
+                       if (io->io_complete) {
                                io->io_complete = 0;
                                g->aio_events[events] = io_u;
                                events++;
@@ -61,22 +61,20 @@ static void fio_gf_io_u_free(struct thread_data *td, struct 
io_u *io_u)
        struct fio_gf_iou *io = io_u->engine_data;
 
        if (io) {
-               if (io->io_complete) {
+               if (io->io_complete)
                        log_err("incomplete IO found.\n");
-               }
                io_u->engine_data = NULL;
                free(io);
        }
-       log_err("issued %lu finished %lu\n", issued, cb_count);
 }
 
 static int fio_gf_io_u_init(struct thread_data *td, struct io_u *io_u)
 {
-       struct fio_gf_iou *io = NULL;
-
        dprint(FD_FILE, "%s\n", __FUNCTION__);
 
        if (!io_u->engine_data) {
+               struct fio_gf_iou *io;
+
                io = malloc(sizeof(struct fio_gf_iou));
                if (!io) {
                        td_verror(td, errno, "malloc");
@@ -91,48 +89,44 @@ static int fio_gf_io_u_init(struct thread_data *td, struct 
io_u *io_u)
 
 static void gf_async_cb(glfs_fd_t * fd, ssize_t ret, void *data)
 {
-       struct io_u *io_u = (struct io_u *)data;
-       struct fio_gf_iou *iou = (struct fio_gf_iou *)io_u->engine_data;
+       struct io_u *io_u = data;
+       struct fio_gf_iou *iou = io_u->engine_data;
 
        dprint(FD_IO, "%s ret %lu\n", __FUNCTION__, ret);
        iou->io_complete = 1;
-       cb_count++;
 }
 
 static int fio_gf_async_queue(struct thread_data fio_unused * td,
                              struct io_u *io_u)
 {
        struct gf_data *g = td->io_ops->data;
-       int r = 0;
+       int r;
 
-       dprint(FD_IO, "%s op %s\n", __FUNCTION__,
-              io_u->ddir == DDIR_READ ? "read" : io_u->ddir ==
-              DDIR_WRITE ? "write" : io_u->ddir ==
-              DDIR_SYNC ? "sync" : "unknown");
+       dprint(FD_IO, "%s op %s\n", __FUNCTION__, io_ddir_name(io_u->ddir));
 
        fio_ro_check(td, io_u);
 
        if (io_u->ddir == DDIR_READ)
                r = glfs_pread_async(g->fd, io_u->xfer_buf, io_u->xfer_buflen,
-                                    io_u->offset, 0, gf_async_cb,
-                                    (void *)io_u);
+                                    io_u->offset, 0, gf_async_cb, io_u);
        else if (io_u->ddir == DDIR_WRITE)
                r = glfs_pwrite_async(g->fd, io_u->xfer_buf, io_u->xfer_buflen,
-                                     io_u->offset, 0, gf_async_cb,
-                                     (void *)io_u);
-       else if (io_u->ddir == DDIR_SYNC) {
-               r = glfs_fsync_async(g->fd, gf_async_cb, (void *)io_u);
-       } else {
-               log_err("unsupported operation.\n");
-               io_u->error = -EINVAL;
-               goto failed;
-       }
+                                     io_u->offset, 0, gf_async_cb, io_u);
+#if defined(CONFIG_GF_TRIM)
+       else if (io_u->ddir == DDIR_TRIM)
+               r = glfs_discard_async(g->fd, io_u->offset, io_u->xfer_buflen,
+                                      gf_async_cb, io_u);
+#endif
+       else if (io_u->ddir == DDIR_SYNC)
+               r = glfs_fsync_async(g->fd, gf_async_cb, io_u);
+       else
+               r = -EINVAL;
+
        if (r) {
-               log_err("glfs failed.\n");
+               log_err("glfs queue failed.\n");
                io_u->error = r;
                goto failed;
        }
-       issued++;
        return FIO_Q_QUEUED;
 
 failed:
@@ -143,29 +137,26 @@ failed:
 
 int fio_gf_async_setup(struct thread_data *td)
 {
-       int r = 0;
-       struct gf_data *g = NULL;
+       struct gf_data *g;
+       int r;
 
 #if defined(NOT_YET)
        log_err("the async interface is still very experimental...\n");
 #endif
        r = fio_gf_setup(td);
-       if (r) {
+       if (r)
                return r;
-       }
+
        td->o.use_thread = 1;
        g = td->io_ops->data;
-       g->aio_events = malloc(td->o.iodepth * sizeof(struct io_u *));
+       g->aio_events = calloc(td->o.iodepth, sizeof(struct io_u *));
        if (!g->aio_events) {
                r = -ENOMEM;
                fio_gf_cleanup(td);
                return r;
        }
 
-       memset(g->aio_events, 0, td->o.iodepth * sizeof(struct io_u *));
-
        return r;
-
 }
 
 static int fio_gf_async_prep(struct thread_data *td, struct io_u *io_u)
diff --git a/engines/rbd.c b/engines/rbd.c
index 5160c32..cf7be0a 100644
--- a/engines/rbd.c
+++ b/engines/rbd.c
@@ -12,7 +12,6 @@
 struct fio_rbd_iou {
        struct io_u *io_u;
        rbd_completion_t completion;
-       int io_complete;
        int io_seen;
 };
 
@@ -24,10 +23,10 @@ struct rbd_data {
 };
 
 struct rbd_options {
-       struct thread_data *td;
        char *rbd_name;
        char *pool_name;
        char *client_name;
+       int busy_poll;
 };
 
 static struct fio_option options[] = {
@@ -41,22 +40,32 @@ static struct fio_option options[] = {
                .group          = FIO_OPT_G_RBD,
        },
        {
-               .name     = "pool",
-               .lname    = "rbd engine pool",
-               .type     = FIO_OPT_STR_STORE,
-               .help     = "Name of the pool hosting the RBD for the RBD 
engine",
-               .off1     = offsetof(struct rbd_options, pool_name),
-               .category = FIO_OPT_C_ENGINE,
-               .group    = FIO_OPT_G_RBD,
+               .name           = "pool",
+               .lname          = "rbd engine pool",
+               .type           = FIO_OPT_STR_STORE,
+               .help           = "Name of the pool hosting the RBD for the RBD 
engine",
+               .off1           = offsetof(struct rbd_options, pool_name),
+               .category       = FIO_OPT_C_ENGINE,
+               .group          = FIO_OPT_G_RBD,
        },
        {
-               .name     = "clientname",
-               .lname    = "rbd engine clientname",
-               .type     = FIO_OPT_STR_STORE,
-               .help     = "Name of the ceph client to access the RBD for the 
RBD engine",
-               .off1     = offsetof(struct rbd_options, client_name),
-               .category = FIO_OPT_C_ENGINE,
-               .group    = FIO_OPT_G_RBD,
+               .name           = "clientname",
+               .lname          = "rbd engine clientname",
+               .type           = FIO_OPT_STR_STORE,
+               .help           = "Name of the ceph client to access the RBD 
for the RBD engine",
+               .off1           = offsetof(struct rbd_options, client_name),
+               .category       = FIO_OPT_C_ENGINE,
+               .group          = FIO_OPT_G_RBD,
+       },
+       {
+               .name           = "busy_poll",
+               .lname          = "Busy poll",
+               .type           = FIO_OPT_BOOL,
+               .help           = "Busy poll for completions instead of 
sleeping",
+               .off1           = offsetof(struct rbd_options, client_name),
+               .def            = "0",
+               .category       = FIO_OPT_C_ENGINE,
+               .group          = FIO_OPT_G_RBD,
        },
        {
                .name = NULL,
@@ -100,7 +109,7 @@ static int _fio_rbd_connect(struct thread_data *td)
        struct rbd_options *o = td->eo;
        int r;
 
-       r = rados_create(&(rbd_data->cluster), o->client_name);
+       r = rados_create(&rbd_data->cluster, o->client_name);
        if (r < 0) {
                log_err("rados_create failed.\n");
                goto failed_early;
@@ -119,13 +128,13 @@ static int _fio_rbd_connect(struct thread_data *td)
        }
 
        r = rados_ioctx_create(rbd_data->cluster, o->pool_name,
-                              &(rbd_data->io_ctx));
+                              &rbd_data->io_ctx);
        if (r < 0) {
                log_err("rados_ioctx_create failed.\n");
                goto failed_shutdown;
        }
 
-       r = rbd_open(rbd_data->io_ctx, o->rbd_name, &(rbd_data->image),
+       r = rbd_open(rbd_data->io_ctx, o->rbd_name, &rbd_data->image,
                     NULL /*snap */ );
        if (r < 0) {
                log_err("rbd_open failed.\n");
@@ -167,12 +176,10 @@ static void _fio_rbd_disconnect(struct rbd_data *rbd_data)
 
 static void _fio_rbd_finish_aiocb(rbd_completion_t comp, void *data)
 {
-       struct io_u *io_u = data;
-       struct fio_rbd_iou *fri = io_u->engine_data;
+       struct fio_rbd_iou *fri = data;
+       struct io_u *io_u = fri->io_u;
        ssize_t ret;
 
-       fri->io_complete = 1;
-
        /*
         * Looks like return value is 0 for success, or < 0 for
         * a specific error. So we have to assume that it can't do
@@ -199,8 +206,7 @@ static inline int fri_check_complete(struct rbd_data 
*rbd_data,
 {
        struct fio_rbd_iou *fri = io_u->engine_data;
 
-       if (fri->io_complete) {
-               fri->io_complete = 0;
+       if (rbd_aio_is_complete(fri->completion)) {
                fri->io_seen = 1;
                rbd_data->aio_events[*events] = io_u;
                (*events)++;
@@ -247,6 +253,7 @@ static int fio_rbd_getevents(struct thread_data *td, 
unsigned int min,
                             unsigned int max, const struct timespec *t)
 {
        unsigned int this_events, events = 0;
+       struct rbd_options *o = td->eo;
        int wait = 0;
 
        do {
@@ -257,7 +264,10 @@ static int fio_rbd_getevents(struct thread_data *td, 
unsigned int min,
                if (this_events)
                        continue;
 
-               wait = 1;
+               if (!o->busy_poll)
+                       wait = 1;
+               else
+                       nop;
        } while (1);
 
        return events;
@@ -271,70 +281,55 @@ static int fio_rbd_queue(struct thread_data *td, struct 
io_u *io_u)
 
        fio_ro_check(td, io_u);
 
-       fri->io_complete = 0;
        fri->io_seen = 0;
 
-       if (io_u->ddir == DDIR_WRITE) {
-               r = rbd_aio_create_completion(io_u, _fio_rbd_finish_aiocb,
+       r = rbd_aio_create_completion(fri, _fio_rbd_finish_aiocb,
                                                &fri->completion);
-               if (r < 0) {
-                       log_err
-                           ("rbd_aio_create_completion for DDIR_WRITE 
failed.\n");
-                       goto failed;
-               }
+       if (r < 0) {
+               log_err("rbd_aio_create_completion failed.\n");
+               goto failed;
+       }
 
+       if (io_u->ddir == DDIR_WRITE) {
                r = rbd_aio_write(rbd_data->image, io_u->offset,
                                  io_u->xfer_buflen, io_u->xfer_buf,
                                  fri->completion);
                if (r < 0) {
                        log_err("rbd_aio_write failed.\n");
-                       rbd_aio_release(fri->completion);
-                       goto failed;
+                       goto failed_comp;
                }
 
        } else if (io_u->ddir == DDIR_READ) {
-               r = rbd_aio_create_completion(io_u, _fio_rbd_finish_aiocb,
-                                               &fri->completion);
-               if (r < 0) {
-                       log_err
-                           ("rbd_aio_create_completion for DDIR_READ 
failed.\n");
-                       goto failed;
-               }
-
                r = rbd_aio_read(rbd_data->image, io_u->offset,
                                 io_u->xfer_buflen, io_u->xfer_buf,
                                 fri->completion);
 
                if (r < 0) {
                        log_err("rbd_aio_read failed.\n");
-                       rbd_aio_release(fri->completion);
-                       goto failed;
+                       goto failed_comp;
                }
-
-       } else if (io_u->ddir == DDIR_SYNC) {
-               r = rbd_aio_create_completion(io_u, _fio_rbd_finish_aiocb,
-                                               &fri->completion);
+       } else if (io_u->ddir == DDIR_TRIM) {
+               r = rbd_aio_discard(rbd_data->image, io_u->offset,
+                                io_u->xfer_buflen, fri->completion);
                if (r < 0) {
-                       log_err
-                           ("rbd_aio_create_completion for DDIR_SYNC 
failed.\n");
-                       goto failed;
+                       log_err("rbd_aio_discard failed.\n");
+                       goto failed_comp;
                }
-
+       } else if (io_u->ddir == DDIR_SYNC) {
                r = rbd_aio_flush(rbd_data->image, fri->completion);
                if (r < 0) {
                        log_err("rbd_flush failed.\n");
-                       rbd_aio_release(fri->completion);
-                       goto failed;
+                       goto failed_comp;
                }
-
        } else {
                dprint(FD_IO, "%s: Warning: unhandled ddir: %d\n", __func__,
                       io_u->ddir);
-               return FIO_Q_COMPLETED;
+               goto failed_comp;
        }
 
        return FIO_Q_QUEUED;
-
+failed_comp:
+       rbd_aio_release(fri->completion);
 failed:
        io_u->error = r;
        td_verror(td, io_u->error, "xfer");
@@ -445,7 +440,13 @@ static int fio_rbd_open(struct thread_data *td, struct 
fio_file *f)
 
 static int fio_rbd_invalidate(struct thread_data *td, struct fio_file *f)
 {
+#if defined(CONFIG_RBD_INVAL)
+       struct rbd_data *rbd_data = td->io_ops->data;
+
+       return rbd_invalidate_cache(rbd_data->image);
+#else
        return 0;
+#endif
 }
 
 static void fio_rbd_io_u_free(struct thread_data *td, struct io_u *io_u)
--
To unsubscribe from this list: send the line "unsubscribe fio" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to