Hi,

On Mon, May 23, 2011 at 02:59:06PM +0200, Arne Jansen wrote:
> +static struct reada_zone *reada_find_zone(struct btrfs_fs_info *fs_info,
> +                                       struct btrfs_device *dev, u64 logical,
> +                                       struct btrfs_multi_bio *multi)
> +{
> +     int ret;
> +     int looped = 0;
> +     struct reada_zone *zone;
> +     struct btrfs_block_group_cache *cache = NULL;
> +     u64 start;
> +     u64 end;
> +     int i;
> +
> +again:
> +     zone = NULL;
> +     spin_lock(&fs_info->reada_lock);
> +     ret = radix_tree_gang_lookup(&dev->reada_zones, (void **)&zone,
> +                                  logical >> PAGE_CACHE_SHIFT, 1);
> +     if (ret == 1)
> +             kref_get(&zone->refcnt);
> +     spin_unlock(&fs_info->reada_lock);
> +
> +     if (ret == 1) {
> +             if (logical >= zone->start && logical < zone->end)
> +                     return zone;
> +             spin_lock(&fs_info->reada_lock);
> +             reada_zone_put(zone);
> +             spin_unlock(&fs_info->reada_lock);
> +     }
> +
> +     if (looped)
> +             return NULL;
> +
> +     cache = btrfs_lookup_block_group(fs_info, logical);
> +     if (!cache)
> +             return NULL;
> +
> +     start = cache->key.objectid;
> +     end = start + cache->key.offset - 1;
> +     btrfs_put_block_group(cache);
> +
> +     zone = kzalloc(sizeof(*zone), GFP_NOFS);
> +     if (!zone)
> +             return NULL;
> +
> +     zone->start = start;
> +     zone->end = end;
> +     INIT_LIST_HEAD(&zone->list);
> +     spin_lock_init(&zone->lock);
> +     zone->locked = 0;
> +     kref_init(&zone->refcnt);
> +     zone->elems = 0;
> +     zone->device = dev; /* our device always sits at index 0 */
> +     for (i = 0; i < multi->num_stripes; ++i) {
> +             /* bounds have already been checked */
> +             zone->devs[i] = multi->stripes[i].dev;
> +     }
> +     zone->ndevs = multi->num_stripes;
> +
> +     spin_lock(&fs_info->reada_lock);
> +     ret = radix_tree_insert(&dev->reada_zones,
> +                             (unsigned long)zone->end >> PAGE_CACHE_SHIFT,
> +                             zone);

this can sleep inside a spinlock, you initialize the radix tree with
GFP_NOFS, which allows __GFP_WAIT.

Options:
1) use GFP_ATOMIC in radix tree init flags
2) do the radix_tree_preload/radix_tree_preload_end, GFP_NOFS outside of the
locked section is ok but __GFP_WAIT has to be masked out (else radix
tree insert will not use the preloaded node)
3) unmask __GFP_WAIT from radix tree init flags

I'd go for 3, as the atomic context is not required, and is easier
than 2 to implement.

> +     spin_unlock(&fs_info->reada_lock);
> +
> +     if (ret) {
> +             kfree(zone);
> +             looped = 1;
> +             goto again;
> +     }
> +
> +     return zone;
> +}
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to