Skip to content

Commit

Permalink
Btrfs: allocator fixes for space balancing update
Browse files Browse the repository at this point in the history
* Reserved extent accounting:  reserved extents have been
allocated in the rbtrees that track free space but have not
been allocated on disk.  They were never properly accounted for
in the past, making it hard to know how much space was really free.

* btrfs_find_block_group used to return NULL for block groups that
had been removed by the space balancing code.  This made it hard
to account for space during the final stages of a balance run.

Signed-off-by: Chris Mason <chris.mason@oracle.com>
  • Loading branch information
Zheng Yan authored and Chris Mason committed Sep 26, 2008
1 parent 24ab9cd commit e856981
Show file tree
Hide file tree
Showing 2 changed files with 67 additions and 71 deletions.
2 changes: 2 additions & 0 deletions fs/btrfs/ctree.h
Original file line number Diff line number Diff line change
Expand Up @@ -498,6 +498,7 @@ struct btrfs_space_info {
u64 total_bytes;
u64 bytes_used;
u64 bytes_pinned;
u64 bytes_reserved;
int full;
int force_alloc;
struct list_head list;
Expand All @@ -519,6 +520,7 @@ struct btrfs_block_group_cache {
struct btrfs_block_group_item item;
spinlock_t lock;
u64 pinned;
u64 reserved;
u64 flags;
int cached;
int ro;
Expand Down
136 changes: 65 additions & 71 deletions fs/btrfs/extent-tree.c
Original file line number Diff line number Diff line change
Expand Up @@ -325,12 +325,9 @@ static int noinline find_free_space(struct btrfs_root *root,
struct btrfs_block_group_cache *cache = *cache_ret;
struct btrfs_free_space *info = NULL;
u64 last;
u64 total_fs_bytes;
u64 search_start = *start_ret;

WARN_ON(!mutex_is_locked(&root->fs_info->alloc_mutex));
total_fs_bytes = btrfs_super_total_bytes(&root->fs_info->super_copy);

if (!cache)
goto out;

Expand All @@ -354,7 +351,7 @@ static int noinline find_free_space(struct btrfs_root *root,
last = cache->key.objectid + cache->key.offset;

cache = btrfs_lookup_first_block_group(root->fs_info, last);
if (!cache || cache->key.objectid >= total_fs_bytes)
if (!cache)
goto out;

*cache_ret = cache;
Expand Down Expand Up @@ -385,7 +382,6 @@ static struct btrfs_space_info *__find_space_info(struct btrfs_fs_info *info,
return found;
}
return NULL;

}

static struct btrfs_block_group_cache *
Expand All @@ -396,7 +392,6 @@ __btrfs_find_block_group(struct btrfs_root *root,
struct btrfs_block_group_cache *cache;
struct btrfs_block_group_cache *found_group = NULL;
struct btrfs_fs_info *info = root->fs_info;
struct btrfs_space_info *sinfo;
u64 used;
u64 last = 0;
u64 free_check;
Expand All @@ -413,7 +408,7 @@ __btrfs_find_block_group(struct btrfs_root *root,
if (shint && block_group_bits(shint, data) && !shint->ro) {
spin_lock(&shint->lock);
used = btrfs_block_group_used(&shint->item);
if (used + shint->pinned <
if (used + shint->pinned + shint->reserved <
div_factor(shint->key.offset, factor)) {
spin_unlock(&shint->lock);
return shint;
Expand All @@ -424,7 +419,7 @@ __btrfs_find_block_group(struct btrfs_root *root,
if (hint && !hint->ro && block_group_bits(hint, data)) {
spin_lock(&hint->lock);
used = btrfs_block_group_used(&hint->item);
if (used + hint->pinned <
if (used + hint->pinned + hint->reserved <
div_factor(hint->key.offset, factor)) {
spin_unlock(&hint->lock);
return hint;
Expand All @@ -437,27 +432,9 @@ __btrfs_find_block_group(struct btrfs_root *root,
else
last = search_start;
}
sinfo = __find_space_info(root->fs_info, data);
if (!sinfo)
goto found;
again:
while(1) {
struct list_head *l;

cache = NULL;

spin_lock(&sinfo->lock);
list_for_each(l, &sinfo->block_groups) {
struct btrfs_block_group_cache *entry;
entry = list_entry(l, struct btrfs_block_group_cache,
list);
if ((entry->key.objectid >= last) &&
(!cache || (entry->key.objectid <
cache->key.objectid)))
cache = entry;
}
spin_unlock(&sinfo->lock);

while (1) {
cache = btrfs_lookup_first_block_group(root->fs_info, last);
if (!cache)
break;

Expand All @@ -467,7 +444,8 @@ __btrfs_find_block_group(struct btrfs_root *root,

if (!cache->ro && block_group_bits(cache, data)) {
free_check = div_factor(cache->key.offset, factor);
if (used + cache->pinned < free_check) {
if (used + cache->pinned + cache->reserved <
free_check) {
found_group = cache;
spin_unlock(&cache->lock);
goto found;
Expand Down Expand Up @@ -1414,6 +1392,7 @@ int btrfs_write_dirty_block_groups(struct btrfs_trans_handle *trans,
if (!cache)
break;

cache->dirty = 0;
last += cache->key.offset;

err = write_one_cache_group(trans, root,
Expand All @@ -1427,8 +1406,6 @@ int btrfs_write_dirty_block_groups(struct btrfs_trans_handle *trans,
werr = err;
continue;
}

cache->dirty = 0;
}
btrfs_free_path(path);
mutex_unlock(&root->fs_info->alloc_mutex);
Expand Down Expand Up @@ -1460,6 +1437,7 @@ static int update_space_info(struct btrfs_fs_info *info, u64 flags,
found->total_bytes = total_bytes;
found->bytes_used = bytes_used;
found->bytes_pinned = 0;
found->bytes_reserved = 0;
found->full = 0;
found->force_alloc = 0;
*space_info = found;
Expand Down Expand Up @@ -1539,8 +1517,8 @@ static int do_chunk_alloc(struct btrfs_trans_handle *trans,

thresh = div_factor(space_info->total_bytes, 6);
if (!force &&
(space_info->bytes_used + space_info->bytes_pinned + alloc_bytes) <
thresh)
(space_info->bytes_used + space_info->bytes_pinned +
space_info->bytes_reserved + alloc_bytes) < thresh)
goto out;

mutex_lock(&extent_root->fs_info->chunk_mutex);
Expand Down Expand Up @@ -1621,7 +1599,6 @@ static u64 first_logical_byte(struct btrfs_root *root, u64 search_start)
return cache->key.objectid;
}


int btrfs_update_pinned_extents(struct btrfs_root *root,
u64 bytenr, u64 num, int pin)
{
Expand All @@ -1639,29 +1616,20 @@ int btrfs_update_pinned_extents(struct btrfs_root *root,
}
while (num > 0) {
cache = btrfs_lookup_block_group(fs_info, bytenr);
if (!cache) {
u64 first = first_logical_byte(root, bytenr);
WARN_ON(first < bytenr);
len = min(first - bytenr, num);
} else {
len = min(num, cache->key.offset -
(bytenr - cache->key.objectid));
}
BUG_ON(!cache);
len = min(num, cache->key.offset -
(bytenr - cache->key.objectid));
if (pin) {
if (cache) {
spin_lock(&cache->lock);
cache->pinned += len;
cache->space_info->bytes_pinned += len;
spin_unlock(&cache->lock);
}
spin_lock(&cache->lock);
cache->pinned += len;
cache->space_info->bytes_pinned += len;
spin_unlock(&cache->lock);
fs_info->total_pinned += len;
} else {
if (cache) {
spin_lock(&cache->lock);
cache->pinned -= len;
cache->space_info->bytes_pinned -= len;
spin_unlock(&cache->lock);
}
spin_lock(&cache->lock);
cache->pinned -= len;
cache->space_info->bytes_pinned -= len;
spin_unlock(&cache->lock);
fs_info->total_pinned -= len;
}
bytenr += len;
Expand All @@ -1670,6 +1638,36 @@ int btrfs_update_pinned_extents(struct btrfs_root *root,
return 0;
}

static int update_reserved_extents(struct btrfs_root *root,
u64 bytenr, u64 num, int reserve)
{
u64 len;
struct btrfs_block_group_cache *cache;
struct btrfs_fs_info *fs_info = root->fs_info;

WARN_ON(!mutex_is_locked(&root->fs_info->alloc_mutex));
while (num > 0) {
cache = btrfs_lookup_block_group(fs_info, bytenr);
BUG_ON(!cache);
len = min(num, cache->key.offset -
(bytenr - cache->key.objectid));
if (reserve) {
spin_lock(&cache->lock);
cache->reserved += len;
cache->space_info->bytes_reserved += len;
spin_unlock(&cache->lock);
} else {
spin_lock(&cache->lock);
cache->reserved -= len;
cache->space_info->bytes_reserved -= len;
spin_unlock(&cache->lock);
}
bytenr += len;
num -= len;
}
return 0;
}

int btrfs_copy_pinned(struct btrfs_root *root, struct extent_io_tree *copy)
{
u64 last = 0;
Expand Down Expand Up @@ -2126,6 +2124,7 @@ static int __btrfs_free_extent(struct btrfs_trans_handle *trans,
cache = btrfs_lookup_block_group(root->fs_info, bytenr);
BUG_ON(!cache);
btrfs_add_free_space(cache, bytenr, num_bytes);
update_reserved_extents(root, bytenr, num_bytes, 0);
return 0;
}
pin = 1;
Expand Down Expand Up @@ -2225,14 +2224,11 @@ static int noinline find_free_extent(struct btrfs_trans_handle *trans,
search_start = max(search_start, first_logical_byte(root, 0));
orig_search_start = search_start;

if (search_end == (u64)-1)
search_end = btrfs_super_total_bytes(&info->super_copy);

search_start = max(search_start, hint_byte);
total_needed += empty_size;

new_group:
block_group = btrfs_lookup_block_group(info, search_start);
block_group = btrfs_lookup_first_block_group(info, search_start);

/*
* Ok this looks a little tricky, buts its really simple. First if we
Expand All @@ -2257,12 +2253,8 @@ static int noinline find_free_extent(struct btrfs_trans_handle *trans,
ret = do_chunk_alloc(trans, root,
num_bytes + 2 * 1024 * 1024,
data, 1);
if (ret < 0) {
struct btrfs_space_info *info;

info = __find_space_info(root->fs_info, data);
if (ret < 0)
goto error;
}
BUG_ON(ret);
chunk_alloc_done = 1;
search_start = orig_search_start;
Expand Down Expand Up @@ -2378,22 +2370,24 @@ static void dump_space_info(struct btrfs_space_info *info, u64 bytes)
struct list_head *l;

printk(KERN_INFO "space_info has %Lu free, is %sfull\n",
info->total_bytes - info->bytes_used - info->bytes_pinned,
(info->full) ? "" : "not ");
info->total_bytes - info->bytes_used - info->bytes_pinned -
info->bytes_reserved, (info->full) ? "" : "not ");

spin_lock(&info->lock);
list_for_each(l, &info->block_groups) {
cache = list_entry(l, struct btrfs_block_group_cache, list);
spin_lock(&cache->lock);
printk(KERN_INFO "block group %Lu has %Lu bytes, %Lu used "
"%Lu pinned\n",
"%Lu pinned %Lu reserved\n",
cache->key.objectid, cache->key.offset,
btrfs_block_group_used(&cache->item), cache->pinned);
btrfs_block_group_used(&cache->item),
cache->pinned, cache->reserved);
btrfs_dump_free_space(cache, bytes);
spin_unlock(&cache->lock);
}
spin_unlock(&info->lock);
}

static int __btrfs_reserve_extent(struct btrfs_trans_handle *trans,
struct btrfs_root *root,
u64 num_bytes, u64 min_alloc_size,
Expand Down Expand Up @@ -2500,6 +2494,7 @@ int btrfs_reserve_extent(struct btrfs_trans_handle *trans,
ret = __btrfs_reserve_extent(trans, root, num_bytes, min_alloc_size,
empty_size, hint_byte, search_end, ins,
data);
update_reserved_extents(root, ins->objectid, ins->offset, 1);
maybe_unlock_mutex(root);
return ret;
}
Expand Down Expand Up @@ -2625,6 +2620,7 @@ int btrfs_alloc_reserved_extent(struct btrfs_trans_handle *trans,
ret = __btrfs_alloc_reserved_extent(trans, root, parent,
root_objectid, ref_generation,
owner, owner_offset, ins);
update_reserved_extents(root, ins->objectid, ins->offset, 0);
maybe_unlock_mutex(root);
return ret;
}
Expand Down Expand Up @@ -2685,6 +2681,8 @@ int btrfs_alloc_extent(struct btrfs_trans_handle *trans,
owner_objectid, owner_offset, ins);
BUG_ON(ret);

} else {
update_reserved_extents(root, ins->objectid, ins->offset, 1);
}
maybe_unlock_mutex(root);
return ret;
Expand Down Expand Up @@ -3974,10 +3972,6 @@ int btrfs_read_block_groups(struct btrfs_root *root)

ret = btrfs_add_block_group_cache(root->fs_info, cache);
BUG_ON(ret);

if (key.objectid >=
btrfs_super_total_bytes(&info->super_copy))
break;
}
ret = 0;
error:
Expand Down

0 comments on commit e856981

Please sign in to comment.