summaryrefslogtreecommitdiffstats
path: root/fs
diff options
context:
space:
mode:
authorZheng Yan <zheng.yan@oracle.com>2008-09-26 10:05:48 -0400
committerChris Mason <chris.mason@oracle.com>2008-09-26 10:05:48 -0400
commite8569813849b5da394a195c7e76b4faa452b12d1 (patch)
tree87b09cd5bfc3dda16e181c247fdd7f77b6c463ea /fs
parent24ab9cd85c11bccacbd0cce7f8e1aebd4930404c (diff)
downloadop-kernel-dev-e8569813849b5da394a195c7e76b4faa452b12d1.zip
op-kernel-dev-e8569813849b5da394a195c7e76b4faa452b12d1.tar.gz
Btrfs: allocator fixes for space balancing update
* Reserved extent accounting: reserved extents have been allocated in the rbtrees that track free space but have not been allocated on disk. They were never properly accounted for in the past, making it hard to know how much space was really free. * btrfs_find_block_group used to return NULL for block groups that had been removed by the space balancing code. This made it hard to account for space during the final stages of a balance run. Signed-off-by: Chris Mason <chris.mason@oracle.com>
Diffstat (limited to 'fs')
-rw-r--r--fs/btrfs/ctree.h2
-rw-r--r--fs/btrfs/extent-tree.c136
2 files changed, 67 insertions, 71 deletions
diff --git a/fs/btrfs/ctree.h b/fs/btrfs/ctree.h
index 3b3c1ca..c683aaa 100644
--- a/fs/btrfs/ctree.h
+++ b/fs/btrfs/ctree.h
@@ -498,6 +498,7 @@ struct btrfs_space_info {
u64 total_bytes;
u64 bytes_used;
u64 bytes_pinned;
+ u64 bytes_reserved;
int full;
int force_alloc;
struct list_head list;
@@ -519,6 +520,7 @@ struct btrfs_block_group_cache {
struct btrfs_block_group_item item;
spinlock_t lock;
u64 pinned;
+ u64 reserved;
u64 flags;
int cached;
int ro;
diff --git a/fs/btrfs/extent-tree.c b/fs/btrfs/extent-tree.c
index fe4e11b..3e2f969d 100644
--- a/fs/btrfs/extent-tree.c
+++ b/fs/btrfs/extent-tree.c
@@ -325,12 +325,9 @@ static int noinline find_free_space(struct btrfs_root *root,
struct btrfs_block_group_cache *cache = *cache_ret;
struct btrfs_free_space *info = NULL;
u64 last;
- u64 total_fs_bytes;
u64 search_start = *start_ret;
WARN_ON(!mutex_is_locked(&root->fs_info->alloc_mutex));
- total_fs_bytes = btrfs_super_total_bytes(&root->fs_info->super_copy);
-
if (!cache)
goto out;
@@ -354,7 +351,7 @@ new_group:
last = cache->key.objectid + cache->key.offset;
cache = btrfs_lookup_first_block_group(root->fs_info, last);
- if (!cache || cache->key.objectid >= total_fs_bytes)
+ if (!cache)
goto out;
*cache_ret = cache;
@@ -385,7 +382,6 @@ static struct btrfs_space_info *__find_space_info(struct btrfs_fs_info *info,
return found;
}
return NULL;
-
}
static struct btrfs_block_group_cache *
@@ -396,7 +392,6 @@ __btrfs_find_block_group(struct btrfs_root *root,
struct btrfs_block_group_cache *cache;
struct btrfs_block_group_cache *found_group = NULL;
struct btrfs_fs_info *info = root->fs_info;
- struct btrfs_space_info *sinfo;
u64 used;
u64 last = 0;
u64 free_check;
@@ -413,7 +408,7 @@ __btrfs_find_block_group(struct btrfs_root *root,
if (shint && block_group_bits(shint, data) && !shint->ro) {
spin_lock(&shint->lock);
used = btrfs_block_group_used(&shint->item);
- if (used + shint->pinned <
+ if (used + shint->pinned + shint->reserved <
div_factor(shint->key.offset, factor)) {
spin_unlock(&shint->lock);
return shint;
@@ -424,7 +419,7 @@ __btrfs_find_block_group(struct btrfs_root *root,
if (hint && !hint->ro && block_group_bits(hint, data)) {
spin_lock(&hint->lock);
used = btrfs_block_group_used(&hint->item);
- if (used + hint->pinned <
+ if (used + hint->pinned + hint->reserved <
div_factor(hint->key.offset, factor)) {
spin_unlock(&hint->lock);
return hint;
@@ -437,27 +432,9 @@ __btrfs_find_block_group(struct btrfs_root *root,
else
last = search_start;
}
- sinfo = __find_space_info(root->fs_info, data);
- if (!sinfo)
- goto found;
again:
- while(1) {
- struct list_head *l;
-
- cache = NULL;
-
- spin_lock(&sinfo->lock);
- list_for_each(l, &sinfo->block_groups) {
- struct btrfs_block_group_cache *entry;
- entry = list_entry(l, struct btrfs_block_group_cache,
- list);
- if ((entry->key.objectid >= last) &&
- (!cache || (entry->key.objectid <
- cache->key.objectid)))
- cache = entry;
- }
- spin_unlock(&sinfo->lock);
-
+ while (1) {
+ cache = btrfs_lookup_first_block_group(root->fs_info, last);
if (!cache)
break;
@@ -467,7 +444,8 @@ again:
if (!cache->ro && block_group_bits(cache, data)) {
free_check = div_factor(cache->key.offset, factor);
- if (used + cache->pinned < free_check) {
+ if (used + cache->pinned + cache->reserved <
+ free_check) {
found_group = cache;
spin_unlock(&cache->lock);
goto found;
@@ -1414,6 +1392,7 @@ int btrfs_write_dirty_block_groups(struct btrfs_trans_handle *trans,
if (!cache)
break;
+ cache->dirty = 0;
last += cache->key.offset;
err = write_one_cache_group(trans, root,
@@ -1427,8 +1406,6 @@ int btrfs_write_dirty_block_groups(struct btrfs_trans_handle *trans,
werr = err;
continue;
}
-
- cache->dirty = 0;
}
btrfs_free_path(path);
mutex_unlock(&root->fs_info->alloc_mutex);
@@ -1460,6 +1437,7 @@ static int update_space_info(struct btrfs_fs_info *info, u64 flags,
found->total_bytes = total_bytes;
found->bytes_used = bytes_used;
found->bytes_pinned = 0;
+ found->bytes_reserved = 0;
found->full = 0;
found->force_alloc = 0;
*space_info = found;
@@ -1539,8 +1517,8 @@ static int do_chunk_alloc(struct btrfs_trans_handle *trans,
thresh = div_factor(space_info->total_bytes, 6);
if (!force &&
- (space_info->bytes_used + space_info->bytes_pinned + alloc_bytes) <
- thresh)
+ (space_info->bytes_used + space_info->bytes_pinned +
+ space_info->bytes_reserved + alloc_bytes) < thresh)
goto out;
mutex_lock(&extent_root->fs_info->chunk_mutex);
@@ -1621,7 +1599,6 @@ static u64 first_logical_byte(struct btrfs_root *root, u64 search_start)
return cache->key.objectid;
}
-
int btrfs_update_pinned_extents(struct btrfs_root *root,
u64 bytenr, u64 num, int pin)
{
@@ -1639,29 +1616,20 @@ int btrfs_update_pinned_extents(struct btrfs_root *root,
}
while (num > 0) {
cache = btrfs_lookup_block_group(fs_info, bytenr);
- if (!cache) {
- u64 first = first_logical_byte(root, bytenr);
- WARN_ON(first < bytenr);
- len = min(first - bytenr, num);
- } else {
- len = min(num, cache->key.offset -
- (bytenr - cache->key.objectid));
- }
+ BUG_ON(!cache);
+ len = min(num, cache->key.offset -
+ (bytenr - cache->key.objectid));
if (pin) {
- if (cache) {
- spin_lock(&cache->lock);
- cache->pinned += len;
- cache->space_info->bytes_pinned += len;
- spin_unlock(&cache->lock);
- }
+ spin_lock(&cache->lock);
+ cache->pinned += len;
+ cache->space_info->bytes_pinned += len;
+ spin_unlock(&cache->lock);
fs_info->total_pinned += len;
} else {
- if (cache) {
- spin_lock(&cache->lock);
- cache->pinned -= len;
- cache->space_info->bytes_pinned -= len;
- spin_unlock(&cache->lock);
- }
+ spin_lock(&cache->lock);
+ cache->pinned -= len;
+ cache->space_info->bytes_pinned -= len;
+ spin_unlock(&cache->lock);
fs_info->total_pinned -= len;
}
bytenr += len;
@@ -1670,6 +1638,36 @@ int btrfs_update_pinned_extents(struct btrfs_root *root,
return 0;
}
+static int update_reserved_extents(struct btrfs_root *root,
+ u64 bytenr, u64 num, int reserve)
+{
+ u64 len;
+ struct btrfs_block_group_cache *cache;
+ struct btrfs_fs_info *fs_info = root->fs_info;
+
+ WARN_ON(!mutex_is_locked(&root->fs_info->alloc_mutex));
+ while (num > 0) {
+ cache = btrfs_lookup_block_group(fs_info, bytenr);
+ BUG_ON(!cache);
+ len = min(num, cache->key.offset -
+ (bytenr - cache->key.objectid));
+ if (reserve) {
+ spin_lock(&cache->lock);
+ cache->reserved += len;
+ cache->space_info->bytes_reserved += len;
+ spin_unlock(&cache->lock);
+ } else {
+ spin_lock(&cache->lock);
+ cache->reserved -= len;
+ cache->space_info->bytes_reserved -= len;
+ spin_unlock(&cache->lock);
+ }
+ bytenr += len;
+ num -= len;
+ }
+ return 0;
+}
+
int btrfs_copy_pinned(struct btrfs_root *root, struct extent_io_tree *copy)
{
u64 last = 0;
@@ -2126,6 +2124,7 @@ static int __btrfs_free_extent(struct btrfs_trans_handle *trans,
cache = btrfs_lookup_block_group(root->fs_info, bytenr);
BUG_ON(!cache);
btrfs_add_free_space(cache, bytenr, num_bytes);
+ update_reserved_extents(root, bytenr, num_bytes, 0);
return 0;
}
pin = 1;
@@ -2225,14 +2224,11 @@ static int noinline find_free_extent(struct btrfs_trans_handle *trans,
search_start = max(search_start, first_logical_byte(root, 0));
orig_search_start = search_start;
- if (search_end == (u64)-1)
- search_end = btrfs_super_total_bytes(&info->super_copy);
-
search_start = max(search_start, hint_byte);
total_needed += empty_size;
new_group:
- block_group = btrfs_lookup_block_group(info, search_start);
+ block_group = btrfs_lookup_first_block_group(info, search_start);
/*
* Ok this looks a little tricky, buts its really simple. First if we
@@ -2257,12 +2253,8 @@ new_group:
ret = do_chunk_alloc(trans, root,
num_bytes + 2 * 1024 * 1024,
data, 1);
- if (ret < 0) {
- struct btrfs_space_info *info;
-
- info = __find_space_info(root->fs_info, data);
+ if (ret < 0)
goto error;
- }
BUG_ON(ret);
chunk_alloc_done = 1;
search_start = orig_search_start;
@@ -2378,22 +2370,24 @@ static void dump_space_info(struct btrfs_space_info *info, u64 bytes)
struct list_head *l;
printk(KERN_INFO "space_info has %Lu free, is %sfull\n",
- info->total_bytes - info->bytes_used - info->bytes_pinned,
- (info->full) ? "" : "not ");
+ info->total_bytes - info->bytes_used - info->bytes_pinned -
+ info->bytes_reserved, (info->full) ? "" : "not ");
spin_lock(&info->lock);
list_for_each(l, &info->block_groups) {
cache = list_entry(l, struct btrfs_block_group_cache, list);
spin_lock(&cache->lock);
printk(KERN_INFO "block group %Lu has %Lu bytes, %Lu used "
- "%Lu pinned\n",
+ "%Lu pinned %Lu reserved\n",
cache->key.objectid, cache->key.offset,
- btrfs_block_group_used(&cache->item), cache->pinned);
+ btrfs_block_group_used(&cache->item),
+ cache->pinned, cache->reserved);
btrfs_dump_free_space(cache, bytes);
spin_unlock(&cache->lock);
}
spin_unlock(&info->lock);
}
+
static int __btrfs_reserve_extent(struct btrfs_trans_handle *trans,
struct btrfs_root *root,
u64 num_bytes, u64 min_alloc_size,
@@ -2500,6 +2494,7 @@ int btrfs_reserve_extent(struct btrfs_trans_handle *trans,
ret = __btrfs_reserve_extent(trans, root, num_bytes, min_alloc_size,
empty_size, hint_byte, search_end, ins,
data);
+ update_reserved_extents(root, ins->objectid, ins->offset, 1);
maybe_unlock_mutex(root);
return ret;
}
@@ -2625,6 +2620,7 @@ int btrfs_alloc_reserved_extent(struct btrfs_trans_handle *trans,
ret = __btrfs_alloc_reserved_extent(trans, root, parent,
root_objectid, ref_generation,
owner, owner_offset, ins);
+ update_reserved_extents(root, ins->objectid, ins->offset, 0);
maybe_unlock_mutex(root);
return ret;
}
@@ -2685,6 +2681,8 @@ int btrfs_alloc_extent(struct btrfs_trans_handle *trans,
owner_objectid, owner_offset, ins);
BUG_ON(ret);
+ } else {
+ update_reserved_extents(root, ins->objectid, ins->offset, 1);
}
maybe_unlock_mutex(root);
return ret;
@@ -3974,10 +3972,6 @@ int btrfs_read_block_groups(struct btrfs_root *root)
ret = btrfs_add_block_group_cache(root->fs_info, cache);
BUG_ON(ret);
-
- if (key.objectid >=
- btrfs_super_total_bytes(&info->super_copy))
- break;
}
ret = 0;
error:
OpenPOWER on IntegriCloud