@@ -357,6 +357,7 @@ struct ext4_new_group_data {
#define EXT4_IOC_ALLOC_DA_BLKS _IO('f', 12)
#define EXT4_IOC_MOVE_EXT _IOWR('f', 15, struct move_extent)
#define EXT4_IOC_ADD_GLOBAL_ALLOC_RULE _IOW('f', 16, struct ext4_alloc_rule)
+#define EXT4_IOC_CLR_GLOBAL_ALLOC_RULE _IOW('f', 17, struct ext4_alloc_rule)
/*
* ioctl commands in 32 bit emulation
@@ -1386,6 +1387,8 @@ extern int ext4_mb_get_buddy_cache_lock(struct super_block *, ext4_group_t);
extern void ext4_mb_put_buddy_cache_lock(struct super_block *,
ext4_group_t, int);
extern int ext4_mb_add_global_arule(struct inode *, struct ext4_alloc_rule *);
+extern int ext4_mb_del_global_arule(struct inode *, struct ext4_alloc_rule *);
+extern void ext4_mb_release_arule_list(struct ext4_sb_info *);
/* inode.c */
int ext4_forget(handle_t *handle, int is_metadata, struct inode *inode,
@@ -276,6 +276,22 @@ setversion_out:
return err;
}
+ case EXT4_IOC_CLR_GLOBAL_ALLOC_RULE: {
+ struct ext4_alloc_rule arule;
+ int err;
+
+ if (copy_from_user(&arule,
+ (struct ext4_alloc_rule __user *)arg,
+ sizeof(arule)))
+ return -EFAULT;
+
+ if (!capable(CAP_SYS_ADMIN))
+ return -EPERM;
+
+ err = ext4_mb_del_global_arule(inode, &arule);
+ return err;
+ }
+
case EXT4_IOC_GROUP_ADD: {
struct ext4_new_group_data input;
struct super_block *sb = inode->i_sb;
@@ -2928,6 +2928,7 @@ int ext4_mb_release(struct super_block *sb)
free_percpu(sbi->s_locality_groups);
ext4_mb_history_release(sb);
+ ext4_mb_release_arule_list(sbi);
return 0;
}
@@ -5036,6 +5037,28 @@ static void ext4_mb_release_tmp_list(struct list_head *list,
return;
}
+void ext4_mb_release_arule_list(struct ext4_sb_info *sbi)
+{
+ struct ext4_bg_alloc_rule_list *bg_arule_list, *tmp_arule_list;
+ struct ext4_bg_alloc_rule *bg_arule, *tmp_arule;
+
+ list_for_each_entry_safe(bg_arule_list, tmp_arule_list,
+ &sbi->s_bg_arule_list, bg_arule_list){
+ list_for_each_entry_safe(bg_arule, tmp_arule,
+ &bg_arule_list->arule_list, arule_list) {
+ list_del(&bg_arule->arule_list);
+ kfree(bg_arule);
+ bg_arule = NULL;
+ }
+
+ list_del(&bg_arule_list->bg_arule_list);
+ kfree(bg_arule_list);
+ bg_arule_list = NULL;
+ }
+
+ return;
+}
+
static int ext4_mb_check_arule(struct inode *inode,
struct ext4_alloc_rule *arule)
{
@@ -5296,3 +5319,166 @@ out:
write_unlock(&sbi->s_bg_arule_lock);
return 0;
}
+
+static int ext4_mb_del_bg_arule(struct super_block *sb,
+ struct ext4_bg_alloc_rule_list *bg_arule_list,
+ struct ext4_bg_alloc_rule *del_arule)
+{
+ struct ext4_sb_info *sbi = EXT4_SB(sb);
+ struct ext4_bg_alloc_rule *bg_arule, *tmp_arule, *add_arule = NULL;
+ struct ext4_buddy e4b;
+ void *bd_bitmap;
+ ext4_grpblk_t freed_blocks = 0;
+ ext4_group_t group_no = bg_arule_list->bg_num;
+ int err;
+
+ err = ext4_mb_load_buddy(sb, group_no, &e4b);
+ if (err)
+ return err;
+
+ ext4_lock_group(sb, group_no);
+ bd_bitmap = e4b.bd_bitmap;
+
+ list_for_each_entry_safe(bg_arule, tmp_arule,
+ &bg_arule_list->arule_list, arule_list) {
+ /*
+ * bg_arule |-----|
+ * del_arule |---------|
+ */
+ if (del_arule->start <= bg_arule->start &&
+ del_arule->end >= bg_arule->end) {
+ freed_blocks = ext4_mb_count_unused_blocks(bd_bitmap,
+ bg_arule->start, bg_arule->end);
+ ext4_mb_calc_restricted(sbi, &bg_arule_list,
+ bg_arule->alloc_flag,
+ -freed_blocks);
+ list_del(&bg_arule->arule_list);
+ kfree(bg_arule);
+ bg_arule = NULL;
+ /*
+ * bg_arule |-----|
+ * del_arule |---|
+ */
+ } else if (del_arule->start <= bg_arule->start &&
+ del_arule->end >= bg_arule->start) {
+ freed_blocks = ext4_mb_count_unused_blocks(bd_bitmap,
+ bg_arule->start, del_arule->end);
+ ext4_mb_calc_restricted(sbi, &bg_arule_list,
+ bg_arule->alloc_flag,
+ -freed_blocks);
+ bg_arule->start = del_arule->end + 1;
+ /*
+ * bg_arule |-----|
+ * del_arule |---|
+ */
+ } else if (del_arule->start <= bg_arule->end &&
+ del_arule->end >= bg_arule->end) {
+ freed_blocks = ext4_mb_count_unused_blocks(bd_bitmap,
+ del_arule->start, bg_arule->end);
+ ext4_mb_calc_restricted(sbi, &bg_arule_list,
+ bg_arule->alloc_flag,
+ -freed_blocks);
+ bg_arule->end = del_arule->start - 1;
+ /*
+ * bg_arule |-----|
+ * del_arule |---|
+ */
+ } else if (del_arule->start > bg_arule->start &&
+ del_arule->end < bg_arule->end) {
+ add_arule = kmalloc(sizeof(struct ext4_bg_alloc_rule),
+ GFP_KERNEL);
+ if (add_arule == NULL) {
+ ext4_unlock_group(sb, group_no);
+ ext4_mb_release_desc(&e4b);
+ return -ENOMEM;
+ }
+ freed_blocks = ext4_mb_count_unused_blocks(bd_bitmap,
+ del_arule->start, del_arule->end);
+ ext4_mb_calc_restricted(sbi, &bg_arule_list,
+ bg_arule->alloc_flag,
+ -freed_blocks);
+
+ INIT_LIST_HEAD(&add_arule->arule_list);
+ INIT_LIST_HEAD(&add_arule->tmp_list);
+ add_arule->start = del_arule->end + 1;
+ add_arule->end = bg_arule->end;
+ add_arule->alloc_flag = bg_arule->alloc_flag;
+ list_add(&add_arule->arule_list,
+ &bg_arule->arule_list);
+ /* fix existed alloc rule */
+ bg_arule->end = del_arule->start - 1;
+ } else if (del_arule->end < bg_arule->start)
+ break;
+ }
+
+ if (list_empty(&bg_arule_list->arule_list))
+ list_del(&bg_arule_list->bg_arule_list);
+
+ ext4_unlock_group(sb, group_no);
+ ext4_mb_release_desc(&e4b);
+
+ return 0;
+}
+
+int ext4_mb_del_global_arule(struct inode *inode,
+ struct ext4_alloc_rule *arule)
+{
+ struct super_block *sb = inode->i_sb;
+ struct ext4_sb_info *sbi = EXT4_SB(sb);
+ struct ext4_bg_alloc_rule_list *bg_arule_list, *tmp_arule_list;
+ struct ext4_bg_alloc_rule del_arule;
+ ext4_fsblk_t start, end;
+ ext4_group_t start_bgnum, end_bgnum;
+ int ret, del_flag = 0;
+ unsigned long bg_size = EXT4_BLOCKS_PER_GROUP(sb);
+
+ ret = ext4_mb_check_arule(inode, arule);
+ if (ret < 0)
+ return ret;
+
+ start = arule->start;
+ end = arule->start + arule->len - 1;
+
+ ext4_get_group_no_and_offset(sb, start, &start_bgnum, NULL);
+ ext4_get_group_no_and_offset(sb, end, &end_bgnum, NULL);
+
+ write_lock(&sbi->s_bg_arule_lock);
+ while (start <= end) {
+ INIT_LIST_HEAD(&del_arule.arule_list);
+ INIT_LIST_HEAD(&del_arule.tmp_list);
+ ext4_get_group_no_and_offset(sb, start, NULL, &del_arule.start);
+ del_arule.alloc_flag = arule->alloc_flag;
+ /* if end is out of bg in start, fix it */
+ if (end_bgnum > start_bgnum)
+ del_arule.end = (ext4_grpblk_t)(bg_size - 1);
+ else
+ ext4_get_group_no_and_offset(sb, end, NULL,
+ &del_arule.end);
+
+ list_for_each_entry_safe(bg_arule_list, tmp_arule_list,
+ &sbi->s_bg_arule_list, bg_arule_list) {
+ if (bg_arule_list->bg_num != start_bgnum)
+ continue;
+
+ ret = ext4_mb_del_bg_arule(sb, bg_arule_list,
+ &del_arule);
+ if (ret < 0)
+ goto out;
+
+ del_flag = 1;
+ break;
+ }
+
+ start_bgnum++;
+ start = ext4_group_first_block_no(sb, start_bgnum);
+ }
+
+ if (!del_flag)
+ ext4_debug("There is no alloc rule: start=%llu len=%llu\n",
+ (ext4_fsblk_t)arule->start,
+ (ext4_fsblk_t)arule->len);
+
+out:
+ write_unlock(&sbi->s_bg_arule_lock);
+ return ret;
+}