@@ -403,33 +403,30 @@ static void __add_ino_entry(struct f2fs_sb_info *sbi, nid_t ino,
struct inode_management *im = &sbi->im[type];
struct ino_entry *e, *tmp;
- tmp = f2fs_kmem_cache_alloc(ino_entry_slab, GFP_NOFS);
-
- radix_tree_preload(GFP_NOFS | __GFP_NOFAIL);
-
- spin_lock(&im->ino_lock);
- e = radix_tree_lookup(&im->ino_root, ino);
- if (!e) {
- e = tmp;
- if (unlikely(radix_tree_insert(&im->ino_root, ino, e)))
- f2fs_bug_on(sbi, 1);
-
- memset(e, 0, sizeof(struct ino_entry));
- e->ino = ino;
-
- list_add_tail(&e->list, &im->ino_list);
- if (type != ORPHAN_INO)
- im->ino_num++;
+ xa_lock(&im->ino_root);
+ e = xa_load(&im->ino_root, ino);
+ if (e)
+ goto found;
+ xa_unlock(&im->ino_root);
+
+ tmp = f2fs_kmem_cache_alloc(ino_entry_slab, GFP_NOFS | __GFP_ZERO);
+ xa_lock(&im->ino_root);
+ e = __xa_cmpxchg(&im->ino_root, ino, NULL, tmp,
+ GFP_NOFS | __GFP_NOFAIL);
+ if (e) {
+ kmem_cache_free(ino_entry_slab, tmp);
+ goto found;
}
+ e = tmp;
+ e->ino = ino;
+ list_add_tail(&e->list, &im->ino_list);
+ if (type != ORPHAN_INO)
+ im->ino_num++;
+found:
if (type == FLUSH_INO)
f2fs_set_bit(devidx, (char *)&e->dirty_device);
-
- spin_unlock(&im->ino_lock);
- radix_tree_preload_end();
-
- if (e != tmp)
- kmem_cache_free(ino_entry_slab, tmp);
+ xa_unlock(&im->ino_root);
}
static void __remove_ino_entry(struct f2fs_sb_info *sbi, nid_t ino, int type)
@@ -437,17 +434,14 @@ static void __remove_ino_entry(struct f2fs_sb_info *sbi, nid_t ino, int type)
struct inode_management *im = &sbi->im[type];
struct ino_entry *e;
- spin_lock(&im->ino_lock);
- e = radix_tree_lookup(&im->ino_root, ino);
+ xa_lock(&im->ino_root);
+ e = __xa_erase(&im->ino_root, ino);
if (e) {
list_del(&e->list);
- radix_tree_delete(&im->ino_root, ino);
im->ino_num--;
- spin_unlock(&im->ino_lock);
kmem_cache_free(ino_entry_slab, e);
- return;
}
- spin_unlock(&im->ino_lock);
+ xa_unlock(&im->ino_root);
}
void add_ino_entry(struct f2fs_sb_info *sbi, nid_t ino, int type)
@@ -466,12 +460,8 @@ void remove_ino_entry(struct f2fs_sb_info *sbi, nid_t ino, int type)
bool exist_written_data(struct f2fs_sb_info *sbi, nid_t ino, int mode)
{
struct inode_management *im = &sbi->im[mode];
- struct ino_entry *e;
- spin_lock(&im->ino_lock);
- e = radix_tree_lookup(&im->ino_root, ino);
- spin_unlock(&im->ino_lock);
- return e ? true : false;
+ return xa_load(&im->ino_root, ino) ? true : false;
}
void release_ino_entry(struct f2fs_sb_info *sbi, bool all)
@@ -482,14 +472,14 @@ void release_ino_entry(struct f2fs_sb_info *sbi, bool all)
for (i = all ? ORPHAN_INO : APPEND_INO; i < MAX_INO_ENTRY; i++) {
struct inode_management *im = &sbi->im[i];
- spin_lock(&im->ino_lock);
+ xa_lock(&im->ino_root);
list_for_each_entry_safe(e, tmp, &im->ino_list, list) {
list_del(&e->list);
- radix_tree_delete(&im->ino_root, e->ino);
+ __xa_erase(&im->ino_root, e->ino);
kmem_cache_free(ino_entry_slab, e);
im->ino_num--;
}
- spin_unlock(&im->ino_lock);
+ xa_unlock(&im->ino_root);
}
}
@@ -506,11 +496,11 @@ bool is_dirty_device(struct f2fs_sb_info *sbi, nid_t ino,
struct ino_entry *e;
bool is_dirty = false;
- spin_lock(&im->ino_lock);
- e = radix_tree_lookup(&im->ino_root, ino);
+ xa_lock(&im->ino_root);
+ e = xa_load(&im->ino_root, ino);
if (e && f2fs_test_bit(devidx, (char *)&e->dirty_device))
is_dirty = true;
- spin_unlock(&im->ino_lock);
+ xa_unlock(&im->ino_root);
return is_dirty;
}
@@ -519,11 +509,11 @@ int acquire_orphan_inode(struct f2fs_sb_info *sbi)
struct inode_management *im = &sbi->im[ORPHAN_INO];
int err = 0;
- spin_lock(&im->ino_lock);
+ xa_lock(&im->ino_root);
#ifdef CONFIG_F2FS_FAULT_INJECTION
if (time_to_inject(sbi, FAULT_ORPHAN)) {
- spin_unlock(&im->ino_lock);
+ xa_unlock(&im->ino_root);
f2fs_show_injection_info(FAULT_ORPHAN);
return -ENOSPC;
}
@@ -532,7 +522,7 @@ int acquire_orphan_inode(struct f2fs_sb_info *sbi)
err = -ENOSPC;
else
im->ino_num++;
- spin_unlock(&im->ino_lock);
+ xa_unlock(&im->ino_root);
return err;
}
@@ -541,10 +531,10 @@ void release_orphan_inode(struct f2fs_sb_info *sbi)
{
struct inode_management *im = &sbi->im[ORPHAN_INO];
- spin_lock(&im->ino_lock);
+ xa_lock(&im->ino_root);
f2fs_bug_on(sbi, im->ino_num == 0);
im->ino_num--;
- spin_unlock(&im->ino_lock);
+ xa_unlock(&im->ino_root);
}
void add_orphan_inode(struct inode *inode)
@@ -677,7 +667,7 @@ static void write_orphan_inodes(struct f2fs_sb_info *sbi, block_t start_blk)
orphan_blocks = GET_ORPHAN_BLOCKS(im->ino_num);
/*
- * we don't need to do spin_lock(&im->ino_lock) here, since all the
+ * we don't need to lock the ino_root here, since all the
* orphan inode operations are covered under f2fs_lock_op().
* And, spin_lock should be avoided due to page operations below.
*/
@@ -1433,8 +1423,7 @@ void init_ino_entry_info(struct f2fs_sb_info *sbi)
for (i = 0; i < MAX_INO_ENTRY; i++) {
struct inode_management *im = &sbi->im[i];
- INIT_RADIX_TREE(&im->ino_root, GFP_ATOMIC);
- spin_lock_init(&im->ino_lock);
+ xa_init(&im->ino_root);
INIT_LIST_HEAD(&im->ino_list);
im->ino_num = 0;
}
@@ -994,8 +994,7 @@ enum inode_type {
/* for inner inode cache management */
struct inode_management {
- struct radix_tree_root ino_root; /* ino entry array */
- spinlock_t ino_lock; /* for ino entry lock */
+ struct xarray ino_root; /* ino entry array */
struct list_head ino_list; /* inode list head */
unsigned long ino_num; /* number of entries */
};