f2fs: updates on 4.15-rc1
[GitHub/MotorolaMobilityLLC/kernel-slsi.git] / fs / f2fs / node.c
index fca87835a1da3f19a36a651f71493397e0fd7f8c..fe1fc662af2a8cf78516ae65b8f95f76997efa28 100644 (file)
@@ -46,7 +46,7 @@ bool available_free_memory(struct f2fs_sb_info *sbi, int type)
         * give 25%, 25%, 50%, 50%, 50% memory for each components respectively
         */
        if (type == FREE_NIDS) {
-               mem_size = (nm_i->nid_cnt[FREE_NID_LIST] *
+               mem_size = (nm_i->nid_cnt[FREE_NID] *
                                sizeof(struct free_nid)) >> PAGE_SHIFT;
                res = mem_size < ((avail_ram * nm_i->ram_thresh / 100) >> 2);
        } else if (type == NAT_ENTRIES) {
@@ -63,7 +63,7 @@ bool available_free_memory(struct f2fs_sb_info *sbi, int type)
        } else if (type == INO_ENTRIES) {
                int i;
 
-               for (i = 0; i <= UPDATE_INO; i++)
+               for (i = 0; i < MAX_INO_ENTRY; i++)
                        mem_size += sbi->im[i].ino_num *
                                                sizeof(struct ino_entry);
                mem_size >>= PAGE_SHIFT;
@@ -74,6 +74,10 @@ bool available_free_memory(struct f2fs_sb_info *sbi, int type)
                                atomic_read(&sbi->total_ext_node) *
                                sizeof(struct extent_node)) >> PAGE_SHIFT;
                res = mem_size < ((avail_ram * nm_i->ram_thresh / 100) >> 1);
+       } else if (type == INMEM_PAGES) {
+               /* it allows 20% / total_ram for inmemory pages */
+               mem_size = get_pages(sbi, F2FS_INMEM_PAGES);
+               res = mem_size < (val.totalram / 5);
        } else {
                if (!sbi->sb->s_bdi->wb.dirty_exceeded)
                        return true;
@@ -134,6 +138,44 @@ static struct page *get_next_nat_page(struct f2fs_sb_info *sbi, nid_t nid)
        return dst_page;
 }
 
+static struct nat_entry *__alloc_nat_entry(nid_t nid, bool no_fail)
+{
+       struct nat_entry *new;
+
+       if (no_fail)
+               new = f2fs_kmem_cache_alloc(nat_entry_slab,
+                                               GFP_NOFS | __GFP_ZERO);
+       else
+               new = kmem_cache_alloc(nat_entry_slab,
+                                               GFP_NOFS | __GFP_ZERO);
+       if (new) {
+               nat_set_nid(new, nid);
+               nat_reset_flag(new);
+       }
+       return new;
+}
+
+static void __free_nat_entry(struct nat_entry *e)
+{
+       kmem_cache_free(nat_entry_slab, e);
+}
+
+/* must be locked by nat_tree_lock */
+static struct nat_entry *__init_nat_entry(struct f2fs_nm_info *nm_i,
+       struct nat_entry *ne, struct f2fs_nat_entry *raw_ne, bool no_fail)
+{
+       if (no_fail)
+               f2fs_radix_tree_insert(&nm_i->nat_root, nat_get_nid(ne), ne);
+       else if (radix_tree_insert(&nm_i->nat_root, nat_get_nid(ne), ne))
+               return NULL;
+
+       if (raw_ne)
+               node_info_from_raw_nat(&ne->ni, raw_ne);
+       list_add_tail(&ne->list, &nm_i->nat_entries);
+       nm_i->nat_cnt++;
+       return ne;
+}
+
 static struct nat_entry *__lookup_nat_cache(struct f2fs_nm_info *nm_i, nid_t n)
 {
        return radix_tree_lookup(&nm_i->nat_root, n);
@@ -150,7 +192,7 @@ static void __del_from_nat_cache(struct f2fs_nm_info *nm_i, struct nat_entry *e)
        list_del(&e->list);
        radix_tree_delete(&nm_i->nat_root, nat_get_nid(e));
        nm_i->nat_cnt--;
-       kmem_cache_free(nat_entry_slab, e);
+       __free_nat_entry(e);
 }
 
 static void __set_nat_cache_dirty(struct f2fs_nm_info *nm_i,
@@ -246,49 +288,29 @@ bool need_inode_block_update(struct f2fs_sb_info *sbi, nid_t ino)
        return need_update;
 }
 
-static struct nat_entry *grab_nat_entry(struct f2fs_nm_info *nm_i, nid_t nid,
-                                                               bool no_fail)
-{
-       struct nat_entry *new;
-
-       if (no_fail) {
-               new = f2fs_kmem_cache_alloc(nat_entry_slab, GFP_NOFS);
-               f2fs_radix_tree_insert(&nm_i->nat_root, nid, new);
-       } else {
-               new = kmem_cache_alloc(nat_entry_slab, GFP_NOFS);
-               if (!new)
-                       return NULL;
-               if (radix_tree_insert(&nm_i->nat_root, nid, new)) {
-                       kmem_cache_free(nat_entry_slab, new);
-                       return NULL;
-               }
-       }
-
-       memset(new, 0, sizeof(struct nat_entry));
-       nat_set_nid(new, nid);
-       nat_reset_flag(new);
-       list_add_tail(&new->list, &nm_i->nat_entries);
-       nm_i->nat_cnt++;
-       return new;
-}
-
+/* must be locked by nat_tree_lock */
 static void cache_nat_entry(struct f2fs_sb_info *sbi, nid_t nid,
                                                struct f2fs_nat_entry *ne)
 {
        struct f2fs_nm_info *nm_i = NM_I(sbi);
-       struct nat_entry *e;
+       struct nat_entry *new, *e;
 
+       new = __alloc_nat_entry(nid, false);
+       if (!new)
+               return;
+
+       down_write(&nm_i->nat_tree_lock);
        e = __lookup_nat_cache(nm_i, nid);
-       if (!e) {
-               e = grab_nat_entry(nm_i, nid, false);
-               if (e)
-                       node_info_from_raw_nat(&e->ni, ne);
-       } else {
+       if (!e)
+               e = __init_nat_entry(nm_i, new, ne, false);
+       else
                f2fs_bug_on(sbi, nat_get_ino(e) != le32_to_cpu(ne->ino) ||
                                nat_get_blkaddr(e) !=
                                        le32_to_cpu(ne->block_addr) ||
                                nat_get_version(e) != ne->version);
-       }
+       up_write(&nm_i->nat_tree_lock);
+       if (e != new)
+               __free_nat_entry(new);
 }
 
 static void set_node_addr(struct f2fs_sb_info *sbi, struct node_info *ni,
@@ -296,11 +318,12 @@ static void set_node_addr(struct f2fs_sb_info *sbi, struct node_info *ni,
 {
        struct f2fs_nm_info *nm_i = NM_I(sbi);
        struct nat_entry *e;
+       struct nat_entry *new = __alloc_nat_entry(ni->nid, true);
 
        down_write(&nm_i->nat_tree_lock);
        e = __lookup_nat_cache(nm_i, ni->nid);
        if (!e) {
-               e = grab_nat_entry(nm_i, ni->nid, true);
+               e = __init_nat_entry(nm_i, new, NULL, true);
                copy_node_info(&e->ni, ni);
                f2fs_bug_on(sbi, ni->blk_addr == NEW_ADDR);
        } else if (new_blkaddr == NEW_ADDR) {
@@ -312,6 +335,9 @@ static void set_node_addr(struct f2fs_sb_info *sbi, struct node_info *ni,
                copy_node_info(&e->ni, ni);
                f2fs_bug_on(sbi, ni->blk_addr != NULL_ADDR);
        }
+       /* let's free early to reduce memory consumption */
+       if (e != new)
+               __free_nat_entry(new);
 
        /* sanity check */
        f2fs_bug_on(sbi, nat_get_blkaddr(e) != ni->blk_addr);
@@ -327,10 +353,6 @@ static void set_node_addr(struct f2fs_sb_info *sbi, struct node_info *ni,
        if (nat_get_blkaddr(e) != NEW_ADDR && new_blkaddr == NULL_ADDR) {
                unsigned char version = nat_get_version(e);
                nat_set_version(e, inc_node_version(version));
-
-               /* in order to reuse the nid */
-               if (nm_i->next_scan_nid > ni->nid)
-                       nm_i->next_scan_nid = ni->nid;
        }
 
        /* change address */
@@ -424,9 +446,7 @@ void get_node_info(struct f2fs_sb_info *sbi, nid_t nid, struct node_info *ni)
        f2fs_put_page(page, 1);
 cache:
        /* cache nat entry */
-       down_write(&nm_i->nat_tree_lock);
        cache_nat_entry(sbi, nid, &ne);
-       up_write(&nm_i->nat_tree_lock);
 }
 
 /*
@@ -962,7 +982,8 @@ fail:
        return err > 0 ? 0 : err;
 }
 
-int truncate_xattr_node(struct inode *inode, struct page *page)
+/* caller must lock inode page */
+int truncate_xattr_node(struct inode *inode)
 {
        struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
        nid_t nid = F2FS_I(inode)->i_xattr_nid;
@@ -978,10 +999,7 @@ int truncate_xattr_node(struct inode *inode, struct page *page)
 
        f2fs_i_xnid_write(inode, 0);
 
-       set_new_dnode(&dn, inode, page, npage, nid);
-
-       if (page)
-               dn.inode_page_locked = true;
+       set_new_dnode(&dn, inode, NULL, npage, nid);
        truncate_node(&dn);
        return 0;
 }
@@ -1000,7 +1018,7 @@ int remove_inode_page(struct inode *inode)
        if (err)
                return err;
 
-       err = truncate_xattr_node(inode, dn.inode_page);
+       err = truncate_xattr_node(inode);
        if (err) {
                f2fs_put_dnode(&dn);
                return err;
@@ -1220,7 +1238,8 @@ static void flush_inline_data(struct f2fs_sb_info *sbi, nid_t ino)
        if (!inode)
                return;
 
-       page = pagecache_get_page(inode->i_mapping, 0, FGP_LOCK|FGP_NOWAIT, 0);
+       page = f2fs_pagecache_get_page(inode->i_mapping, 0,
+                                       FGP_LOCK|FGP_NOWAIT, 0);
        if (!page)
                goto iput_out;
 
@@ -1244,37 +1263,6 @@ iput_out:
        iput(inode);
 }
 
-void move_node_page(struct page *node_page, int gc_type)
-{
-       if (gc_type == FG_GC) {
-               struct f2fs_sb_info *sbi = F2FS_P_SB(node_page);
-               struct writeback_control wbc = {
-                       .sync_mode = WB_SYNC_ALL,
-                       .nr_to_write = 1,
-                       .for_reclaim = 0,
-               };
-
-               set_page_dirty(node_page);
-               f2fs_wait_on_page_writeback(node_page, NODE, true);
-
-               f2fs_bug_on(sbi, PageWriteback(node_page));
-               if (!clear_page_dirty_for_io(node_page))
-                       goto out_page;
-
-               if (NODE_MAPPING(sbi)->a_ops->writepage(node_page, &wbc))
-                       unlock_page(node_page);
-               goto release_page;
-       } else {
-               /* set page dirty and write it */
-               if (!PageWriteback(node_page))
-                       set_page_dirty(node_page);
-       }
-out_page:
-       unlock_page(node_page);
-release_page:
-       f2fs_put_page(node_page, 0);
-}
-
 static struct page *last_fsync_dnode(struct f2fs_sb_info *sbi, nid_t ino)
 {
        pgoff_t index, end;
@@ -1344,6 +1332,7 @@ static int __write_node_page(struct page *page, bool atomic, bool *submitted,
        struct node_info ni;
        struct f2fs_io_info fio = {
                .sbi = sbi,
+               .ino = ino_of_node(page),
                .type = NODE,
                .op = REQ_OP_WRITE,
                .op_flags = wbc_to_write_flags(wbc),
@@ -1416,6 +1405,37 @@ redirty_out:
        return AOP_WRITEPAGE_ACTIVATE;
 }
 
+void move_node_page(struct page *node_page, int gc_type)
+{
+       if (gc_type == FG_GC) {
+               struct writeback_control wbc = {
+                       .sync_mode = WB_SYNC_ALL,
+                       .nr_to_write = 1,
+                       .for_reclaim = 0,
+               };
+
+               set_page_dirty(node_page);
+               f2fs_wait_on_page_writeback(node_page, NODE, true);
+
+               f2fs_bug_on(F2FS_P_SB(node_page), PageWriteback(node_page));
+               if (!clear_page_dirty_for_io(node_page))
+                       goto out_page;
+
+               if (__write_node_page(node_page, false, NULL,
+                                       &wbc, false, FS_GC_NODE_IO))
+                       unlock_page(node_page);
+               goto release_page;
+       } else {
+               /* set page dirty and write it */
+               if (!PageWriteback(node_page))
+                       set_page_dirty(node_page);
+       }
+out_page:
+       unlock_page(node_page);
+release_page:
+       f2fs_put_page(node_page, 0);
+}
+
 static int f2fs_write_node_page(struct page *page,
                                struct writeback_control *wbc)
 {
@@ -1761,35 +1781,54 @@ static struct free_nid *__lookup_free_nid_list(struct f2fs_nm_info *nm_i,
        return radix_tree_lookup(&nm_i->free_nid_root, n);
 }
 
-static int __insert_nid_to_list(struct f2fs_sb_info *sbi,
-                       struct free_nid *i, enum nid_list list, bool new)
+static int __insert_free_nid(struct f2fs_sb_info *sbi,
+                       struct free_nid *i, enum nid_state state)
 {
        struct f2fs_nm_info *nm_i = NM_I(sbi);
 
-       if (new) {
-               int err = radix_tree_insert(&nm_i->free_nid_root, i->nid, i);
-               if (err)
-                       return err;
-       }
+       int err = radix_tree_insert(&nm_i->free_nid_root, i->nid, i);
+       if (err)
+               return err;
 
-       f2fs_bug_on(sbi, list == FREE_NID_LIST ? i->state != NID_NEW :
-                                               i->state != NID_ALLOC);
-       nm_i->nid_cnt[list]++;
-       list_add_tail(&i->list, &nm_i->nid_list[list]);
+       f2fs_bug_on(sbi, state != i->state);
+       nm_i->nid_cnt[state]++;
+       if (state == FREE_NID)
+               list_add_tail(&i->list, &nm_i->free_nid_list);
        return 0;
 }
 
-static void __remove_nid_from_list(struct f2fs_sb_info *sbi,
-                       struct free_nid *i, enum nid_list list, bool reuse)
+static void __remove_free_nid(struct f2fs_sb_info *sbi,
+                       struct free_nid *i, enum nid_state state)
+{
+       struct f2fs_nm_info *nm_i = NM_I(sbi);
+
+       f2fs_bug_on(sbi, state != i->state);
+       nm_i->nid_cnt[state]--;
+       if (state == FREE_NID)
+               list_del(&i->list);
+       radix_tree_delete(&nm_i->free_nid_root, i->nid);
+}
+
+static void __move_free_nid(struct f2fs_sb_info *sbi, struct free_nid *i,
+                       enum nid_state org_state, enum nid_state dst_state)
 {
        struct f2fs_nm_info *nm_i = NM_I(sbi);
 
-       f2fs_bug_on(sbi, list == FREE_NID_LIST ? i->state != NID_NEW :
-                                               i->state != NID_ALLOC);
-       nm_i->nid_cnt[list]--;
-       list_del(&i->list);
-       if (!reuse)
-               radix_tree_delete(&nm_i->free_nid_root, i->nid);
+       f2fs_bug_on(sbi, org_state != i->state);
+       i->state = dst_state;
+       nm_i->nid_cnt[org_state]--;
+       nm_i->nid_cnt[dst_state]++;
+
+       switch (dst_state) {
+       case PREALLOC_NID:
+               list_del(&i->list);
+               break;
+       case FREE_NID:
+               list_add_tail(&i->list, &nm_i->free_nid_list);
+               break;
+       default:
+               BUG_ON(1);
+       }
 }
 
 /* return if the nid is recognized as free */
@@ -1807,7 +1846,7 @@ static bool add_free_nid(struct f2fs_sb_info *sbi, nid_t nid, bool build)
 
        i = f2fs_kmem_cache_alloc(free_nid_slab, GFP_NOFS);
        i->nid = nid;
-       i->state = NID_NEW;
+       i->state = FREE_NID;
 
        if (radix_tree_preload(GFP_NOFS))
                goto err;
@@ -1820,7 +1859,7 @@ static bool add_free_nid(struct f2fs_sb_info *sbi, nid_t nid, bool build)
                 *  - f2fs_create
                 *   - f2fs_new_inode
                 *    - alloc_nid
-                *     - __insert_nid_to_list(ALLOC_NID_LIST)
+                *     - __insert_nid_to_list(PREALLOC_NID)
                 *                     - f2fs_balance_fs_bg
                 *                      - build_free_nids
                 *                       - __build_free_nids
@@ -1833,8 +1872,8 @@ static bool add_free_nid(struct f2fs_sb_info *sbi, nid_t nid, bool build)
                 *     - new_node_page
                 *      - set_node_addr
                 *  - alloc_nid_done
-                *   - __remove_nid_from_list(ALLOC_NID_LIST)
-                *                         - __insert_nid_to_list(FREE_NID_LIST)
+                *   - __remove_nid_from_list(PREALLOC_NID)
+                *                         - __insert_nid_to_list(FREE_NID)
                 */
                ne = __lookup_nat_cache(nm_i, nid);
                if (ne && (!get_nat_flag(ne, IS_CHECKPOINTED) ||
@@ -1843,13 +1882,13 @@ static bool add_free_nid(struct f2fs_sb_info *sbi, nid_t nid, bool build)
 
                e = __lookup_free_nid_list(nm_i, nid);
                if (e) {
-                       if (e->state == NID_NEW)
+                       if (e->state == FREE_NID)
                                ret = true;
                        goto err_out;
                }
        }
        ret = true;
-       err = __insert_nid_to_list(sbi, i, FREE_NID_LIST, true);
+       err = __insert_free_nid(sbi, i, FREE_NID);
 err_out:
        spin_unlock(&nm_i->nid_list_lock);
        radix_tree_preload_end();
@@ -1867,8 +1906,8 @@ static void remove_free_nid(struct f2fs_sb_info *sbi, nid_t nid)
 
        spin_lock(&nm_i->nid_list_lock);
        i = __lookup_free_nid_list(nm_i, nid);
-       if (i && i->state == NID_NEW) {
-               __remove_nid_from_list(sbi, i, FREE_NID_LIST, false);
+       if (i && i->state == FREE_NID) {
+               __remove_free_nid(sbi, i, FREE_NID);
                need_free = true;
        }
        spin_unlock(&nm_i->nid_list_lock);
@@ -1887,15 +1926,18 @@ static void update_free_nid_bitmap(struct f2fs_sb_info *sbi, nid_t nid,
        if (!test_bit_le(nat_ofs, nm_i->nat_block_bitmap))
                return;
 
-       if (set)
+       if (set) {
+               if (test_bit_le(nid_ofs, nm_i->free_nid_bitmap[nat_ofs]))
+                       return;
                __set_bit_le(nid_ofs, nm_i->free_nid_bitmap[nat_ofs]);
-       else
-               __clear_bit_le(nid_ofs, nm_i->free_nid_bitmap[nat_ofs]);
-
-       if (set)
                nm_i->free_nid_count[nat_ofs]++;
-       else if (!build)
-               nm_i->free_nid_count[nat_ofs]--;
+       } else {
+               if (!test_bit_le(nid_ofs, nm_i->free_nid_bitmap[nat_ofs]))
+                       return;
+               __clear_bit_le(nid_ofs, nm_i->free_nid_bitmap[nat_ofs]);
+               if (!build)
+                       nm_i->free_nid_count[nat_ofs]--;
+       }
 }
 
 static void scan_nat_page(struct f2fs_sb_info *sbi,
@@ -1930,12 +1972,32 @@ static void scan_nat_page(struct f2fs_sb_info *sbi,
        }
 }
 
-static void scan_free_nid_bits(struct f2fs_sb_info *sbi)
+static void scan_curseg_cache(struct f2fs_sb_info *sbi)
 {
-       struct f2fs_nm_info *nm_i = NM_I(sbi);
        struct curseg_info *curseg = CURSEG_I(sbi, CURSEG_HOT_DATA);
        struct f2fs_journal *journal = curseg->journal;
+       int i;
+
+       down_read(&curseg->journal_rwsem);
+       for (i = 0; i < nats_in_cursum(journal); i++) {
+               block_t addr;
+               nid_t nid;
+
+               addr = le32_to_cpu(nat_in_journal(journal, i).block_addr);
+               nid = le32_to_cpu(nid_in_journal(journal, i));
+               if (addr == NULL_ADDR)
+                       add_free_nid(sbi, nid, true);
+               else
+                       remove_free_nid(sbi, nid);
+       }
+       up_read(&curseg->journal_rwsem);
+}
+
+static void scan_free_nid_bits(struct f2fs_sb_info *sbi)
+{
+       struct f2fs_nm_info *nm_i = NM_I(sbi);
        unsigned int i, idx;
+       nid_t nid;
 
        down_read(&nm_i->nat_tree_lock);
 
@@ -1945,40 +2007,27 @@ static void scan_free_nid_bits(struct f2fs_sb_info *sbi)
                if (!nm_i->free_nid_count[i])
                        continue;
                for (idx = 0; idx < NAT_ENTRY_PER_BLOCK; idx++) {
-                       nid_t nid;
-
-                       if (!test_bit_le(idx, nm_i->free_nid_bitmap[i]))
-                               continue;
+                       idx = find_next_bit_le(nm_i->free_nid_bitmap[i],
+                                               NAT_ENTRY_PER_BLOCK, idx);
+                       if (idx >= NAT_ENTRY_PER_BLOCK)
+                               break;
 
                        nid = i * NAT_ENTRY_PER_BLOCK + idx;
                        add_free_nid(sbi, nid, true);
 
-                       if (nm_i->nid_cnt[FREE_NID_LIST] >= MAX_FREE_NIDS)
+                       if (nm_i->nid_cnt[FREE_NID] >= MAX_FREE_NIDS)
                                goto out;
                }
        }
 out:
-       down_read(&curseg->journal_rwsem);
-       for (i = 0; i < nats_in_cursum(journal); i++) {
-               block_t addr;
-               nid_t nid;
+       scan_curseg_cache(sbi);
 
-               addr = le32_to_cpu(nat_in_journal(journal, i).block_addr);
-               nid = le32_to_cpu(nid_in_journal(journal, i));
-               if (addr == NULL_ADDR)
-                       add_free_nid(sbi, nid, true);
-               else
-                       remove_free_nid(sbi, nid);
-       }
-       up_read(&curseg->journal_rwsem);
        up_read(&nm_i->nat_tree_lock);
 }
 
 static void __build_free_nids(struct f2fs_sb_info *sbi, bool sync, bool mount)
 {
        struct f2fs_nm_info *nm_i = NM_I(sbi);
-       struct curseg_info *curseg = CURSEG_I(sbi, CURSEG_HOT_DATA);
-       struct f2fs_journal *journal = curseg->journal;
        int i = 0;
        nid_t nid = nm_i->next_scan_nid;
 
@@ -1986,7 +2035,7 @@ static void __build_free_nids(struct f2fs_sb_info *sbi, bool sync, bool mount)
                nid = 0;
 
        /* Enough entries */
-       if (nm_i->nid_cnt[FREE_NID_LIST] >= NAT_ENTRY_PER_BLOCK)
+       if (nm_i->nid_cnt[FREE_NID] >= NAT_ENTRY_PER_BLOCK)
                return;
 
        if (!sync && !available_free_memory(sbi, FREE_NIDS))
@@ -1996,7 +2045,7 @@ static void __build_free_nids(struct f2fs_sb_info *sbi, bool sync, bool mount)
                /* try to find free nids in free_nid_bitmap */
                scan_free_nid_bits(sbi);
 
-               if (nm_i->nid_cnt[FREE_NID_LIST])
+               if (nm_i->nid_cnt[FREE_NID] >= NAT_ENTRY_PER_BLOCK)
                        return;
        }
 
@@ -2024,18 +2073,8 @@ static void __build_free_nids(struct f2fs_sb_info *sbi, bool sync, bool mount)
        nm_i->next_scan_nid = nid;
 
        /* find free nids from current sum_pages */
-       down_read(&curseg->journal_rwsem);
-       for (i = 0; i < nats_in_cursum(journal); i++) {
-               block_t addr;
+       scan_curseg_cache(sbi);
 
-               addr = le32_to_cpu(nat_in_journal(journal, i).block_addr);
-               nid = le32_to_cpu(nid_in_journal(journal, i));
-               if (addr == NULL_ADDR)
-                       add_free_nid(sbi, nid, true);
-               else
-                       remove_free_nid(sbi, nid);
-       }
-       up_read(&curseg->journal_rwsem);
        up_read(&nm_i->nat_tree_lock);
 
        ra_meta_pages(sbi, NAT_BLOCK_OFFSET(nm_i->next_scan_nid),
@@ -2073,15 +2112,13 @@ retry:
        }
 
        /* We should not use stale free nids created by build_free_nids */
-       if (nm_i->nid_cnt[FREE_NID_LIST] && !on_build_free_nids(nm_i)) {
-               f2fs_bug_on(sbi, list_empty(&nm_i->nid_list[FREE_NID_LIST]));
-               i = list_first_entry(&nm_i->nid_list[FREE_NID_LIST],
+       if (nm_i->nid_cnt[FREE_NID] && !on_build_free_nids(nm_i)) {
+               f2fs_bug_on(sbi, list_empty(&nm_i->free_nid_list));
+               i = list_first_entry(&nm_i->free_nid_list,
                                        struct free_nid, list);
                *nid = i->nid;
 
-               __remove_nid_from_list(sbi, i, FREE_NID_LIST, true);
-               i->state = NID_ALLOC;
-               __insert_nid_to_list(sbi, i, ALLOC_NID_LIST, false);
+               __move_free_nid(sbi, i, FREE_NID, PREALLOC_NID);
                nm_i->available_nids--;
 
                update_free_nid_bitmap(sbi, *nid, false, false);
@@ -2107,7 +2144,7 @@ void alloc_nid_done(struct f2fs_sb_info *sbi, nid_t nid)
        spin_lock(&nm_i->nid_list_lock);
        i = __lookup_free_nid_list(nm_i, nid);
        f2fs_bug_on(sbi, !i);
-       __remove_nid_from_list(sbi, i, ALLOC_NID_LIST, false);
+       __remove_free_nid(sbi, i, PREALLOC_NID);
        spin_unlock(&nm_i->nid_list_lock);
 
        kmem_cache_free(free_nid_slab, i);
@@ -2130,12 +2167,10 @@ void alloc_nid_failed(struct f2fs_sb_info *sbi, nid_t nid)
        f2fs_bug_on(sbi, !i);
 
        if (!available_free_memory(sbi, FREE_NIDS)) {
-               __remove_nid_from_list(sbi, i, ALLOC_NID_LIST, false);
+               __remove_free_nid(sbi, i, PREALLOC_NID);
                need_free = true;
        } else {
-               __remove_nid_from_list(sbi, i, ALLOC_NID_LIST, true);
-               i->state = NID_NEW;
-               __insert_nid_to_list(sbi, i, FREE_NID_LIST, false);
+               __move_free_nid(sbi, i, PREALLOC_NID, FREE_NID);
        }
 
        nm_i->available_nids++;
@@ -2154,20 +2189,19 @@ int try_to_free_nids(struct f2fs_sb_info *sbi, int nr_shrink)
        struct free_nid *i, *next;
        int nr = nr_shrink;
 
-       if (nm_i->nid_cnt[FREE_NID_LIST] <= MAX_FREE_NIDS)
+       if (nm_i->nid_cnt[FREE_NID] <= MAX_FREE_NIDS)
                return 0;
 
        if (!mutex_trylock(&nm_i->build_lock))
                return 0;
 
        spin_lock(&nm_i->nid_list_lock);
-       list_for_each_entry_safe(i, next, &nm_i->nid_list[FREE_NID_LIST],
-                                                                       list) {
+       list_for_each_entry_safe(i, next, &nm_i->free_nid_list, list) {
                if (nr_shrink <= 0 ||
-                               nm_i->nid_cnt[FREE_NID_LIST] <= MAX_FREE_NIDS)
+                               nm_i->nid_cnt[FREE_NID] <= MAX_FREE_NIDS)
                        break;
 
-               __remove_nid_from_list(sbi, i, FREE_NID_LIST, false);
+               __remove_free_nid(sbi, i, FREE_NID);
                kmem_cache_free(free_nid_slab, i);
                nr_shrink--;
        }
@@ -2193,8 +2227,8 @@ void recover_inline_xattr(struct inode *inode, struct page *page)
                goto update_inode;
        }
 
-       dst_addr = inline_xattr_addr(ipage);
-       src_addr = inline_xattr_addr(page);
+       dst_addr = inline_xattr_addr(inode, ipage);
+       src_addr = inline_xattr_addr(inode, page);
        inline_size = inline_xattr_size(inode);
 
        f2fs_wait_on_page_writeback(ipage, NODE, true);
@@ -2283,6 +2317,12 @@ retry:
        dst->i_inline = src->i_inline & (F2FS_INLINE_XATTR | F2FS_EXTRA_ATTR);
        if (dst->i_inline & F2FS_EXTRA_ATTR) {
                dst->i_extra_isize = src->i_extra_isize;
+
+               if (f2fs_sb_has_flexible_inline_xattr(sbi->sb) &&
+                       F2FS_FITS_IN_INODE(src, le16_to_cpu(src->i_extra_isize),
+                                                       i_inline_xattr_size))
+                       dst->i_inline_xattr_size = src->i_inline_xattr_size;
+
                if (f2fs_sb_has_project_quota(sbi->sb) &&
                        F2FS_FITS_IN_INODE(src, le16_to_cpu(src->i_extra_isize),
                                                                i_projid))
@@ -2354,8 +2394,8 @@ static void remove_nats_in_journal(struct f2fs_sb_info *sbi)
 
                ne = __lookup_nat_cache(nm_i, nid);
                if (!ne) {
-                       ne = grab_nat_entry(nm_i, nid, true);
-                       node_info_from_raw_nat(&ne->ni, &raw_ne);
+                       ne = __alloc_nat_entry(nid, true);
+                       __init_nat_entry(nm_i, ne, &raw_ne, true);
                }
 
                /*
@@ -2401,15 +2441,17 @@ static void __update_nat_bits(struct f2fs_sb_info *sbi, nid_t start_nid,
        unsigned int nat_index = start_nid / NAT_ENTRY_PER_BLOCK;
        struct f2fs_nat_block *nat_blk = page_address(page);
        int valid = 0;
-       int i;
+       int i = 0;
 
        if (!enabled_nat_bits(sbi, NULL))
                return;
 
-       for (i = 0; i < NAT_ENTRY_PER_BLOCK; i++) {
-               if (start_nid == 0 && i == 0)
-                       valid++;
-               if (nat_blk->entries[i].block_addr)
+       if (nat_index == 0) {
+               valid = 1;
+               i = 1;
+       }
+       for (; i < NAT_ENTRY_PER_BLOCK; i++) {
+               if (nat_blk->entries[i].block_addr != NULL_ADDR)
                        valid++;
        }
        if (valid == 0) {
@@ -2604,7 +2646,7 @@ static inline void load_free_nid_bitmap(struct f2fs_sb_info *sbi)
                __set_bit_le(i, nm_i->nat_block_bitmap);
 
                nid = i * NAT_ENTRY_PER_BLOCK;
-               last_nid = (i + 1) * NAT_ENTRY_PER_BLOCK;
+               last_nid = nid + NAT_ENTRY_PER_BLOCK;
 
                spin_lock(&NM_I(sbi)->nid_list_lock);
                for (; nid < last_nid; nid++)
@@ -2639,16 +2681,15 @@ static int init_node_manager(struct f2fs_sb_info *sbi)
        /* not used nids: 0, node, meta, (and root counted as valid node) */
        nm_i->available_nids = nm_i->max_nid - sbi->total_valid_node_count -
                                                        F2FS_RESERVED_NODE_NUM;
-       nm_i->nid_cnt[FREE_NID_LIST] = 0;
-       nm_i->nid_cnt[ALLOC_NID_LIST] = 0;
+       nm_i->nid_cnt[FREE_NID] = 0;
+       nm_i->nid_cnt[PREALLOC_NID] = 0;
        nm_i->nat_cnt = 0;
        nm_i->ram_thresh = DEF_RAM_THRESHOLD;
        nm_i->ra_nid_pages = DEF_RA_NID_PAGES;
        nm_i->dirty_nats_ratio = DEF_DIRTY_NAT_RATIO_THRESHOLD;
 
        INIT_RADIX_TREE(&nm_i->free_nid_root, GFP_ATOMIC);
-       INIT_LIST_HEAD(&nm_i->nid_list[FREE_NID_LIST]);
-       INIT_LIST_HEAD(&nm_i->nid_list[ALLOC_NID_LIST]);
+       INIT_LIST_HEAD(&nm_i->free_nid_list);
        INIT_RADIX_TREE(&nm_i->nat_root, GFP_NOIO);
        INIT_RADIX_TREE(&nm_i->nat_set_root, GFP_NOIO);
        INIT_LIST_HEAD(&nm_i->nat_entries);
@@ -2740,16 +2781,15 @@ void destroy_node_manager(struct f2fs_sb_info *sbi)
 
        /* destroy free nid list */
        spin_lock(&nm_i->nid_list_lock);
-       list_for_each_entry_safe(i, next_i, &nm_i->nid_list[FREE_NID_LIST],
-                                                                       list) {
-               __remove_nid_from_list(sbi, i, FREE_NID_LIST, false);
+       list_for_each_entry_safe(i, next_i, &nm_i->free_nid_list, list) {
+               __remove_free_nid(sbi, i, FREE_NID);
                spin_unlock(&nm_i->nid_list_lock);
                kmem_cache_free(free_nid_slab, i);
                spin_lock(&nm_i->nid_list_lock);
        }
-       f2fs_bug_on(sbi, nm_i->nid_cnt[FREE_NID_LIST]);
-       f2fs_bug_on(sbi, nm_i->nid_cnt[ALLOC_NID_LIST]);
-       f2fs_bug_on(sbi, !list_empty(&nm_i->nid_list[ALLOC_NID_LIST]));
+       f2fs_bug_on(sbi, nm_i->nid_cnt[FREE_NID]);
+       f2fs_bug_on(sbi, nm_i->nid_cnt[PREALLOC_NID]);
+       f2fs_bug_on(sbi, !list_empty(&nm_i->free_nid_list));
        spin_unlock(&nm_i->nid_list_lock);
 
        /* destroy nat cache */