Message ID | 20230603150327.3596033-17-shikemeng@huaweicloud.com |
---|---|
State | Awaiting Upstream |
Headers | show |
Series | Fixes, cleanups and unit test for mballoc | expand |
On Sat, Jun 03, 2023 at 11:03:24PM +0800, Kemeng Shi wrote: > call ext4_mb_mark_group_bb in ext4_mb_clear_bb to remove repeat code > to update block bitmap and group descriptor on disk. > > Note: ext4_mb_clear_bb will update buddy and bitmap in two critical sections > instead of update in the same critical section. > > Original lock behavior introduced in 7a2fcbf7f857 ("ext4: don't use > blocks freed but not yet committed in buddy cache init") to avoid > race betwwen ext4_mb_free_blocks and ext4_mb_init_cache: > ext4_mb_load_buddy_gfp > ext4_lock_group > mb_clear_bits(bitmap_bh, ...) > mb_free_blocks/ext4_mb_free_metadata > ext4_unlock_group > ext4_mb_unload_buddy > > New lock behavior in this patch: > ext4_mb_load_buddy_gfp > ext4_lock_group > mb_clear_bits(bitmap_bh, ...) > ext4_unlock_group > > /* no ext4_mb_init_cache for the same group will be called as > ext4_mb_load_buddy_gfp will ensure buddy page is update-to-date */ > > ext4_lock_group > mb_free_blocks/ext4_mb_free_metadata > ext4_unlock_group > ext4_mb_unload_buddy > > As buddy page for group is always update-to-date between > ext4_mb_load_buddy_gfp and ext4_mb_unload_buddy. Then no > ext4_mb_init_cache will be called for the same group concurrentlly when > we update bitmap and buddy page betwwen buddy load and unload. > > Signed-off-by: Kemeng Shi <shikemeng@huaweicloud.com> Hi Kemeng, Sorry for the late reply I was trying to understand the codepath properly. So I have a question here: With the changes you've made in the patch, the flow would look something like: ext4_mb_clear_bb(): ext4_mb_mark_group_bb(): ext4_group_lock() - Mark bitmap free - Modify gdp ext4_group_unlock() ext4_handle_dirty_metadata() - I understand this will add the bitmap and gdp buffers to journal's dirty metadata list ... ext4_group_lock() ext4_mb_free_metadata() - Add ext4_free_data entries to sbi->s_freed_data_list. (On commit ext4_journal_commit_callback() will then free the buddy for these) ext4_group_unlock() My question is what happens if journal commits between ext4_handle_dirty_metadata() and ext4_mb_free_metadata() call (Possible?). Then we might never end up freeing the metadata in the buddy bitmap because the commit callback wont be able to find the ext4_free_data entries in sbi->s_freed_data_list. Regards, ojaswin > --- > fs/ext4/mballoc.c | 90 ++++++++++++----------------------------------- > 1 file changed, 23 insertions(+), 67 deletions(-) > > diff --git a/fs/ext4/mballoc.c b/fs/ext4/mballoc.c > index 46b37f5c9223..e4f1b34448e3 100644 > --- a/fs/ext4/mballoc.c > +++ b/fs/ext4/mballoc.c > @@ -6135,19 +6135,21 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, > ext4_fsblk_t block, unsigned long count, > int flags) > { > - struct buffer_head *bitmap_bh = NULL; > + struct ext4_mark_context mc = { > + .handle = handle, > + .sb = inode->i_sb, > + .state = 0, > + }; > struct super_block *sb = inode->i_sb; > - struct ext4_group_desc *gdp; > struct ext4_group_info *grp; > unsigned int overflow; > ext4_grpblk_t bit; > - struct buffer_head *gd_bh; > ext4_group_t block_group; > struct ext4_sb_info *sbi; > struct ext4_buddy e4b; > unsigned int count_clusters; > int err = 0; > - int ret; > + int mark_flags = 0; > > sbi = EXT4_SB(sb); > > @@ -6179,18 +6181,6 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, > /* The range changed so it's no longer validated */ > flags &= ~EXT4_FREE_BLOCKS_VALIDATED; > } > - count_clusters = EXT4_NUM_B2C(sbi, count); > - bitmap_bh = ext4_read_block_bitmap(sb, block_group); > - if (IS_ERR(bitmap_bh)) { > - err = PTR_ERR(bitmap_bh); > - bitmap_bh = NULL; > - goto error_return; > - } > - gdp = ext4_get_group_desc(sb, block_group, &gd_bh); > - if (!gdp) { > - err = -EIO; > - goto error_return; > - } > > if (!(flags & EXT4_FREE_BLOCKS_VALIDATED) && > !ext4_inode_block_valid(inode, block, count)) { > @@ -6200,28 +6190,7 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, > goto error_return; > } > > - BUFFER_TRACE(bitmap_bh, "getting write access"); > - err = ext4_journal_get_write_access(handle, sb, bitmap_bh, > - EXT4_JTR_NONE); > - if (err) > - goto error_return; > - > - /* > - * We are about to modify some metadata. Call the journal APIs > - * to unshare ->b_data if a currently-committing transaction is > - * using it > - */ > - BUFFER_TRACE(gd_bh, "get_write_access"); > - err = ext4_journal_get_write_access(handle, sb, gd_bh, EXT4_JTR_NONE); > - if (err) > - goto error_return; > -#ifdef AGGRESSIVE_CHECK > - { > - int i; > - for (i = 0; i < count_clusters; i++) > - BUG_ON(!mb_test_bit(bit + i, bitmap_bh->b_data)); > - } > -#endif > + count_clusters = EXT4_NUM_B2C(sbi, count); > trace_ext4_mballoc_free(sb, inode, block_group, bit, count_clusters); > > /* __GFP_NOFAIL: retry infinitely, ignore TIF_MEMDIE and memcg limit. */ > @@ -6230,6 +6199,22 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, > if (err) > goto error_return; > > +#ifdef AGGRESSIVE_CHECK > + mark_flags |= EXT4_MB_BITMAP_MARKED_CHECK; > +#endif > + err = ext4_mb_mark_group_bb(&mc, block_group, bit, count_clusters, > + mark_flags); > + > + > + if (err && mc.changed == 0) { > + ext4_mb_unload_buddy(&e4b); > + goto error_return; > + } > + > +#ifdef AGGRESSIVE_CHECK > + BUG_ON(mc.changed != count_clusters); > +#endif > + > /* > * We need to make sure we don't reuse the freed block until after the > * transaction is committed. We make an exception if the inode is to be > @@ -6252,13 +6237,8 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, > new_entry->efd_tid = handle->h_transaction->t_tid; > > ext4_lock_group(sb, block_group); > - mb_clear_bits(bitmap_bh->b_data, bit, count_clusters); > ext4_mb_free_metadata(handle, &e4b, new_entry); > } else { > - /* need to update group_info->bb_free and bitmap > - * with group lock held. generate_buddy look at > - * them with group lock_held > - */ > if (test_opt(sb, DISCARD)) { > err = ext4_issue_discard(sb, block_group, bit, > count_clusters, NULL); > @@ -6271,23 +6251,11 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, > EXT4_MB_GRP_CLEAR_TRIMMED(e4b.bd_info); > > ext4_lock_group(sb, block_group); > - mb_clear_bits(bitmap_bh->b_data, bit, count_clusters); > mb_free_blocks(inode, &e4b, bit, count_clusters); > } > > - ret = ext4_free_group_clusters(sb, gdp) + count_clusters; > - ext4_free_group_clusters_set(sb, gdp, ret); > - ext4_block_bitmap_csum_set(sb, gdp, bitmap_bh); > - ext4_group_desc_csum_set(sb, block_group, gdp); > ext4_unlock_group(sb, block_group); > > - if (sbi->s_log_groups_per_flex) { > - ext4_group_t flex_group = ext4_flex_group(sbi, block_group); > - atomic64_add(count_clusters, > - &sbi_array_rcu_deref(sbi, s_flex_groups, > - flex_group)->free_clusters); > - } > - > /* > * on a bigalloc file system, defer the s_freeclusters_counter > * update to the caller (ext4_remove_space and friends) so they > @@ -6302,26 +6270,14 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, > > ext4_mb_unload_buddy(&e4b); > > - /* We dirtied the bitmap block */ > - BUFFER_TRACE(bitmap_bh, "dirtied bitmap block"); > - err = ext4_handle_dirty_metadata(handle, NULL, bitmap_bh); > - > - /* And the group descriptor block */ > - BUFFER_TRACE(gd_bh, "dirtied group descriptor block"); > - ret = ext4_handle_dirty_metadata(handle, NULL, gd_bh); > - if (!err) > - err = ret; > - > if (overflow && !err) { > block += count; > count = overflow; > - put_bh(bitmap_bh); > /* The range changed so it's no longer validated */ > flags &= ~EXT4_FREE_BLOCKS_VALIDATED; > goto do_more; > } > error_return: > - brelse(bitmap_bh); > ext4_std_error(sb, err); > return; > } > -- > 2.30.0 >
on 6/6/2023 5:39 PM, Ojaswin Mujoo wrote: > On Sat, Jun 03, 2023 at 11:03:24PM +0800, Kemeng Shi wrote: >> call ext4_mb_mark_group_bb in ext4_mb_clear_bb to remove repeat code >> to update block bitmap and group descriptor on disk. >> >> Note: ext4_mb_clear_bb will update buddy and bitmap in two critical sections >> instead of update in the same critical section. >> >> Original lock behavior introduced in 7a2fcbf7f857 ("ext4: don't use >> blocks freed but not yet committed in buddy cache init") to avoid >> race betwwen ext4_mb_free_blocks and ext4_mb_init_cache: >> ext4_mb_load_buddy_gfp >> ext4_lock_group >> mb_clear_bits(bitmap_bh, ...) >> mb_free_blocks/ext4_mb_free_metadata >> ext4_unlock_group >> ext4_mb_unload_buddy >> >> New lock behavior in this patch: >> ext4_mb_load_buddy_gfp >> ext4_lock_group >> mb_clear_bits(bitmap_bh, ...) >> ext4_unlock_group >> >> /* no ext4_mb_init_cache for the same group will be called as >> ext4_mb_load_buddy_gfp will ensure buddy page is update-to-date */ >> >> ext4_lock_group >> mb_free_blocks/ext4_mb_free_metadata >> ext4_unlock_group >> ext4_mb_unload_buddy >> >> As buddy page for group is always update-to-date between >> ext4_mb_load_buddy_gfp and ext4_mb_unload_buddy. Then no >> ext4_mb_init_cache will be called for the same group concurrentlly when >> we update bitmap and buddy page betwwen buddy load and unload. >> >> Signed-off-by: Kemeng Shi <shikemeng@huaweicloud.com> > > Hi Kemeng, > > Sorry for the late reply I was trying to understand the codepath > properly. So I have a question here: > > With the changes you've made in the patch, the flow would look something > like: > > ext4_mb_clear_bb(): > ext4_mb_mark_group_bb(): > ext4_group_lock() > - Mark bitmap free > - Modify gdp > ext4_group_unlock() > ext4_handle_dirty_metadata() > - I understand this will add the bitmap and gdp buffers to journal's > dirty metadata list > ... > ext4_group_lock() > ext4_mb_free_metadata() > - Add ext4_free_data entries to sbi->s_freed_data_list. (On commit > ext4_journal_commit_callback() will then free the buddy for these) > ext4_group_unlock() > > My question is what happens if journal commits between > ext4_handle_dirty_metadata() and ext4_mb_free_metadata() call (Possible?). Then we might > never end up freeing the metadata in the buddy bitmap because the commit callback wont > be able to find the ext4_free_data entries in sbi->s_freed_data_list. > > Regards, > ojaswin > Hi Ojaswin, thanks for the reply. To my knowledge, commit should be normally done after handle is stopped as following: ext4_journal_start_sb start_this_handle read_lock(&journal->j_state_lock); atomic_inc(&transaction->t_updates); read_unlock(&journal->j_state_lock); ext4_journal_stop jbd2_journal_stop stop_this_handle if (atomic_dec_and_test(&transaction->t_updates)) wake_up(&journal->j_wait_updates); jbd2_journal_commit_transaction jbd2_journal_wait_updates while (1) if (!atomic_read(&transaction->t_updates)) /* break loop */ ... if (journal->j_commit_callback) journal->j_commit_callback(journal, commit_transaction); So no commit of transaction should not happen between ext4_handle_dirty_metadata and ext4_mb_free_metadata. >> --- >> fs/ext4/mballoc.c | 90 ++++++++++++----------------------------------- >> 1 file changed, 23 insertions(+), 67 deletions(-) >> >> diff --git a/fs/ext4/mballoc.c b/fs/ext4/mballoc.c >> index 46b37f5c9223..e4f1b34448e3 100644 >> --- a/fs/ext4/mballoc.c >> +++ b/fs/ext4/mballoc.c >> @@ -6135,19 +6135,21 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, >> ext4_fsblk_t block, unsigned long count, >> int flags) >> { >> - struct buffer_head *bitmap_bh = NULL; >> + struct ext4_mark_context mc = { >> + .handle = handle, >> + .sb = inode->i_sb, >> + .state = 0, >> + }; >> struct super_block *sb = inode->i_sb; >> - struct ext4_group_desc *gdp; >> struct ext4_group_info *grp; >> unsigned int overflow; >> ext4_grpblk_t bit; >> - struct buffer_head *gd_bh; >> ext4_group_t block_group; >> struct ext4_sb_info *sbi; >> struct ext4_buddy e4b; >> unsigned int count_clusters; >> int err = 0; >> - int ret; >> + int mark_flags = 0; >> >> sbi = EXT4_SB(sb); >> >> @@ -6179,18 +6181,6 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, >> /* The range changed so it's no longer validated */ >> flags &= ~EXT4_FREE_BLOCKS_VALIDATED; >> } >> - count_clusters = EXT4_NUM_B2C(sbi, count); >> - bitmap_bh = ext4_read_block_bitmap(sb, block_group); >> - if (IS_ERR(bitmap_bh)) { >> - err = PTR_ERR(bitmap_bh); >> - bitmap_bh = NULL; >> - goto error_return; >> - } >> - gdp = ext4_get_group_desc(sb, block_group, &gd_bh); >> - if (!gdp) { >> - err = -EIO; >> - goto error_return; >> - } >> >> if (!(flags & EXT4_FREE_BLOCKS_VALIDATED) && >> !ext4_inode_block_valid(inode, block, count)) { >> @@ -6200,28 +6190,7 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, >> goto error_return; >> } >> >> - BUFFER_TRACE(bitmap_bh, "getting write access"); >> - err = ext4_journal_get_write_access(handle, sb, bitmap_bh, >> - EXT4_JTR_NONE); >> - if (err) >> - goto error_return; >> - >> - /* >> - * We are about to modify some metadata. Call the journal APIs >> - * to unshare ->b_data if a currently-committing transaction is >> - * using it >> - */ >> - BUFFER_TRACE(gd_bh, "get_write_access"); >> - err = ext4_journal_get_write_access(handle, sb, gd_bh, EXT4_JTR_NONE); >> - if (err) >> - goto error_return; >> -#ifdef AGGRESSIVE_CHECK >> - { >> - int i; >> - for (i = 0; i < count_clusters; i++) >> - BUG_ON(!mb_test_bit(bit + i, bitmap_bh->b_data)); >> - } >> -#endif >> + count_clusters = EXT4_NUM_B2C(sbi, count); >> trace_ext4_mballoc_free(sb, inode, block_group, bit, count_clusters); >> >> /* __GFP_NOFAIL: retry infinitely, ignore TIF_MEMDIE and memcg limit. */ >> @@ -6230,6 +6199,22 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, >> if (err) >> goto error_return; >> >> +#ifdef AGGRESSIVE_CHECK >> + mark_flags |= EXT4_MB_BITMAP_MARKED_CHECK; >> +#endif >> + err = ext4_mb_mark_group_bb(&mc, block_group, bit, count_clusters, >> + mark_flags); >> + >> + >> + if (err && mc.changed == 0) { >> + ext4_mb_unload_buddy(&e4b); >> + goto error_return; >> + } >> + >> +#ifdef AGGRESSIVE_CHECK >> + BUG_ON(mc.changed != count_clusters); >> +#endif >> + >> /* >> * We need to make sure we don't reuse the freed block until after the >> * transaction is committed. We make an exception if the inode is to be >> @@ -6252,13 +6237,8 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, >> new_entry->efd_tid = handle->h_transaction->t_tid; >> >> ext4_lock_group(sb, block_group); >> - mb_clear_bits(bitmap_bh->b_data, bit, count_clusters); >> ext4_mb_free_metadata(handle, &e4b, new_entry); >> } else { >> - /* need to update group_info->bb_free and bitmap >> - * with group lock held. generate_buddy look at >> - * them with group lock_held >> - */ >> if (test_opt(sb, DISCARD)) { >> err = ext4_issue_discard(sb, block_group, bit, >> count_clusters, NULL); >> @@ -6271,23 +6251,11 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, >> EXT4_MB_GRP_CLEAR_TRIMMED(e4b.bd_info); >> >> ext4_lock_group(sb, block_group); >> - mb_clear_bits(bitmap_bh->b_data, bit, count_clusters); >> mb_free_blocks(inode, &e4b, bit, count_clusters); >> } >> >> - ret = ext4_free_group_clusters(sb, gdp) + count_clusters; >> - ext4_free_group_clusters_set(sb, gdp, ret); >> - ext4_block_bitmap_csum_set(sb, gdp, bitmap_bh); >> - ext4_group_desc_csum_set(sb, block_group, gdp); >> ext4_unlock_group(sb, block_group); >> >> - if (sbi->s_log_groups_per_flex) { >> - ext4_group_t flex_group = ext4_flex_group(sbi, block_group); >> - atomic64_add(count_clusters, >> - &sbi_array_rcu_deref(sbi, s_flex_groups, >> - flex_group)->free_clusters); >> - } >> - >> /* >> * on a bigalloc file system, defer the s_freeclusters_counter >> * update to the caller (ext4_remove_space and friends) so they >> @@ -6302,26 +6270,14 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, >> >> ext4_mb_unload_buddy(&e4b); >> >> - /* We dirtied the bitmap block */ >> - BUFFER_TRACE(bitmap_bh, "dirtied bitmap block"); >> - err = ext4_handle_dirty_metadata(handle, NULL, bitmap_bh); >> - >> - /* And the group descriptor block */ >> - BUFFER_TRACE(gd_bh, "dirtied group descriptor block"); >> - ret = ext4_handle_dirty_metadata(handle, NULL, gd_bh); >> - if (!err) >> - err = ret; >> - >> if (overflow && !err) { >> block += count; >> count = overflow; >> - put_bh(bitmap_bh); >> /* The range changed so it's no longer validated */ >> flags &= ~EXT4_FREE_BLOCKS_VALIDATED; >> goto do_more; >> } >> error_return: >> - brelse(bitmap_bh); >> ext4_std_error(sb, err); >> return; >> } >> -- >> 2.30.0 >> >
On Tue, Jun 06, 2023 at 10:16:38PM +0800, Kemeng Shi wrote: > > > on 6/6/2023 5:39 PM, Ojaswin Mujoo wrote: > > On Sat, Jun 03, 2023 at 11:03:24PM +0800, Kemeng Shi wrote: > >> call ext4_mb_mark_group_bb in ext4_mb_clear_bb to remove repeat code > >> to update block bitmap and group descriptor on disk. > >> > >> Note: ext4_mb_clear_bb will update buddy and bitmap in two critical sections > >> instead of update in the same critical section. > >> > >> Original lock behavior introduced in 7a2fcbf7f857 ("ext4: don't use > >> blocks freed but not yet committed in buddy cache init") to avoid > >> race betwwen ext4_mb_free_blocks and ext4_mb_init_cache: > >> ext4_mb_load_buddy_gfp > >> ext4_lock_group > >> mb_clear_bits(bitmap_bh, ...) > >> mb_free_blocks/ext4_mb_free_metadata > >> ext4_unlock_group > >> ext4_mb_unload_buddy > >> > >> New lock behavior in this patch: > >> ext4_mb_load_buddy_gfp > >> ext4_lock_group > >> mb_clear_bits(bitmap_bh, ...) > >> ext4_unlock_group > >> > >> /* no ext4_mb_init_cache for the same group will be called as > >> ext4_mb_load_buddy_gfp will ensure buddy page is update-to-date */ > >> > >> ext4_lock_group > >> mb_free_blocks/ext4_mb_free_metadata > >> ext4_unlock_group > >> ext4_mb_unload_buddy > >> > >> As buddy page for group is always update-to-date between > >> ext4_mb_load_buddy_gfp and ext4_mb_unload_buddy. Then no > >> ext4_mb_init_cache will be called for the same group concurrentlly when > >> we update bitmap and buddy page betwwen buddy load and unload. > >> > >> Signed-off-by: Kemeng Shi <shikemeng@huaweicloud.com> > > > > Hi Kemeng, > > > > Sorry for the late reply I was trying to understand the codepath > > properly. So I have a question here: > > > > With the changes you've made in the patch, the flow would look something > > like: > > > > ext4_mb_clear_bb(): > > ext4_mb_mark_group_bb(): > > ext4_group_lock() > > - Mark bitmap free > > - Modify gdp > > ext4_group_unlock() > > ext4_handle_dirty_metadata() > > - I understand this will add the bitmap and gdp buffers to journal's > > dirty metadata list > > ... > > ext4_group_lock() > > ext4_mb_free_metadata() > > - Add ext4_free_data entries to sbi->s_freed_data_list. (On commit > > ext4_journal_commit_callback() will then free the buddy for these) > > ext4_group_unlock() > > > > My question is what happens if journal commits between > > ext4_handle_dirty_metadata() and ext4_mb_free_metadata() call (Possible?). Then we might > > never end up freeing the metadata in the buddy bitmap because the commit callback wont > > be able to find the ext4_free_data entries in sbi->s_freed_data_list. > > > > Regards, > > ojaswin > > > Hi Ojaswin, thanks for the reply. To my knowledge, commit should be normally done after handle > is stopped as following: > ext4_journal_start_sb > start_this_handle > read_lock(&journal->j_state_lock); > atomic_inc(&transaction->t_updates); > read_unlock(&journal->j_state_lock); > > ext4_journal_stop > jbd2_journal_stop > stop_this_handle > if (atomic_dec_and_test(&transaction->t_updates)) > wake_up(&journal->j_wait_updates); > > jbd2_journal_commit_transaction > jbd2_journal_wait_updates > while (1) > if (!atomic_read(&transaction->t_updates)) > /* break loop */ > ... > > if (journal->j_commit_callback) > journal->j_commit_callback(journal, commit_transaction); > > So no commit of transaction should not happen between ext4_handle_dirty_metadata and > ext4_mb_free_metadata. Hi Kemeng, Okay makes sense. Thanks for the explanation :) Regards, ojaswin > > >> --- > >> fs/ext4/mballoc.c | 90 ++++++++++++----------------------------------- > >> 1 file changed, 23 insertions(+), 67 deletions(-) > >> > >> diff --git a/fs/ext4/mballoc.c b/fs/ext4/mballoc.c > >> index 46b37f5c9223..e4f1b34448e3 100644 > >> --- a/fs/ext4/mballoc.c > >> +++ b/fs/ext4/mballoc.c > >> @@ -6135,19 +6135,21 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, > >> ext4_fsblk_t block, unsigned long count, > >> int flags) > >> { > >> - struct buffer_head *bitmap_bh = NULL; > >> + struct ext4_mark_context mc = { > >> + .handle = handle, > >> + .sb = inode->i_sb, > >> + .state = 0, > >> + }; > >> struct super_block *sb = inode->i_sb; > >> - struct ext4_group_desc *gdp; > >> struct ext4_group_info *grp; > >> unsigned int overflow; > >> ext4_grpblk_t bit; > >> - struct buffer_head *gd_bh; > >> ext4_group_t block_group; > >> struct ext4_sb_info *sbi; > >> struct ext4_buddy e4b; > >> unsigned int count_clusters; > >> int err = 0; > >> - int ret; > >> + int mark_flags = 0; > >> > >> sbi = EXT4_SB(sb); > >> > >> @@ -6179,18 +6181,6 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, > >> /* The range changed so it's no longer validated */ > >> flags &= ~EXT4_FREE_BLOCKS_VALIDATED; > >> } > >> - count_clusters = EXT4_NUM_B2C(sbi, count); > >> - bitmap_bh = ext4_read_block_bitmap(sb, block_group); > >> - if (IS_ERR(bitmap_bh)) { > >> - err = PTR_ERR(bitmap_bh); > >> - bitmap_bh = NULL; > >> - goto error_return; > >> - } > >> - gdp = ext4_get_group_desc(sb, block_group, &gd_bh); > >> - if (!gdp) { > >> - err = -EIO; > >> - goto error_return; > >> - } > >> > >> if (!(flags & EXT4_FREE_BLOCKS_VALIDATED) && > >> !ext4_inode_block_valid(inode, block, count)) { > >> @@ -6200,28 +6190,7 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, > >> goto error_return; > >> } > >> > >> - BUFFER_TRACE(bitmap_bh, "getting write access"); > >> - err = ext4_journal_get_write_access(handle, sb, bitmap_bh, > >> - EXT4_JTR_NONE); > >> - if (err) > >> - goto error_return; > >> - > >> - /* > >> - * We are about to modify some metadata. Call the journal APIs > >> - * to unshare ->b_data if a currently-committing transaction is > >> - * using it > >> - */ > >> - BUFFER_TRACE(gd_bh, "get_write_access"); > >> - err = ext4_journal_get_write_access(handle, sb, gd_bh, EXT4_JTR_NONE); > >> - if (err) > >> - goto error_return; > >> -#ifdef AGGRESSIVE_CHECK > >> - { > >> - int i; > >> - for (i = 0; i < count_clusters; i++) > >> - BUG_ON(!mb_test_bit(bit + i, bitmap_bh->b_data)); > >> - } > >> -#endif > >> + count_clusters = EXT4_NUM_B2C(sbi, count); > >> trace_ext4_mballoc_free(sb, inode, block_group, bit, count_clusters); > >> > >> /* __GFP_NOFAIL: retry infinitely, ignore TIF_MEMDIE and memcg limit. */ > >> @@ -6230,6 +6199,22 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, > >> if (err) > >> goto error_return; > >> > >> +#ifdef AGGRESSIVE_CHECK > >> + mark_flags |= EXT4_MB_BITMAP_MARKED_CHECK; > >> +#endif > >> + err = ext4_mb_mark_group_bb(&mc, block_group, bit, count_clusters, > >> + mark_flags); > >> + > >> + > >> + if (err && mc.changed == 0) { > >> + ext4_mb_unload_buddy(&e4b); > >> + goto error_return; > >> + } > >> + > >> +#ifdef AGGRESSIVE_CHECK > >> + BUG_ON(mc.changed != count_clusters); > >> +#endif > >> + > >> /* > >> * We need to make sure we don't reuse the freed block until after the > >> * transaction is committed. We make an exception if the inode is to be > >> @@ -6252,13 +6237,8 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, > >> new_entry->efd_tid = handle->h_transaction->t_tid; > >> > >> ext4_lock_group(sb, block_group); > >> - mb_clear_bits(bitmap_bh->b_data, bit, count_clusters); > >> ext4_mb_free_metadata(handle, &e4b, new_entry); > >> } else { > >> - /* need to update group_info->bb_free and bitmap > >> - * with group lock held. generate_buddy look at > >> - * them with group lock_held > >> - */ > >> if (test_opt(sb, DISCARD)) { > >> err = ext4_issue_discard(sb, block_group, bit, > >> count_clusters, NULL); > >> @@ -6271,23 +6251,11 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, > >> EXT4_MB_GRP_CLEAR_TRIMMED(e4b.bd_info); > >> > >> ext4_lock_group(sb, block_group); > >> - mb_clear_bits(bitmap_bh->b_data, bit, count_clusters); > >> mb_free_blocks(inode, &e4b, bit, count_clusters); > >> } > >> > >> - ret = ext4_free_group_clusters(sb, gdp) + count_clusters; > >> - ext4_free_group_clusters_set(sb, gdp, ret); > >> - ext4_block_bitmap_csum_set(sb, gdp, bitmap_bh); > >> - ext4_group_desc_csum_set(sb, block_group, gdp); > >> ext4_unlock_group(sb, block_group); > >> > >> - if (sbi->s_log_groups_per_flex) { > >> - ext4_group_t flex_group = ext4_flex_group(sbi, block_group); > >> - atomic64_add(count_clusters, > >> - &sbi_array_rcu_deref(sbi, s_flex_groups, > >> - flex_group)->free_clusters); > >> - } > >> - > >> /* > >> * on a bigalloc file system, defer the s_freeclusters_counter > >> * update to the caller (ext4_remove_space and friends) so they > >> @@ -6302,26 +6270,14 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, > >> > >> ext4_mb_unload_buddy(&e4b); > >> > >> - /* We dirtied the bitmap block */ > >> - BUFFER_TRACE(bitmap_bh, "dirtied bitmap block"); > >> - err = ext4_handle_dirty_metadata(handle, NULL, bitmap_bh); > >> - > >> - /* And the group descriptor block */ > >> - BUFFER_TRACE(gd_bh, "dirtied group descriptor block"); > >> - ret = ext4_handle_dirty_metadata(handle, NULL, gd_bh); > >> - if (!err) > >> - err = ret; > >> - > >> if (overflow && !err) { > >> block += count; > >> count = overflow; > >> - put_bh(bitmap_bh); > >> /* The range changed so it's no longer validated */ > >> flags &= ~EXT4_FREE_BLOCKS_VALIDATED; > >> goto do_more; > >> } > >> error_return: > >> - brelse(bitmap_bh); > >> ext4_std_error(sb, err); > >> return; > >> } > >> -- > >> 2.30.0 > >> > > > > -- > Best wishes > Kemeng Shi >
diff --git a/fs/ext4/mballoc.c b/fs/ext4/mballoc.c index 46b37f5c9223..e4f1b34448e3 100644 --- a/fs/ext4/mballoc.c +++ b/fs/ext4/mballoc.c @@ -6135,19 +6135,21 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, ext4_fsblk_t block, unsigned long count, int flags) { - struct buffer_head *bitmap_bh = NULL; + struct ext4_mark_context mc = { + .handle = handle, + .sb = inode->i_sb, + .state = 0, + }; struct super_block *sb = inode->i_sb; - struct ext4_group_desc *gdp; struct ext4_group_info *grp; unsigned int overflow; ext4_grpblk_t bit; - struct buffer_head *gd_bh; ext4_group_t block_group; struct ext4_sb_info *sbi; struct ext4_buddy e4b; unsigned int count_clusters; int err = 0; - int ret; + int mark_flags = 0; sbi = EXT4_SB(sb); @@ -6179,18 +6181,6 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, /* The range changed so it's no longer validated */ flags &= ~EXT4_FREE_BLOCKS_VALIDATED; } - count_clusters = EXT4_NUM_B2C(sbi, count); - bitmap_bh = ext4_read_block_bitmap(sb, block_group); - if (IS_ERR(bitmap_bh)) { - err = PTR_ERR(bitmap_bh); - bitmap_bh = NULL; - goto error_return; - } - gdp = ext4_get_group_desc(sb, block_group, &gd_bh); - if (!gdp) { - err = -EIO; - goto error_return; - } if (!(flags & EXT4_FREE_BLOCKS_VALIDATED) && !ext4_inode_block_valid(inode, block, count)) { @@ -6200,28 +6190,7 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, goto error_return; } - BUFFER_TRACE(bitmap_bh, "getting write access"); - err = ext4_journal_get_write_access(handle, sb, bitmap_bh, - EXT4_JTR_NONE); - if (err) - goto error_return; - - /* - * We are about to modify some metadata. Call the journal APIs - * to unshare ->b_data if a currently-committing transaction is - * using it - */ - BUFFER_TRACE(gd_bh, "get_write_access"); - err = ext4_journal_get_write_access(handle, sb, gd_bh, EXT4_JTR_NONE); - if (err) - goto error_return; -#ifdef AGGRESSIVE_CHECK - { - int i; - for (i = 0; i < count_clusters; i++) - BUG_ON(!mb_test_bit(bit + i, bitmap_bh->b_data)); - } -#endif + count_clusters = EXT4_NUM_B2C(sbi, count); trace_ext4_mballoc_free(sb, inode, block_group, bit, count_clusters); /* __GFP_NOFAIL: retry infinitely, ignore TIF_MEMDIE and memcg limit. */ @@ -6230,6 +6199,22 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, if (err) goto error_return; +#ifdef AGGRESSIVE_CHECK + mark_flags |= EXT4_MB_BITMAP_MARKED_CHECK; +#endif + err = ext4_mb_mark_group_bb(&mc, block_group, bit, count_clusters, + mark_flags); + + + if (err && mc.changed == 0) { + ext4_mb_unload_buddy(&e4b); + goto error_return; + } + +#ifdef AGGRESSIVE_CHECK + BUG_ON(mc.changed != count_clusters); +#endif + /* * We need to make sure we don't reuse the freed block until after the * transaction is committed. We make an exception if the inode is to be @@ -6252,13 +6237,8 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, new_entry->efd_tid = handle->h_transaction->t_tid; ext4_lock_group(sb, block_group); - mb_clear_bits(bitmap_bh->b_data, bit, count_clusters); ext4_mb_free_metadata(handle, &e4b, new_entry); } else { - /* need to update group_info->bb_free and bitmap - * with group lock held. generate_buddy look at - * them with group lock_held - */ if (test_opt(sb, DISCARD)) { err = ext4_issue_discard(sb, block_group, bit, count_clusters, NULL); @@ -6271,23 +6251,11 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, EXT4_MB_GRP_CLEAR_TRIMMED(e4b.bd_info); ext4_lock_group(sb, block_group); - mb_clear_bits(bitmap_bh->b_data, bit, count_clusters); mb_free_blocks(inode, &e4b, bit, count_clusters); } - ret = ext4_free_group_clusters(sb, gdp) + count_clusters; - ext4_free_group_clusters_set(sb, gdp, ret); - ext4_block_bitmap_csum_set(sb, gdp, bitmap_bh); - ext4_group_desc_csum_set(sb, block_group, gdp); ext4_unlock_group(sb, block_group); - if (sbi->s_log_groups_per_flex) { - ext4_group_t flex_group = ext4_flex_group(sbi, block_group); - atomic64_add(count_clusters, - &sbi_array_rcu_deref(sbi, s_flex_groups, - flex_group)->free_clusters); - } - /* * on a bigalloc file system, defer the s_freeclusters_counter * update to the caller (ext4_remove_space and friends) so they @@ -6302,26 +6270,14 @@ static void ext4_mb_clear_bb(handle_t *handle, struct inode *inode, ext4_mb_unload_buddy(&e4b); - /* We dirtied the bitmap block */ - BUFFER_TRACE(bitmap_bh, "dirtied bitmap block"); - err = ext4_handle_dirty_metadata(handle, NULL, bitmap_bh); - - /* And the group descriptor block */ - BUFFER_TRACE(gd_bh, "dirtied group descriptor block"); - ret = ext4_handle_dirty_metadata(handle, NULL, gd_bh); - if (!err) - err = ret; - if (overflow && !err) { block += count; count = overflow; - put_bh(bitmap_bh); /* The range changed so it's no longer validated */ flags &= ~EXT4_FREE_BLOCKS_VALIDATED; goto do_more; } error_return: - brelse(bitmap_bh); ext4_std_error(sb, err); return; }
call ext4_mb_mark_group_bb in ext4_mb_clear_bb to remove repeat code to update block bitmap and group descriptor on disk. Note: ext4_mb_clear_bb will update buddy and bitmap in two critical sections instead of update in the same critical section. Original lock behavior introduced in 7a2fcbf7f857 ("ext4: don't use blocks freed but not yet committed in buddy cache init") to avoid race betwwen ext4_mb_free_blocks and ext4_mb_init_cache: ext4_mb_load_buddy_gfp ext4_lock_group mb_clear_bits(bitmap_bh, ...) mb_free_blocks/ext4_mb_free_metadata ext4_unlock_group ext4_mb_unload_buddy New lock behavior in this patch: ext4_mb_load_buddy_gfp ext4_lock_group mb_clear_bits(bitmap_bh, ...) ext4_unlock_group /* no ext4_mb_init_cache for the same group will be called as ext4_mb_load_buddy_gfp will ensure buddy page is update-to-date */ ext4_lock_group mb_free_blocks/ext4_mb_free_metadata ext4_unlock_group ext4_mb_unload_buddy As buddy page for group is always update-to-date between ext4_mb_load_buddy_gfp and ext4_mb_unload_buddy. Then no ext4_mb_init_cache will be called for the same group concurrentlly when we update bitmap and buddy page betwwen buddy load and unload. Signed-off-by: Kemeng Shi <shikemeng@huaweicloud.com> --- fs/ext4/mballoc.c | 90 ++++++++++++----------------------------------- 1 file changed, 23 insertions(+), 67 deletions(-)