@@ -333,7 +333,7 @@ static noinline int lock_delalloc_folios(struct inode *inode,
333333 goto out ;
334334 }
335335 range_start = max_t (u64 , folio_pos (folio ), start );
336- range_len = min_t (u64 , folio_end (folio ), end + 1 ) - range_start ;
336+ range_len = min_t (u64 , folio_next_pos (folio ), end + 1 ) - range_start ;
337337 btrfs_folio_set_lock (fs_info , folio , range_start , range_len );
338338
339339 processed_end = range_start + range_len - 1 ;
@@ -387,7 +387,7 @@ noinline_for_stack bool find_lock_delalloc_range(struct inode *inode,
387387 ASSERT (orig_end > orig_start );
388388
389389 /* The range should at least cover part of the folio */
390- ASSERT (!(orig_start >= folio_end (locked_folio ) ||
390+ ASSERT (!(orig_start >= folio_next_pos (locked_folio ) ||
391391 orig_end <= folio_pos (locked_folio )));
392392again :
393393 /* step one, find a bunch of delalloc bytes starting at start */
@@ -493,7 +493,7 @@ static void end_folio_read(struct folio *folio, bool uptodate, u64 start, u32 le
493493 struct btrfs_fs_info * fs_info = folio_to_fs_info (folio );
494494
495495 ASSERT (folio_pos (folio ) <= start &&
496- start + len <= folio_end (folio ));
496+ start + len <= folio_next_pos (folio ));
497497
498498 if (uptodate && btrfs_verify_folio (folio , start , len ))
499499 btrfs_folio_set_uptodate (fs_info , folio , start , len );
@@ -1201,7 +1201,7 @@ static bool can_skip_one_ordered_range(struct btrfs_inode *inode,
12011201 * finished our folio read and unlocked the folio.
12021202 */
12031203 if (btrfs_folio_test_dirty (fs_info , folio , cur , blocksize )) {
1204- u64 range_len = min ( folio_end (folio ),
1204+ u64 range_len = umin ( folio_next_pos (folio ),
12051205 ordered -> file_offset + ordered -> num_bytes ) - cur ;
12061206
12071207 ret = true;
@@ -1223,7 +1223,7 @@ static bool can_skip_one_ordered_range(struct btrfs_inode *inode,
12231223 * So we return true and update @next_ret to the OE/folio boundary.
12241224 */
12251225 if (btrfs_folio_test_uptodate (fs_info , folio , cur , blocksize )) {
1226- u64 range_len = min ( folio_end (folio ),
1226+ u64 range_len = umin ( folio_next_pos (folio ),
12271227 ordered -> file_offset + ordered -> num_bytes ) - cur ;
12281228
12291229 /*
@@ -2215,7 +2215,7 @@ static noinline_for_stack void write_one_eb(struct extent_buffer *eb,
22152215 for (int i = 0 ; i < num_extent_folios (eb ); i ++ ) {
22162216 struct folio * folio = eb -> folios [i ];
22172217 u64 range_start = max_t (u64 , eb -> start , folio_pos (folio ));
2218- u32 range_len = min_t (u64 , folio_end (folio ),
2218+ u32 range_len = min_t (u64 , folio_next_pos (folio ),
22192219 eb -> start + eb -> len ) - range_start ;
22202220
22212221 folio_lock (folio );
@@ -2619,7 +2619,7 @@ void extent_write_locked_range(struct inode *inode, const struct folio *locked_f
26192619 continue ;
26202620 }
26212621
2622- cur_end = min_t (u64 , folio_end (folio ) - 1 , end );
2622+ cur_end = min_t (u64 , folio_next_pos (folio ) - 1 , end );
26232623 cur_len = cur_end + 1 - cur ;
26242624
26252625 ASSERT (folio_test_locked (folio ));
@@ -3860,7 +3860,7 @@ int read_extent_buffer_pages_nowait(struct extent_buffer *eb, int mirror_num,
38603860 for (int i = 0 ; i < num_extent_folios (eb ); i ++ ) {
38613861 struct folio * folio = eb -> folios [i ];
38623862 u64 range_start = max_t (u64 , eb -> start , folio_pos (folio ));
3863- u32 range_len = min_t (u64 , folio_end (folio ),
3863+ u32 range_len = min_t (u64 , folio_next_pos (folio ),
38643864 eb -> start + eb -> len ) - range_start ;
38653865
38663866 bio_add_folio_nofail (& bbio -> bio , folio , range_len ,
0 commit comments