@@ -28,37 +28,37 @@ xfs_bnobt_dup_cursor(
2828 struct xfs_btree_cur * cur )
2929{
3030 return xfs_bnobt_init_cursor (cur -> bc_mp , cur -> bc_tp , cur -> bc_ag .agbp ,
31- cur -> bc_ag . pag );
31+ to_perag ( cur -> bc_group ) );
3232}
3333
3434STATIC struct xfs_btree_cur *
3535xfs_cntbt_dup_cursor (
3636 struct xfs_btree_cur * cur )
3737{
3838 return xfs_cntbt_init_cursor (cur -> bc_mp , cur -> bc_tp , cur -> bc_ag .agbp ,
39- cur -> bc_ag . pag );
39+ to_perag ( cur -> bc_group ) );
4040}
4141
42-
4342STATIC void
4443xfs_allocbt_set_root (
4544 struct xfs_btree_cur * cur ,
4645 const union xfs_btree_ptr * ptr ,
4746 int inc )
4847{
49- struct xfs_buf * agbp = cur -> bc_ag .agbp ;
50- struct xfs_agf * agf = agbp -> b_addr ;
48+ struct xfs_perag * pag = to_perag (cur -> bc_group );
49+ struct xfs_buf * agbp = cur -> bc_ag .agbp ;
50+ struct xfs_agf * agf = agbp -> b_addr ;
5151
5252 ASSERT (ptr -> s != 0 );
5353
5454 if (xfs_btree_is_bno (cur -> bc_ops )) {
5555 agf -> agf_bno_root = ptr -> s ;
5656 be32_add_cpu (& agf -> agf_bno_level , inc );
57- cur -> bc_ag . pag -> pagf_bno_level += inc ;
57+ pag -> pagf_bno_level += inc ;
5858 } else {
5959 agf -> agf_cnt_root = ptr -> s ;
6060 be32_add_cpu (& agf -> agf_cnt_level , inc );
61- cur -> bc_ag . pag -> pagf_cnt_level += inc ;
61+ pag -> pagf_cnt_level += inc ;
6262 }
6363
6464 xfs_alloc_log_agf (cur -> bc_tp , agbp , XFS_AGF_ROOTS | XFS_AGF_LEVELS );
@@ -75,7 +75,7 @@ xfs_allocbt_alloc_block(
7575 xfs_agblock_t bno ;
7676
7777 /* Allocate the new block from the freelist. If we can't, give up. */
78- error = xfs_alloc_get_freelist (cur -> bc_ag . pag , cur -> bc_tp ,
78+ error = xfs_alloc_get_freelist (to_perag ( cur -> bc_group ) , cur -> bc_tp ,
7979 cur -> bc_ag .agbp , & bno , 1 );
8080 if (error )
8181 return error ;
@@ -86,7 +86,7 @@ xfs_allocbt_alloc_block(
8686 }
8787
8888 atomic64_inc (& cur -> bc_mp -> m_allocbt_blks );
89- xfs_extent_busy_reuse (pag_group ( cur -> bc_ag . pag ) , bno , 1 , false);
89+ xfs_extent_busy_reuse (cur -> bc_group , bno , 1 , false);
9090
9191 new -> s = cpu_to_be32 (bno );
9292
@@ -104,8 +104,8 @@ xfs_allocbt_free_block(
104104 int error ;
105105
106106 bno = xfs_daddr_to_agbno (cur -> bc_mp , xfs_buf_daddr (bp ));
107- error = xfs_alloc_put_freelist (cur -> bc_ag . pag , cur -> bc_tp , agbp , NULL ,
108- bno , 1 );
107+ error = xfs_alloc_put_freelist (to_perag ( cur -> bc_group ) , cur -> bc_tp ,
108+ agbp , NULL , bno , 1 );
109109 if (error )
110110 return error ;
111111
@@ -178,7 +178,7 @@ xfs_allocbt_init_ptr_from_cur(
178178{
179179 struct xfs_agf * agf = cur -> bc_ag .agbp -> b_addr ;
180180
181- ASSERT (pag_agno ( cur -> bc_ag . pag ) == be32_to_cpu (agf -> agf_seqno ));
181+ ASSERT (cur -> bc_group -> xg_gno == be32_to_cpu (agf -> agf_seqno ));
182182
183183 if (xfs_btree_is_bno (cur -> bc_ops ))
184184 ptr -> s = agf -> agf_bno_root ;
@@ -492,7 +492,7 @@ xfs_bnobt_init_cursor(
492492
493493 cur = xfs_btree_alloc_cursor (mp , tp , & xfs_bnobt_ops ,
494494 mp -> m_alloc_maxlevels , xfs_allocbt_cur_cache );
495- cur -> bc_ag . pag = xfs_perag_hold ( pag );
495+ cur -> bc_group = xfs_group_hold ( pag_group ( pag ) );
496496 cur -> bc_ag .agbp = agbp ;
497497 if (agbp ) {
498498 struct xfs_agf * agf = agbp -> b_addr ;
@@ -518,7 +518,7 @@ xfs_cntbt_init_cursor(
518518
519519 cur = xfs_btree_alloc_cursor (mp , tp , & xfs_cntbt_ops ,
520520 mp -> m_alloc_maxlevels , xfs_allocbt_cur_cache );
521- cur -> bc_ag . pag = xfs_perag_hold ( pag );
521+ cur -> bc_group = xfs_group_hold ( pag_group ( pag ) );
522522 cur -> bc_ag .agbp = agbp ;
523523 if (agbp ) {
524524 struct xfs_agf * agf = agbp -> b_addr ;
0 commit comments