diff options
| -rw-r--r-- | fs/xfs/libxfs/xfs_bmap.c | 40 | ||||
| -rw-r--r-- | fs/xfs/libxfs/xfs_format.h | 62 | ||||
| -rw-r--r-- | fs/xfs/libxfs/xfs_ialloc.c | 6 | ||||
| -rw-r--r-- | fs/xfs/libxfs/xfs_sb.c | 12 | ||||
| -rw-r--r-- | fs/xfs/xfs_fsops.c | 20 | ||||
| -rw-r--r-- | fs/xfs/xfs_iomap.c | 3 | ||||
| -rw-r--r-- | fs/xfs/xfs_linux.h | 9 | ||||
| -rw-r--r-- | fs/xfs/xfs_log_recover.c | 4 | ||||
| -rw-r--r-- | fs/xfs/xfs_mount.c | 918 | ||||
| -rw-r--r-- | fs/xfs/xfs_mount.h | 95 | ||||
| -rw-r--r-- | fs/xfs/xfs_super.c | 105 | ||||
| -rw-r--r-- | fs/xfs/xfs_super.h | 2 | ||||
| -rw-r--r-- | fs/xfs/xfs_trans.c | 234 |
13 files changed, 381 insertions, 1129 deletions
diff --git a/fs/xfs/libxfs/xfs_bmap.c b/fs/xfs/libxfs/xfs_bmap.c index 60cfa90163b8..8ae3775a9b63 100644 --- a/fs/xfs/libxfs/xfs_bmap.c +++ b/fs/xfs/libxfs/xfs_bmap.c | |||
| @@ -2215,9 +2215,8 @@ xfs_bmap_add_extent_delay_real( | |||
| 2215 | diff = (int)(temp + temp2 - startblockval(PREV.br_startblock) - | 2215 | diff = (int)(temp + temp2 - startblockval(PREV.br_startblock) - |
| 2216 | (bma->cur ? bma->cur->bc_private.b.allocated : 0)); | 2216 | (bma->cur ? bma->cur->bc_private.b.allocated : 0)); |
| 2217 | if (diff > 0) { | 2217 | if (diff > 0) { |
| 2218 | error = xfs_icsb_modify_counters(bma->ip->i_mount, | 2218 | error = xfs_mod_fdblocks(bma->ip->i_mount, |
| 2219 | XFS_SBS_FDBLOCKS, | 2219 | -((int64_t)diff), false); |
| 2220 | -((int64_t)diff), 0); | ||
| 2221 | ASSERT(!error); | 2220 | ASSERT(!error); |
| 2222 | if (error) | 2221 | if (error) |
| 2223 | goto done; | 2222 | goto done; |
| @@ -2268,9 +2267,8 @@ xfs_bmap_add_extent_delay_real( | |||
| 2268 | temp += bma->cur->bc_private.b.allocated; | 2267 | temp += bma->cur->bc_private.b.allocated; |
| 2269 | ASSERT(temp <= da_old); | 2268 | ASSERT(temp <= da_old); |
| 2270 | if (temp < da_old) | 2269 | if (temp < da_old) |
| 2271 | xfs_icsb_modify_counters(bma->ip->i_mount, | 2270 | xfs_mod_fdblocks(bma->ip->i_mount, |
| 2272 | XFS_SBS_FDBLOCKS, | 2271 | (int64_t)(da_old - temp), false); |
| 2273 | (int64_t)(da_old - temp), 0); | ||
| 2274 | } | 2272 | } |
| 2275 | 2273 | ||
| 2276 | /* clear out the allocated field, done with it now in any case. */ | 2274 | /* clear out the allocated field, done with it now in any case. */ |
| @@ -2948,8 +2946,8 @@ xfs_bmap_add_extent_hole_delay( | |||
| 2948 | } | 2946 | } |
| 2949 | if (oldlen != newlen) { | 2947 | if (oldlen != newlen) { |
| 2950 | ASSERT(oldlen > newlen); | 2948 | ASSERT(oldlen > newlen); |
| 2951 | xfs_icsb_modify_counters(ip->i_mount, XFS_SBS_FDBLOCKS, | 2949 | xfs_mod_fdblocks(ip->i_mount, (int64_t)(oldlen - newlen), |
| 2952 | (int64_t)(oldlen - newlen), 0); | 2950 | false); |
| 2953 | /* | 2951 | /* |
| 2954 | * Nothing to do for disk quota accounting here. | 2952 | * Nothing to do for disk quota accounting here. |
| 2955 | */ | 2953 | */ |
| @@ -4166,18 +4164,15 @@ xfs_bmapi_reserve_delalloc( | |||
| 4166 | ASSERT(indlen > 0); | 4164 | ASSERT(indlen > 0); |
| 4167 | 4165 | ||
| 4168 | if (rt) { | 4166 | if (rt) { |
| 4169 | error = xfs_mod_incore_sb(mp, XFS_SBS_FREXTENTS, | 4167 | error = xfs_mod_frextents(mp, -((int64_t)extsz)); |
| 4170 | -((int64_t)extsz), 0); | ||
| 4171 | } else { | 4168 | } else { |
| 4172 | error = xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, | 4169 | error = xfs_mod_fdblocks(mp, -((int64_t)alen), false); |
| 4173 | -((int64_t)alen), 0); | ||
| 4174 | } | 4170 | } |
| 4175 | 4171 | ||
| 4176 | if (error) | 4172 | if (error) |
| 4177 | goto out_unreserve_quota; | 4173 | goto out_unreserve_quota; |
| 4178 | 4174 | ||
| 4179 | error = xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, | 4175 | error = xfs_mod_fdblocks(mp, -((int64_t)indlen), false); |
| 4180 | -((int64_t)indlen), 0); | ||
| 4181 | if (error) | 4176 | if (error) |
| 4182 | goto out_unreserve_blocks; | 4177 | goto out_unreserve_blocks; |
| 4183 | 4178 | ||
| @@ -4204,9 +4199,9 @@ xfs_bmapi_reserve_delalloc( | |||
| 4204 | 4199 | ||
| 4205 | out_unreserve_blocks: | 4200 | out_unreserve_blocks: |
| 4206 | if (rt) | 4201 | if (rt) |
| 4207 | xfs_mod_incore_sb(mp, XFS_SBS_FREXTENTS, extsz, 0); | 4202 | xfs_mod_frextents(mp, extsz); |
| 4208 | else | 4203 | else |
| 4209 | xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, alen, 0); | 4204 | xfs_mod_fdblocks(mp, alen, false); |
| 4210 | out_unreserve_quota: | 4205 | out_unreserve_quota: |
| 4211 | if (XFS_IS_QUOTA_ON(mp)) | 4206 | if (XFS_IS_QUOTA_ON(mp)) |
| 4212 | xfs_trans_unreserve_quota_nblks(NULL, ip, (long)alen, 0, rt ? | 4207 | xfs_trans_unreserve_quota_nblks(NULL, ip, (long)alen, 0, rt ? |
| @@ -5019,10 +5014,8 @@ xfs_bmap_del_extent( | |||
| 5019 | * Nothing to do for disk quota accounting here. | 5014 | * Nothing to do for disk quota accounting here. |
| 5020 | */ | 5015 | */ |
| 5021 | ASSERT(da_old >= da_new); | 5016 | ASSERT(da_old >= da_new); |
| 5022 | if (da_old > da_new) { | 5017 | if (da_old > da_new) |
| 5023 | xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, | 5018 | xfs_mod_fdblocks(mp, (int64_t)(da_old - da_new), false); |
| 5024 | (int64_t)(da_old - da_new), 0); | ||
| 5025 | } | ||
| 5026 | done: | 5019 | done: |
| 5027 | *logflagsp = flags; | 5020 | *logflagsp = flags; |
| 5028 | return error; | 5021 | return error; |
| @@ -5291,14 +5284,13 @@ xfs_bunmapi( | |||
| 5291 | 5284 | ||
| 5292 | rtexts = XFS_FSB_TO_B(mp, del.br_blockcount); | 5285 | rtexts = XFS_FSB_TO_B(mp, del.br_blockcount); |
| 5293 | do_div(rtexts, mp->m_sb.sb_rextsize); | 5286 | do_div(rtexts, mp->m_sb.sb_rextsize); |
| 5294 | xfs_mod_incore_sb(mp, XFS_SBS_FREXTENTS, | 5287 | xfs_mod_frextents(mp, (int64_t)rtexts); |
| 5295 | (int64_t)rtexts, 0); | ||
| 5296 | (void)xfs_trans_reserve_quota_nblks(NULL, | 5288 | (void)xfs_trans_reserve_quota_nblks(NULL, |
| 5297 | ip, -((long)del.br_blockcount), 0, | 5289 | ip, -((long)del.br_blockcount), 0, |
| 5298 | XFS_QMOPT_RES_RTBLKS); | 5290 | XFS_QMOPT_RES_RTBLKS); |
| 5299 | } else { | 5291 | } else { |
| 5300 | xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, | 5292 | xfs_mod_fdblocks(mp, (int64_t)del.br_blockcount, |
| 5301 | (int64_t)del.br_blockcount, 0); | 5293 | false); |
| 5302 | (void)xfs_trans_reserve_quota_nblks(NULL, | 5294 | (void)xfs_trans_reserve_quota_nblks(NULL, |
| 5303 | ip, -((long)del.br_blockcount), 0, | 5295 | ip, -((long)del.br_blockcount), 0, |
| 5304 | XFS_QMOPT_RES_REGBLKS); | 5296 | XFS_QMOPT_RES_REGBLKS); |
diff --git a/fs/xfs/libxfs/xfs_format.h b/fs/xfs/libxfs/xfs_format.h index 8eb718979383..4daaa662337b 100644 --- a/fs/xfs/libxfs/xfs_format.h +++ b/fs/xfs/libxfs/xfs_format.h | |||
| @@ -264,68 +264,6 @@ typedef struct xfs_dsb { | |||
| 264 | /* must be padded to 64 bit alignment */ | 264 | /* must be padded to 64 bit alignment */ |
| 265 | } xfs_dsb_t; | 265 | } xfs_dsb_t; |
| 266 | 266 | ||
| 267 | /* | ||
| 268 | * Sequence number values for the fields. | ||
| 269 | */ | ||
| 270 | typedef enum { | ||
| 271 | XFS_SBS_MAGICNUM, XFS_SBS_BLOCKSIZE, XFS_SBS_DBLOCKS, XFS_SBS_RBLOCKS, | ||
| 272 | XFS_SBS_REXTENTS, XFS_SBS_UUID, XFS_SBS_LOGSTART, XFS_SBS_ROOTINO, | ||
| 273 | XFS_SBS_RBMINO, XFS_SBS_RSUMINO, XFS_SBS_REXTSIZE, XFS_SBS_AGBLOCKS, | ||
| 274 | XFS_SBS_AGCOUNT, XFS_SBS_RBMBLOCKS, XFS_SBS_LOGBLOCKS, | ||
| 275 | XFS_SBS_VERSIONNUM, XFS_SBS_SECTSIZE, XFS_SBS_INODESIZE, | ||
| 276 | XFS_SBS_INOPBLOCK, XFS_SBS_FNAME, XFS_SBS_BLOCKLOG, | ||
| 277 | XFS_SBS_SECTLOG, XFS_SBS_INODELOG, XFS_SBS_INOPBLOG, XFS_SBS_AGBLKLOG, | ||
| 278 | XFS_SBS_REXTSLOG, XFS_SBS_INPROGRESS, XFS_SBS_IMAX_PCT, XFS_SBS_ICOUNT, | ||
| 279 | XFS_SBS_IFREE, XFS_SBS_FDBLOCKS, XFS_SBS_FREXTENTS, XFS_SBS_UQUOTINO, | ||
| 280 | XFS_SBS_GQUOTINO, XFS_SBS_QFLAGS, XFS_SBS_FLAGS, XFS_SBS_SHARED_VN, | ||
| 281 | XFS_SBS_INOALIGNMT, XFS_SBS_UNIT, XFS_SBS_WIDTH, XFS_SBS_DIRBLKLOG, | ||
| 282 | XFS_SBS_LOGSECTLOG, XFS_SBS_LOGSECTSIZE, XFS_SBS_LOGSUNIT, | ||
| 283 | XFS_SBS_FEATURES2, XFS_SBS_BAD_FEATURES2, XFS_SBS_FEATURES_COMPAT, | ||
| 284 | XFS_SBS_FEATURES_RO_COMPAT, XFS_SBS_FEATURES_INCOMPAT, | ||
| 285 | XFS_SBS_FEATURES_LOG_INCOMPAT, XFS_SBS_CRC, XFS_SBS_PAD, | ||
| 286 | XFS_SBS_PQUOTINO, XFS_SBS_LSN, | ||
| 287 | XFS_SBS_FIELDCOUNT | ||
| 288 | } xfs_sb_field_t; | ||
| 289 | |||
| 290 | /* | ||
| 291 | * Mask values, defined based on the xfs_sb_field_t values. | ||
| 292 | * Only define the ones we're using. | ||
| 293 | */ | ||
| 294 | #define XFS_SB_MVAL(x) (1LL << XFS_SBS_ ## x) | ||
| 295 | #define XFS_SB_UUID XFS_SB_MVAL(UUID) | ||
| 296 | #define XFS_SB_FNAME XFS_SB_MVAL(FNAME) | ||
| 297 | #define XFS_SB_ROOTINO XFS_SB_MVAL(ROOTINO) | ||
| 298 | #define XFS_SB_RBMINO XFS_SB_MVAL(RBMINO) | ||
| 299 | #define XFS_SB_RSUMINO XFS_SB_MVAL(RSUMINO) | ||
| 300 | #define XFS_SB_VERSIONNUM XFS_SB_MVAL(VERSIONNUM) | ||
| 301 | #define XFS_SB_UQUOTINO XFS_SB_MVAL(UQUOTINO) | ||
| 302 | #define XFS_SB_GQUOTINO XFS_SB_MVAL(GQUOTINO) | ||
| 303 | #define XFS_SB_QFLAGS XFS_SB_MVAL(QFLAGS) | ||
| 304 | #define XFS_SB_SHARED_VN XFS_SB_MVAL(SHARED_VN) | ||
| 305 | #define XFS_SB_UNIT XFS_SB_MVAL(UNIT) | ||
| 306 | #define XFS_SB_WIDTH XFS_SB_MVAL(WIDTH) | ||
| 307 | #define XFS_SB_ICOUNT XFS_SB_MVAL(ICOUNT) | ||
| 308 | #define XFS_SB_IFREE XFS_SB_MVAL(IFREE) | ||
| 309 | #define XFS_SB_FDBLOCKS XFS_SB_MVAL(FDBLOCKS) | ||
| 310 | #define XFS_SB_FEATURES2 (XFS_SB_MVAL(FEATURES2) | \ | ||
| 311 | XFS_SB_MVAL(BAD_FEATURES2)) | ||
| 312 | #define XFS_SB_FEATURES_COMPAT XFS_SB_MVAL(FEATURES_COMPAT) | ||
| 313 | #define XFS_SB_FEATURES_RO_COMPAT XFS_SB_MVAL(FEATURES_RO_COMPAT) | ||
| 314 | #define XFS_SB_FEATURES_INCOMPAT XFS_SB_MVAL(FEATURES_INCOMPAT) | ||
| 315 | #define XFS_SB_FEATURES_LOG_INCOMPAT XFS_SB_MVAL(FEATURES_LOG_INCOMPAT) | ||
| 316 | #define XFS_SB_CRC XFS_SB_MVAL(CRC) | ||
| 317 | #define XFS_SB_PQUOTINO XFS_SB_MVAL(PQUOTINO) | ||
| 318 | #define XFS_SB_NUM_BITS ((int)XFS_SBS_FIELDCOUNT) | ||
| 319 | #define XFS_SB_ALL_BITS ((1LL << XFS_SB_NUM_BITS) - 1) | ||
| 320 | #define XFS_SB_MOD_BITS \ | ||
| 321 | (XFS_SB_UUID | XFS_SB_ROOTINO | XFS_SB_RBMINO | XFS_SB_RSUMINO | \ | ||
| 322 | XFS_SB_VERSIONNUM | XFS_SB_UQUOTINO | XFS_SB_GQUOTINO | \ | ||
| 323 | XFS_SB_QFLAGS | XFS_SB_SHARED_VN | XFS_SB_UNIT | XFS_SB_WIDTH | \ | ||
| 324 | XFS_SB_ICOUNT | XFS_SB_IFREE | XFS_SB_FDBLOCKS | XFS_SB_FEATURES2 | \ | ||
| 325 | XFS_SB_FEATURES_COMPAT | XFS_SB_FEATURES_RO_COMPAT | \ | ||
| 326 | XFS_SB_FEATURES_INCOMPAT | XFS_SB_FEATURES_LOG_INCOMPAT | \ | ||
| 327 | XFS_SB_PQUOTINO) | ||
| 328 | |||
| 329 | 267 | ||
| 330 | /* | 268 | /* |
| 331 | * Misc. Flags - warning - these will be cleared by xfs_repair unless | 269 | * Misc. Flags - warning - these will be cleared by xfs_repair unless |
diff --git a/fs/xfs/libxfs/xfs_ialloc.c b/fs/xfs/libxfs/xfs_ialloc.c index db0444893e96..07349a183a11 100644 --- a/fs/xfs/libxfs/xfs_ialloc.c +++ b/fs/xfs/libxfs/xfs_ialloc.c | |||
| @@ -376,7 +376,8 @@ xfs_ialloc_ag_alloc( | |||
| 376 | */ | 376 | */ |
| 377 | newlen = args.mp->m_ialloc_inos; | 377 | newlen = args.mp->m_ialloc_inos; |
| 378 | if (args.mp->m_maxicount && | 378 | if (args.mp->m_maxicount && |
| 379 | args.mp->m_sb.sb_icount + newlen > args.mp->m_maxicount) | 379 | percpu_counter_read(&args.mp->m_icount) + newlen > |
| 380 | args.mp->m_maxicount) | ||
| 380 | return -ENOSPC; | 381 | return -ENOSPC; |
| 381 | args.minlen = args.maxlen = args.mp->m_ialloc_blks; | 382 | args.minlen = args.maxlen = args.mp->m_ialloc_blks; |
| 382 | /* | 383 | /* |
| @@ -1340,7 +1341,8 @@ xfs_dialloc( | |||
| 1340 | * inode. | 1341 | * inode. |
| 1341 | */ | 1342 | */ |
| 1342 | if (mp->m_maxicount && | 1343 | if (mp->m_maxicount && |
| 1343 | mp->m_sb.sb_icount + mp->m_ialloc_inos > mp->m_maxicount) { | 1344 | percpu_counter_read(&mp->m_icount) + mp->m_ialloc_inos > |
| 1345 | mp->m_maxicount) { | ||
| 1344 | noroom = 1; | 1346 | noroom = 1; |
| 1345 | okalloc = 0; | 1347 | okalloc = 0; |
| 1346 | } | 1348 | } |
diff --git a/fs/xfs/libxfs/xfs_sb.c b/fs/xfs/libxfs/xfs_sb.c index f3ea02bf893e..dc4bfc5d88fc 100644 --- a/fs/xfs/libxfs/xfs_sb.c +++ b/fs/xfs/libxfs/xfs_sb.c | |||
| @@ -735,17 +735,15 @@ xfs_initialize_perag_data( | |||
| 735 | btree += pag->pagf_btreeblks; | 735 | btree += pag->pagf_btreeblks; |
| 736 | xfs_perag_put(pag); | 736 | xfs_perag_put(pag); |
| 737 | } | 737 | } |
| 738 | /* | 738 | |
| 739 | * Overwrite incore superblock counters with just-read data | 739 | /* Overwrite incore superblock counters with just-read data */ |
| 740 | */ | ||
| 741 | spin_lock(&mp->m_sb_lock); | 740 | spin_lock(&mp->m_sb_lock); |
| 742 | sbp->sb_ifree = ifree; | 741 | sbp->sb_ifree = ifree; |
| 743 | sbp->sb_icount = ialloc; | 742 | sbp->sb_icount = ialloc; |
| 744 | sbp->sb_fdblocks = bfree + bfreelst + btree; | 743 | sbp->sb_fdblocks = bfree + bfreelst + btree; |
| 745 | spin_unlock(&mp->m_sb_lock); | 744 | spin_unlock(&mp->m_sb_lock); |
| 746 | 745 | ||
| 747 | /* Fixup the per-cpu counters as well. */ | 746 | xfs_reinit_percpu_counters(mp); |
| 748 | xfs_icsb_reinit_counters(mp); | ||
| 749 | 747 | ||
| 750 | return 0; | 748 | return 0; |
| 751 | } | 749 | } |
| @@ -763,6 +761,10 @@ xfs_log_sb( | |||
| 763 | struct xfs_mount *mp = tp->t_mountp; | 761 | struct xfs_mount *mp = tp->t_mountp; |
| 764 | struct xfs_buf *bp = xfs_trans_getsb(tp, mp, 0); | 762 | struct xfs_buf *bp = xfs_trans_getsb(tp, mp, 0); |
| 765 | 763 | ||
| 764 | mp->m_sb.sb_icount = percpu_counter_sum(&mp->m_icount); | ||
| 765 | mp->m_sb.sb_ifree = percpu_counter_sum(&mp->m_ifree); | ||
| 766 | mp->m_sb.sb_fdblocks = percpu_counter_sum(&mp->m_fdblocks); | ||
| 767 | |||
| 766 | xfs_sb_to_disk(XFS_BUF_TO_SBP(bp), &mp->m_sb); | 768 | xfs_sb_to_disk(XFS_BUF_TO_SBP(bp), &mp->m_sb); |
| 767 | xfs_trans_buf_set_type(tp, bp, XFS_BLFT_SB_BUF); | 769 | xfs_trans_buf_set_type(tp, bp, XFS_BLFT_SB_BUF); |
| 768 | xfs_trans_log_buf(tp, bp, 0, sizeof(struct xfs_dsb)); | 770 | xfs_trans_log_buf(tp, bp, 0, sizeof(struct xfs_dsb)); |
diff --git a/fs/xfs/xfs_fsops.c b/fs/xfs/xfs_fsops.c index 74efe5b760dc..cb7e8a29dfb6 100644 --- a/fs/xfs/xfs_fsops.c +++ b/fs/xfs/xfs_fsops.c | |||
| @@ -637,12 +637,13 @@ xfs_fs_counts( | |||
| 637 | xfs_mount_t *mp, | 637 | xfs_mount_t *mp, |
| 638 | xfs_fsop_counts_t *cnt) | 638 | xfs_fsop_counts_t *cnt) |
| 639 | { | 639 | { |
| 640 | xfs_icsb_sync_counters(mp, XFS_ICSB_LAZY_COUNT); | 640 | cnt->allocino = percpu_counter_read_positive(&mp->m_icount); |
| 641 | cnt->freeino = percpu_counter_read_positive(&mp->m_ifree); | ||
| 642 | cnt->freedata = percpu_counter_read_positive(&mp->m_fdblocks) - | ||
| 643 | XFS_ALLOC_SET_ASIDE(mp); | ||
| 644 | |||
| 641 | spin_lock(&mp->m_sb_lock); | 645 | spin_lock(&mp->m_sb_lock); |
| 642 | cnt->freedata = mp->m_sb.sb_fdblocks - XFS_ALLOC_SET_ASIDE(mp); | ||
| 643 | cnt->freertx = mp->m_sb.sb_frextents; | 646 | cnt->freertx = mp->m_sb.sb_frextents; |
| 644 | cnt->freeino = mp->m_sb.sb_ifree; | ||
| 645 | cnt->allocino = mp->m_sb.sb_icount; | ||
| 646 | spin_unlock(&mp->m_sb_lock); | 647 | spin_unlock(&mp->m_sb_lock); |
| 647 | return 0; | 648 | return 0; |
| 648 | } | 649 | } |
| @@ -692,14 +693,9 @@ xfs_reserve_blocks( | |||
| 692 | * what to do. This means that the amount of free space can | 693 | * what to do. This means that the amount of free space can |
| 693 | * change while we do this, so we need to retry if we end up | 694 | * change while we do this, so we need to retry if we end up |
| 694 | * trying to reserve more space than is available. | 695 | * trying to reserve more space than is available. |
| 695 | * | ||
| 696 | * We also use the xfs_mod_incore_sb() interface so that we | ||
| 697 | * don't have to care about whether per cpu counter are | ||
| 698 | * enabled, disabled or even compiled in.... | ||
| 699 | */ | 696 | */ |
| 700 | retry: | 697 | retry: |
| 701 | spin_lock(&mp->m_sb_lock); | 698 | spin_lock(&mp->m_sb_lock); |
| 702 | xfs_icsb_sync_counters_locked(mp, 0); | ||
| 703 | 699 | ||
| 704 | /* | 700 | /* |
| 705 | * If our previous reservation was larger than the current value, | 701 | * If our previous reservation was larger than the current value, |
| @@ -716,7 +712,8 @@ retry: | |||
| 716 | } else { | 712 | } else { |
| 717 | __int64_t free; | 713 | __int64_t free; |
| 718 | 714 | ||
| 719 | free = mp->m_sb.sb_fdblocks - XFS_ALLOC_SET_ASIDE(mp); | 715 | free = percpu_counter_sum(&mp->m_fdblocks) - |
| 716 | XFS_ALLOC_SET_ASIDE(mp); | ||
| 720 | if (!free) | 717 | if (!free) |
| 721 | goto out; /* ENOSPC and fdblks_delta = 0 */ | 718 | goto out; /* ENOSPC and fdblks_delta = 0 */ |
| 722 | 719 | ||
| @@ -755,8 +752,7 @@ out: | |||
| 755 | * the extra reserve blocks from the reserve..... | 752 | * the extra reserve blocks from the reserve..... |
| 756 | */ | 753 | */ |
| 757 | int error; | 754 | int error; |
| 758 | error = xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, | 755 | error = xfs_mod_fdblocks(mp, fdblks_delta, 0); |
| 759 | fdblks_delta, 0); | ||
| 760 | if (error == -ENOSPC) | 756 | if (error == -ENOSPC) |
| 761 | goto retry; | 757 | goto retry; |
| 762 | } | 758 | } |
diff --git a/fs/xfs/xfs_iomap.c b/fs/xfs/xfs_iomap.c index ccb1dd0d509e..38e633bad8c2 100644 --- a/fs/xfs/xfs_iomap.c +++ b/fs/xfs/xfs_iomap.c | |||
| @@ -460,8 +460,7 @@ xfs_iomap_prealloc_size( | |||
| 460 | alloc_blocks = XFS_FILEOFF_MIN(roundup_pow_of_two(MAXEXTLEN), | 460 | alloc_blocks = XFS_FILEOFF_MIN(roundup_pow_of_two(MAXEXTLEN), |
| 461 | alloc_blocks); | 461 | alloc_blocks); |
| 462 | 462 | ||
| 463 | xfs_icsb_sync_counters(mp, XFS_ICSB_LAZY_COUNT); | 463 | freesp = percpu_counter_read_positive(&mp->m_fdblocks); |
| 464 | freesp = mp->m_sb.sb_fdblocks; | ||
| 465 | if (freesp < mp->m_low_space[XFS_LOWSP_5_PCNT]) { | 464 | if (freesp < mp->m_low_space[XFS_LOWSP_5_PCNT]) { |
| 466 | shift = 2; | 465 | shift = 2; |
| 467 | if (freesp < mp->m_low_space[XFS_LOWSP_4_PCNT]) | 466 | if (freesp < mp->m_low_space[XFS_LOWSP_4_PCNT]) |
diff --git a/fs/xfs/xfs_linux.h b/fs/xfs/xfs_linux.h index c31d2c2eadc4..7c7842c85a08 100644 --- a/fs/xfs/xfs_linux.h +++ b/fs/xfs/xfs_linux.h | |||
| @@ -116,15 +116,6 @@ typedef __uint64_t __psunsigned_t; | |||
| 116 | #undef XFS_NATIVE_HOST | 116 | #undef XFS_NATIVE_HOST |
| 117 | #endif | 117 | #endif |
| 118 | 118 | ||
| 119 | /* | ||
| 120 | * Feature macros (disable/enable) | ||
| 121 | */ | ||
| 122 | #ifdef CONFIG_SMP | ||
| 123 | #define HAVE_PERCPU_SB /* per cpu superblock counters are a 2.6 feature */ | ||
| 124 | #else | ||
| 125 | #undef HAVE_PERCPU_SB /* per cpu superblock counters are a 2.6 feature */ | ||
| 126 | #endif | ||
| 127 | |||
| 128 | #define irix_sgid_inherit xfs_params.sgid_inherit.val | 119 | #define irix_sgid_inherit xfs_params.sgid_inherit.val |
| 129 | #define irix_symlink_mode xfs_params.symlink_mode.val | 120 | #define irix_symlink_mode xfs_params.symlink_mode.val |
| 130 | #define xfs_panic_mask xfs_params.panic_mask.val | 121 | #define xfs_panic_mask xfs_params.panic_mask.val |
diff --git a/fs/xfs/xfs_log_recover.c b/fs/xfs/xfs_log_recover.c index a5a945fc3bdc..4f5784f85a5b 100644 --- a/fs/xfs/xfs_log_recover.c +++ b/fs/xfs/xfs_log_recover.c | |||
| @@ -4463,10 +4463,10 @@ xlog_do_recover( | |||
| 4463 | xfs_sb_from_disk(sbp, XFS_BUF_TO_SBP(bp)); | 4463 | xfs_sb_from_disk(sbp, XFS_BUF_TO_SBP(bp)); |
| 4464 | ASSERT(sbp->sb_magicnum == XFS_SB_MAGIC); | 4464 | ASSERT(sbp->sb_magicnum == XFS_SB_MAGIC); |
| 4465 | ASSERT(xfs_sb_good_version(sbp)); | 4465 | ASSERT(xfs_sb_good_version(sbp)); |
| 4466 | xfs_reinit_percpu_counters(log->l_mp); | ||
| 4467 | |||
| 4466 | xfs_buf_relse(bp); | 4468 | xfs_buf_relse(bp); |
| 4467 | 4469 | ||
| 4468 | /* We've re-read the superblock so re-initialize per-cpu counters */ | ||
| 4469 | xfs_icsb_reinit_counters(log->l_mp); | ||
| 4470 | 4470 | ||
| 4471 | xlog_recover_check_summary(log); | 4471 | xlog_recover_check_summary(log); |
| 4472 | 4472 | ||
diff --git a/fs/xfs/xfs_mount.c b/fs/xfs/xfs_mount.c index 4fa80e63eea2..2ce7ee3b4ec1 100644 --- a/fs/xfs/xfs_mount.c +++ b/fs/xfs/xfs_mount.c | |||
| @@ -43,18 +43,6 @@ | |||
| 43 | #include "xfs_sysfs.h" | 43 | #include "xfs_sysfs.h" |
| 44 | 44 | ||
| 45 | 45 | ||
| 46 | #ifdef HAVE_PERCPU_SB | ||
| 47 | STATIC void xfs_icsb_balance_counter(xfs_mount_t *, xfs_sb_field_t, | ||
| 48 | int); | ||
| 49 | STATIC void xfs_icsb_balance_counter_locked(xfs_mount_t *, xfs_sb_field_t, | ||
| 50 | int); | ||
| 51 | STATIC void xfs_icsb_disable_counter(xfs_mount_t *, xfs_sb_field_t); | ||
| 52 | #else | ||
| 53 | |||
| 54 | #define xfs_icsb_balance_counter(mp, a, b) do { } while (0) | ||
| 55 | #define xfs_icsb_balance_counter_locked(mp, a, b) do { } while (0) | ||
| 56 | #endif | ||
| 57 | |||
| 58 | static DEFINE_MUTEX(xfs_uuid_table_mutex); | 46 | static DEFINE_MUTEX(xfs_uuid_table_mutex); |
| 59 | static int xfs_uuid_table_size; | 47 | static int xfs_uuid_table_size; |
| 60 | static uuid_t *xfs_uuid_table; | 48 | static uuid_t *xfs_uuid_table; |
| @@ -347,8 +335,7 @@ reread: | |||
| 347 | goto reread; | 335 | goto reread; |
| 348 | } | 336 | } |
| 349 | 337 | ||
| 350 | /* Initialize per-cpu counters */ | 338 | xfs_reinit_percpu_counters(mp); |
| 351 | xfs_icsb_reinit_counters(mp); | ||
| 352 | 339 | ||
| 353 | /* no need to be quiet anymore, so reset the buf ops */ | 340 | /* no need to be quiet anymore, so reset the buf ops */ |
| 354 | bp->b_ops = &xfs_sb_buf_ops; | 341 | bp->b_ops = &xfs_sb_buf_ops; |
| @@ -1087,8 +1074,6 @@ xfs_log_sbcount(xfs_mount_t *mp) | |||
| 1087 | if (!xfs_fs_writable(mp, SB_FREEZE_COMPLETE)) | 1074 | if (!xfs_fs_writable(mp, SB_FREEZE_COMPLETE)) |
| 1088 | return 0; | 1075 | return 0; |
| 1089 | 1076 | ||
| 1090 | xfs_icsb_sync_counters(mp, 0); | ||
| 1091 | |||
| 1092 | /* | 1077 | /* |
| 1093 | * we don't need to do this if we are updating the superblock | 1078 | * we don't need to do this if we are updating the superblock |
| 1094 | * counters on every modification. | 1079 | * counters on every modification. |
| @@ -1099,253 +1084,136 @@ xfs_log_sbcount(xfs_mount_t *mp) | |||
| 1099 | return xfs_sync_sb(mp, true); | 1084 | return xfs_sync_sb(mp, true); |
| 1100 | } | 1085 | } |
| 1101 | 1086 | ||
| 1102 | /* | 1087 | int |
| 1103 | * xfs_mod_incore_sb_unlocked() is a utility routine commonly used to apply | 1088 | xfs_mod_icount( |
| 1104 | * a delta to a specified field in the in-core superblock. Simply | 1089 | struct xfs_mount *mp, |
| 1105 | * switch on the field indicated and apply the delta to that field. | 1090 | int64_t delta) |
| 1106 | * Fields are not allowed to dip below zero, so if the delta would | ||
| 1107 | * do this do not apply it and return EINVAL. | ||
| 1108 | * | ||
| 1109 | * The m_sb_lock must be held when this routine is called. | ||
| 1110 | */ | ||
| 1111 | STATIC int | ||
| 1112 | xfs_mod_incore_sb_unlocked( | ||
| 1113 | xfs_mount_t *mp, | ||
| 1114 | xfs_sb_field_t field, | ||
| 1115 | int64_t delta, | ||
| 1116 | int rsvd) | ||
| 1117 | { | 1091 | { |
| 1118 | int scounter; /* short counter for 32 bit fields */ | 1092 | /* deltas are +/-64, hence the large batch size of 128. */ |
| 1119 | long long lcounter; /* long counter for 64 bit fields */ | 1093 | __percpu_counter_add(&mp->m_icount, delta, 128); |
| 1120 | long long res_used, rem; | 1094 | if (percpu_counter_compare(&mp->m_icount, 0) < 0) { |
| 1121 | |||
| 1122 | /* | ||
| 1123 | * With the in-core superblock spin lock held, switch | ||
| 1124 | * on the indicated field. Apply the delta to the | ||
| 1125 | * proper field. If the fields value would dip below | ||
| 1126 | * 0, then do not apply the delta and return EINVAL. | ||
| 1127 | */ | ||
| 1128 | switch (field) { | ||
| 1129 | case XFS_SBS_ICOUNT: | ||
| 1130 | lcounter = (long long)mp->m_sb.sb_icount; | ||
| 1131 | lcounter += delta; | ||
| 1132 | if (lcounter < 0) { | ||
| 1133 | ASSERT(0); | ||
| 1134 | return -EINVAL; | ||
| 1135 | } | ||
| 1136 | mp->m_sb.sb_icount = lcounter; | ||
| 1137 | return 0; | ||
| 1138 | case XFS_SBS_IFREE: | ||
| 1139 | lcounter = (long long)mp->m_sb.sb_ifree; | ||
| 1140 | lcounter += delta; | ||
| 1141 | if (lcounter < 0) { | ||
| 1142 | ASSERT(0); | ||
| 1143 | return -EINVAL; | ||
| 1144 | } | ||
| 1145 | mp->m_sb.sb_ifree = lcounter; | ||
| 1146 | return 0; | ||
| 1147 | case XFS_SBS_FDBLOCKS: | ||
| 1148 | lcounter = (long long) | ||
| 1149 | mp->m_sb.sb_fdblocks - XFS_ALLOC_SET_ASIDE(mp); | ||
| 1150 | res_used = (long long)(mp->m_resblks - mp->m_resblks_avail); | ||
| 1151 | |||
| 1152 | if (delta > 0) { /* Putting blocks back */ | ||
| 1153 | if (res_used > delta) { | ||
| 1154 | mp->m_resblks_avail += delta; | ||
| 1155 | } else { | ||
| 1156 | rem = delta - res_used; | ||
| 1157 | mp->m_resblks_avail = mp->m_resblks; | ||
| 1158 | lcounter += rem; | ||
| 1159 | } | ||
| 1160 | } else { /* Taking blocks away */ | ||
| 1161 | lcounter += delta; | ||
| 1162 | if (lcounter >= 0) { | ||
| 1163 | mp->m_sb.sb_fdblocks = lcounter + | ||
| 1164 | XFS_ALLOC_SET_ASIDE(mp); | ||
| 1165 | return 0; | ||
| 1166 | } | ||
| 1167 | |||
| 1168 | /* | ||
| 1169 | * We are out of blocks, use any available reserved | ||
| 1170 | * blocks if were allowed to. | ||
| 1171 | */ | ||
| 1172 | if (!rsvd) | ||
| 1173 | return -ENOSPC; | ||
| 1174 | |||
| 1175 | lcounter = (long long)mp->m_resblks_avail + delta; | ||
| 1176 | if (lcounter >= 0) { | ||
| 1177 | mp->m_resblks_avail = lcounter; | ||
| 1178 | return 0; | ||
| 1179 | } | ||
| 1180 | printk_once(KERN_WARNING | ||
| 1181 | "Filesystem \"%s\": reserve blocks depleted! " | ||
| 1182 | "Consider increasing reserve pool size.", | ||
| 1183 | mp->m_fsname); | ||
| 1184 | return -ENOSPC; | ||
| 1185 | } | ||
| 1186 | |||
| 1187 | mp->m_sb.sb_fdblocks = lcounter + XFS_ALLOC_SET_ASIDE(mp); | ||
| 1188 | return 0; | ||
| 1189 | case XFS_SBS_FREXTENTS: | ||
| 1190 | lcounter = (long long)mp->m_sb.sb_frextents; | ||
| 1191 | lcounter += delta; | ||
| 1192 | if (lcounter < 0) { | ||
| 1193 | return -ENOSPC; | ||
| 1194 | } | ||
| 1195 | mp->m_sb.sb_frextents = lcounter; | ||
| 1196 | return 0; | ||
| 1197 | case XFS_SBS_DBLOCKS: | ||
| 1198 | lcounter = (long long)mp->m_sb.sb_dblocks; | ||
| 1199 | lcounter += delta; | ||
| 1200 | if (lcounter < 0) { | ||
| 1201 | ASSERT(0); | ||
| 1202 | return -EINVAL; | ||
| 1203 | } | ||
| 1204 | mp->m_sb.sb_dblocks = lcounter; | ||
| 1205 | return 0; | ||
| 1206 | case XFS_SBS_AGCOUNT: | ||
| 1207 | scounter = mp->m_sb.sb_agcount; | ||
| 1208 | scounter += delta; | ||
| 1209 | if (scounter < 0) { | ||
| 1210 | ASSERT(0); | ||
| 1211 | return -EINVAL; | ||
| 1212 | } | ||
| 1213 | mp->m_sb.sb_agcount = scounter; | ||
| 1214 | return 0; | ||
| 1215 | case XFS_SBS_IMAX_PCT: | ||
| 1216 | scounter = mp->m_sb.sb_imax_pct; | ||
| 1217 | scounter += delta; | ||
| 1218 | if (scounter < 0) { | ||
| 1219 | ASSERT(0); | ||
| 1220 | return -EINVAL; | ||
| 1221 | } | ||
| 1222 | mp->m_sb.sb_imax_pct = scounter; | ||
| 1223 | return 0; | ||
| 1224 | case XFS_SBS_REXTSIZE: | ||
| 1225 | scounter = mp->m_sb.sb_rextsize; | ||
| 1226 | scounter += delta; | ||
| 1227 | if (scounter < 0) { | ||
| 1228 | ASSERT(0); | ||
| 1229 | return -EINVAL; | ||
| 1230 | } | ||
| 1231 | mp->m_sb.sb_rextsize = scounter; | ||
| 1232 | return 0; | ||
| 1233 | case XFS_SBS_RBMBLOCKS: | ||
| 1234 | scounter = mp->m_sb.sb_rbmblocks; | ||
| 1235 | scounter += delta; | ||
| 1236 | if (scounter < 0) { | ||
| 1237 | ASSERT(0); | ||
| 1238 | return -EINVAL; | ||
| 1239 | } | ||
| 1240 | mp->m_sb.sb_rbmblocks = scounter; | ||
| 1241 | return 0; | ||
| 1242 | case XFS_SBS_RBLOCKS: | ||
| 1243 | lcounter = (long long)mp->m_sb.sb_rblocks; | ||
| 1244 | lcounter += delta; | ||
| 1245 | if (lcounter < 0) { | ||
| 1246 | ASSERT(0); | ||
| 1247 | return -EINVAL; | ||
| 1248 | } | ||
| 1249 | mp->m_sb.sb_rblocks = lcounter; | ||
| 1250 | return 0; | ||
| 1251 | case XFS_SBS_REXTENTS: | ||
| 1252 | lcounter = (long long)mp->m_sb.sb_rextents; | ||
| 1253 | lcounter += delta; | ||
| 1254 | if (lcounter < 0) { | ||
| 1255 | ASSERT(0); | ||
| 1256 | return -EINVAL; | ||
| 1257 | } | ||
| 1258 | mp->m_sb.sb_rextents = lcounter; | ||
| 1259 | return 0; | ||
| 1260 | case XFS_SBS_REXTSLOG: | ||
| 1261 | scounter = mp->m_sb.sb_rextslog; | ||
| 1262 | scounter += delta; | ||
| 1263 | if (scounter < 0) { | ||
| 1264 | ASSERT(0); | ||
| 1265 | return -EINVAL; | ||
| 1266 | } | ||
| 1267 | mp->m_sb.sb_rextslog = scounter; | ||
| 1268 | return 0; | ||
| 1269 | default: | ||
| 1270 | ASSERT(0); | 1095 | ASSERT(0); |
| 1096 | percpu_counter_add(&mp->m_icount, -delta); | ||
| 1271 | return -EINVAL; | 1097 | return -EINVAL; |
| 1272 | } | 1098 | } |
| 1099 | return 0; | ||
| 1273 | } | 1100 | } |
| 1274 | 1101 | ||
| 1275 | /* | ||
| 1276 | * xfs_mod_incore_sb() is used to change a field in the in-core | ||
| 1277 | * superblock structure by the specified delta. This modification | ||
| 1278 | * is protected by the m_sb_lock. Just use the xfs_mod_incore_sb_unlocked() | ||
| 1279 | * routine to do the work. | ||
| 1280 | */ | ||
| 1281 | int | 1102 | int |
| 1282 | xfs_mod_incore_sb( | 1103 | xfs_mod_ifree( |
| 1283 | struct xfs_mount *mp, | 1104 | struct xfs_mount *mp, |
| 1284 | xfs_sb_field_t field, | 1105 | int64_t delta) |
| 1285 | int64_t delta, | ||
| 1286 | int rsvd) | ||
| 1287 | { | 1106 | { |
| 1288 | int status; | 1107 | percpu_counter_add(&mp->m_ifree, delta); |
| 1289 | 1108 | if (percpu_counter_compare(&mp->m_ifree, 0) < 0) { | |
| 1290 | #ifdef HAVE_PERCPU_SB | 1109 | ASSERT(0); |
| 1291 | ASSERT(field < XFS_SBS_ICOUNT || field > XFS_SBS_FDBLOCKS); | 1110 | percpu_counter_add(&mp->m_ifree, -delta); |
| 1292 | #endif | 1111 | return -EINVAL; |
| 1293 | spin_lock(&mp->m_sb_lock); | 1112 | } |
| 1294 | status = xfs_mod_incore_sb_unlocked(mp, field, delta, rsvd); | 1113 | return 0; |
| 1295 | spin_unlock(&mp->m_sb_lock); | ||
| 1296 | |||
| 1297 | return status; | ||
| 1298 | } | 1114 | } |
| 1299 | 1115 | ||
| 1300 | /* | ||
| 1301 | * Change more than one field in the in-core superblock structure at a time. | ||
| 1302 | * | ||
| 1303 | * The fields and changes to those fields are specified in the array of | ||
| 1304 | * xfs_mod_sb structures passed in. Either all of the specified deltas | ||
| 1305 | * will be applied or none of them will. If any modified field dips below 0, | ||
| 1306 | * then all modifications will be backed out and EINVAL will be returned. | ||
| 1307 | * | ||
| 1308 | * Note that this function may not be used for the superblock values that | ||
| 1309 | * are tracked with the in-memory per-cpu counters - a direct call to | ||
| 1310 | * xfs_icsb_modify_counters is required for these. | ||
| 1311 | */ | ||
| 1312 | int | 1116 | int |
| 1313 | xfs_mod_incore_sb_batch( | 1117 | xfs_mod_fdblocks( |
| 1314 | struct xfs_mount *mp, | 1118 | struct xfs_mount *mp, |
| 1315 | xfs_mod_sb_t *msb, | 1119 | int64_t delta, |
| 1316 | uint nmsb, | 1120 | bool rsvd) |
| 1317 | int rsvd) | ||
| 1318 | { | 1121 | { |
| 1319 | xfs_mod_sb_t *msbp; | 1122 | int64_t lcounter; |
| 1320 | int error = 0; | 1123 | long long res_used; |
| 1124 | s32 batch; | ||
| 1125 | |||
| 1126 | if (delta > 0) { | ||
| 1127 | /* | ||
| 1128 | * If the reserve pool is depleted, put blocks back into it | ||
| 1129 | * first. Most of the time the pool is full. | ||
| 1130 | */ | ||
| 1131 | if (likely(mp->m_resblks == mp->m_resblks_avail)) { | ||
| 1132 | percpu_counter_add(&mp->m_fdblocks, delta); | ||
| 1133 | return 0; | ||
| 1134 | } | ||
| 1135 | |||
| 1136 | spin_lock(&mp->m_sb_lock); | ||
| 1137 | res_used = (long long)(mp->m_resblks - mp->m_resblks_avail); | ||
| 1138 | |||
| 1139 | if (res_used > delta) { | ||
| 1140 | mp->m_resblks_avail += delta; | ||
| 1141 | } else { | ||
| 1142 | delta -= res_used; | ||
| 1143 | mp->m_resblks_avail = mp->m_resblks; | ||
| 1144 | percpu_counter_add(&mp->m_fdblocks, delta); | ||
| 1145 | } | ||
| 1146 | spin_unlock(&mp->m_sb_lock); | ||
| 1147 | return 0; | ||
| 1148 | } | ||
| 1321 | 1149 | ||
| 1322 | /* | 1150 | /* |
| 1323 | * Loop through the array of mod structures and apply each individually. | 1151 | * Taking blocks away, need to be more accurate the closer we |
| 1324 | * If any fail, then back out all those which have already been applied. | 1152 | * are to zero. |
| 1325 | * Do all of this within the scope of the m_sb_lock so that all of the | 1153 | * |
| 1326 | * changes will be atomic. | 1154 | * batch size is set to a maximum of 1024 blocks - if we are |
| 1155 | * allocating of freeing extents larger than this then we aren't | ||
| 1156 | * going to be hammering the counter lock so a lock per update | ||
| 1157 | * is not a problem. | ||
| 1158 | * | ||
| 1159 | * If the counter has a value of less than 2 * max batch size, | ||
| 1160 | * then make everything serialise as we are real close to | ||
| 1161 | * ENOSPC. | ||
| 1162 | */ | ||
| 1163 | #define __BATCH 1024 | ||
| 1164 | if (percpu_counter_compare(&mp->m_fdblocks, 2 * __BATCH) < 0) | ||
| 1165 | batch = 1; | ||
| 1166 | else | ||
| 1167 | batch = __BATCH; | ||
| 1168 | #undef __BATCH | ||
| 1169 | |||
| 1170 | __percpu_counter_add(&mp->m_fdblocks, delta, batch); | ||
| 1171 | if (percpu_counter_compare(&mp->m_fdblocks, | ||
| 1172 | XFS_ALLOC_SET_ASIDE(mp)) >= 0) { | ||
| 1173 | /* we had space! */ | ||
| 1174 | return 0; | ||
| 1175 | } | ||
| 1176 | |||
| 1177 | /* | ||
| 1178 | * lock up the sb for dipping into reserves before releasing the space | ||
| 1179 | * that took us to ENOSPC. | ||
| 1327 | */ | 1180 | */ |
| 1328 | spin_lock(&mp->m_sb_lock); | 1181 | spin_lock(&mp->m_sb_lock); |
| 1329 | for (msbp = msb; msbp < (msb + nmsb); msbp++) { | 1182 | percpu_counter_add(&mp->m_fdblocks, -delta); |
| 1330 | ASSERT(msbp->msb_field < XFS_SBS_ICOUNT || | 1183 | if (!rsvd) |
| 1331 | msbp->msb_field > XFS_SBS_FDBLOCKS); | 1184 | goto fdblocks_enospc; |
| 1332 | 1185 | ||
| 1333 | error = xfs_mod_incore_sb_unlocked(mp, msbp->msb_field, | 1186 | lcounter = (long long)mp->m_resblks_avail + delta; |
| 1334 | msbp->msb_delta, rsvd); | 1187 | if (lcounter >= 0) { |
| 1335 | if (error) | 1188 | mp->m_resblks_avail = lcounter; |
| 1336 | goto unwind; | 1189 | spin_unlock(&mp->m_sb_lock); |
| 1190 | return 0; | ||
| 1337 | } | 1191 | } |
| 1192 | printk_once(KERN_WARNING | ||
| 1193 | "Filesystem \"%s\": reserve blocks depleted! " | ||
| 1194 | "Consider increasing reserve pool size.", | ||
| 1195 | mp->m_fsname); | ||
| 1196 | fdblocks_enospc: | ||
| 1338 | spin_unlock(&mp->m_sb_lock); | 1197 | spin_unlock(&mp->m_sb_lock); |
| 1339 | return 0; | 1198 | return -ENOSPC; |
| 1199 | } | ||
| 1340 | 1200 | ||
| 1341 | unwind: | 1201 | int |
| 1342 | while (--msbp >= msb) { | 1202 | xfs_mod_frextents( |
| 1343 | error = xfs_mod_incore_sb_unlocked(mp, msbp->msb_field, | 1203 | struct xfs_mount *mp, |
| 1344 | -msbp->msb_delta, rsvd); | 1204 | int64_t delta) |
| 1345 | ASSERT(error == 0); | 1205 | { |
| 1346 | } | 1206 | int64_t lcounter; |
| 1207 | int ret = 0; | ||
| 1208 | |||
| 1209 | spin_lock(&mp->m_sb_lock); | ||
| 1210 | lcounter = mp->m_sb.sb_frextents + delta; | ||
| 1211 | if (lcounter < 0) | ||
| 1212 | ret = -ENOSPC; | ||
| 1213 | else | ||
| 1214 | mp->m_sb.sb_frextents = lcounter; | ||
| 1347 | spin_unlock(&mp->m_sb_lock); | 1215 | spin_unlock(&mp->m_sb_lock); |
| 1348 | return error; | 1216 | return ret; |
| 1349 | } | 1217 | } |
| 1350 | 1218 | ||
| 1351 | /* | 1219 | /* |
| @@ -1407,573 +1275,3 @@ xfs_dev_is_read_only( | |||
| 1407 | } | 1275 | } |
| 1408 | return 0; | 1276 | return 0; |
| 1409 | } | 1277 | } |
| 1410 | |||
| 1411 | #ifdef HAVE_PERCPU_SB | ||
| 1412 | /* | ||
| 1413 | * Per-cpu incore superblock counters | ||
| 1414 | * | ||
| 1415 | * Simple concept, difficult implementation | ||
| 1416 | * | ||
| 1417 | * Basically, replace the incore superblock counters with a distributed per cpu | ||
| 1418 | * counter for contended fields (e.g. free block count). | ||
| 1419 | * | ||
| 1420 | * Difficulties arise in that the incore sb is used for ENOSPC checking, and | ||
| 1421 | * hence needs to be accurately read when we are running low on space. Hence | ||
| 1422 | * there is a method to enable and disable the per-cpu counters based on how | ||
| 1423 | * much "stuff" is available in them. | ||
| 1424 | * | ||
| 1425 | * Basically, a counter is enabled if there is enough free resource to justify | ||
| 1426 | * running a per-cpu fast-path. If the per-cpu counter runs out (i.e. a local | ||
| 1427 | * ENOSPC), then we disable the counters to synchronise all callers and | ||
| 1428 | * re-distribute the available resources. | ||
| 1429 | * | ||
| 1430 | * If, once we redistributed the available resources, we still get a failure, | ||
| 1431 | * we disable the per-cpu counter and go through the slow path. | ||
| 1432 | * | ||
| 1433 | * The slow path is the current xfs_mod_incore_sb() function. This means that | ||
| 1434 | * when we disable a per-cpu counter, we need to drain its resources back to | ||
| 1435 | * the global superblock. We do this after disabling the counter to prevent | ||
| 1436 | * more threads from queueing up on the counter. | ||
| 1437 | * | ||
| 1438 | * Essentially, this means that we still need a lock in the fast path to enable | ||
| 1439 | * synchronisation between the global counters and the per-cpu counters. This | ||
| 1440 | * is not a problem because the lock will be local to a CPU almost all the time | ||
| 1441 | * and have little contention except when we get to ENOSPC conditions. | ||
| 1442 | * | ||
| 1443 | * Basically, this lock becomes a barrier that enables us to lock out the fast | ||
| 1444 | * path while we do things like enabling and disabling counters and | ||
| 1445 | * synchronising the counters. | ||
| 1446 | * | ||
| 1447 | * Locking rules: | ||
| 1448 | * | ||
| 1449 | * 1. m_sb_lock before picking up per-cpu locks | ||
| 1450 | * 2. per-cpu locks always picked up via for_each_online_cpu() order | ||
| 1451 | * 3. accurate counter sync requires m_sb_lock + per cpu locks | ||
| 1452 | * 4. modifying per-cpu counters requires holding per-cpu lock | ||
| 1453 | * 5. modifying global counters requires holding m_sb_lock | ||
| 1454 | * 6. enabling or disabling a counter requires holding the m_sb_lock | ||
| 1455 | * and _none_ of the per-cpu locks. | ||
| 1456 | * | ||
| 1457 | * Disabled counters are only ever re-enabled by a balance operation | ||
| 1458 | * that results in more free resources per CPU than a given threshold. | ||
| 1459 | * To ensure counters don't remain disabled, they are rebalanced when | ||
| 1460 | * the global resource goes above a higher threshold (i.e. some hysteresis | ||
| 1461 | * is present to prevent thrashing). | ||
| 1462 | */ | ||
| 1463 | |||
| 1464 | #ifdef CONFIG_HOTPLUG_CPU | ||
| 1465 | /* | ||
| 1466 | * hot-plug CPU notifier support. | ||
| 1467 | * | ||
| 1468 | * We need a notifier per filesystem as we need to be able to identify | ||
| 1469 | * the filesystem to balance the counters out. This is achieved by | ||
| 1470 | * having a notifier block embedded in the xfs_mount_t and doing pointer | ||
| 1471 | * magic to get the mount pointer from the notifier block address. | ||
| 1472 | */ | ||
| 1473 | STATIC int | ||
| 1474 | xfs_icsb_cpu_notify( | ||
| 1475 | struct notifier_block *nfb, | ||
| 1476 | unsigned long action, | ||
| 1477 | void *hcpu) | ||
| 1478 | { | ||
| 1479 | xfs_icsb_cnts_t *cntp; | ||
| 1480 | xfs_mount_t *mp; | ||
| 1481 | |||
| 1482 | mp = (xfs_mount_t *)container_of(nfb, xfs_mount_t, m_icsb_notifier); | ||
| 1483 | cntp = (xfs_icsb_cnts_t *) | ||
| 1484 | per_cpu_ptr(mp->m_sb_cnts, (unsigned long)hcpu); | ||
| 1485 | switch (action) { | ||
| 1486 | case CPU_UP_PREPARE: | ||
| 1487 | case CPU_UP_PREPARE_FROZEN: | ||
| 1488 | /* Easy Case - initialize the area and locks, and | ||
| 1489 | * then rebalance when online does everything else for us. */ | ||
| 1490 | memset(cntp, 0, sizeof(xfs_icsb_cnts_t)); | ||
| 1491 | break; | ||
| 1492 | case CPU_ONLINE: | ||
| 1493 | case CPU_ONLINE_FROZEN: | ||
| 1494 | xfs_icsb_lock(mp); | ||
| 1495 | xfs_icsb_balance_counter(mp, XFS_SBS_ICOUNT, 0); | ||
| 1496 | xfs_icsb_balance_counter(mp, XFS_SBS_IFREE, 0); | ||
| 1497 | xfs_icsb_balance_counter(mp, XFS_SBS_FDBLOCKS, 0); | ||
| 1498 | xfs_icsb_unlock(mp); | ||
| 1499 | break; | ||
| 1500 | case CPU_DEAD: | ||
| 1501 | case CPU_DEAD_FROZEN: | ||
| 1502 | /* Disable all the counters, then fold the dead cpu's | ||
| 1503 | * count into the total on the global superblock and | ||
| 1504 | * re-enable the counters. */ | ||
| 1505 | xfs_icsb_lock(mp); | ||
| 1506 | spin_lock(&mp->m_sb_lock); | ||
| 1507 | xfs_icsb_disable_counter(mp, XFS_SBS_ICOUNT); | ||
| 1508 | xfs_icsb_disable_counter(mp, XFS_SBS_IFREE); | ||
| 1509 | xfs_icsb_disable_counter(mp, XFS_SBS_FDBLOCKS); | ||
| 1510 | |||
| 1511 | mp->m_sb.sb_icount += cntp->icsb_icount; | ||
| 1512 | mp->m_sb.sb_ifree += cntp->icsb_ifree; | ||
| 1513 | mp->m_sb.sb_fdblocks += cntp->icsb_fdblocks; | ||
| 1514 | |||
| 1515 | memset(cntp, 0, sizeof(xfs_icsb_cnts_t)); | ||
| 1516 | |||
| 1517 | xfs_icsb_balance_counter_locked(mp, XFS_SBS_ICOUNT, 0); | ||
| 1518 | xfs_icsb_balance_counter_locked(mp, XFS_SBS_IFREE, 0); | ||
| 1519 | xfs_icsb_balance_counter_locked(mp, XFS_SBS_FDBLOCKS, 0); | ||
| 1520 | spin_unlock(&mp->m_sb_lock); | ||
| 1521 | xfs_icsb_unlock(mp); | ||
| 1522 | break; | ||
| 1523 | } | ||
| 1524 | |||
| 1525 | return NOTIFY_OK; | ||
| 1526 | } | ||
| 1527 | #endif /* CONFIG_HOTPLUG_CPU */ | ||
| 1528 | |||
| 1529 | int | ||
| 1530 | xfs_icsb_init_counters( | ||
| 1531 | xfs_mount_t *mp) | ||
| 1532 | { | ||
| 1533 | xfs_icsb_cnts_t *cntp; | ||
| 1534 | int i; | ||
| 1535 | |||
| 1536 | mp->m_sb_cnts = alloc_percpu(xfs_icsb_cnts_t); | ||
| 1537 | if (mp->m_sb_cnts == NULL) | ||
| 1538 | return -ENOMEM; | ||
| 1539 | |||
| 1540 | for_each_online_cpu(i) { | ||
| 1541 | cntp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, i); | ||
| 1542 | memset(cntp, 0, sizeof(xfs_icsb_cnts_t)); | ||
| 1543 | } | ||
| 1544 | |||
| 1545 | mutex_init(&mp->m_icsb_mutex); | ||
| 1546 | |||
| 1547 | /* | ||
| 1548 | * start with all counters disabled so that the | ||
| 1549 | * initial balance kicks us off correctly | ||
| 1550 | */ | ||
| 1551 | mp->m_icsb_counters = -1; | ||
| 1552 | |||
| 1553 | #ifdef CONFIG_HOTPLUG_CPU | ||
| 1554 | mp->m_icsb_notifier.notifier_call = xfs_icsb_cpu_notify; | ||
| 1555 | mp->m_icsb_notifier.priority = 0; | ||
| 1556 | register_hotcpu_notifier(&mp->m_icsb_notifier); | ||
| 1557 | #endif /* CONFIG_HOTPLUG_CPU */ | ||
| 1558 | |||
| 1559 | return 0; | ||
| 1560 | } | ||
| 1561 | |||
| 1562 | void | ||
| 1563 | xfs_icsb_reinit_counters( | ||
| 1564 | xfs_mount_t *mp) | ||
| 1565 | { | ||
| 1566 | xfs_icsb_lock(mp); | ||
| 1567 | /* | ||
| 1568 | * start with all counters disabled so that the | ||
| 1569 | * initial balance kicks us off correctly | ||
| 1570 | */ | ||
| 1571 | mp->m_icsb_counters = -1; | ||
| 1572 | xfs_icsb_balance_counter(mp, XFS_SBS_ICOUNT, 0); | ||
| 1573 | xfs_icsb_balance_counter(mp, XFS_SBS_IFREE, 0); | ||
| 1574 | xfs_icsb_balance_counter(mp, XFS_SBS_FDBLOCKS, 0); | ||
| 1575 | xfs_icsb_unlock(mp); | ||
| 1576 | } | ||
| 1577 | |||
| 1578 | void | ||
| 1579 | xfs_icsb_destroy_counters( | ||
| 1580 | xfs_mount_t *mp) | ||
| 1581 | { | ||
| 1582 | if (mp->m_sb_cnts) { | ||
| 1583 | unregister_hotcpu_notifier(&mp->m_icsb_notifier); | ||
| 1584 | free_percpu(mp->m_sb_cnts); | ||
| 1585 | } | ||
| 1586 | mutex_destroy(&mp->m_icsb_mutex); | ||
| 1587 | } | ||
| 1588 | |||
| 1589 | STATIC void | ||
| 1590 | xfs_icsb_lock_cntr( | ||
| 1591 | xfs_icsb_cnts_t *icsbp) | ||
| 1592 | { | ||
| 1593 | while (test_and_set_bit(XFS_ICSB_FLAG_LOCK, &icsbp->icsb_flags)) { | ||
| 1594 | ndelay(1000); | ||
| 1595 | } | ||
| 1596 | } | ||
| 1597 | |||
| 1598 | STATIC void | ||
| 1599 | xfs_icsb_unlock_cntr( | ||
| 1600 | xfs_icsb_cnts_t *icsbp) | ||
| 1601 | { | ||
| 1602 | clear_bit(XFS_ICSB_FLAG_LOCK, &icsbp->icsb_flags); | ||
| 1603 | } | ||
| 1604 | |||
| 1605 | |||
| 1606 | STATIC void | ||
| 1607 | xfs_icsb_lock_all_counters( | ||
| 1608 | xfs_mount_t *mp) | ||
| 1609 | { | ||
| 1610 | xfs_icsb_cnts_t *cntp; | ||
| 1611 | int i; | ||
| 1612 | |||
| 1613 | for_each_online_cpu(i) { | ||
| 1614 | cntp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, i); | ||
| 1615 | xfs_icsb_lock_cntr(cntp); | ||
| 1616 | } | ||
| 1617 | } | ||
| 1618 | |||
| 1619 | STATIC void | ||
| 1620 | xfs_icsb_unlock_all_counters( | ||
| 1621 | xfs_mount_t *mp) | ||
| 1622 | { | ||
| 1623 | xfs_icsb_cnts_t *cntp; | ||
| 1624 | int i; | ||
| 1625 | |||
| 1626 | for_each_online_cpu(i) { | ||
| 1627 | cntp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, i); | ||
| 1628 | xfs_icsb_unlock_cntr(cntp); | ||
| 1629 | } | ||
| 1630 | } | ||
| 1631 | |||
| 1632 | STATIC void | ||
| 1633 | xfs_icsb_count( | ||
| 1634 | xfs_mount_t *mp, | ||
| 1635 | xfs_icsb_cnts_t *cnt, | ||
| 1636 | int flags) | ||
| 1637 | { | ||
| 1638 | xfs_icsb_cnts_t *cntp; | ||
| 1639 | int i; | ||
| 1640 | |||
| 1641 | memset(cnt, 0, sizeof(xfs_icsb_cnts_t)); | ||
| 1642 | |||
| 1643 | if (!(flags & XFS_ICSB_LAZY_COUNT)) | ||
| 1644 | xfs_icsb_lock_all_counters(mp); | ||
| 1645 | |||
| 1646 | for_each_online_cpu(i) { | ||
| 1647 | cntp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, i); | ||
| 1648 | cnt->icsb_icount += cntp->icsb_icount; | ||
| 1649 | cnt->icsb_ifree += cntp->icsb_ifree; | ||
| 1650 | cnt->icsb_fdblocks += cntp->icsb_fdblocks; | ||
| 1651 | } | ||
| 1652 | |||
| 1653 | if (!(flags & XFS_ICSB_LAZY_COUNT)) | ||
| 1654 | xfs_icsb_unlock_all_counters(mp); | ||
| 1655 | } | ||
| 1656 | |||
| 1657 | STATIC int | ||
| 1658 | xfs_icsb_counter_disabled( | ||
| 1659 | xfs_mount_t *mp, | ||
| 1660 | xfs_sb_field_t field) | ||
| 1661 | { | ||
| 1662 | ASSERT((field >= XFS_SBS_ICOUNT) && (field <= XFS_SBS_FDBLOCKS)); | ||
| 1663 | return test_bit(field, &mp->m_icsb_counters); | ||
| 1664 | } | ||
| 1665 | |||
| 1666 | STATIC void | ||
| 1667 | xfs_icsb_disable_counter( | ||
| 1668 | xfs_mount_t *mp, | ||
| 1669 | xfs_sb_field_t field) | ||
| 1670 | { | ||
| 1671 | xfs_icsb_cnts_t cnt; | ||
| 1672 | |||
| 1673 | ASSERT((field >= XFS_SBS_ICOUNT) && (field <= XFS_SBS_FDBLOCKS)); | ||
| 1674 | |||
| 1675 | /* | ||
| 1676 | * If we are already disabled, then there is nothing to do | ||
| 1677 | * here. We check before locking all the counters to avoid | ||
| 1678 | * the expensive lock operation when being called in the | ||
| 1679 | * slow path and the counter is already disabled. This is | ||
| 1680 | * safe because the only time we set or clear this state is under | ||
| 1681 | * the m_icsb_mutex. | ||
| 1682 | */ | ||
| 1683 | if (xfs_icsb_counter_disabled(mp, field)) | ||
| 1684 | return; | ||
| 1685 | |||
| 1686 | xfs_icsb_lock_all_counters(mp); | ||
| 1687 | if (!test_and_set_bit(field, &mp->m_icsb_counters)) { | ||
| 1688 | /* drain back to superblock */ | ||
| 1689 | |||
| 1690 | xfs_icsb_count(mp, &cnt, XFS_ICSB_LAZY_COUNT); | ||
| 1691 | switch(field) { | ||
| 1692 | case XFS_SBS_ICOUNT: | ||
| 1693 | mp->m_sb.sb_icount = cnt.icsb_icount; | ||
| 1694 | break; | ||
| 1695 | case XFS_SBS_IFREE: | ||
| 1696 | mp->m_sb.sb_ifree = cnt.icsb_ifree; | ||
| 1697 | break; | ||
| 1698 | case XFS_SBS_FDBLOCKS: | ||
| 1699 | mp->m_sb.sb_fdblocks = cnt.icsb_fdblocks; | ||
| 1700 | break; | ||
| 1701 | default: | ||
| 1702 | BUG(); | ||
| 1703 | } | ||
| 1704 | } | ||
| 1705 | |||
| 1706 | xfs_icsb_unlock_all_counters(mp); | ||
| 1707 | } | ||
| 1708 | |||
| 1709 | STATIC void | ||
| 1710 | xfs_icsb_enable_counter( | ||
| 1711 | xfs_mount_t *mp, | ||
| 1712 | xfs_sb_field_t field, | ||
| 1713 | uint64_t count, | ||
| 1714 | uint64_t resid) | ||
| 1715 | { | ||
| 1716 | xfs_icsb_cnts_t *cntp; | ||
| 1717 | int i; | ||
| 1718 | |||
| 1719 | ASSERT((field >= XFS_SBS_ICOUNT) && (field <= XFS_SBS_FDBLOCKS)); | ||
| 1720 | |||
| 1721 | xfs_icsb_lock_all_counters(mp); | ||
| 1722 | for_each_online_cpu(i) { | ||
| 1723 | cntp = per_cpu_ptr(mp->m_sb_cnts, i); | ||
| 1724 | switch (field) { | ||
| 1725 | case XFS_SBS_ICOUNT: | ||
| 1726 | cntp->icsb_icount = count + resid; | ||
| 1727 | break; | ||
| 1728 | case XFS_SBS_IFREE: | ||
| 1729 | cntp->icsb_ifree = count + resid; | ||
| 1730 | break; | ||
| 1731 | case XFS_SBS_FDBLOCKS: | ||
| 1732 | cntp->icsb_fdblocks = count + resid; | ||
| 1733 | break; | ||
| 1734 | default: | ||
| 1735 | BUG(); | ||
| 1736 | break; | ||
| 1737 | } | ||
| 1738 | resid = 0; | ||
| 1739 | } | ||
| 1740 | clear_bit(field, &mp->m_icsb_counters); | ||
| 1741 | xfs_icsb_unlock_all_counters(mp); | ||
| 1742 | } | ||
| 1743 | |||
| 1744 | void | ||
| 1745 | xfs_icsb_sync_counters_locked( | ||
| 1746 | xfs_mount_t *mp, | ||
| 1747 | int flags) | ||
| 1748 | { | ||
| 1749 | xfs_icsb_cnts_t cnt; | ||
| 1750 | |||
| 1751 | xfs_icsb_count(mp, &cnt, flags); | ||
| 1752 | |||
| 1753 | if (!xfs_icsb_counter_disabled(mp, XFS_SBS_ICOUNT)) | ||
| 1754 | mp->m_sb.sb_icount = cnt.icsb_icount; | ||
| 1755 | if (!xfs_icsb_counter_disabled(mp, XFS_SBS_IFREE)) | ||
| 1756 | mp->m_sb.sb_ifree = cnt.icsb_ifree; | ||
| 1757 | if (!xfs_icsb_counter_disabled(mp, XFS_SBS_FDBLOCKS)) | ||
| 1758 | mp->m_sb.sb_fdblocks = cnt.icsb_fdblocks; | ||
| 1759 | } | ||
| 1760 | |||
| 1761 | /* | ||
| 1762 | * Accurate update of per-cpu counters to incore superblock | ||
| 1763 | */ | ||
| 1764 | void | ||
| 1765 | xfs_icsb_sync_counters( | ||
| 1766 | xfs_mount_t *mp, | ||
| 1767 | int flags) | ||
| 1768 | { | ||
| 1769 | spin_lock(&mp->m_sb_lock); | ||
| 1770 | xfs_icsb_sync_counters_locked(mp, flags); | ||
| 1771 | spin_unlock(&mp->m_sb_lock); | ||
| 1772 | } | ||
| 1773 | |||
| 1774 | /* | ||
| 1775 | * Balance and enable/disable counters as necessary. | ||
| 1776 | * | ||
| 1777 | * Thresholds for re-enabling counters are somewhat magic. inode counts are | ||
| 1778 | * chosen to be the same number as single on disk allocation chunk per CPU, and | ||
| 1779 | * free blocks is something far enough zero that we aren't going thrash when we | ||
| 1780 | * get near ENOSPC. We also need to supply a minimum we require per cpu to | ||
| 1781 | * prevent looping endlessly when xfs_alloc_space asks for more than will | ||
| 1782 | * be distributed to a single CPU but each CPU has enough blocks to be | ||
| 1783 | * reenabled. | ||
| 1784 | * | ||
| 1785 | * Note that we can be called when counters are already disabled. | ||
| 1786 | * xfs_icsb_disable_counter() optimises the counter locking in this case to | ||
| 1787 | * prevent locking every per-cpu counter needlessly. | ||
| 1788 | */ | ||
| 1789 | |||
| 1790 | #define XFS_ICSB_INO_CNTR_REENABLE (uint64_t)64 | ||
| 1791 | #define XFS_ICSB_FDBLK_CNTR_REENABLE(mp) \ | ||
| 1792 | (uint64_t)(512 + XFS_ALLOC_SET_ASIDE(mp)) | ||
| 1793 | STATIC void | ||
| 1794 | xfs_icsb_balance_counter_locked( | ||
| 1795 | xfs_mount_t *mp, | ||
| 1796 | xfs_sb_field_t field, | ||
| 1797 | int min_per_cpu) | ||
| 1798 | { | ||
| 1799 | uint64_t count, resid; | ||
| 1800 | int weight = num_online_cpus(); | ||
| 1801 | uint64_t min = (uint64_t)min_per_cpu; | ||
| 1802 | |||
| 1803 | /* disable counter and sync counter */ | ||
| 1804 | xfs_icsb_disable_counter(mp, field); | ||
| 1805 | |||
| 1806 | /* update counters - first CPU gets residual*/ | ||
| 1807 | switch (field) { | ||
| 1808 | case XFS_SBS_ICOUNT: | ||
| 1809 | count = mp->m_sb.sb_icount; | ||
| 1810 | resid = do_div(count, weight); | ||
| 1811 | if (count < max(min, XFS_ICSB_INO_CNTR_REENABLE)) | ||
| 1812 | return; | ||
| 1813 | break; | ||
| 1814 | case XFS_SBS_IFREE: | ||
| 1815 | count = mp->m_sb.sb_ifree; | ||
| 1816 | resid = do_div(count, weight); | ||
| 1817 | if (count < max(min, XFS_ICSB_INO_CNTR_REENABLE)) | ||
| 1818 | return; | ||
| 1819 | break; | ||
| 1820 | case XFS_SBS_FDBLOCKS: | ||
| 1821 | count = mp->m_sb.sb_fdblocks; | ||
| 1822 | resid = do_div(count, weight); | ||
| 1823 | if (count < max(min, XFS_ICSB_FDBLK_CNTR_REENABLE(mp))) | ||
| 1824 | return; | ||
| 1825 | break; | ||
| 1826 | default: | ||
| 1827 | BUG(); | ||
| 1828 | count = resid = 0; /* quiet, gcc */ | ||
| 1829 | break; | ||
| 1830 | } | ||
| 1831 | |||
| 1832 | xfs_icsb_enable_counter(mp, field, count, resid); | ||
| 1833 | } | ||
| 1834 | |||
| 1835 | STATIC void | ||
| 1836 | xfs_icsb_balance_counter( | ||
| 1837 | xfs_mount_t *mp, | ||
| 1838 | xfs_sb_field_t fields, | ||
| 1839 | int min_per_cpu) | ||
| 1840 | { | ||
| 1841 | spin_lock(&mp->m_sb_lock); | ||
| 1842 | xfs_icsb_balance_counter_locked(mp, fields, min_per_cpu); | ||
| 1843 | spin_unlock(&mp->m_sb_lock); | ||
| 1844 | } | ||
| 1845 | |||
| 1846 | int | ||
| 1847 | xfs_icsb_modify_counters( | ||
| 1848 | xfs_mount_t *mp, | ||
| 1849 | xfs_sb_field_t field, | ||
| 1850 | int64_t delta, | ||
| 1851 | int rsvd) | ||
| 1852 | { | ||
| 1853 | xfs_icsb_cnts_t *icsbp; | ||
| 1854 | long long lcounter; /* long counter for 64 bit fields */ | ||
| 1855 | int ret = 0; | ||
| 1856 | |||
| 1857 | might_sleep(); | ||
| 1858 | again: | ||
| 1859 | preempt_disable(); | ||
| 1860 | icsbp = this_cpu_ptr(mp->m_sb_cnts); | ||
| 1861 | |||
| 1862 | /* | ||
| 1863 | * if the counter is disabled, go to slow path | ||
| 1864 | */ | ||
| 1865 | if (unlikely(xfs_icsb_counter_disabled(mp, field))) | ||
| 1866 | goto slow_path; | ||
| 1867 | xfs_icsb_lock_cntr(icsbp); | ||
| 1868 | if (unlikely(xfs_icsb_counter_disabled(mp, field))) { | ||
| 1869 | xfs_icsb_unlock_cntr(icsbp); | ||
| 1870 | goto slow_path; | ||
| 1871 | } | ||
| 1872 | |||
| 1873 | switch (field) { | ||
| 1874 | case XFS_SBS_ICOUNT: | ||
| 1875 | lcounter = icsbp->icsb_icount; | ||
| 1876 | lcounter += delta; | ||
| 1877 | if (unlikely(lcounter < 0)) | ||
| 1878 | goto balance_counter; | ||
| 1879 | icsbp->icsb_icount = lcounter; | ||
| 1880 | break; | ||
| 1881 | |||
| 1882 | case XFS_SBS_IFREE: | ||
| 1883 | lcounter = icsbp->icsb_ifree; | ||
| 1884 | lcounter += delta; | ||
| 1885 | if (unlikely(lcounter < 0)) | ||
| 1886 | goto balance_counter; | ||
| 1887 | icsbp->icsb_ifree = lcounter; | ||
| 1888 | break; | ||
| 1889 | |||
| 1890 | case XFS_SBS_FDBLOCKS: | ||
| 1891 | BUG_ON((mp->m_resblks - mp->m_resblks_avail) != 0); | ||
| 1892 | |||
| 1893 | lcounter = icsbp->icsb_fdblocks - XFS_ALLOC_SET_ASIDE(mp); | ||
| 1894 | lcounter += delta; | ||
| 1895 | if (unlikely(lcounter < 0)) | ||
| 1896 | goto balance_counter; | ||
| 1897 | icsbp->icsb_fdblocks = lcounter + XFS_ALLOC_SET_ASIDE(mp); | ||
| 1898 | break; | ||
| 1899 | default: | ||
| 1900 | BUG(); | ||
| 1901 | break; | ||
| 1902 | } | ||
| 1903 | xfs_icsb_unlock_cntr(icsbp); | ||
| 1904 | preempt_enable(); | ||
| 1905 | return 0; | ||
| 1906 | |||
| 1907 | slow_path: | ||
| 1908 | preempt_enable(); | ||
| 1909 | |||
| 1910 | /* | ||
| 1911 | * serialise with a mutex so we don't burn lots of cpu on | ||
| 1912 | * the superblock lock. We still need to hold the superblock | ||
| 1913 | * lock, however, when we modify the global structures. | ||
| 1914 | */ | ||
| 1915 | xfs_icsb_lock(mp); | ||
| 1916 | |||
| 1917 | /* | ||
| 1918 | * Now running atomically. | ||
| 1919 | * | ||
| 1920 | * If the counter is enabled, someone has beaten us to rebalancing. | ||
| 1921 | * Drop the lock and try again in the fast path.... | ||
| 1922 | */ | ||
| 1923 | if (!(xfs_icsb_counter_disabled(mp, field))) { | ||
| 1924 | xfs_icsb_unlock(mp); | ||
| 1925 | goto again; | ||
| 1926 | } | ||
| 1927 | |||
| 1928 | /* | ||
| 1929 | * The counter is currently disabled. Because we are | ||
| 1930 | * running atomically here, we know a rebalance cannot | ||
| 1931 | * be in progress. Hence we can go straight to operating | ||
| 1932 | * on the global superblock. We do not call xfs_mod_incore_sb() | ||
| 1933 | * here even though we need to get the m_sb_lock. Doing so | ||
| 1934 | * will cause us to re-enter this function and deadlock. | ||
| 1935 | * Hence we get the m_sb_lock ourselves and then call | ||
| 1936 | * xfs_mod_incore_sb_unlocked() as the unlocked path operates | ||
| 1937 | * directly on the global counters. | ||
| 1938 | */ | ||
| 1939 | spin_lock(&mp->m_sb_lock); | ||
| 1940 | ret = xfs_mod_incore_sb_unlocked(mp, field, delta, rsvd); | ||
| 1941 | spin_unlock(&mp->m_sb_lock); | ||
| 1942 | |||
| 1943 | /* | ||
| 1944 | * Now that we've modified the global superblock, we | ||
| 1945 | * may be able to re-enable the distributed counters | ||
| 1946 | * (e.g. lots of space just got freed). After that | ||
| 1947 | * we are done. | ||
| 1948 | */ | ||
| 1949 | if (ret != -ENOSPC) | ||
| 1950 | xfs_icsb_balance_counter(mp, field, 0); | ||
| 1951 | xfs_icsb_unlock(mp); | ||
| 1952 | return ret; | ||
| 1953 | |||
| 1954 | balance_counter: | ||
| 1955 | xfs_icsb_unlock_cntr(icsbp); | ||
| 1956 | preempt_enable(); | ||
| 1957 | |||
| 1958 | /* | ||
| 1959 | * We may have multiple threads here if multiple per-cpu | ||
| 1960 | * counters run dry at the same time. This will mean we can | ||
| 1961 | * do more balances than strictly necessary but it is not | ||
| 1962 | * the common slowpath case. | ||
| 1963 | */ | ||
| 1964 | xfs_icsb_lock(mp); | ||
| 1965 | |||
| 1966 | /* | ||
| 1967 | * running atomically. | ||
| 1968 | * | ||
| 1969 | * This will leave the counter in the correct state for future | ||
| 1970 | * accesses. After the rebalance, we simply try again and our retry | ||
| 1971 | * will either succeed through the fast path or slow path without | ||
| 1972 | * another balance operation being required. | ||
| 1973 | */ | ||
| 1974 | xfs_icsb_balance_counter(mp, field, delta); | ||
| 1975 | xfs_icsb_unlock(mp); | ||
| 1976 | goto again; | ||
| 1977 | } | ||
| 1978 | |||
| 1979 | #endif | ||
diff --git a/fs/xfs/xfs_mount.h b/fs/xfs/xfs_mount.h index 0d8abd6364d9..8c995a2ccb6f 100644 --- a/fs/xfs/xfs_mount.h +++ b/fs/xfs/xfs_mount.h | |||
| @@ -18,8 +18,6 @@ | |||
| 18 | #ifndef __XFS_MOUNT_H__ | 18 | #ifndef __XFS_MOUNT_H__ |
| 19 | #define __XFS_MOUNT_H__ | 19 | #define __XFS_MOUNT_H__ |
| 20 | 20 | ||
| 21 | #ifdef __KERNEL__ | ||
| 22 | |||
| 23 | struct xlog; | 21 | struct xlog; |
| 24 | struct xfs_inode; | 22 | struct xfs_inode; |
| 25 | struct xfs_mru_cache; | 23 | struct xfs_mru_cache; |
| @@ -29,44 +27,6 @@ struct xfs_quotainfo; | |||
| 29 | struct xfs_dir_ops; | 27 | struct xfs_dir_ops; |
| 30 | struct xfs_da_geometry; | 28 | struct xfs_da_geometry; |
| 31 | 29 | ||
| 32 | #ifdef HAVE_PERCPU_SB | ||
| 33 | |||
| 34 | /* | ||
| 35 | * Valid per-cpu incore superblock counters. Note that if you add new counters, | ||
| 36 | * you may need to define new counter disabled bit field descriptors as there | ||
| 37 | * are more possible fields in the superblock that can fit in a bitfield on a | ||
| 38 | * 32 bit platform. The XFS_SBS_* values for the current current counters just | ||
| 39 | * fit. | ||
| 40 | */ | ||
| 41 | typedef struct xfs_icsb_cnts { | ||
| 42 | uint64_t icsb_fdblocks; | ||
| 43 | uint64_t icsb_ifree; | ||
| 44 | uint64_t icsb_icount; | ||
| 45 | unsigned long icsb_flags; | ||
| 46 | } xfs_icsb_cnts_t; | ||
| 47 | |||
| 48 | #define XFS_ICSB_FLAG_LOCK (1 << 0) /* counter lock bit */ | ||
| 49 | |||
| 50 | #define XFS_ICSB_LAZY_COUNT (1 << 1) /* accuracy not needed */ | ||
| 51 | |||
| 52 | extern int xfs_icsb_init_counters(struct xfs_mount *); | ||
| 53 | extern void xfs_icsb_reinit_counters(struct xfs_mount *); | ||
| 54 | extern void xfs_icsb_destroy_counters(struct xfs_mount *); | ||
| 55 | extern void xfs_icsb_sync_counters(struct xfs_mount *, int); | ||
| 56 | extern void xfs_icsb_sync_counters_locked(struct xfs_mount *, int); | ||
| 57 | extern int xfs_icsb_modify_counters(struct xfs_mount *, xfs_sb_field_t, | ||
| 58 | int64_t, int); | ||
| 59 | |||
| 60 | #else | ||
| 61 | #define xfs_icsb_init_counters(mp) (0) | ||
| 62 | #define xfs_icsb_destroy_counters(mp) do { } while (0) | ||
| 63 | #define xfs_icsb_reinit_counters(mp) do { } while (0) | ||
| 64 | #define xfs_icsb_sync_counters(mp, flags) do { } while (0) | ||
| 65 | #define xfs_icsb_sync_counters_locked(mp, flags) do { } while (0) | ||
| 66 | #define xfs_icsb_modify_counters(mp, field, delta, rsvd) \ | ||
| 67 | xfs_mod_incore_sb(mp, field, delta, rsvd) | ||
| 68 | #endif | ||
| 69 | |||
| 70 | /* dynamic preallocation free space thresholds, 5% down to 1% */ | 30 | /* dynamic preallocation free space thresholds, 5% down to 1% */ |
| 71 | enum { | 31 | enum { |
| 72 | XFS_LOWSP_1_PCNT = 0, | 32 | XFS_LOWSP_1_PCNT = 0, |
| @@ -81,8 +41,13 @@ typedef struct xfs_mount { | |||
| 81 | struct super_block *m_super; | 41 | struct super_block *m_super; |
| 82 | xfs_tid_t m_tid; /* next unused tid for fs */ | 42 | xfs_tid_t m_tid; /* next unused tid for fs */ |
| 83 | struct xfs_ail *m_ail; /* fs active log item list */ | 43 | struct xfs_ail *m_ail; /* fs active log item list */ |
| 84 | xfs_sb_t m_sb; /* copy of fs superblock */ | 44 | |
| 45 | struct xfs_sb m_sb; /* copy of fs superblock */ | ||
| 85 | spinlock_t m_sb_lock; /* sb counter lock */ | 46 | spinlock_t m_sb_lock; /* sb counter lock */ |
| 47 | struct percpu_counter m_icount; /* allocated inodes counter */ | ||
| 48 | struct percpu_counter m_ifree; /* free inodes counter */ | ||
| 49 | struct percpu_counter m_fdblocks; /* free block counter */ | ||
| 50 | |||
| 86 | struct xfs_buf *m_sb_bp; /* buffer for superblock */ | 51 | struct xfs_buf *m_sb_bp; /* buffer for superblock */ |
| 87 | char *m_fsname; /* filesystem name */ | 52 | char *m_fsname; /* filesystem name */ |
| 88 | int m_fsname_len; /* strlen of fs name */ | 53 | int m_fsname_len; /* strlen of fs name */ |
| @@ -152,12 +117,6 @@ typedef struct xfs_mount { | |||
| 152 | const struct xfs_dir_ops *m_nondir_inode_ops; /* !dir inode ops */ | 117 | const struct xfs_dir_ops *m_nondir_inode_ops; /* !dir inode ops */ |
| 153 | uint m_chsize; /* size of next field */ | 118 | uint m_chsize; /* size of next field */ |
| 154 | atomic_t m_active_trans; /* number trans frozen */ | 119 | atomic_t m_active_trans; /* number trans frozen */ |
| 155 | #ifdef HAVE_PERCPU_SB | ||
| 156 | xfs_icsb_cnts_t __percpu *m_sb_cnts; /* per-cpu superblock counters */ | ||
| 157 | unsigned long m_icsb_counters; /* disabled per-cpu counters */ | ||
| 158 | struct notifier_block m_icsb_notifier; /* hotplug cpu notifier */ | ||
| 159 | struct mutex m_icsb_mutex; /* balancer sync lock */ | ||
| 160 | #endif | ||
| 161 | struct xfs_mru_cache *m_filestream; /* per-mount filestream data */ | 120 | struct xfs_mru_cache *m_filestream; /* per-mount filestream data */ |
| 162 | struct delayed_work m_reclaim_work; /* background inode reclaim */ | 121 | struct delayed_work m_reclaim_work; /* background inode reclaim */ |
| 163 | struct delayed_work m_eofblocks_work; /* background eof blocks | 122 | struct delayed_work m_eofblocks_work; /* background eof blocks |
| @@ -301,35 +260,6 @@ xfs_daddr_to_agbno(struct xfs_mount *mp, xfs_daddr_t d) | |||
| 301 | } | 260 | } |
| 302 | 261 | ||
| 303 | /* | 262 | /* |
| 304 | * Per-cpu superblock locking functions | ||
| 305 | */ | ||
| 306 | #ifdef HAVE_PERCPU_SB | ||
| 307 | static inline void | ||
| 308 | xfs_icsb_lock(xfs_mount_t *mp) | ||
| 309 | { | ||
| 310 | mutex_lock(&mp->m_icsb_mutex); | ||
| 311 | } | ||
| 312 | |||
| 313 | static inline void | ||
| 314 | xfs_icsb_unlock(xfs_mount_t *mp) | ||
| 315 | { | ||
| 316 | mutex_unlock(&mp->m_icsb_mutex); | ||
| 317 | } | ||
| 318 | #else | ||
| 319 | #define xfs_icsb_lock(mp) | ||
| 320 | #define xfs_icsb_unlock(mp) | ||
| 321 | #endif | ||
| 322 | |||
| 323 | /* | ||
| 324 | * This structure is for use by the xfs_mod_incore_sb_batch() routine. | ||
| 325 | * xfs_growfs can specify a few fields which are more than int limit | ||
| 326 | */ | ||
| 327 | typedef struct xfs_mod_sb { | ||
| 328 | xfs_sb_field_t msb_field; /* Field to modify, see below */ | ||
| 329 | int64_t msb_delta; /* Change to make to specified field */ | ||
| 330 | } xfs_mod_sb_t; | ||
| 331 | |||
| 332 | /* | ||
| 333 | * Per-ag incore structure, copies of information in agf and agi, to improve the | 263 | * Per-ag incore structure, copies of information in agf and agi, to improve the |
| 334 | * performance of allocation group selection. | 264 | * performance of allocation group selection. |
| 335 | */ | 265 | */ |
| @@ -383,11 +313,14 @@ extern __uint64_t xfs_default_resblks(xfs_mount_t *mp); | |||
| 383 | extern int xfs_mountfs(xfs_mount_t *mp); | 313 | extern int xfs_mountfs(xfs_mount_t *mp); |
| 384 | extern int xfs_initialize_perag(xfs_mount_t *mp, xfs_agnumber_t agcount, | 314 | extern int xfs_initialize_perag(xfs_mount_t *mp, xfs_agnumber_t agcount, |
| 385 | xfs_agnumber_t *maxagi); | 315 | xfs_agnumber_t *maxagi); |
| 386 | |||
| 387 | extern void xfs_unmountfs(xfs_mount_t *); | 316 | extern void xfs_unmountfs(xfs_mount_t *); |
| 388 | extern int xfs_mod_incore_sb(xfs_mount_t *, xfs_sb_field_t, int64_t, int); | 317 | |
| 389 | extern int xfs_mod_incore_sb_batch(xfs_mount_t *, xfs_mod_sb_t *, | 318 | extern int xfs_mod_icount(struct xfs_mount *mp, int64_t delta); |
| 390 | uint, int); | 319 | extern int xfs_mod_ifree(struct xfs_mount *mp, int64_t delta); |
| 320 | extern int xfs_mod_fdblocks(struct xfs_mount *mp, int64_t delta, | ||
| 321 | bool reserved); | ||
| 322 | extern int xfs_mod_frextents(struct xfs_mount *mp, int64_t delta); | ||
| 323 | |||
| 391 | extern int xfs_mount_log_sb(xfs_mount_t *); | 324 | extern int xfs_mount_log_sb(xfs_mount_t *); |
| 392 | extern struct xfs_buf *xfs_getsb(xfs_mount_t *, int); | 325 | extern struct xfs_buf *xfs_getsb(xfs_mount_t *, int); |
| 393 | extern int xfs_readsb(xfs_mount_t *, int); | 326 | extern int xfs_readsb(xfs_mount_t *, int); |
| @@ -399,6 +332,4 @@ extern int xfs_dev_is_read_only(struct xfs_mount *, char *); | |||
| 399 | 332 | ||
| 400 | extern void xfs_set_low_space_thresholds(struct xfs_mount *); | 333 | extern void xfs_set_low_space_thresholds(struct xfs_mount *); |
| 401 | 334 | ||
| 402 | #endif /* __KERNEL__ */ | ||
| 403 | |||
| 404 | #endif /* __XFS_MOUNT_H__ */ | 335 | #endif /* __XFS_MOUNT_H__ */ |
diff --git a/fs/xfs/xfs_super.c b/fs/xfs/xfs_super.c index 58453e3255f8..f32ad64c4d05 100644 --- a/fs/xfs/xfs_super.c +++ b/fs/xfs/xfs_super.c | |||
| @@ -1013,24 +1013,6 @@ xfs_free_fsname( | |||
| 1013 | kfree(mp->m_logname); | 1013 | kfree(mp->m_logname); |
| 1014 | } | 1014 | } |
| 1015 | 1015 | ||
| 1016 | STATIC void | ||
| 1017 | xfs_fs_put_super( | ||
| 1018 | struct super_block *sb) | ||
| 1019 | { | ||
| 1020 | struct xfs_mount *mp = XFS_M(sb); | ||
| 1021 | |||
| 1022 | xfs_notice(mp, "Unmounting Filesystem"); | ||
| 1023 | xfs_filestream_unmount(mp); | ||
| 1024 | xfs_unmountfs(mp); | ||
| 1025 | |||
| 1026 | xfs_freesb(mp); | ||
| 1027 | xfs_icsb_destroy_counters(mp); | ||
| 1028 | xfs_destroy_mount_workqueues(mp); | ||
| 1029 | xfs_close_devices(mp); | ||
| 1030 | xfs_free_fsname(mp); | ||
| 1031 | kfree(mp); | ||
| 1032 | } | ||
| 1033 | |||
| 1034 | STATIC int | 1016 | STATIC int |
| 1035 | xfs_fs_sync_fs( | 1017 | xfs_fs_sync_fs( |
| 1036 | struct super_block *sb, | 1018 | struct super_block *sb, |
| @@ -1066,6 +1048,9 @@ xfs_fs_statfs( | |||
| 1066 | xfs_sb_t *sbp = &mp->m_sb; | 1048 | xfs_sb_t *sbp = &mp->m_sb; |
| 1067 | struct xfs_inode *ip = XFS_I(dentry->d_inode); | 1049 | struct xfs_inode *ip = XFS_I(dentry->d_inode); |
| 1068 | __uint64_t fakeinos, id; | 1050 | __uint64_t fakeinos, id; |
| 1051 | __uint64_t icount; | ||
| 1052 | __uint64_t ifree; | ||
| 1053 | __uint64_t fdblocks; | ||
| 1069 | xfs_extlen_t lsize; | 1054 | xfs_extlen_t lsize; |
| 1070 | __int64_t ffree; | 1055 | __int64_t ffree; |
| 1071 | 1056 | ||
| @@ -1076,17 +1061,21 @@ xfs_fs_statfs( | |||
| 1076 | statp->f_fsid.val[0] = (u32)id; | 1061 | statp->f_fsid.val[0] = (u32)id; |
| 1077 | statp->f_fsid.val[1] = (u32)(id >> 32); | 1062 | statp->f_fsid.val[1] = (u32)(id >> 32); |
| 1078 | 1063 | ||
| 1079 | xfs_icsb_sync_counters(mp, XFS_ICSB_LAZY_COUNT); | 1064 | icount = percpu_counter_sum(&mp->m_icount); |
| 1065 | ifree = percpu_counter_sum(&mp->m_ifree); | ||
| 1066 | fdblocks = percpu_counter_sum(&mp->m_fdblocks); | ||
| 1080 | 1067 | ||
| 1081 | spin_lock(&mp->m_sb_lock); | 1068 | spin_lock(&mp->m_sb_lock); |
| 1082 | statp->f_bsize = sbp->sb_blocksize; | 1069 | statp->f_bsize = sbp->sb_blocksize; |
| 1083 | lsize = sbp->sb_logstart ? sbp->sb_logblocks : 0; | 1070 | lsize = sbp->sb_logstart ? sbp->sb_logblocks : 0; |
| 1084 | statp->f_blocks = sbp->sb_dblocks - lsize; | 1071 | statp->f_blocks = sbp->sb_dblocks - lsize; |
| 1085 | statp->f_bfree = statp->f_bavail = | 1072 | spin_unlock(&mp->m_sb_lock); |
| 1086 | sbp->sb_fdblocks - XFS_ALLOC_SET_ASIDE(mp); | 1073 | |
| 1074 | statp->f_bfree = fdblocks - XFS_ALLOC_SET_ASIDE(mp); | ||
| 1075 | statp->f_bavail = statp->f_bfree; | ||
| 1076 | |||
| 1087 | fakeinos = statp->f_bfree << sbp->sb_inopblog; | 1077 | fakeinos = statp->f_bfree << sbp->sb_inopblog; |
| 1088 | statp->f_files = | 1078 | statp->f_files = MIN(icount + fakeinos, (__uint64_t)XFS_MAXINUMBER); |
| 1089 | MIN(sbp->sb_icount + fakeinos, (__uint64_t)XFS_MAXINUMBER); | ||
| 1090 | if (mp->m_maxicount) | 1079 | if (mp->m_maxicount) |
| 1091 | statp->f_files = min_t(typeof(statp->f_files), | 1080 | statp->f_files = min_t(typeof(statp->f_files), |
| 1092 | statp->f_files, | 1081 | statp->f_files, |
| @@ -1098,10 +1087,9 @@ xfs_fs_statfs( | |||
| 1098 | sbp->sb_icount); | 1087 | sbp->sb_icount); |
| 1099 | 1088 | ||
| 1100 | /* make sure statp->f_ffree does not underflow */ | 1089 | /* make sure statp->f_ffree does not underflow */ |
| 1101 | ffree = statp->f_files - (sbp->sb_icount - sbp->sb_ifree); | 1090 | ffree = statp->f_files - (icount - ifree); |
| 1102 | statp->f_ffree = max_t(__int64_t, ffree, 0); | 1091 | statp->f_ffree = max_t(__int64_t, ffree, 0); |
| 1103 | 1092 | ||
| 1104 | spin_unlock(&mp->m_sb_lock); | ||
| 1105 | 1093 | ||
| 1106 | if ((ip->i_d.di_flags & XFS_DIFLAG_PROJINHERIT) && | 1094 | if ((ip->i_d.di_flags & XFS_DIFLAG_PROJINHERIT) && |
| 1107 | ((mp->m_qflags & (XFS_PQUOTA_ACCT|XFS_PQUOTA_ENFD))) == | 1095 | ((mp->m_qflags & (XFS_PQUOTA_ACCT|XFS_PQUOTA_ENFD))) == |
| @@ -1382,6 +1370,51 @@ xfs_finish_flags( | |||
| 1382 | return 0; | 1370 | return 0; |
| 1383 | } | 1371 | } |
| 1384 | 1372 | ||
| 1373 | static int | ||
| 1374 | xfs_init_percpu_counters( | ||
| 1375 | struct xfs_mount *mp) | ||
| 1376 | { | ||
| 1377 | int error; | ||
| 1378 | |||
| 1379 | error = percpu_counter_init(&mp->m_icount, 0, GFP_KERNEL); | ||
| 1380 | if (error) | ||
| 1381 | return ENOMEM; | ||
| 1382 | |||
| 1383 | error = percpu_counter_init(&mp->m_ifree, 0, GFP_KERNEL); | ||
| 1384 | if (error) | ||
| 1385 | goto free_icount; | ||
| 1386 | |||
| 1387 | error = percpu_counter_init(&mp->m_fdblocks, 0, GFP_KERNEL); | ||
| 1388 | if (error) | ||
| 1389 | goto free_ifree; | ||
| 1390 | |||
| 1391 | return 0; | ||
| 1392 | |||
| 1393 | free_ifree: | ||
| 1394 | percpu_counter_destroy(&mp->m_ifree); | ||
| 1395 | free_icount: | ||
| 1396 | percpu_counter_destroy(&mp->m_icount); | ||
| 1397 | return -ENOMEM; | ||
| 1398 | } | ||
| 1399 | |||
| 1400 | void | ||
| 1401 | xfs_reinit_percpu_counters( | ||
| 1402 | struct xfs_mount *mp) | ||
| 1403 | { | ||
| 1404 | percpu_counter_set(&mp->m_icount, mp->m_sb.sb_icount); | ||
| 1405 | percpu_counter_set(&mp->m_ifree, mp->m_sb.sb_ifree); | ||
| 1406 | percpu_counter_set(&mp->m_fdblocks, mp->m_sb.sb_fdblocks); | ||
| 1407 | } | ||
| 1408 | |||
| 1409 | static void | ||
| 1410 | xfs_destroy_percpu_counters( | ||
| 1411 | struct xfs_mount *mp) | ||
| 1412 | { | ||
| 1413 | percpu_counter_destroy(&mp->m_icount); | ||
| 1414 | percpu_counter_destroy(&mp->m_ifree); | ||
| 1415 | percpu_counter_destroy(&mp->m_fdblocks); | ||
| 1416 | } | ||
| 1417 | |||
| 1385 | STATIC int | 1418 | STATIC int |
| 1386 | xfs_fs_fill_super( | 1419 | xfs_fs_fill_super( |
| 1387 | struct super_block *sb, | 1420 | struct super_block *sb, |
| @@ -1430,7 +1463,7 @@ xfs_fs_fill_super( | |||
| 1430 | if (error) | 1463 | if (error) |
| 1431 | goto out_close_devices; | 1464 | goto out_close_devices; |
| 1432 | 1465 | ||
| 1433 | error = xfs_icsb_init_counters(mp); | 1466 | error = xfs_init_percpu_counters(mp); |
| 1434 | if (error) | 1467 | if (error) |
| 1435 | goto out_destroy_workqueues; | 1468 | goto out_destroy_workqueues; |
| 1436 | 1469 | ||
| @@ -1488,7 +1521,7 @@ xfs_fs_fill_super( | |||
| 1488 | out_free_sb: | 1521 | out_free_sb: |
| 1489 | xfs_freesb(mp); | 1522 | xfs_freesb(mp); |
| 1490 | out_destroy_counters: | 1523 | out_destroy_counters: |
| 1491 | xfs_icsb_destroy_counters(mp); | 1524 | xfs_destroy_percpu_counters(mp); |
| 1492 | out_destroy_workqueues: | 1525 | out_destroy_workqueues: |
| 1493 | xfs_destroy_mount_workqueues(mp); | 1526 | xfs_destroy_mount_workqueues(mp); |
| 1494 | out_close_devices: | 1527 | out_close_devices: |
| @@ -1505,6 +1538,24 @@ out_destroy_workqueues: | |||
| 1505 | goto out_free_sb; | 1538 | goto out_free_sb; |
| 1506 | } | 1539 | } |
| 1507 | 1540 | ||
| 1541 | STATIC void | ||
| 1542 | xfs_fs_put_super( | ||
| 1543 | struct super_block *sb) | ||
| 1544 | { | ||
| 1545 | struct xfs_mount *mp = XFS_M(sb); | ||
| 1546 | |||
| 1547 | xfs_notice(mp, "Unmounting Filesystem"); | ||
| 1548 | xfs_filestream_unmount(mp); | ||
| 1549 | xfs_unmountfs(mp); | ||
| 1550 | |||
| 1551 | xfs_freesb(mp); | ||
| 1552 | xfs_destroy_percpu_counters(mp); | ||
| 1553 | xfs_destroy_mount_workqueues(mp); | ||
| 1554 | xfs_close_devices(mp); | ||
| 1555 | xfs_free_fsname(mp); | ||
| 1556 | kfree(mp); | ||
| 1557 | } | ||
| 1558 | |||
| 1508 | STATIC struct dentry * | 1559 | STATIC struct dentry * |
| 1509 | xfs_fs_mount( | 1560 | xfs_fs_mount( |
| 1510 | struct file_system_type *fs_type, | 1561 | struct file_system_type *fs_type, |
diff --git a/fs/xfs/xfs_super.h b/fs/xfs/xfs_super.h index 2b830c2f322e..499058fea303 100644 --- a/fs/xfs/xfs_super.h +++ b/fs/xfs/xfs_super.h | |||
| @@ -72,6 +72,8 @@ extern const struct export_operations xfs_export_operations; | |||
| 72 | extern const struct xattr_handler *xfs_xattr_handlers[]; | 72 | extern const struct xattr_handler *xfs_xattr_handlers[]; |
| 73 | extern const struct quotactl_ops xfs_quotactl_operations; | 73 | extern const struct quotactl_ops xfs_quotactl_operations; |
| 74 | 74 | ||
| 75 | extern void xfs_reinit_percpu_counters(struct xfs_mount *mp); | ||
| 76 | |||
| 75 | #define XFS_M(sb) ((struct xfs_mount *)((sb)->s_fs_info)) | 77 | #define XFS_M(sb) ((struct xfs_mount *)((sb)->s_fs_info)) |
| 76 | 78 | ||
| 77 | #endif /* __XFS_SUPER_H__ */ | 79 | #endif /* __XFS_SUPER_H__ */ |
diff --git a/fs/xfs/xfs_trans.c b/fs/xfs/xfs_trans.c index eb90cd59a0ec..220ef2c906b2 100644 --- a/fs/xfs/xfs_trans.c +++ b/fs/xfs/xfs_trans.c | |||
| @@ -173,7 +173,7 @@ xfs_trans_reserve( | |||
| 173 | uint rtextents) | 173 | uint rtextents) |
| 174 | { | 174 | { |
| 175 | int error = 0; | 175 | int error = 0; |
| 176 | int rsvd = (tp->t_flags & XFS_TRANS_RESERVE) != 0; | 176 | bool rsvd = (tp->t_flags & XFS_TRANS_RESERVE) != 0; |
| 177 | 177 | ||
| 178 | /* Mark this thread as being in a transaction */ | 178 | /* Mark this thread as being in a transaction */ |
| 179 | current_set_flags_nested(&tp->t_pflags, PF_FSTRANS); | 179 | current_set_flags_nested(&tp->t_pflags, PF_FSTRANS); |
| @@ -184,8 +184,7 @@ xfs_trans_reserve( | |||
| 184 | * fail if the count would go below zero. | 184 | * fail if the count would go below zero. |
| 185 | */ | 185 | */ |
| 186 | if (blocks > 0) { | 186 | if (blocks > 0) { |
| 187 | error = xfs_icsb_modify_counters(tp->t_mountp, XFS_SBS_FDBLOCKS, | 187 | error = xfs_mod_fdblocks(tp->t_mountp, -((int64_t)blocks), rsvd); |
| 188 | -((int64_t)blocks), rsvd); | ||
| 189 | if (error != 0) { | 188 | if (error != 0) { |
| 190 | current_restore_flags_nested(&tp->t_pflags, PF_FSTRANS); | 189 | current_restore_flags_nested(&tp->t_pflags, PF_FSTRANS); |
| 191 | return -ENOSPC; | 190 | return -ENOSPC; |
| @@ -236,8 +235,7 @@ xfs_trans_reserve( | |||
| 236 | * fail if the count would go below zero. | 235 | * fail if the count would go below zero. |
| 237 | */ | 236 | */ |
| 238 | if (rtextents > 0) { | 237 | if (rtextents > 0) { |
| 239 | error = xfs_mod_incore_sb(tp->t_mountp, XFS_SBS_FREXTENTS, | 238 | error = xfs_mod_frextents(tp->t_mountp, -((int64_t)rtextents)); |
| 240 | -((int64_t)rtextents), rsvd); | ||
| 241 | if (error) { | 239 | if (error) { |
| 242 | error = -ENOSPC; | 240 | error = -ENOSPC; |
| 243 | goto undo_log; | 241 | goto undo_log; |
| @@ -268,8 +266,7 @@ undo_log: | |||
| 268 | 266 | ||
| 269 | undo_blocks: | 267 | undo_blocks: |
| 270 | if (blocks > 0) { | 268 | if (blocks > 0) { |
| 271 | xfs_icsb_modify_counters(tp->t_mountp, XFS_SBS_FDBLOCKS, | 269 | xfs_mod_fdblocks(tp->t_mountp, -((int64_t)blocks), rsvd); |
| 272 | (int64_t)blocks, rsvd); | ||
| 273 | tp->t_blk_res = 0; | 270 | tp->t_blk_res = 0; |
| 274 | } | 271 | } |
| 275 | 272 | ||
| @@ -488,6 +485,54 @@ xfs_trans_apply_sb_deltas( | |||
| 488 | sizeof(sbp->sb_frextents) - 1); | 485 | sizeof(sbp->sb_frextents) - 1); |
| 489 | } | 486 | } |
| 490 | 487 | ||
| 488 | STATIC int | ||
| 489 | xfs_sb_mod8( | ||
| 490 | uint8_t *field, | ||
| 491 | int8_t delta) | ||
| 492 | { | ||
| 493 | int8_t counter = *field; | ||
| 494 | |||
| 495 | counter += delta; | ||
| 496 | if (counter < 0) { | ||
| 497 | ASSERT(0); | ||
| 498 | return -EINVAL; | ||
| 499 | } | ||
| 500 | *field = counter; | ||
| 501 | return 0; | ||
| 502 | } | ||
| 503 | |||
| 504 | STATIC int | ||
| 505 | xfs_sb_mod32( | ||
| 506 | uint32_t *field, | ||
| 507 | int32_t delta) | ||
| 508 | { | ||
| 509 | int32_t counter = *field; | ||
| 510 | |||
| 511 | counter += delta; | ||
| 512 | if (counter < 0) { | ||
| 513 | ASSERT(0); | ||
| 514 | return -EINVAL; | ||
| 515 | } | ||
| 516 | *field = counter; | ||
| 517 | return 0; | ||
| 518 | } | ||
| 519 | |||
| 520 | STATIC int | ||
| 521 | xfs_sb_mod64( | ||
| 522 | uint64_t *field, | ||
| 523 | int64_t delta) | ||
| 524 | { | ||
| 525 | int64_t counter = *field; | ||
| 526 | |||
| 527 | counter += delta; | ||
| 528 | if (counter < 0) { | ||
| 529 | ASSERT(0); | ||
| 530 | return -EINVAL; | ||
| 531 | } | ||
| 532 | *field = counter; | ||
| 533 | return 0; | ||
| 534 | } | ||
| 535 | |||
| 491 | /* | 536 | /* |
| 492 | * xfs_trans_unreserve_and_mod_sb() is called to release unused reservations | 537 | * xfs_trans_unreserve_and_mod_sb() is called to release unused reservations |
| 493 | * and apply superblock counter changes to the in-core superblock. The | 538 | * and apply superblock counter changes to the in-core superblock. The |
| @@ -495,13 +540,6 @@ xfs_trans_apply_sb_deltas( | |||
| 495 | * applied to the in-core superblock. The idea is that that has already been | 540 | * applied to the in-core superblock. The idea is that that has already been |
| 496 | * done. | 541 | * done. |
| 497 | * | 542 | * |
| 498 | * This is done efficiently with a single call to xfs_mod_incore_sb_batch(). | ||
| 499 | * However, we have to ensure that we only modify each superblock field only | ||
| 500 | * once because the application of the delta values may not be atomic. That can | ||
| 501 | * lead to ENOSPC races occurring if we have two separate modifcations of the | ||
| 502 | * free space counter to put back the entire reservation and then take away | ||
| 503 | * what we used. | ||
| 504 | * | ||
| 505 | * If we are not logging superblock counters, then the inode allocated/free and | 543 | * If we are not logging superblock counters, then the inode allocated/free and |
| 506 | * used block counts are not updated in the on disk superblock. In this case, | 544 | * used block counts are not updated in the on disk superblock. In this case, |
| 507 | * XFS_TRANS_SB_DIRTY will not be set when the transaction is updated but we | 545 | * XFS_TRANS_SB_DIRTY will not be set when the transaction is updated but we |
| @@ -509,21 +547,15 @@ xfs_trans_apply_sb_deltas( | |||
| 509 | */ | 547 | */ |
| 510 | void | 548 | void |
| 511 | xfs_trans_unreserve_and_mod_sb( | 549 | xfs_trans_unreserve_and_mod_sb( |
| 512 | xfs_trans_t *tp) | 550 | struct xfs_trans *tp) |
| 513 | { | 551 | { |
| 514 | xfs_mod_sb_t msb[9]; /* If you add cases, add entries */ | 552 | struct xfs_mount *mp = tp->t_mountp; |
| 515 | xfs_mod_sb_t *msbp; | 553 | bool rsvd = (tp->t_flags & XFS_TRANS_RESERVE) != 0; |
| 516 | xfs_mount_t *mp = tp->t_mountp; | 554 | int64_t blkdelta = 0; |
| 517 | /* REFERENCED */ | 555 | int64_t rtxdelta = 0; |
| 518 | int error; | 556 | int64_t idelta = 0; |
| 519 | int rsvd; | 557 | int64_t ifreedelta = 0; |
| 520 | int64_t blkdelta = 0; | 558 | int error; |
| 521 | int64_t rtxdelta = 0; | ||
| 522 | int64_t idelta = 0; | ||
| 523 | int64_t ifreedelta = 0; | ||
| 524 | |||
| 525 | msbp = msb; | ||
| 526 | rsvd = (tp->t_flags & XFS_TRANS_RESERVE) != 0; | ||
| 527 | 559 | ||
| 528 | /* calculate deltas */ | 560 | /* calculate deltas */ |
| 529 | if (tp->t_blk_res > 0) | 561 | if (tp->t_blk_res > 0) |
| @@ -547,97 +579,115 @@ xfs_trans_unreserve_and_mod_sb( | |||
| 547 | 579 | ||
| 548 | /* apply the per-cpu counters */ | 580 | /* apply the per-cpu counters */ |
| 549 | if (blkdelta) { | 581 | if (blkdelta) { |
| 550 | error = xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, | 582 | error = xfs_mod_fdblocks(mp, blkdelta, rsvd); |
| 551 | blkdelta, rsvd); | ||
| 552 | if (error) | 583 | if (error) |
| 553 | goto out; | 584 | goto out; |
| 554 | } | 585 | } |
| 555 | 586 | ||
| 556 | if (idelta) { | 587 | if (idelta) { |
| 557 | error = xfs_icsb_modify_counters(mp, XFS_SBS_ICOUNT, | 588 | error = xfs_mod_icount(mp, idelta); |
| 558 | idelta, rsvd); | ||
| 559 | if (error) | 589 | if (error) |
| 560 | goto out_undo_fdblocks; | 590 | goto out_undo_fdblocks; |
| 561 | } | 591 | } |
| 562 | 592 | ||
| 563 | if (ifreedelta) { | 593 | if (ifreedelta) { |
| 564 | error = xfs_icsb_modify_counters(mp, XFS_SBS_IFREE, | 594 | error = xfs_mod_ifree(mp, ifreedelta); |
| 565 | ifreedelta, rsvd); | ||
| 566 | if (error) | 595 | if (error) |
| 567 | goto out_undo_icount; | 596 | goto out_undo_icount; |
| 568 | } | 597 | } |
| 569 | 598 | ||
| 599 | if (rtxdelta == 0 && !(tp->t_flags & XFS_TRANS_SB_DIRTY)) | ||
| 600 | return; | ||
| 601 | |||
| 570 | /* apply remaining deltas */ | 602 | /* apply remaining deltas */ |
| 571 | if (rtxdelta != 0) { | 603 | spin_lock(&mp->m_sb_lock); |
| 572 | msbp->msb_field = XFS_SBS_FREXTENTS; | 604 | if (rtxdelta) { |
| 573 | msbp->msb_delta = rtxdelta; | 605 | error = xfs_sb_mod64(&mp->m_sb.sb_frextents, rtxdelta); |
| 574 | msbp++; | 606 | if (error) |
| 607 | goto out_undo_ifree; | ||
| 575 | } | 608 | } |
| 576 | 609 | ||
| 577 | if (tp->t_flags & XFS_TRANS_SB_DIRTY) { | 610 | if (tp->t_dblocks_delta != 0) { |
| 578 | if (tp->t_dblocks_delta != 0) { | 611 | error = xfs_sb_mod64(&mp->m_sb.sb_dblocks, tp->t_dblocks_delta); |
| 579 | msbp->msb_field = XFS_SBS_DBLOCKS; | 612 | if (error) |
| 580 | msbp->msb_delta = tp->t_dblocks_delta; | 613 | goto out_undo_frextents; |
| 581 | msbp++; | ||
| 582 | } | ||
| 583 | if (tp->t_agcount_delta != 0) { | ||
| 584 | msbp->msb_field = XFS_SBS_AGCOUNT; | ||
| 585 | msbp->msb_delta = tp->t_agcount_delta; | ||
| 586 | msbp++; | ||
| 587 | } | ||
| 588 | if (tp->t_imaxpct_delta != 0) { | ||
| 589 | msbp->msb_field = XFS_SBS_IMAX_PCT; | ||
| 590 | msbp->msb_delta = tp->t_imaxpct_delta; | ||
| 591 | msbp++; | ||
| 592 | } | ||
| 593 | if (tp->t_rextsize_delta != 0) { | ||
| 594 | msbp->msb_field = XFS_SBS_REXTSIZE; | ||
| 595 | msbp->msb_delta = tp->t_rextsize_delta; | ||
| 596 | msbp++; | ||
| 597 | } | ||
| 598 | if (tp->t_rbmblocks_delta != 0) { | ||
| 599 | msbp->msb_field = XFS_SBS_RBMBLOCKS; | ||
| 600 | msbp->msb_delta = tp->t_rbmblocks_delta; | ||
| 601 | msbp++; | ||
| 602 | } | ||
| 603 | if (tp->t_rblocks_delta != 0) { | ||
| 604 | msbp->msb_field = XFS_SBS_RBLOCKS; | ||
| 605 | msbp->msb_delta = tp->t_rblocks_delta; | ||
| 606 | msbp++; | ||
| 607 | } | ||
| 608 | if (tp->t_rextents_delta != 0) { | ||
| 609 | msbp->msb_field = XFS_SBS_REXTENTS; | ||
| 610 | msbp->msb_delta = tp->t_rextents_delta; | ||
| 611 | msbp++; | ||
| 612 | } | ||
| 613 | if (tp->t_rextslog_delta != 0) { | ||
| 614 | msbp->msb_field = XFS_SBS_REXTSLOG; | ||
| 615 | msbp->msb_delta = tp->t_rextslog_delta; | ||
| 616 | msbp++; | ||
| 617 | } | ||
| 618 | } | 614 | } |
| 619 | 615 | if (tp->t_agcount_delta != 0) { | |
| 620 | /* | 616 | error = xfs_sb_mod32(&mp->m_sb.sb_agcount, tp->t_agcount_delta); |
| 621 | * If we need to change anything, do it. | ||
| 622 | */ | ||
| 623 | if (msbp > msb) { | ||
| 624 | error = xfs_mod_incore_sb_batch(tp->t_mountp, msb, | ||
| 625 | (uint)(msbp - msb), rsvd); | ||
| 626 | if (error) | 617 | if (error) |
| 627 | goto out_undo_ifreecount; | 618 | goto out_undo_dblocks; |
| 628 | } | 619 | } |
| 629 | 620 | if (tp->t_imaxpct_delta != 0) { | |
| 621 | error = xfs_sb_mod8(&mp->m_sb.sb_imax_pct, tp->t_imaxpct_delta); | ||
| 622 | if (error) | ||
| 623 | goto out_undo_agcount; | ||
| 624 | } | ||
| 625 | if (tp->t_rextsize_delta != 0) { | ||
| 626 | error = xfs_sb_mod32(&mp->m_sb.sb_rextsize, | ||
| 627 | tp->t_rextsize_delta); | ||
| 628 | if (error) | ||
| 629 | goto out_undo_imaxpct; | ||
| 630 | } | ||
| 631 | if (tp->t_rbmblocks_delta != 0) { | ||
| 632 | error = xfs_sb_mod32(&mp->m_sb.sb_rbmblocks, | ||
| 633 | tp->t_rbmblocks_delta); | ||
| 634 | if (error) | ||
| 635 | goto out_undo_rextsize; | ||
| 636 | } | ||
| 637 | if (tp->t_rblocks_delta != 0) { | ||
| 638 | error = xfs_sb_mod64(&mp->m_sb.sb_rblocks, tp->t_rblocks_delta); | ||
| 639 | if (error) | ||
| 640 | goto out_undo_rbmblocks; | ||
| 641 | } | ||
| 642 | if (tp->t_rextents_delta != 0) { | ||
| 643 | error = xfs_sb_mod64(&mp->m_sb.sb_rextents, | ||
| 644 | tp->t_rextents_delta); | ||
| 645 | if (error) | ||
| 646 | goto out_undo_rblocks; | ||
| 647 | } | ||
| 648 | if (tp->t_rextslog_delta != 0) { | ||
| 649 | error = xfs_sb_mod8(&mp->m_sb.sb_rextslog, | ||
| 650 | tp->t_rextslog_delta); | ||
| 651 | if (error) | ||
| 652 | goto out_undo_rextents; | ||
| 653 | } | ||
| 654 | spin_unlock(&mp->m_sb_lock); | ||
| 630 | return; | 655 | return; |
| 631 | 656 | ||
| 632 | out_undo_ifreecount: | 657 | out_undo_rextents: |
| 658 | if (tp->t_rextents_delta) | ||
| 659 | xfs_sb_mod64(&mp->m_sb.sb_rextents, -tp->t_rextents_delta); | ||
| 660 | out_undo_rblocks: | ||
| 661 | if (tp->t_rblocks_delta) | ||
| 662 | xfs_sb_mod64(&mp->m_sb.sb_rblocks, -tp->t_rblocks_delta); | ||
| 663 | out_undo_rbmblocks: | ||
| 664 | if (tp->t_rbmblocks_delta) | ||
| 665 | xfs_sb_mod32(&mp->m_sb.sb_rbmblocks, -tp->t_rbmblocks_delta); | ||
| 666 | out_undo_rextsize: | ||
| 667 | if (tp->t_rextsize_delta) | ||
| 668 | xfs_sb_mod32(&mp->m_sb.sb_rextsize, -tp->t_rextsize_delta); | ||
| 669 | out_undo_imaxpct: | ||
| 670 | if (tp->t_rextsize_delta) | ||
| 671 | xfs_sb_mod8(&mp->m_sb.sb_imax_pct, -tp->t_imaxpct_delta); | ||
| 672 | out_undo_agcount: | ||
| 673 | if (tp->t_agcount_delta) | ||
| 674 | xfs_sb_mod32(&mp->m_sb.sb_agcount, -tp->t_agcount_delta); | ||
| 675 | out_undo_dblocks: | ||
| 676 | if (tp->t_dblocks_delta) | ||
| 677 | xfs_sb_mod64(&mp->m_sb.sb_dblocks, -tp->t_dblocks_delta); | ||
| 678 | out_undo_frextents: | ||
| 679 | if (rtxdelta) | ||
| 680 | xfs_sb_mod64(&mp->m_sb.sb_frextents, -rtxdelta); | ||
| 681 | out_undo_ifree: | ||
| 682 | spin_unlock(&mp->m_sb_lock); | ||
| 633 | if (ifreedelta) | 683 | if (ifreedelta) |
| 634 | xfs_icsb_modify_counters(mp, XFS_SBS_IFREE, -ifreedelta, rsvd); | 684 | xfs_mod_ifree(mp, -ifreedelta); |
| 635 | out_undo_icount: | 685 | out_undo_icount: |
| 636 | if (idelta) | 686 | if (idelta) |
| 637 | xfs_icsb_modify_counters(mp, XFS_SBS_ICOUNT, -idelta, rsvd); | 687 | xfs_mod_icount(mp, -idelta); |
| 638 | out_undo_fdblocks: | 688 | out_undo_fdblocks: |
| 639 | if (blkdelta) | 689 | if (blkdelta) |
| 640 | xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, -blkdelta, rsvd); | 690 | xfs_mod_fdblocks(mp, -blkdelta, rsvd); |
| 641 | out: | 691 | out: |
| 642 | ASSERT(error == 0); | 692 | ASSERT(error == 0); |
| 643 | return; | 693 | return; |
