diff options
author | Dave Chinner <david@fromorbit.com> | 2015-02-23 18:27:28 -0500 |
---|---|---|
committer | Dave Chinner <david@fromorbit.com> | 2015-02-23 18:27:28 -0500 |
commit | 4225441a1eec45241efe529d23403d8ca3d1d71b (patch) | |
tree | 12ed2dead95f32531cccebf31d95d67280c6e361 /fs/xfs | |
parent | 3cabb836d801c3ad791c2dc6be07ec5819ab0a37 (diff) | |
parent | 964aa8d9e4d36e0e54a88683d14c7d5b9d02aed8 (diff) |
Merge branch 'xfs-generic-sb-counters' into for-next
Conflicts:
fs/xfs/xfs_super.c
Diffstat (limited to 'fs/xfs')
-rw-r--r-- | fs/xfs/libxfs/xfs_bmap.c | 40 | ||||
-rw-r--r-- | fs/xfs/libxfs/xfs_format.h | 62 | ||||
-rw-r--r-- | fs/xfs/libxfs/xfs_ialloc.c | 6 | ||||
-rw-r--r-- | fs/xfs/libxfs/xfs_sb.c | 12 | ||||
-rw-r--r-- | fs/xfs/xfs_fsops.c | 20 | ||||
-rw-r--r-- | fs/xfs/xfs_iomap.c | 3 | ||||
-rw-r--r-- | fs/xfs/xfs_linux.h | 9 | ||||
-rw-r--r-- | fs/xfs/xfs_log_recover.c | 4 | ||||
-rw-r--r-- | fs/xfs/xfs_mount.c | 918 | ||||
-rw-r--r-- | fs/xfs/xfs_mount.h | 95 | ||||
-rw-r--r-- | fs/xfs/xfs_super.c | 105 | ||||
-rw-r--r-- | fs/xfs/xfs_super.h | 2 | ||||
-rw-r--r-- | fs/xfs/xfs_trans.c | 234 |
13 files changed, 381 insertions, 1129 deletions
diff --git a/fs/xfs/libxfs/xfs_bmap.c b/fs/xfs/libxfs/xfs_bmap.c index 60cfa90163b8..8ae3775a9b63 100644 --- a/fs/xfs/libxfs/xfs_bmap.c +++ b/fs/xfs/libxfs/xfs_bmap.c | |||
@@ -2215,9 +2215,8 @@ xfs_bmap_add_extent_delay_real( | |||
2215 | diff = (int)(temp + temp2 - startblockval(PREV.br_startblock) - | 2215 | diff = (int)(temp + temp2 - startblockval(PREV.br_startblock) - |
2216 | (bma->cur ? bma->cur->bc_private.b.allocated : 0)); | 2216 | (bma->cur ? bma->cur->bc_private.b.allocated : 0)); |
2217 | if (diff > 0) { | 2217 | if (diff > 0) { |
2218 | error = xfs_icsb_modify_counters(bma->ip->i_mount, | 2218 | error = xfs_mod_fdblocks(bma->ip->i_mount, |
2219 | XFS_SBS_FDBLOCKS, | 2219 | -((int64_t)diff), false); |
2220 | -((int64_t)diff), 0); | ||
2221 | ASSERT(!error); | 2220 | ASSERT(!error); |
2222 | if (error) | 2221 | if (error) |
2223 | goto done; | 2222 | goto done; |
@@ -2268,9 +2267,8 @@ xfs_bmap_add_extent_delay_real( | |||
2268 | temp += bma->cur->bc_private.b.allocated; | 2267 | temp += bma->cur->bc_private.b.allocated; |
2269 | ASSERT(temp <= da_old); | 2268 | ASSERT(temp <= da_old); |
2270 | if (temp < da_old) | 2269 | if (temp < da_old) |
2271 | xfs_icsb_modify_counters(bma->ip->i_mount, | 2270 | xfs_mod_fdblocks(bma->ip->i_mount, |
2272 | XFS_SBS_FDBLOCKS, | 2271 | (int64_t)(da_old - temp), false); |
2273 | (int64_t)(da_old - temp), 0); | ||
2274 | } | 2272 | } |
2275 | 2273 | ||
2276 | /* clear out the allocated field, done with it now in any case. */ | 2274 | /* clear out the allocated field, done with it now in any case. */ |
@@ -2948,8 +2946,8 @@ xfs_bmap_add_extent_hole_delay( | |||
2948 | } | 2946 | } |
2949 | if (oldlen != newlen) { | 2947 | if (oldlen != newlen) { |
2950 | ASSERT(oldlen > newlen); | 2948 | ASSERT(oldlen > newlen); |
2951 | xfs_icsb_modify_counters(ip->i_mount, XFS_SBS_FDBLOCKS, | 2949 | xfs_mod_fdblocks(ip->i_mount, (int64_t)(oldlen - newlen), |
2952 | (int64_t)(oldlen - newlen), 0); | 2950 | false); |
2953 | /* | 2951 | /* |
2954 | * Nothing to do for disk quota accounting here. | 2952 | * Nothing to do for disk quota accounting here. |
2955 | */ | 2953 | */ |
@@ -4166,18 +4164,15 @@ xfs_bmapi_reserve_delalloc( | |||
4166 | ASSERT(indlen > 0); | 4164 | ASSERT(indlen > 0); |
4167 | 4165 | ||
4168 | if (rt) { | 4166 | if (rt) { |
4169 | error = xfs_mod_incore_sb(mp, XFS_SBS_FREXTENTS, | 4167 | error = xfs_mod_frextents(mp, -((int64_t)extsz)); |
4170 | -((int64_t)extsz), 0); | ||
4171 | } else { | 4168 | } else { |
4172 | error = xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, | 4169 | error = xfs_mod_fdblocks(mp, -((int64_t)alen), false); |
4173 | -((int64_t)alen), 0); | ||
4174 | } | 4170 | } |
4175 | 4171 | ||
4176 | if (error) | 4172 | if (error) |
4177 | goto out_unreserve_quota; | 4173 | goto out_unreserve_quota; |
4178 | 4174 | ||
4179 | error = xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, | 4175 | error = xfs_mod_fdblocks(mp, -((int64_t)indlen), false); |
4180 | -((int64_t)indlen), 0); | ||
4181 | if (error) | 4176 | if (error) |
4182 | goto out_unreserve_blocks; | 4177 | goto out_unreserve_blocks; |
4183 | 4178 | ||
@@ -4204,9 +4199,9 @@ xfs_bmapi_reserve_delalloc( | |||
4204 | 4199 | ||
4205 | out_unreserve_blocks: | 4200 | out_unreserve_blocks: |
4206 | if (rt) | 4201 | if (rt) |
4207 | xfs_mod_incore_sb(mp, XFS_SBS_FREXTENTS, extsz, 0); | 4202 | xfs_mod_frextents(mp, extsz); |
4208 | else | 4203 | else |
4209 | xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, alen, 0); | 4204 | xfs_mod_fdblocks(mp, alen, false); |
4210 | out_unreserve_quota: | 4205 | out_unreserve_quota: |
4211 | if (XFS_IS_QUOTA_ON(mp)) | 4206 | if (XFS_IS_QUOTA_ON(mp)) |
4212 | xfs_trans_unreserve_quota_nblks(NULL, ip, (long)alen, 0, rt ? | 4207 | xfs_trans_unreserve_quota_nblks(NULL, ip, (long)alen, 0, rt ? |
@@ -5019,10 +5014,8 @@ xfs_bmap_del_extent( | |||
5019 | * Nothing to do for disk quota accounting here. | 5014 | * Nothing to do for disk quota accounting here. |
5020 | */ | 5015 | */ |
5021 | ASSERT(da_old >= da_new); | 5016 | ASSERT(da_old >= da_new); |
5022 | if (da_old > da_new) { | 5017 | if (da_old > da_new) |
5023 | xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, | 5018 | xfs_mod_fdblocks(mp, (int64_t)(da_old - da_new), false); |
5024 | (int64_t)(da_old - da_new), 0); | ||
5025 | } | ||
5026 | done: | 5019 | done: |
5027 | *logflagsp = flags; | 5020 | *logflagsp = flags; |
5028 | return error; | 5021 | return error; |
@@ -5291,14 +5284,13 @@ xfs_bunmapi( | |||
5291 | 5284 | ||
5292 | rtexts = XFS_FSB_TO_B(mp, del.br_blockcount); | 5285 | rtexts = XFS_FSB_TO_B(mp, del.br_blockcount); |
5293 | do_div(rtexts, mp->m_sb.sb_rextsize); | 5286 | do_div(rtexts, mp->m_sb.sb_rextsize); |
5294 | xfs_mod_incore_sb(mp, XFS_SBS_FREXTENTS, | 5287 | xfs_mod_frextents(mp, (int64_t)rtexts); |
5295 | (int64_t)rtexts, 0); | ||
5296 | (void)xfs_trans_reserve_quota_nblks(NULL, | 5288 | (void)xfs_trans_reserve_quota_nblks(NULL, |
5297 | ip, -((long)del.br_blockcount), 0, | 5289 | ip, -((long)del.br_blockcount), 0, |
5298 | XFS_QMOPT_RES_RTBLKS); | 5290 | XFS_QMOPT_RES_RTBLKS); |
5299 | } else { | 5291 | } else { |
5300 | xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, | 5292 | xfs_mod_fdblocks(mp, (int64_t)del.br_blockcount, |
5301 | (int64_t)del.br_blockcount, 0); | 5293 | false); |
5302 | (void)xfs_trans_reserve_quota_nblks(NULL, | 5294 | (void)xfs_trans_reserve_quota_nblks(NULL, |
5303 | ip, -((long)del.br_blockcount), 0, | 5295 | ip, -((long)del.br_blockcount), 0, |
5304 | XFS_QMOPT_RES_REGBLKS); | 5296 | XFS_QMOPT_RES_REGBLKS); |
diff --git a/fs/xfs/libxfs/xfs_format.h b/fs/xfs/libxfs/xfs_format.h index 8eb718979383..4daaa662337b 100644 --- a/fs/xfs/libxfs/xfs_format.h +++ b/fs/xfs/libxfs/xfs_format.h | |||
@@ -264,68 +264,6 @@ typedef struct xfs_dsb { | |||
264 | /* must be padded to 64 bit alignment */ | 264 | /* must be padded to 64 bit alignment */ |
265 | } xfs_dsb_t; | 265 | } xfs_dsb_t; |
266 | 266 | ||
267 | /* | ||
268 | * Sequence number values for the fields. | ||
269 | */ | ||
270 | typedef enum { | ||
271 | XFS_SBS_MAGICNUM, XFS_SBS_BLOCKSIZE, XFS_SBS_DBLOCKS, XFS_SBS_RBLOCKS, | ||
272 | XFS_SBS_REXTENTS, XFS_SBS_UUID, XFS_SBS_LOGSTART, XFS_SBS_ROOTINO, | ||
273 | XFS_SBS_RBMINO, XFS_SBS_RSUMINO, XFS_SBS_REXTSIZE, XFS_SBS_AGBLOCKS, | ||
274 | XFS_SBS_AGCOUNT, XFS_SBS_RBMBLOCKS, XFS_SBS_LOGBLOCKS, | ||
275 | XFS_SBS_VERSIONNUM, XFS_SBS_SECTSIZE, XFS_SBS_INODESIZE, | ||
276 | XFS_SBS_INOPBLOCK, XFS_SBS_FNAME, XFS_SBS_BLOCKLOG, | ||
277 | XFS_SBS_SECTLOG, XFS_SBS_INODELOG, XFS_SBS_INOPBLOG, XFS_SBS_AGBLKLOG, | ||
278 | XFS_SBS_REXTSLOG, XFS_SBS_INPROGRESS, XFS_SBS_IMAX_PCT, XFS_SBS_ICOUNT, | ||
279 | XFS_SBS_IFREE, XFS_SBS_FDBLOCKS, XFS_SBS_FREXTENTS, XFS_SBS_UQUOTINO, | ||
280 | XFS_SBS_GQUOTINO, XFS_SBS_QFLAGS, XFS_SBS_FLAGS, XFS_SBS_SHARED_VN, | ||
281 | XFS_SBS_INOALIGNMT, XFS_SBS_UNIT, XFS_SBS_WIDTH, XFS_SBS_DIRBLKLOG, | ||
282 | XFS_SBS_LOGSECTLOG, XFS_SBS_LOGSECTSIZE, XFS_SBS_LOGSUNIT, | ||
283 | XFS_SBS_FEATURES2, XFS_SBS_BAD_FEATURES2, XFS_SBS_FEATURES_COMPAT, | ||
284 | XFS_SBS_FEATURES_RO_COMPAT, XFS_SBS_FEATURES_INCOMPAT, | ||
285 | XFS_SBS_FEATURES_LOG_INCOMPAT, XFS_SBS_CRC, XFS_SBS_PAD, | ||
286 | XFS_SBS_PQUOTINO, XFS_SBS_LSN, | ||
287 | XFS_SBS_FIELDCOUNT | ||
288 | } xfs_sb_field_t; | ||
289 | |||
290 | /* | ||
291 | * Mask values, defined based on the xfs_sb_field_t values. | ||
292 | * Only define the ones we're using. | ||
293 | */ | ||
294 | #define XFS_SB_MVAL(x) (1LL << XFS_SBS_ ## x) | ||
295 | #define XFS_SB_UUID XFS_SB_MVAL(UUID) | ||
296 | #define XFS_SB_FNAME XFS_SB_MVAL(FNAME) | ||
297 | #define XFS_SB_ROOTINO XFS_SB_MVAL(ROOTINO) | ||
298 | #define XFS_SB_RBMINO XFS_SB_MVAL(RBMINO) | ||
299 | #define XFS_SB_RSUMINO XFS_SB_MVAL(RSUMINO) | ||
300 | #define XFS_SB_VERSIONNUM XFS_SB_MVAL(VERSIONNUM) | ||
301 | #define XFS_SB_UQUOTINO XFS_SB_MVAL(UQUOTINO) | ||
302 | #define XFS_SB_GQUOTINO XFS_SB_MVAL(GQUOTINO) | ||
303 | #define XFS_SB_QFLAGS XFS_SB_MVAL(QFLAGS) | ||
304 | #define XFS_SB_SHARED_VN XFS_SB_MVAL(SHARED_VN) | ||
305 | #define XFS_SB_UNIT XFS_SB_MVAL(UNIT) | ||
306 | #define XFS_SB_WIDTH XFS_SB_MVAL(WIDTH) | ||
307 | #define XFS_SB_ICOUNT XFS_SB_MVAL(ICOUNT) | ||
308 | #define XFS_SB_IFREE XFS_SB_MVAL(IFREE) | ||
309 | #define XFS_SB_FDBLOCKS XFS_SB_MVAL(FDBLOCKS) | ||
310 | #define XFS_SB_FEATURES2 (XFS_SB_MVAL(FEATURES2) | \ | ||
311 | XFS_SB_MVAL(BAD_FEATURES2)) | ||
312 | #define XFS_SB_FEATURES_COMPAT XFS_SB_MVAL(FEATURES_COMPAT) | ||
313 | #define XFS_SB_FEATURES_RO_COMPAT XFS_SB_MVAL(FEATURES_RO_COMPAT) | ||
314 | #define XFS_SB_FEATURES_INCOMPAT XFS_SB_MVAL(FEATURES_INCOMPAT) | ||
315 | #define XFS_SB_FEATURES_LOG_INCOMPAT XFS_SB_MVAL(FEATURES_LOG_INCOMPAT) | ||
316 | #define XFS_SB_CRC XFS_SB_MVAL(CRC) | ||
317 | #define XFS_SB_PQUOTINO XFS_SB_MVAL(PQUOTINO) | ||
318 | #define XFS_SB_NUM_BITS ((int)XFS_SBS_FIELDCOUNT) | ||
319 | #define XFS_SB_ALL_BITS ((1LL << XFS_SB_NUM_BITS) - 1) | ||
320 | #define XFS_SB_MOD_BITS \ | ||
321 | (XFS_SB_UUID | XFS_SB_ROOTINO | XFS_SB_RBMINO | XFS_SB_RSUMINO | \ | ||
322 | XFS_SB_VERSIONNUM | XFS_SB_UQUOTINO | XFS_SB_GQUOTINO | \ | ||
323 | XFS_SB_QFLAGS | XFS_SB_SHARED_VN | XFS_SB_UNIT | XFS_SB_WIDTH | \ | ||
324 | XFS_SB_ICOUNT | XFS_SB_IFREE | XFS_SB_FDBLOCKS | XFS_SB_FEATURES2 | \ | ||
325 | XFS_SB_FEATURES_COMPAT | XFS_SB_FEATURES_RO_COMPAT | \ | ||
326 | XFS_SB_FEATURES_INCOMPAT | XFS_SB_FEATURES_LOG_INCOMPAT | \ | ||
327 | XFS_SB_PQUOTINO) | ||
328 | |||
329 | 267 | ||
330 | /* | 268 | /* |
331 | * Misc. Flags - warning - these will be cleared by xfs_repair unless | 269 | * Misc. Flags - warning - these will be cleared by xfs_repair unless |
diff --git a/fs/xfs/libxfs/xfs_ialloc.c b/fs/xfs/libxfs/xfs_ialloc.c index db0444893e96..07349a183a11 100644 --- a/fs/xfs/libxfs/xfs_ialloc.c +++ b/fs/xfs/libxfs/xfs_ialloc.c | |||
@@ -376,7 +376,8 @@ xfs_ialloc_ag_alloc( | |||
376 | */ | 376 | */ |
377 | newlen = args.mp->m_ialloc_inos; | 377 | newlen = args.mp->m_ialloc_inos; |
378 | if (args.mp->m_maxicount && | 378 | if (args.mp->m_maxicount && |
379 | args.mp->m_sb.sb_icount + newlen > args.mp->m_maxicount) | 379 | percpu_counter_read(&args.mp->m_icount) + newlen > |
380 | args.mp->m_maxicount) | ||
380 | return -ENOSPC; | 381 | return -ENOSPC; |
381 | args.minlen = args.maxlen = args.mp->m_ialloc_blks; | 382 | args.minlen = args.maxlen = args.mp->m_ialloc_blks; |
382 | /* | 383 | /* |
@@ -1340,7 +1341,8 @@ xfs_dialloc( | |||
1340 | * inode. | 1341 | * inode. |
1341 | */ | 1342 | */ |
1342 | if (mp->m_maxicount && | 1343 | if (mp->m_maxicount && |
1343 | mp->m_sb.sb_icount + mp->m_ialloc_inos > mp->m_maxicount) { | 1344 | percpu_counter_read(&mp->m_icount) + mp->m_ialloc_inos > |
1345 | mp->m_maxicount) { | ||
1344 | noroom = 1; | 1346 | noroom = 1; |
1345 | okalloc = 0; | 1347 | okalloc = 0; |
1346 | } | 1348 | } |
diff --git a/fs/xfs/libxfs/xfs_sb.c b/fs/xfs/libxfs/xfs_sb.c index f3ea02bf893e..dc4bfc5d88fc 100644 --- a/fs/xfs/libxfs/xfs_sb.c +++ b/fs/xfs/libxfs/xfs_sb.c | |||
@@ -735,17 +735,15 @@ xfs_initialize_perag_data( | |||
735 | btree += pag->pagf_btreeblks; | 735 | btree += pag->pagf_btreeblks; |
736 | xfs_perag_put(pag); | 736 | xfs_perag_put(pag); |
737 | } | 737 | } |
738 | /* | 738 | |
739 | * Overwrite incore superblock counters with just-read data | 739 | /* Overwrite incore superblock counters with just-read data */ |
740 | */ | ||
741 | spin_lock(&mp->m_sb_lock); | 740 | spin_lock(&mp->m_sb_lock); |
742 | sbp->sb_ifree = ifree; | 741 | sbp->sb_ifree = ifree; |
743 | sbp->sb_icount = ialloc; | 742 | sbp->sb_icount = ialloc; |
744 | sbp->sb_fdblocks = bfree + bfreelst + btree; | 743 | sbp->sb_fdblocks = bfree + bfreelst + btree; |
745 | spin_unlock(&mp->m_sb_lock); | 744 | spin_unlock(&mp->m_sb_lock); |
746 | 745 | ||
747 | /* Fixup the per-cpu counters as well. */ | 746 | xfs_reinit_percpu_counters(mp); |
748 | xfs_icsb_reinit_counters(mp); | ||
749 | 747 | ||
750 | return 0; | 748 | return 0; |
751 | } | 749 | } |
@@ -763,6 +761,10 @@ xfs_log_sb( | |||
763 | struct xfs_mount *mp = tp->t_mountp; | 761 | struct xfs_mount *mp = tp->t_mountp; |
764 | struct xfs_buf *bp = xfs_trans_getsb(tp, mp, 0); | 762 | struct xfs_buf *bp = xfs_trans_getsb(tp, mp, 0); |
765 | 763 | ||
764 | mp->m_sb.sb_icount = percpu_counter_sum(&mp->m_icount); | ||
765 | mp->m_sb.sb_ifree = percpu_counter_sum(&mp->m_ifree); | ||
766 | mp->m_sb.sb_fdblocks = percpu_counter_sum(&mp->m_fdblocks); | ||
767 | |||
766 | xfs_sb_to_disk(XFS_BUF_TO_SBP(bp), &mp->m_sb); | 768 | xfs_sb_to_disk(XFS_BUF_TO_SBP(bp), &mp->m_sb); |
767 | xfs_trans_buf_set_type(tp, bp, XFS_BLFT_SB_BUF); | 769 | xfs_trans_buf_set_type(tp, bp, XFS_BLFT_SB_BUF); |
768 | xfs_trans_log_buf(tp, bp, 0, sizeof(struct xfs_dsb)); | 770 | xfs_trans_log_buf(tp, bp, 0, sizeof(struct xfs_dsb)); |
diff --git a/fs/xfs/xfs_fsops.c b/fs/xfs/xfs_fsops.c index 74efe5b760dc..cb7e8a29dfb6 100644 --- a/fs/xfs/xfs_fsops.c +++ b/fs/xfs/xfs_fsops.c | |||
@@ -637,12 +637,13 @@ xfs_fs_counts( | |||
637 | xfs_mount_t *mp, | 637 | xfs_mount_t *mp, |
638 | xfs_fsop_counts_t *cnt) | 638 | xfs_fsop_counts_t *cnt) |
639 | { | 639 | { |
640 | xfs_icsb_sync_counters(mp, XFS_ICSB_LAZY_COUNT); | 640 | cnt->allocino = percpu_counter_read_positive(&mp->m_icount); |
641 | cnt->freeino = percpu_counter_read_positive(&mp->m_ifree); | ||
642 | cnt->freedata = percpu_counter_read_positive(&mp->m_fdblocks) - | ||
643 | XFS_ALLOC_SET_ASIDE(mp); | ||
644 | |||
641 | spin_lock(&mp->m_sb_lock); | 645 | spin_lock(&mp->m_sb_lock); |
642 | cnt->freedata = mp->m_sb.sb_fdblocks - XFS_ALLOC_SET_ASIDE(mp); | ||
643 | cnt->freertx = mp->m_sb.sb_frextents; | 646 | cnt->freertx = mp->m_sb.sb_frextents; |
644 | cnt->freeino = mp->m_sb.sb_ifree; | ||
645 | cnt->allocino = mp->m_sb.sb_icount; | ||
646 | spin_unlock(&mp->m_sb_lock); | 647 | spin_unlock(&mp->m_sb_lock); |
647 | return 0; | 648 | return 0; |
648 | } | 649 | } |
@@ -692,14 +693,9 @@ xfs_reserve_blocks( | |||
692 | * what to do. This means that the amount of free space can | 693 | * what to do. This means that the amount of free space can |
693 | * change while we do this, so we need to retry if we end up | 694 | * change while we do this, so we need to retry if we end up |
694 | * trying to reserve more space than is available. | 695 | * trying to reserve more space than is available. |
695 | * | ||
696 | * We also use the xfs_mod_incore_sb() interface so that we | ||
697 | * don't have to care about whether per cpu counter are | ||
698 | * enabled, disabled or even compiled in.... | ||
699 | */ | 696 | */ |
700 | retry: | 697 | retry: |
701 | spin_lock(&mp->m_sb_lock); | 698 | spin_lock(&mp->m_sb_lock); |
702 | xfs_icsb_sync_counters_locked(mp, 0); | ||
703 | 699 | ||
704 | /* | 700 | /* |
705 | * If our previous reservation was larger than the current value, | 701 | * If our previous reservation was larger than the current value, |
@@ -716,7 +712,8 @@ retry: | |||
716 | } else { | 712 | } else { |
717 | __int64_t free; | 713 | __int64_t free; |
718 | 714 | ||
719 | free = mp->m_sb.sb_fdblocks - XFS_ALLOC_SET_ASIDE(mp); | 715 | free = percpu_counter_sum(&mp->m_fdblocks) - |
716 | XFS_ALLOC_SET_ASIDE(mp); | ||
720 | if (!free) | 717 | if (!free) |
721 | goto out; /* ENOSPC and fdblks_delta = 0 */ | 718 | goto out; /* ENOSPC and fdblks_delta = 0 */ |
722 | 719 | ||
@@ -755,8 +752,7 @@ out: | |||
755 | * the extra reserve blocks from the reserve..... | 752 | * the extra reserve blocks from the reserve..... |
756 | */ | 753 | */ |
757 | int error; | 754 | int error; |
758 | error = xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, | 755 | error = xfs_mod_fdblocks(mp, fdblks_delta, 0); |
759 | fdblks_delta, 0); | ||
760 | if (error == -ENOSPC) | 756 | if (error == -ENOSPC) |
761 | goto retry; | 757 | goto retry; |
762 | } | 758 | } |
diff --git a/fs/xfs/xfs_iomap.c b/fs/xfs/xfs_iomap.c index ccb1dd0d509e..38e633bad8c2 100644 --- a/fs/xfs/xfs_iomap.c +++ b/fs/xfs/xfs_iomap.c | |||
@@ -460,8 +460,7 @@ xfs_iomap_prealloc_size( | |||
460 | alloc_blocks = XFS_FILEOFF_MIN(roundup_pow_of_two(MAXEXTLEN), | 460 | alloc_blocks = XFS_FILEOFF_MIN(roundup_pow_of_two(MAXEXTLEN), |
461 | alloc_blocks); | 461 | alloc_blocks); |
462 | 462 | ||
463 | xfs_icsb_sync_counters(mp, XFS_ICSB_LAZY_COUNT); | 463 | freesp = percpu_counter_read_positive(&mp->m_fdblocks); |
464 | freesp = mp->m_sb.sb_fdblocks; | ||
465 | if (freesp < mp->m_low_space[XFS_LOWSP_5_PCNT]) { | 464 | if (freesp < mp->m_low_space[XFS_LOWSP_5_PCNT]) { |
466 | shift = 2; | 465 | shift = 2; |
467 | if (freesp < mp->m_low_space[XFS_LOWSP_4_PCNT]) | 466 | if (freesp < mp->m_low_space[XFS_LOWSP_4_PCNT]) |
diff --git a/fs/xfs/xfs_linux.h b/fs/xfs/xfs_linux.h index c31d2c2eadc4..7c7842c85a08 100644 --- a/fs/xfs/xfs_linux.h +++ b/fs/xfs/xfs_linux.h | |||
@@ -116,15 +116,6 @@ typedef __uint64_t __psunsigned_t; | |||
116 | #undef XFS_NATIVE_HOST | 116 | #undef XFS_NATIVE_HOST |
117 | #endif | 117 | #endif |
118 | 118 | ||
119 | /* | ||
120 | * Feature macros (disable/enable) | ||
121 | */ | ||
122 | #ifdef CONFIG_SMP | ||
123 | #define HAVE_PERCPU_SB /* per cpu superblock counters are a 2.6 feature */ | ||
124 | #else | ||
125 | #undef HAVE_PERCPU_SB /* per cpu superblock counters are a 2.6 feature */ | ||
126 | #endif | ||
127 | |||
128 | #define irix_sgid_inherit xfs_params.sgid_inherit.val | 119 | #define irix_sgid_inherit xfs_params.sgid_inherit.val |
129 | #define irix_symlink_mode xfs_params.symlink_mode.val | 120 | #define irix_symlink_mode xfs_params.symlink_mode.val |
130 | #define xfs_panic_mask xfs_params.panic_mask.val | 121 | #define xfs_panic_mask xfs_params.panic_mask.val |
diff --git a/fs/xfs/xfs_log_recover.c b/fs/xfs/xfs_log_recover.c index a5a945fc3bdc..4f5784f85a5b 100644 --- a/fs/xfs/xfs_log_recover.c +++ b/fs/xfs/xfs_log_recover.c | |||
@@ -4463,10 +4463,10 @@ xlog_do_recover( | |||
4463 | xfs_sb_from_disk(sbp, XFS_BUF_TO_SBP(bp)); | 4463 | xfs_sb_from_disk(sbp, XFS_BUF_TO_SBP(bp)); |
4464 | ASSERT(sbp->sb_magicnum == XFS_SB_MAGIC); | 4464 | ASSERT(sbp->sb_magicnum == XFS_SB_MAGIC); |
4465 | ASSERT(xfs_sb_good_version(sbp)); | 4465 | ASSERT(xfs_sb_good_version(sbp)); |
4466 | xfs_reinit_percpu_counters(log->l_mp); | ||
4467 | |||
4466 | xfs_buf_relse(bp); | 4468 | xfs_buf_relse(bp); |
4467 | 4469 | ||
4468 | /* We've re-read the superblock so re-initialize per-cpu counters */ | ||
4469 | xfs_icsb_reinit_counters(log->l_mp); | ||
4470 | 4470 | ||
4471 | xlog_recover_check_summary(log); | 4471 | xlog_recover_check_summary(log); |
4472 | 4472 | ||
diff --git a/fs/xfs/xfs_mount.c b/fs/xfs/xfs_mount.c index 4fa80e63eea2..2ce7ee3b4ec1 100644 --- a/fs/xfs/xfs_mount.c +++ b/fs/xfs/xfs_mount.c | |||
@@ -43,18 +43,6 @@ | |||
43 | #include "xfs_sysfs.h" | 43 | #include "xfs_sysfs.h" |
44 | 44 | ||
45 | 45 | ||
46 | #ifdef HAVE_PERCPU_SB | ||
47 | STATIC void xfs_icsb_balance_counter(xfs_mount_t *, xfs_sb_field_t, | ||
48 | int); | ||
49 | STATIC void xfs_icsb_balance_counter_locked(xfs_mount_t *, xfs_sb_field_t, | ||
50 | int); | ||
51 | STATIC void xfs_icsb_disable_counter(xfs_mount_t *, xfs_sb_field_t); | ||
52 | #else | ||
53 | |||
54 | #define xfs_icsb_balance_counter(mp, a, b) do { } while (0) | ||
55 | #define xfs_icsb_balance_counter_locked(mp, a, b) do { } while (0) | ||
56 | #endif | ||
57 | |||
58 | static DEFINE_MUTEX(xfs_uuid_table_mutex); | 46 | static DEFINE_MUTEX(xfs_uuid_table_mutex); |
59 | static int xfs_uuid_table_size; | 47 | static int xfs_uuid_table_size; |
60 | static uuid_t *xfs_uuid_table; | 48 | static uuid_t *xfs_uuid_table; |
@@ -347,8 +335,7 @@ reread: | |||
347 | goto reread; | 335 | goto reread; |
348 | } | 336 | } |
349 | 337 | ||
350 | /* Initialize per-cpu counters */ | 338 | xfs_reinit_percpu_counters(mp); |
351 | xfs_icsb_reinit_counters(mp); | ||
352 | 339 | ||
353 | /* no need to be quiet anymore, so reset the buf ops */ | 340 | /* no need to be quiet anymore, so reset the buf ops */ |
354 | bp->b_ops = &xfs_sb_buf_ops; | 341 | bp->b_ops = &xfs_sb_buf_ops; |
@@ -1087,8 +1074,6 @@ xfs_log_sbcount(xfs_mount_t *mp) | |||
1087 | if (!xfs_fs_writable(mp, SB_FREEZE_COMPLETE)) | 1074 | if (!xfs_fs_writable(mp, SB_FREEZE_COMPLETE)) |
1088 | return 0; | 1075 | return 0; |
1089 | 1076 | ||
1090 | xfs_icsb_sync_counters(mp, 0); | ||
1091 | |||
1092 | /* | 1077 | /* |
1093 | * we don't need to do this if we are updating the superblock | 1078 | * we don't need to do this if we are updating the superblock |
1094 | * counters on every modification. | 1079 | * counters on every modification. |
@@ -1099,253 +1084,136 @@ xfs_log_sbcount(xfs_mount_t *mp) | |||
1099 | return xfs_sync_sb(mp, true); | 1084 | return xfs_sync_sb(mp, true); |
1100 | } | 1085 | } |
1101 | 1086 | ||
1102 | /* | 1087 | int |
1103 | * xfs_mod_incore_sb_unlocked() is a utility routine commonly used to apply | 1088 | xfs_mod_icount( |
1104 | * a delta to a specified field in the in-core superblock. Simply | 1089 | struct xfs_mount *mp, |
1105 | * switch on the field indicated and apply the delta to that field. | 1090 | int64_t delta) |
1106 | * Fields are not allowed to dip below zero, so if the delta would | ||
1107 | * do this do not apply it and return EINVAL. | ||
1108 | * | ||
1109 | * The m_sb_lock must be held when this routine is called. | ||
1110 | */ | ||
1111 | STATIC int | ||
1112 | xfs_mod_incore_sb_unlocked( | ||
1113 | xfs_mount_t *mp, | ||
1114 | xfs_sb_field_t field, | ||
1115 | int64_t delta, | ||
1116 | int rsvd) | ||
1117 | { | 1091 | { |
1118 | int scounter; /* short counter for 32 bit fields */ | 1092 | /* deltas are +/-64, hence the large batch size of 128. */ |
1119 | long long lcounter; /* long counter for 64 bit fields */ | 1093 | __percpu_counter_add(&mp->m_icount, delta, 128); |
1120 | long long res_used, rem; | 1094 | if (percpu_counter_compare(&mp->m_icount, 0) < 0) { |
1121 | |||
1122 | /* | ||
1123 | * With the in-core superblock spin lock held, switch | ||
1124 | * on the indicated field. Apply the delta to the | ||
1125 | * proper field. If the fields value would dip below | ||
1126 | * 0, then do not apply the delta and return EINVAL. | ||
1127 | */ | ||
1128 | switch (field) { | ||
1129 | case XFS_SBS_ICOUNT: | ||
1130 | lcounter = (long long)mp->m_sb.sb_icount; | ||
1131 | lcounter += delta; | ||
1132 | if (lcounter < 0) { | ||
1133 | ASSERT(0); | ||
1134 | return -EINVAL; | ||
1135 | } | ||
1136 | mp->m_sb.sb_icount = lcounter; | ||
1137 | return 0; | ||
1138 | case XFS_SBS_IFREE: | ||
1139 | lcounter = (long long)mp->m_sb.sb_ifree; | ||
1140 | lcounter += delta; | ||
1141 | if (lcounter < 0) { | ||
1142 | ASSERT(0); | ||
1143 | return -EINVAL; | ||
1144 | } | ||
1145 | mp->m_sb.sb_ifree = lcounter; | ||
1146 | return 0; | ||
1147 | case XFS_SBS_FDBLOCKS: | ||
1148 | lcounter = (long long) | ||
1149 | mp->m_sb.sb_fdblocks - XFS_ALLOC_SET_ASIDE(mp); | ||
1150 | res_used = (long long)(mp->m_resblks - mp->m_resblks_avail); | ||
1151 | |||
1152 | if (delta > 0) { /* Putting blocks back */ | ||
1153 | if (res_used > delta) { | ||
1154 | mp->m_resblks_avail += delta; | ||
1155 | } else { | ||
1156 | rem = delta - res_used; | ||
1157 | mp->m_resblks_avail = mp->m_resblks; | ||
1158 | lcounter += rem; | ||
1159 | } | ||
1160 | } else { /* Taking blocks away */ | ||
1161 | lcounter += delta; | ||
1162 | if (lcounter >= 0) { | ||
1163 | mp->m_sb.sb_fdblocks = lcounter + | ||
1164 | XFS_ALLOC_SET_ASIDE(mp); | ||
1165 | return 0; | ||
1166 | } | ||
1167 | |||
1168 | /* | ||
1169 | * We are out of blocks, use any available reserved | ||
1170 | * blocks if were allowed to. | ||
1171 | */ | ||
1172 | if (!rsvd) | ||
1173 | return -ENOSPC; | ||
1174 | |||
1175 | lcounter = (long long)mp->m_resblks_avail + delta; | ||
1176 | if (lcounter >= 0) { | ||
1177 | mp->m_resblks_avail = lcounter; | ||
1178 | return 0; | ||
1179 | } | ||
1180 | printk_once(KERN_WARNING | ||
1181 | "Filesystem \"%s\": reserve blocks depleted! " | ||
1182 | "Consider increasing reserve pool size.", | ||
1183 | mp->m_fsname); | ||
1184 | return -ENOSPC; | ||
1185 | } | ||
1186 | |||
1187 | mp->m_sb.sb_fdblocks = lcounter + XFS_ALLOC_SET_ASIDE(mp); | ||
1188 | return 0; | ||
1189 | case XFS_SBS_FREXTENTS: | ||
1190 | lcounter = (long long)mp->m_sb.sb_frextents; | ||
1191 | lcounter += delta; | ||
1192 | if (lcounter < 0) { | ||
1193 | return -ENOSPC; | ||
1194 | } | ||
1195 | mp->m_sb.sb_frextents = lcounter; | ||
1196 | return 0; | ||
1197 | case XFS_SBS_DBLOCKS: | ||
1198 | lcounter = (long long)mp->m_sb.sb_dblocks; | ||
1199 | lcounter += delta; | ||
1200 | if (lcounter < 0) { | ||
1201 | ASSERT(0); | ||
1202 | return -EINVAL; | ||
1203 | } | ||
1204 | mp->m_sb.sb_dblocks = lcounter; | ||
1205 | return 0; | ||
1206 | case XFS_SBS_AGCOUNT: | ||
1207 | scounter = mp->m_sb.sb_agcount; | ||
1208 | scounter += delta; | ||
1209 | if (scounter < 0) { | ||
1210 | ASSERT(0); | ||
1211 | return -EINVAL; | ||
1212 | } | ||
1213 | mp->m_sb.sb_agcount = scounter; | ||
1214 | return 0; | ||
1215 | case XFS_SBS_IMAX_PCT: | ||
1216 | scounter = mp->m_sb.sb_imax_pct; | ||
1217 | scounter += delta; | ||
1218 | if (scounter < 0) { | ||
1219 | ASSERT(0); | ||
1220 | return -EINVAL; | ||
1221 | } | ||
1222 | mp->m_sb.sb_imax_pct = scounter; | ||
1223 | return 0; | ||
1224 | case XFS_SBS_REXTSIZE: | ||
1225 | scounter = mp->m_sb.sb_rextsize; | ||
1226 | scounter += delta; | ||
1227 | if (scounter < 0) { | ||
1228 | ASSERT(0); | ||
1229 | return -EINVAL; | ||
1230 | } | ||
1231 | mp->m_sb.sb_rextsize = scounter; | ||
1232 | return 0; | ||
1233 | case XFS_SBS_RBMBLOCKS: | ||
1234 | scounter = mp->m_sb.sb_rbmblocks; | ||
1235 | scounter += delta; | ||
1236 | if (scounter < 0) { | ||
1237 | ASSERT(0); | ||
1238 | return -EINVAL; | ||
1239 | } | ||
1240 | mp->m_sb.sb_rbmblocks = scounter; | ||
1241 | return 0; | ||
1242 | case XFS_SBS_RBLOCKS: | ||
1243 | lcounter = (long long)mp->m_sb.sb_rblocks; | ||
1244 | lcounter += delta; | ||
1245 | if (lcounter < 0) { | ||
1246 | ASSERT(0); | ||
1247 | return -EINVAL; | ||
1248 | } | ||
1249 | mp->m_sb.sb_rblocks = lcounter; | ||
1250 | return 0; | ||
1251 | case XFS_SBS_REXTENTS: | ||
1252 | lcounter = (long long)mp->m_sb.sb_rextents; | ||
1253 | lcounter += delta; | ||
1254 | if (lcounter < 0) { | ||
1255 | ASSERT(0); | ||
1256 | return -EINVAL; | ||
1257 | } | ||
1258 | mp->m_sb.sb_rextents = lcounter; | ||
1259 | return 0; | ||
1260 | case XFS_SBS_REXTSLOG: | ||
1261 | scounter = mp->m_sb.sb_rextslog; | ||
1262 | scounter += delta; | ||
1263 | if (scounter < 0) { | ||
1264 | ASSERT(0); | ||
1265 | return -EINVAL; | ||
1266 | } | ||
1267 | mp->m_sb.sb_rextslog = scounter; | ||
1268 | return 0; | ||
1269 | default: | ||
1270 | ASSERT(0); | 1095 | ASSERT(0); |
1096 | percpu_counter_add(&mp->m_icount, -delta); | ||
1271 | return -EINVAL; | 1097 | return -EINVAL; |
1272 | } | 1098 | } |
1099 | return 0; | ||
1273 | } | 1100 | } |
1274 | 1101 | ||
1275 | /* | ||
1276 | * xfs_mod_incore_sb() is used to change a field in the in-core | ||
1277 | * superblock structure by the specified delta. This modification | ||
1278 | * is protected by the m_sb_lock. Just use the xfs_mod_incore_sb_unlocked() | ||
1279 | * routine to do the work. | ||
1280 | */ | ||
1281 | int | 1102 | int |
1282 | xfs_mod_incore_sb( | 1103 | xfs_mod_ifree( |
1283 | struct xfs_mount *mp, | 1104 | struct xfs_mount *mp, |
1284 | xfs_sb_field_t field, | 1105 | int64_t delta) |
1285 | int64_t delta, | ||
1286 | int rsvd) | ||
1287 | { | 1106 | { |
1288 | int status; | 1107 | percpu_counter_add(&mp->m_ifree, delta); |
1289 | 1108 | if (percpu_counter_compare(&mp->m_ifree, 0) < 0) { | |
1290 | #ifdef HAVE_PERCPU_SB | 1109 | ASSERT(0); |
1291 | ASSERT(field < XFS_SBS_ICOUNT || field > XFS_SBS_FDBLOCKS); | 1110 | percpu_counter_add(&mp->m_ifree, -delta); |
1292 | #endif | 1111 | return -EINVAL; |
1293 | spin_lock(&mp->m_sb_lock); | 1112 | } |
1294 | status = xfs_mod_incore_sb_unlocked(mp, field, delta, rsvd); | 1113 | return 0; |
1295 | spin_unlock(&mp->m_sb_lock); | ||
1296 | |||
1297 | return status; | ||
1298 | } | 1114 | } |
1299 | 1115 | ||
1300 | /* | ||
1301 | * Change more than one field in the in-core superblock structure at a time. | ||
1302 | * | ||
1303 | * The fields and changes to those fields are specified in the array of | ||
1304 | * xfs_mod_sb structures passed in. Either all of the specified deltas | ||
1305 | * will be applied or none of them will. If any modified field dips below 0, | ||
1306 | * then all modifications will be backed out and EINVAL will be returned. | ||
1307 | * | ||
1308 | * Note that this function may not be used for the superblock values that | ||
1309 | * are tracked with the in-memory per-cpu counters - a direct call to | ||
1310 | * xfs_icsb_modify_counters is required for these. | ||
1311 | */ | ||
1312 | int | 1116 | int |
1313 | xfs_mod_incore_sb_batch( | 1117 | xfs_mod_fdblocks( |
1314 | struct xfs_mount *mp, | 1118 | struct xfs_mount *mp, |
1315 | xfs_mod_sb_t *msb, | 1119 | int64_t delta, |
1316 | uint nmsb, | 1120 | bool rsvd) |
1317 | int rsvd) | ||
1318 | { | 1121 | { |
1319 | xfs_mod_sb_t *msbp; | 1122 | int64_t lcounter; |
1320 | int error = 0; | 1123 | long long res_used; |
1124 | s32 batch; | ||
1125 | |||
1126 | if (delta > 0) { | ||
1127 | /* | ||
1128 | * If the reserve pool is depleted, put blocks back into it | ||
1129 | * first. Most of the time the pool is full. | ||
1130 | */ | ||
1131 | if (likely(mp->m_resblks == mp->m_resblks_avail)) { | ||
1132 | percpu_counter_add(&mp->m_fdblocks, delta); | ||
1133 | return 0; | ||
1134 | } | ||
1135 | |||
1136 | spin_lock(&mp->m_sb_lock); | ||
1137 | res_used = (long long)(mp->m_resblks - mp->m_resblks_avail); | ||
1138 | |||
1139 | if (res_used > delta) { | ||
1140 | mp->m_resblks_avail += delta; | ||
1141 | } else { | ||
1142 | delta -= res_used; | ||
1143 | mp->m_resblks_avail = mp->m_resblks; | ||
1144 | percpu_counter_add(&mp->m_fdblocks, delta); | ||
1145 | } | ||
1146 | spin_unlock(&mp->m_sb_lock); | ||
1147 | return 0; | ||
1148 | } | ||
1321 | 1149 | ||
1322 | /* | 1150 | /* |
1323 | * Loop through the array of mod structures and apply each individually. | 1151 | * Taking blocks away, need to be more accurate the closer we |
1324 | * If any fail, then back out all those which have already been applied. | 1152 | * are to zero. |
1325 | * Do all of this within the scope of the m_sb_lock so that all of the | 1153 | * |
1326 | * changes will be atomic. | 1154 | * batch size is set to a maximum of 1024 blocks - if we are |
1155 | * allocating of freeing extents larger than this then we aren't | ||
1156 | * going to be hammering the counter lock so a lock per update | ||
1157 | * is not a problem. | ||
1158 | * | ||
1159 | * If the counter has a value of less than 2 * max batch size, | ||
1160 | * then make everything serialise as we are real close to | ||
1161 | * ENOSPC. | ||
1162 | */ | ||
1163 | #define __BATCH 1024 | ||
1164 | if (percpu_counter_compare(&mp->m_fdblocks, 2 * __BATCH) < 0) | ||
1165 | batch = 1; | ||
1166 | else | ||
1167 | batch = __BATCH; | ||
1168 | #undef __BATCH | ||
1169 | |||
1170 | __percpu_counter_add(&mp->m_fdblocks, delta, batch); | ||
1171 | if (percpu_counter_compare(&mp->m_fdblocks, | ||
1172 | XFS_ALLOC_SET_ASIDE(mp)) >= 0) { | ||
1173 | /* we had space! */ | ||
1174 | return 0; | ||
1175 | } | ||
1176 | |||
1177 | /* | ||
1178 | * lock up the sb for dipping into reserves before releasing the space | ||
1179 | * that took us to ENOSPC. | ||
1327 | */ | 1180 | */ |
1328 | spin_lock(&mp->m_sb_lock); | 1181 | spin_lock(&mp->m_sb_lock); |
1329 | for (msbp = msb; msbp < (msb + nmsb); msbp++) { | 1182 | percpu_counter_add(&mp->m_fdblocks, -delta); |
1330 | ASSERT(msbp->msb_field < XFS_SBS_ICOUNT || | 1183 | if (!rsvd) |
1331 | msbp->msb_field > XFS_SBS_FDBLOCKS); | 1184 | goto fdblocks_enospc; |
1332 | 1185 | ||
1333 | error = xfs_mod_incore_sb_unlocked(mp, msbp->msb_field, | 1186 | lcounter = (long long)mp->m_resblks_avail + delta; |
1334 | msbp->msb_delta, rsvd); | 1187 | if (lcounter >= 0) { |
1335 | if (error) | 1188 | mp->m_resblks_avail = lcounter; |
1336 | goto unwind; | 1189 | spin_unlock(&mp->m_sb_lock); |
1190 | return 0; | ||
1337 | } | 1191 | } |
1192 | printk_once(KERN_WARNING | ||
1193 | "Filesystem \"%s\": reserve blocks depleted! " | ||
1194 | "Consider increasing reserve pool size.", | ||
1195 | mp->m_fsname); | ||
1196 | fdblocks_enospc: | ||
1338 | spin_unlock(&mp->m_sb_lock); | 1197 | spin_unlock(&mp->m_sb_lock); |
1339 | return 0; | 1198 | return -ENOSPC; |
1199 | } | ||
1340 | 1200 | ||
1341 | unwind: | 1201 | int |
1342 | while (--msbp >= msb) { | 1202 | xfs_mod_frextents( |
1343 | error = xfs_mod_incore_sb_unlocked(mp, msbp->msb_field, | 1203 | struct xfs_mount *mp, |
1344 | -msbp->msb_delta, rsvd); | 1204 | int64_t delta) |
1345 | ASSERT(error == 0); | 1205 | { |
1346 | } | 1206 | int64_t lcounter; |
1207 | int ret = 0; | ||
1208 | |||
1209 | spin_lock(&mp->m_sb_lock); | ||
1210 | lcounter = mp->m_sb.sb_frextents + delta; | ||
1211 | if (lcounter < 0) | ||
1212 | ret = -ENOSPC; | ||
1213 | else | ||
1214 | mp->m_sb.sb_frextents = lcounter; | ||
1347 | spin_unlock(&mp->m_sb_lock); | 1215 | spin_unlock(&mp->m_sb_lock); |
1348 | return error; | 1216 | return ret; |
1349 | } | 1217 | } |
1350 | 1218 | ||
1351 | /* | 1219 | /* |
@@ -1407,573 +1275,3 @@ xfs_dev_is_read_only( | |||
1407 | } | 1275 | } |
1408 | return 0; | 1276 | return 0; |
1409 | } | 1277 | } |
1410 | |||
1411 | #ifdef HAVE_PERCPU_SB | ||
1412 | /* | ||
1413 | * Per-cpu incore superblock counters | ||
1414 | * | ||
1415 | * Simple concept, difficult implementation | ||
1416 | * | ||
1417 | * Basically, replace the incore superblock counters with a distributed per cpu | ||
1418 | * counter for contended fields (e.g. free block count). | ||
1419 | * | ||
1420 | * Difficulties arise in that the incore sb is used for ENOSPC checking, and | ||
1421 | * hence needs to be accurately read when we are running low on space. Hence | ||
1422 | * there is a method to enable and disable the per-cpu counters based on how | ||
1423 | * much "stuff" is available in them. | ||
1424 | * | ||
1425 | * Basically, a counter is enabled if there is enough free resource to justify | ||
1426 | * running a per-cpu fast-path. If the per-cpu counter runs out (i.e. a local | ||
1427 | * ENOSPC), then we disable the counters to synchronise all callers and | ||
1428 | * re-distribute the available resources. | ||
1429 | * | ||
1430 | * If, once we redistributed the available resources, we still get a failure, | ||
1431 | * we disable the per-cpu counter and go through the slow path. | ||
1432 | * | ||
1433 | * The slow path is the current xfs_mod_incore_sb() function. This means that | ||
1434 | * when we disable a per-cpu counter, we need to drain its resources back to | ||
1435 | * the global superblock. We do this after disabling the counter to prevent | ||
1436 | * more threads from queueing up on the counter. | ||
1437 | * | ||
1438 | * Essentially, this means that we still need a lock in the fast path to enable | ||
1439 | * synchronisation between the global counters and the per-cpu counters. This | ||
1440 | * is not a problem because the lock will be local to a CPU almost all the time | ||
1441 | * and have little contention except when we get to ENOSPC conditions. | ||
1442 | * | ||
1443 | * Basically, this lock becomes a barrier that enables us to lock out the fast | ||
1444 | * path while we do things like enabling and disabling counters and | ||
1445 | * synchronising the counters. | ||
1446 | * | ||
1447 | * Locking rules: | ||
1448 | * | ||
1449 | * 1. m_sb_lock before picking up per-cpu locks | ||
1450 | * 2. per-cpu locks always picked up via for_each_online_cpu() order | ||
1451 | * 3. accurate counter sync requires m_sb_lock + per cpu locks | ||
1452 | * 4. modifying per-cpu counters requires holding per-cpu lock | ||
1453 | * 5. modifying global counters requires holding m_sb_lock | ||
1454 | * 6. enabling or disabling a counter requires holding the m_sb_lock | ||
1455 | * and _none_ of the per-cpu locks. | ||
1456 | * | ||
1457 | * Disabled counters are only ever re-enabled by a balance operation | ||
1458 | * that results in more free resources per CPU than a given threshold. | ||
1459 | * To ensure counters don't remain disabled, they are rebalanced when | ||
1460 | * the global resource goes above a higher threshold (i.e. some hysteresis | ||
1461 | * is present to prevent thrashing). | ||
1462 | */ | ||
1463 | |||
1464 | #ifdef CONFIG_HOTPLUG_CPU | ||
1465 | /* | ||
1466 | * hot-plug CPU notifier support. | ||
1467 | * | ||
1468 | * We need a notifier per filesystem as we need to be able to identify | ||
1469 | * the filesystem to balance the counters out. This is achieved by | ||
1470 | * having a notifier block embedded in the xfs_mount_t and doing pointer | ||
1471 | * magic to get the mount pointer from the notifier block address. | ||
1472 | */ | ||
1473 | STATIC int | ||
1474 | xfs_icsb_cpu_notify( | ||
1475 | struct notifier_block *nfb, | ||
1476 | unsigned long action, | ||
1477 | void *hcpu) | ||
1478 | { | ||
1479 | xfs_icsb_cnts_t *cntp; | ||
1480 | xfs_mount_t *mp; | ||
1481 | |||
1482 | mp = (xfs_mount_t *)container_of(nfb, xfs_mount_t, m_icsb_notifier); | ||
1483 | cntp = (xfs_icsb_cnts_t *) | ||
1484 | per_cpu_ptr(mp->m_sb_cnts, (unsigned long)hcpu); | ||
1485 | switch (action) { | ||
1486 | case CPU_UP_PREPARE: | ||
1487 | case CPU_UP_PREPARE_FROZEN: | ||
1488 | /* Easy Case - initialize the area and locks, and | ||
1489 | * then rebalance when online does everything else for us. */ | ||
1490 | memset(cntp, 0, sizeof(xfs_icsb_cnts_t)); | ||
1491 | break; | ||
1492 | case CPU_ONLINE: | ||
1493 | case CPU_ONLINE_FROZEN: | ||
1494 | xfs_icsb_lock(mp); | ||
1495 | xfs_icsb_balance_counter(mp, XFS_SBS_ICOUNT, 0); | ||
1496 | xfs_icsb_balance_counter(mp, XFS_SBS_IFREE, 0); | ||
1497 | xfs_icsb_balance_counter(mp, XFS_SBS_FDBLOCKS, 0); | ||
1498 | xfs_icsb_unlock(mp); | ||
1499 | break; | ||
1500 | case CPU_DEAD: | ||
1501 | case CPU_DEAD_FROZEN: | ||
1502 | /* Disable all the counters, then fold the dead cpu's | ||
1503 | * count into the total on the global superblock and | ||
1504 | * re-enable the counters. */ | ||
1505 | xfs_icsb_lock(mp); | ||
1506 | spin_lock(&mp->m_sb_lock); | ||
1507 | xfs_icsb_disable_counter(mp, XFS_SBS_ICOUNT); | ||
1508 | xfs_icsb_disable_counter(mp, XFS_SBS_IFREE); | ||
1509 | xfs_icsb_disable_counter(mp, XFS_SBS_FDBLOCKS); | ||
1510 | |||
1511 | mp->m_sb.sb_icount += cntp->icsb_icount; | ||
1512 | mp->m_sb.sb_ifree += cntp->icsb_ifree; | ||
1513 | mp->m_sb.sb_fdblocks += cntp->icsb_fdblocks; | ||
1514 | |||
1515 | memset(cntp, 0, sizeof(xfs_icsb_cnts_t)); | ||
1516 | |||
1517 | xfs_icsb_balance_counter_locked(mp, XFS_SBS_ICOUNT, 0); | ||
1518 | xfs_icsb_balance_counter_locked(mp, XFS_SBS_IFREE, 0); | ||
1519 | xfs_icsb_balance_counter_locked(mp, XFS_SBS_FDBLOCKS, 0); | ||
1520 | spin_unlock(&mp->m_sb_lock); | ||
1521 | xfs_icsb_unlock(mp); | ||
1522 | break; | ||
1523 | } | ||
1524 | |||
1525 | return NOTIFY_OK; | ||
1526 | } | ||
1527 | #endif /* CONFIG_HOTPLUG_CPU */ | ||
1528 | |||
1529 | int | ||
1530 | xfs_icsb_init_counters( | ||
1531 | xfs_mount_t *mp) | ||
1532 | { | ||
1533 | xfs_icsb_cnts_t *cntp; | ||
1534 | int i; | ||
1535 | |||
1536 | mp->m_sb_cnts = alloc_percpu(xfs_icsb_cnts_t); | ||
1537 | if (mp->m_sb_cnts == NULL) | ||
1538 | return -ENOMEM; | ||
1539 | |||
1540 | for_each_online_cpu(i) { | ||
1541 | cntp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, i); | ||
1542 | memset(cntp, 0, sizeof(xfs_icsb_cnts_t)); | ||
1543 | } | ||
1544 | |||
1545 | mutex_init(&mp->m_icsb_mutex); | ||
1546 | |||
1547 | /* | ||
1548 | * start with all counters disabled so that the | ||
1549 | * initial balance kicks us off correctly | ||
1550 | */ | ||
1551 | mp->m_icsb_counters = -1; | ||
1552 | |||
1553 | #ifdef CONFIG_HOTPLUG_CPU | ||
1554 | mp->m_icsb_notifier.notifier_call = xfs_icsb_cpu_notify; | ||
1555 | mp->m_icsb_notifier.priority = 0; | ||
1556 | register_hotcpu_notifier(&mp->m_icsb_notifier); | ||
1557 | #endif /* CONFIG_HOTPLUG_CPU */ | ||
1558 | |||
1559 | return 0; | ||
1560 | } | ||
1561 | |||
1562 | void | ||
1563 | xfs_icsb_reinit_counters( | ||
1564 | xfs_mount_t *mp) | ||
1565 | { | ||
1566 | xfs_icsb_lock(mp); | ||
1567 | /* | ||
1568 | * start with all counters disabled so that the | ||
1569 | * initial balance kicks us off correctly | ||
1570 | */ | ||
1571 | mp->m_icsb_counters = -1; | ||
1572 | xfs_icsb_balance_counter(mp, XFS_SBS_ICOUNT, 0); | ||
1573 | xfs_icsb_balance_counter(mp, XFS_SBS_IFREE, 0); | ||
1574 | xfs_icsb_balance_counter(mp, XFS_SBS_FDBLOCKS, 0); | ||
1575 | xfs_icsb_unlock(mp); | ||
1576 | } | ||
1577 | |||
1578 | void | ||
1579 | xfs_icsb_destroy_counters( | ||
1580 | xfs_mount_t *mp) | ||
1581 | { | ||
1582 | if (mp->m_sb_cnts) { | ||
1583 | unregister_hotcpu_notifier(&mp->m_icsb_notifier); | ||
1584 | free_percpu(mp->m_sb_cnts); | ||
1585 | } | ||
1586 | mutex_destroy(&mp->m_icsb_mutex); | ||
1587 | } | ||
1588 | |||
1589 | STATIC void | ||
1590 | xfs_icsb_lock_cntr( | ||
1591 | xfs_icsb_cnts_t *icsbp) | ||
1592 | { | ||
1593 | while (test_and_set_bit(XFS_ICSB_FLAG_LOCK, &icsbp->icsb_flags)) { | ||
1594 | ndelay(1000); | ||
1595 | } | ||
1596 | } | ||
1597 | |||
1598 | STATIC void | ||
1599 | xfs_icsb_unlock_cntr( | ||
1600 | xfs_icsb_cnts_t *icsbp) | ||
1601 | { | ||
1602 | clear_bit(XFS_ICSB_FLAG_LOCK, &icsbp->icsb_flags); | ||
1603 | } | ||
1604 | |||
1605 | |||
1606 | STATIC void | ||
1607 | xfs_icsb_lock_all_counters( | ||
1608 | xfs_mount_t *mp) | ||
1609 | { | ||
1610 | xfs_icsb_cnts_t *cntp; | ||
1611 | int i; | ||
1612 | |||
1613 | for_each_online_cpu(i) { | ||
1614 | cntp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, i); | ||
1615 | xfs_icsb_lock_cntr(cntp); | ||
1616 | } | ||
1617 | } | ||
1618 | |||
1619 | STATIC void | ||
1620 | xfs_icsb_unlock_all_counters( | ||
1621 | xfs_mount_t *mp) | ||
1622 | { | ||
1623 | xfs_icsb_cnts_t *cntp; | ||
1624 | int i; | ||
1625 | |||
1626 | for_each_online_cpu(i) { | ||
1627 | cntp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, i); | ||
1628 | xfs_icsb_unlock_cntr(cntp); | ||
1629 | } | ||
1630 | } | ||
1631 | |||
1632 | STATIC void | ||
1633 | xfs_icsb_count( | ||
1634 | xfs_mount_t *mp, | ||
1635 | xfs_icsb_cnts_t *cnt, | ||
1636 | int flags) | ||
1637 | { | ||
1638 | xfs_icsb_cnts_t *cntp; | ||
1639 | int i; | ||
1640 | |||
1641 | memset(cnt, 0, sizeof(xfs_icsb_cnts_t)); | ||
1642 | |||
1643 | if (!(flags & XFS_ICSB_LAZY_COUNT)) | ||
1644 | xfs_icsb_lock_all_counters(mp); | ||
1645 | |||
1646 | for_each_online_cpu(i) { | ||
1647 | cntp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, i); | ||
1648 | cnt->icsb_icount += cntp->icsb_icount; | ||
1649 | cnt->icsb_ifree += cntp->icsb_ifree; | ||
1650 | cnt->icsb_fdblocks += cntp->icsb_fdblocks; | ||
1651 | } | ||
1652 | |||
1653 | if (!(flags & XFS_ICSB_LAZY_COUNT)) | ||
1654 | xfs_icsb_unlock_all_counters(mp); | ||
1655 | } | ||
1656 | |||
1657 | STATIC int | ||
1658 | xfs_icsb_counter_disabled( | ||
1659 | xfs_mount_t *mp, | ||
1660 | xfs_sb_field_t field) | ||
1661 | { | ||
1662 | ASSERT((field >= XFS_SBS_ICOUNT) && (field <= XFS_SBS_FDBLOCKS)); | ||
1663 | return test_bit(field, &mp->m_icsb_counters); | ||
1664 | } | ||
1665 | |||
1666 | STATIC void | ||
1667 | xfs_icsb_disable_counter( | ||
1668 | xfs_mount_t *mp, | ||
1669 | xfs_sb_field_t field) | ||
1670 | { | ||
1671 | xfs_icsb_cnts_t cnt; | ||
1672 | |||
1673 | ASSERT((field >= XFS_SBS_ICOUNT) && (field <= XFS_SBS_FDBLOCKS)); | ||
1674 | |||
1675 | /* | ||
1676 | * If we are already disabled, then there is nothing to do | ||
1677 | * here. We check before locking all the counters to avoid | ||
1678 | * the expensive lock operation when being called in the | ||
1679 | * slow path and the counter is already disabled. This is | ||
1680 | * safe because the only time we set or clear this state is under | ||
1681 | * the m_icsb_mutex. | ||
1682 | */ | ||
1683 | if (xfs_icsb_counter_disabled(mp, field)) | ||
1684 | return; | ||
1685 | |||
1686 | xfs_icsb_lock_all_counters(mp); | ||
1687 | if (!test_and_set_bit(field, &mp->m_icsb_counters)) { | ||
1688 | /* drain back to superblock */ | ||
1689 | |||
1690 | xfs_icsb_count(mp, &cnt, XFS_ICSB_LAZY_COUNT); | ||
1691 | switch(field) { | ||
1692 | case XFS_SBS_ICOUNT: | ||
1693 | mp->m_sb.sb_icount = cnt.icsb_icount; | ||
1694 | break; | ||
1695 | case XFS_SBS_IFREE: | ||
1696 | mp->m_sb.sb_ifree = cnt.icsb_ifree; | ||
1697 | break; | ||
1698 | case XFS_SBS_FDBLOCKS: | ||
1699 | mp->m_sb.sb_fdblocks = cnt.icsb_fdblocks; | ||
1700 | break; | ||
1701 | default: | ||
1702 | BUG(); | ||
1703 | } | ||
1704 | } | ||
1705 | |||
1706 | xfs_icsb_unlock_all_counters(mp); | ||
1707 | } | ||
1708 | |||
1709 | STATIC void | ||
1710 | xfs_icsb_enable_counter( | ||
1711 | xfs_mount_t *mp, | ||
1712 | xfs_sb_field_t field, | ||
1713 | uint64_t count, | ||
1714 | uint64_t resid) | ||
1715 | { | ||
1716 | xfs_icsb_cnts_t *cntp; | ||
1717 | int i; | ||
1718 | |||
1719 | ASSERT((field >= XFS_SBS_ICOUNT) && (field <= XFS_SBS_FDBLOCKS)); | ||
1720 | |||
1721 | xfs_icsb_lock_all_counters(mp); | ||
1722 | for_each_online_cpu(i) { | ||
1723 | cntp = per_cpu_ptr(mp->m_sb_cnts, i); | ||
1724 | switch (field) { | ||
1725 | case XFS_SBS_ICOUNT: | ||
1726 | cntp->icsb_icount = count + resid; | ||
1727 | break; | ||
1728 | case XFS_SBS_IFREE: | ||
1729 | cntp->icsb_ifree = count + resid; | ||
1730 | break; | ||
1731 | case XFS_SBS_FDBLOCKS: | ||
1732 | cntp->icsb_fdblocks = count + resid; | ||
1733 | break; | ||
1734 | default: | ||
1735 | BUG(); | ||
1736 | break; | ||
1737 | } | ||
1738 | resid = 0; | ||
1739 | } | ||
1740 | clear_bit(field, &mp->m_icsb_counters); | ||
1741 | xfs_icsb_unlock_all_counters(mp); | ||
1742 | } | ||
1743 | |||
1744 | void | ||
1745 | xfs_icsb_sync_counters_locked( | ||
1746 | xfs_mount_t *mp, | ||
1747 | int flags) | ||
1748 | { | ||
1749 | xfs_icsb_cnts_t cnt; | ||
1750 | |||
1751 | xfs_icsb_count(mp, &cnt, flags); | ||
1752 | |||
1753 | if (!xfs_icsb_counter_disabled(mp, XFS_SBS_ICOUNT)) | ||
1754 | mp->m_sb.sb_icount = cnt.icsb_icount; | ||
1755 | if (!xfs_icsb_counter_disabled(mp, XFS_SBS_IFREE)) | ||
1756 | mp->m_sb.sb_ifree = cnt.icsb_ifree; | ||
1757 | if (!xfs_icsb_counter_disabled(mp, XFS_SBS_FDBLOCKS)) | ||
1758 | mp->m_sb.sb_fdblocks = cnt.icsb_fdblocks; | ||
1759 | } | ||
1760 | |||
1761 | /* | ||
1762 | * Accurate update of per-cpu counters to incore superblock | ||
1763 | */ | ||
1764 | void | ||
1765 | xfs_icsb_sync_counters( | ||
1766 | xfs_mount_t *mp, | ||
1767 | int flags) | ||
1768 | { | ||
1769 | spin_lock(&mp->m_sb_lock); | ||
1770 | xfs_icsb_sync_counters_locked(mp, flags); | ||
1771 | spin_unlock(&mp->m_sb_lock); | ||
1772 | } | ||
1773 | |||
1774 | /* | ||
1775 | * Balance and enable/disable counters as necessary. | ||
1776 | * | ||
1777 | * Thresholds for re-enabling counters are somewhat magic. inode counts are | ||
1778 | * chosen to be the same number as single on disk allocation chunk per CPU, and | ||
1779 | * free blocks is something far enough zero that we aren't going thrash when we | ||
1780 | * get near ENOSPC. We also need to supply a minimum we require per cpu to | ||
1781 | * prevent looping endlessly when xfs_alloc_space asks for more than will | ||
1782 | * be distributed to a single CPU but each CPU has enough blocks to be | ||
1783 | * reenabled. | ||
1784 | * | ||
1785 | * Note that we can be called when counters are already disabled. | ||
1786 | * xfs_icsb_disable_counter() optimises the counter locking in this case to | ||
1787 | * prevent locking every per-cpu counter needlessly. | ||
1788 | */ | ||
1789 | |||
1790 | #define XFS_ICSB_INO_CNTR_REENABLE (uint64_t)64 | ||
1791 | #define XFS_ICSB_FDBLK_CNTR_REENABLE(mp) \ | ||
1792 | (uint64_t)(512 + XFS_ALLOC_SET_ASIDE(mp)) | ||
1793 | STATIC void | ||
1794 | xfs_icsb_balance_counter_locked( | ||
1795 | xfs_mount_t *mp, | ||
1796 | xfs_sb_field_t field, | ||
1797 | int min_per_cpu) | ||
1798 | { | ||
1799 | uint64_t count, resid; | ||
1800 | int weight = num_online_cpus(); | ||
1801 | uint64_t min = (uint64_t)min_per_cpu; | ||
1802 | |||
1803 | /* disable counter and sync counter */ | ||
1804 | xfs_icsb_disable_counter(mp, field); | ||
1805 | |||
1806 | /* update counters - first CPU gets residual*/ | ||
1807 | switch (field) { | ||
1808 | case XFS_SBS_ICOUNT: | ||
1809 | count = mp->m_sb.sb_icount; | ||
1810 | resid = do_div(count, weight); | ||
1811 | if (count < max(min, XFS_ICSB_INO_CNTR_REENABLE)) | ||
1812 | return; | ||
1813 | break; | ||
1814 | case XFS_SBS_IFREE: | ||
1815 | count = mp->m_sb.sb_ifree; | ||
1816 | resid = do_div(count, weight); | ||
1817 | if (count < max(min, XFS_ICSB_INO_CNTR_REENABLE)) | ||
1818 | return; | ||
1819 | break; | ||
1820 | case XFS_SBS_FDBLOCKS: | ||
1821 | count = mp->m_sb.sb_fdblocks; | ||
1822 | resid = do_div(count, weight); | ||
1823 | if (count < max(min, XFS_ICSB_FDBLK_CNTR_REENABLE(mp))) | ||
1824 | return; | ||
1825 | break; | ||
1826 | default: | ||
1827 | BUG(); | ||
1828 | count = resid = 0; /* quiet, gcc */ | ||
1829 | break; | ||
1830 | } | ||
1831 | |||
1832 | xfs_icsb_enable_counter(mp, field, count, resid); | ||
1833 | } | ||
1834 | |||
1835 | STATIC void | ||
1836 | xfs_icsb_balance_counter( | ||
1837 | xfs_mount_t *mp, | ||
1838 | xfs_sb_field_t fields, | ||
1839 | int min_per_cpu) | ||
1840 | { | ||
1841 | spin_lock(&mp->m_sb_lock); | ||
1842 | xfs_icsb_balance_counter_locked(mp, fields, min_per_cpu); | ||
1843 | spin_unlock(&mp->m_sb_lock); | ||
1844 | } | ||
1845 | |||
1846 | int | ||
1847 | xfs_icsb_modify_counters( | ||
1848 | xfs_mount_t *mp, | ||
1849 | xfs_sb_field_t field, | ||
1850 | int64_t delta, | ||
1851 | int rsvd) | ||
1852 | { | ||
1853 | xfs_icsb_cnts_t *icsbp; | ||
1854 | long long lcounter; /* long counter for 64 bit fields */ | ||
1855 | int ret = 0; | ||
1856 | |||
1857 | might_sleep(); | ||
1858 | again: | ||
1859 | preempt_disable(); | ||
1860 | icsbp = this_cpu_ptr(mp->m_sb_cnts); | ||
1861 | |||
1862 | /* | ||
1863 | * if the counter is disabled, go to slow path | ||
1864 | */ | ||
1865 | if (unlikely(xfs_icsb_counter_disabled(mp, field))) | ||
1866 | goto slow_path; | ||
1867 | xfs_icsb_lock_cntr(icsbp); | ||
1868 | if (unlikely(xfs_icsb_counter_disabled(mp, field))) { | ||
1869 | xfs_icsb_unlock_cntr(icsbp); | ||
1870 | goto slow_path; | ||
1871 | } | ||
1872 | |||
1873 | switch (field) { | ||
1874 | case XFS_SBS_ICOUNT: | ||
1875 | lcounter = icsbp->icsb_icount; | ||
1876 | lcounter += delta; | ||
1877 | if (unlikely(lcounter < 0)) | ||
1878 | goto balance_counter; | ||
1879 | icsbp->icsb_icount = lcounter; | ||
1880 | break; | ||
1881 | |||
1882 | case XFS_SBS_IFREE: | ||
1883 | lcounter = icsbp->icsb_ifree; | ||
1884 | lcounter += delta; | ||
1885 | if (unlikely(lcounter < 0)) | ||
1886 | goto balance_counter; | ||
1887 | icsbp->icsb_ifree = lcounter; | ||
1888 | break; | ||
1889 | |||
1890 | case XFS_SBS_FDBLOCKS: | ||
1891 | BUG_ON((mp->m_resblks - mp->m_resblks_avail) != 0); | ||
1892 | |||
1893 | lcounter = icsbp->icsb_fdblocks - XFS_ALLOC_SET_ASIDE(mp); | ||
1894 | lcounter += delta; | ||
1895 | if (unlikely(lcounter < 0)) | ||
1896 | goto balance_counter; | ||
1897 | icsbp->icsb_fdblocks = lcounter + XFS_ALLOC_SET_ASIDE(mp); | ||
1898 | break; | ||
1899 | default: | ||
1900 | BUG(); | ||
1901 | break; | ||
1902 | } | ||
1903 | xfs_icsb_unlock_cntr(icsbp); | ||
1904 | preempt_enable(); | ||
1905 | return 0; | ||
1906 | |||
1907 | slow_path: | ||
1908 | preempt_enable(); | ||
1909 | |||
1910 | /* | ||
1911 | * serialise with a mutex so we don't burn lots of cpu on | ||
1912 | * the superblock lock. We still need to hold the superblock | ||
1913 | * lock, however, when we modify the global structures. | ||
1914 | */ | ||
1915 | xfs_icsb_lock(mp); | ||
1916 | |||
1917 | /* | ||
1918 | * Now running atomically. | ||
1919 | * | ||
1920 | * If the counter is enabled, someone has beaten us to rebalancing. | ||
1921 | * Drop the lock and try again in the fast path.... | ||
1922 | */ | ||
1923 | if (!(xfs_icsb_counter_disabled(mp, field))) { | ||
1924 | xfs_icsb_unlock(mp); | ||
1925 | goto again; | ||
1926 | } | ||
1927 | |||
1928 | /* | ||
1929 | * The counter is currently disabled. Because we are | ||
1930 | * running atomically here, we know a rebalance cannot | ||
1931 | * be in progress. Hence we can go straight to operating | ||
1932 | * on the global superblock. We do not call xfs_mod_incore_sb() | ||
1933 | * here even though we need to get the m_sb_lock. Doing so | ||
1934 | * will cause us to re-enter this function and deadlock. | ||
1935 | * Hence we get the m_sb_lock ourselves and then call | ||
1936 | * xfs_mod_incore_sb_unlocked() as the unlocked path operates | ||
1937 | * directly on the global counters. | ||
1938 | */ | ||
1939 | spin_lock(&mp->m_sb_lock); | ||
1940 | ret = xfs_mod_incore_sb_unlocked(mp, field, delta, rsvd); | ||
1941 | spin_unlock(&mp->m_sb_lock); | ||
1942 | |||
1943 | /* | ||
1944 | * Now that we've modified the global superblock, we | ||
1945 | * may be able to re-enable the distributed counters | ||
1946 | * (e.g. lots of space just got freed). After that | ||
1947 | * we are done. | ||
1948 | */ | ||
1949 | if (ret != -ENOSPC) | ||
1950 | xfs_icsb_balance_counter(mp, field, 0); | ||
1951 | xfs_icsb_unlock(mp); | ||
1952 | return ret; | ||
1953 | |||
1954 | balance_counter: | ||
1955 | xfs_icsb_unlock_cntr(icsbp); | ||
1956 | preempt_enable(); | ||
1957 | |||
1958 | /* | ||
1959 | * We may have multiple threads here if multiple per-cpu | ||
1960 | * counters run dry at the same time. This will mean we can | ||
1961 | * do more balances than strictly necessary but it is not | ||
1962 | * the common slowpath case. | ||
1963 | */ | ||
1964 | xfs_icsb_lock(mp); | ||
1965 | |||
1966 | /* | ||
1967 | * running atomically. | ||
1968 | * | ||
1969 | * This will leave the counter in the correct state for future | ||
1970 | * accesses. After the rebalance, we simply try again and our retry | ||
1971 | * will either succeed through the fast path or slow path without | ||
1972 | * another balance operation being required. | ||
1973 | */ | ||
1974 | xfs_icsb_balance_counter(mp, field, delta); | ||
1975 | xfs_icsb_unlock(mp); | ||
1976 | goto again; | ||
1977 | } | ||
1978 | |||
1979 | #endif | ||
diff --git a/fs/xfs/xfs_mount.h b/fs/xfs/xfs_mount.h index 0d8abd6364d9..8c995a2ccb6f 100644 --- a/fs/xfs/xfs_mount.h +++ b/fs/xfs/xfs_mount.h | |||
@@ -18,8 +18,6 @@ | |||
18 | #ifndef __XFS_MOUNT_H__ | 18 | #ifndef __XFS_MOUNT_H__ |
19 | #define __XFS_MOUNT_H__ | 19 | #define __XFS_MOUNT_H__ |
20 | 20 | ||
21 | #ifdef __KERNEL__ | ||
22 | |||
23 | struct xlog; | 21 | struct xlog; |
24 | struct xfs_inode; | 22 | struct xfs_inode; |
25 | struct xfs_mru_cache; | 23 | struct xfs_mru_cache; |
@@ -29,44 +27,6 @@ struct xfs_quotainfo; | |||
29 | struct xfs_dir_ops; | 27 | struct xfs_dir_ops; |
30 | struct xfs_da_geometry; | 28 | struct xfs_da_geometry; |
31 | 29 | ||
32 | #ifdef HAVE_PERCPU_SB | ||
33 | |||
34 | /* | ||
35 | * Valid per-cpu incore superblock counters. Note that if you add new counters, | ||
36 | * you may need to define new counter disabled bit field descriptors as there | ||
37 | * are more possible fields in the superblock that can fit in a bitfield on a | ||
38 | * 32 bit platform. The XFS_SBS_* values for the current current counters just | ||
39 | * fit. | ||
40 | */ | ||
41 | typedef struct xfs_icsb_cnts { | ||
42 | uint64_t icsb_fdblocks; | ||
43 | uint64_t icsb_ifree; | ||
44 | uint64_t icsb_icount; | ||
45 | unsigned long icsb_flags; | ||
46 | } xfs_icsb_cnts_t; | ||
47 | |||
48 | #define XFS_ICSB_FLAG_LOCK (1 << 0) /* counter lock bit */ | ||
49 | |||
50 | #define XFS_ICSB_LAZY_COUNT (1 << 1) /* accuracy not needed */ | ||
51 | |||
52 | extern int xfs_icsb_init_counters(struct xfs_mount *); | ||
53 | extern void xfs_icsb_reinit_counters(struct xfs_mount *); | ||
54 | extern void xfs_icsb_destroy_counters(struct xfs_mount *); | ||
55 | extern void xfs_icsb_sync_counters(struct xfs_mount *, int); | ||
56 | extern void xfs_icsb_sync_counters_locked(struct xfs_mount *, int); | ||
57 | extern int xfs_icsb_modify_counters(struct xfs_mount *, xfs_sb_field_t, | ||
58 | int64_t, int); | ||
59 | |||
60 | #else | ||
61 | #define xfs_icsb_init_counters(mp) (0) | ||
62 | #define xfs_icsb_destroy_counters(mp) do { } while (0) | ||
63 | #define xfs_icsb_reinit_counters(mp) do { } while (0) | ||
64 | #define xfs_icsb_sync_counters(mp, flags) do { } while (0) | ||
65 | #define xfs_icsb_sync_counters_locked(mp, flags) do { } while (0) | ||
66 | #define xfs_icsb_modify_counters(mp, field, delta, rsvd) \ | ||
67 | xfs_mod_incore_sb(mp, field, delta, rsvd) | ||
68 | #endif | ||
69 | |||
70 | /* dynamic preallocation free space thresholds, 5% down to 1% */ | 30 | /* dynamic preallocation free space thresholds, 5% down to 1% */ |
71 | enum { | 31 | enum { |
72 | XFS_LOWSP_1_PCNT = 0, | 32 | XFS_LOWSP_1_PCNT = 0, |
@@ -81,8 +41,13 @@ typedef struct xfs_mount { | |||
81 | struct super_block *m_super; | 41 | struct super_block *m_super; |
82 | xfs_tid_t m_tid; /* next unused tid for fs */ | 42 | xfs_tid_t m_tid; /* next unused tid for fs */ |
83 | struct xfs_ail *m_ail; /* fs active log item list */ | 43 | struct xfs_ail *m_ail; /* fs active log item list */ |
84 | xfs_sb_t m_sb; /* copy of fs superblock */ | 44 | |
45 | struct xfs_sb m_sb; /* copy of fs superblock */ | ||
85 | spinlock_t m_sb_lock; /* sb counter lock */ | 46 | spinlock_t m_sb_lock; /* sb counter lock */ |
47 | struct percpu_counter m_icount; /* allocated inodes counter */ | ||
48 | struct percpu_counter m_ifree; /* free inodes counter */ | ||
49 | struct percpu_counter m_fdblocks; /* free block counter */ | ||
50 | |||
86 | struct xfs_buf *m_sb_bp; /* buffer for superblock */ | 51 | struct xfs_buf *m_sb_bp; /* buffer for superblock */ |
87 | char *m_fsname; /* filesystem name */ | 52 | char *m_fsname; /* filesystem name */ |
88 | int m_fsname_len; /* strlen of fs name */ | 53 | int m_fsname_len; /* strlen of fs name */ |
@@ -152,12 +117,6 @@ typedef struct xfs_mount { | |||
152 | const struct xfs_dir_ops *m_nondir_inode_ops; /* !dir inode ops */ | 117 | const struct xfs_dir_ops *m_nondir_inode_ops; /* !dir inode ops */ |
153 | uint m_chsize; /* size of next field */ | 118 | uint m_chsize; /* size of next field */ |
154 | atomic_t m_active_trans; /* number trans frozen */ | 119 | atomic_t m_active_trans; /* number trans frozen */ |
155 | #ifdef HAVE_PERCPU_SB | ||
156 | xfs_icsb_cnts_t __percpu *m_sb_cnts; /* per-cpu superblock counters */ | ||
157 | unsigned long m_icsb_counters; /* disabled per-cpu counters */ | ||
158 | struct notifier_block m_icsb_notifier; /* hotplug cpu notifier */ | ||
159 | struct mutex m_icsb_mutex; /* balancer sync lock */ | ||
160 | #endif | ||
161 | struct xfs_mru_cache *m_filestream; /* per-mount filestream data */ | 120 | struct xfs_mru_cache *m_filestream; /* per-mount filestream data */ |
162 | struct delayed_work m_reclaim_work; /* background inode reclaim */ | 121 | struct delayed_work m_reclaim_work; /* background inode reclaim */ |
163 | struct delayed_work m_eofblocks_work; /* background eof blocks | 122 | struct delayed_work m_eofblocks_work; /* background eof blocks |
@@ -301,35 +260,6 @@ xfs_daddr_to_agbno(struct xfs_mount *mp, xfs_daddr_t d) | |||
301 | } | 260 | } |
302 | 261 | ||
303 | /* | 262 | /* |
304 | * Per-cpu superblock locking functions | ||
305 | */ | ||
306 | #ifdef HAVE_PERCPU_SB | ||
307 | static inline void | ||
308 | xfs_icsb_lock(xfs_mount_t *mp) | ||
309 | { | ||
310 | mutex_lock(&mp->m_icsb_mutex); | ||
311 | } | ||
312 | |||
313 | static inline void | ||
314 | xfs_icsb_unlock(xfs_mount_t *mp) | ||
315 | { | ||
316 | mutex_unlock(&mp->m_icsb_mutex); | ||
317 | } | ||
318 | #else | ||
319 | #define xfs_icsb_lock(mp) | ||
320 | #define xfs_icsb_unlock(mp) | ||
321 | #endif | ||
322 | |||
323 | /* | ||
324 | * This structure is for use by the xfs_mod_incore_sb_batch() routine. | ||
325 | * xfs_growfs can specify a few fields which are more than int limit | ||
326 | */ | ||
327 | typedef struct xfs_mod_sb { | ||
328 | xfs_sb_field_t msb_field; /* Field to modify, see below */ | ||
329 | int64_t msb_delta; /* Change to make to specified field */ | ||
330 | } xfs_mod_sb_t; | ||
331 | |||
332 | /* | ||
333 | * Per-ag incore structure, copies of information in agf and agi, to improve the | 263 | * Per-ag incore structure, copies of information in agf and agi, to improve the |
334 | * performance of allocation group selection. | 264 | * performance of allocation group selection. |
335 | */ | 265 | */ |
@@ -383,11 +313,14 @@ extern __uint64_t xfs_default_resblks(xfs_mount_t *mp); | |||
383 | extern int xfs_mountfs(xfs_mount_t *mp); | 313 | extern int xfs_mountfs(xfs_mount_t *mp); |
384 | extern int xfs_initialize_perag(xfs_mount_t *mp, xfs_agnumber_t agcount, | 314 | extern int xfs_initialize_perag(xfs_mount_t *mp, xfs_agnumber_t agcount, |
385 | xfs_agnumber_t *maxagi); | 315 | xfs_agnumber_t *maxagi); |
386 | |||
387 | extern void xfs_unmountfs(xfs_mount_t *); | 316 | extern void xfs_unmountfs(xfs_mount_t *); |
388 | extern int xfs_mod_incore_sb(xfs_mount_t *, xfs_sb_field_t, int64_t, int); | 317 | |
389 | extern int xfs_mod_incore_sb_batch(xfs_mount_t *, xfs_mod_sb_t *, | 318 | extern int xfs_mod_icount(struct xfs_mount *mp, int64_t delta); |
390 | uint, int); | 319 | extern int xfs_mod_ifree(struct xfs_mount *mp, int64_t delta); |
320 | extern int xfs_mod_fdblocks(struct xfs_mount *mp, int64_t delta, | ||
321 | bool reserved); | ||
322 | extern int xfs_mod_frextents(struct xfs_mount *mp, int64_t delta); | ||
323 | |||
391 | extern int xfs_mount_log_sb(xfs_mount_t *); | 324 | extern int xfs_mount_log_sb(xfs_mount_t *); |
392 | extern struct xfs_buf *xfs_getsb(xfs_mount_t *, int); | 325 | extern struct xfs_buf *xfs_getsb(xfs_mount_t *, int); |
393 | extern int xfs_readsb(xfs_mount_t *, int); | 326 | extern int xfs_readsb(xfs_mount_t *, int); |
@@ -399,6 +332,4 @@ extern int xfs_dev_is_read_only(struct xfs_mount *, char *); | |||
399 | 332 | ||
400 | extern void xfs_set_low_space_thresholds(struct xfs_mount *); | 333 | extern void xfs_set_low_space_thresholds(struct xfs_mount *); |
401 | 334 | ||
402 | #endif /* __KERNEL__ */ | ||
403 | |||
404 | #endif /* __XFS_MOUNT_H__ */ | 335 | #endif /* __XFS_MOUNT_H__ */ |
diff --git a/fs/xfs/xfs_super.c b/fs/xfs/xfs_super.c index 58453e3255f8..f32ad64c4d05 100644 --- a/fs/xfs/xfs_super.c +++ b/fs/xfs/xfs_super.c | |||
@@ -1013,24 +1013,6 @@ xfs_free_fsname( | |||
1013 | kfree(mp->m_logname); | 1013 | kfree(mp->m_logname); |
1014 | } | 1014 | } |
1015 | 1015 | ||
1016 | STATIC void | ||
1017 | xfs_fs_put_super( | ||
1018 | struct super_block *sb) | ||
1019 | { | ||
1020 | struct xfs_mount *mp = XFS_M(sb); | ||
1021 | |||
1022 | xfs_notice(mp, "Unmounting Filesystem"); | ||
1023 | xfs_filestream_unmount(mp); | ||
1024 | xfs_unmountfs(mp); | ||
1025 | |||
1026 | xfs_freesb(mp); | ||
1027 | xfs_icsb_destroy_counters(mp); | ||
1028 | xfs_destroy_mount_workqueues(mp); | ||
1029 | xfs_close_devices(mp); | ||
1030 | xfs_free_fsname(mp); | ||
1031 | kfree(mp); | ||
1032 | } | ||
1033 | |||
1034 | STATIC int | 1016 | STATIC int |
1035 | xfs_fs_sync_fs( | 1017 | xfs_fs_sync_fs( |
1036 | struct super_block *sb, | 1018 | struct super_block *sb, |
@@ -1066,6 +1048,9 @@ xfs_fs_statfs( | |||
1066 | xfs_sb_t *sbp = &mp->m_sb; | 1048 | xfs_sb_t *sbp = &mp->m_sb; |
1067 | struct xfs_inode *ip = XFS_I(dentry->d_inode); | 1049 | struct xfs_inode *ip = XFS_I(dentry->d_inode); |
1068 | __uint64_t fakeinos, id; | 1050 | __uint64_t fakeinos, id; |
1051 | __uint64_t icount; | ||
1052 | __uint64_t ifree; | ||
1053 | __uint64_t fdblocks; | ||
1069 | xfs_extlen_t lsize; | 1054 | xfs_extlen_t lsize; |
1070 | __int64_t ffree; | 1055 | __int64_t ffree; |
1071 | 1056 | ||
@@ -1076,17 +1061,21 @@ xfs_fs_statfs( | |||
1076 | statp->f_fsid.val[0] = (u32)id; | 1061 | statp->f_fsid.val[0] = (u32)id; |
1077 | statp->f_fsid.val[1] = (u32)(id >> 32); | 1062 | statp->f_fsid.val[1] = (u32)(id >> 32); |
1078 | 1063 | ||
1079 | xfs_icsb_sync_counters(mp, XFS_ICSB_LAZY_COUNT); | 1064 | icount = percpu_counter_sum(&mp->m_icount); |
1065 | ifree = percpu_counter_sum(&mp->m_ifree); | ||
1066 | fdblocks = percpu_counter_sum(&mp->m_fdblocks); | ||
1080 | 1067 | ||
1081 | spin_lock(&mp->m_sb_lock); | 1068 | spin_lock(&mp->m_sb_lock); |
1082 | statp->f_bsize = sbp->sb_blocksize; | 1069 | statp->f_bsize = sbp->sb_blocksize; |
1083 | lsize = sbp->sb_logstart ? sbp->sb_logblocks : 0; | 1070 | lsize = sbp->sb_logstart ? sbp->sb_logblocks : 0; |
1084 | statp->f_blocks = sbp->sb_dblocks - lsize; | 1071 | statp->f_blocks = sbp->sb_dblocks - lsize; |
1085 | statp->f_bfree = statp->f_bavail = | 1072 | spin_unlock(&mp->m_sb_lock); |
1086 | sbp->sb_fdblocks - XFS_ALLOC_SET_ASIDE(mp); | 1073 | |
1074 | statp->f_bfree = fdblocks - XFS_ALLOC_SET_ASIDE(mp); | ||
1075 | statp->f_bavail = statp->f_bfree; | ||
1076 | |||
1087 | fakeinos = statp->f_bfree << sbp->sb_inopblog; | 1077 | fakeinos = statp->f_bfree << sbp->sb_inopblog; |
1088 | statp->f_files = | 1078 | statp->f_files = MIN(icount + fakeinos, (__uint64_t)XFS_MAXINUMBER); |
1089 | MIN(sbp->sb_icount + fakeinos, (__uint64_t)XFS_MAXINUMBER); | ||
1090 | if (mp->m_maxicount) | 1079 | if (mp->m_maxicount) |
1091 | statp->f_files = min_t(typeof(statp->f_files), | 1080 | statp->f_files = min_t(typeof(statp->f_files), |
1092 | statp->f_files, | 1081 | statp->f_files, |
@@ -1098,10 +1087,9 @@ xfs_fs_statfs( | |||
1098 | sbp->sb_icount); | 1087 | sbp->sb_icount); |
1099 | 1088 | ||
1100 | /* make sure statp->f_ffree does not underflow */ | 1089 | /* make sure statp->f_ffree does not underflow */ |
1101 | ffree = statp->f_files - (sbp->sb_icount - sbp->sb_ifree); | 1090 | ffree = statp->f_files - (icount - ifree); |
1102 | statp->f_ffree = max_t(__int64_t, ffree, 0); | 1091 | statp->f_ffree = max_t(__int64_t, ffree, 0); |
1103 | 1092 | ||
1104 | spin_unlock(&mp->m_sb_lock); | ||
1105 | 1093 | ||
1106 | if ((ip->i_d.di_flags & XFS_DIFLAG_PROJINHERIT) && | 1094 | if ((ip->i_d.di_flags & XFS_DIFLAG_PROJINHERIT) && |
1107 | ((mp->m_qflags & (XFS_PQUOTA_ACCT|XFS_PQUOTA_ENFD))) == | 1095 | ((mp->m_qflags & (XFS_PQUOTA_ACCT|XFS_PQUOTA_ENFD))) == |
@@ -1382,6 +1370,51 @@ xfs_finish_flags( | |||
1382 | return 0; | 1370 | return 0; |
1383 | } | 1371 | } |
1384 | 1372 | ||
1373 | static int | ||
1374 | xfs_init_percpu_counters( | ||
1375 | struct xfs_mount *mp) | ||
1376 | { | ||
1377 | int error; | ||
1378 | |||
1379 | error = percpu_counter_init(&mp->m_icount, 0, GFP_KERNEL); | ||
1380 | if (error) | ||
1381 | return ENOMEM; | ||
1382 | |||
1383 | error = percpu_counter_init(&mp->m_ifree, 0, GFP_KERNEL); | ||
1384 | if (error) | ||
1385 | goto free_icount; | ||
1386 | |||
1387 | error = percpu_counter_init(&mp->m_fdblocks, 0, GFP_KERNEL); | ||
1388 | if (error) | ||
1389 | goto free_ifree; | ||
1390 | |||
1391 | return 0; | ||
1392 | |||
1393 | free_ifree: | ||
1394 | percpu_counter_destroy(&mp->m_ifree); | ||
1395 | free_icount: | ||
1396 | percpu_counter_destroy(&mp->m_icount); | ||
1397 | return -ENOMEM; | ||
1398 | } | ||
1399 | |||
1400 | void | ||
1401 | xfs_reinit_percpu_counters( | ||
1402 | struct xfs_mount *mp) | ||
1403 | { | ||
1404 | percpu_counter_set(&mp->m_icount, mp->m_sb.sb_icount); | ||
1405 | percpu_counter_set(&mp->m_ifree, mp->m_sb.sb_ifree); | ||
1406 | percpu_counter_set(&mp->m_fdblocks, mp->m_sb.sb_fdblocks); | ||
1407 | } | ||
1408 | |||
1409 | static void | ||
1410 | xfs_destroy_percpu_counters( | ||
1411 | struct xfs_mount *mp) | ||
1412 | { | ||
1413 | percpu_counter_destroy(&mp->m_icount); | ||
1414 | percpu_counter_destroy(&mp->m_ifree); | ||
1415 | percpu_counter_destroy(&mp->m_fdblocks); | ||
1416 | } | ||
1417 | |||
1385 | STATIC int | 1418 | STATIC int |
1386 | xfs_fs_fill_super( | 1419 | xfs_fs_fill_super( |
1387 | struct super_block *sb, | 1420 | struct super_block *sb, |
@@ -1430,7 +1463,7 @@ xfs_fs_fill_super( | |||
1430 | if (error) | 1463 | if (error) |
1431 | goto out_close_devices; | 1464 | goto out_close_devices; |
1432 | 1465 | ||
1433 | error = xfs_icsb_init_counters(mp); | 1466 | error = xfs_init_percpu_counters(mp); |
1434 | if (error) | 1467 | if (error) |
1435 | goto out_destroy_workqueues; | 1468 | goto out_destroy_workqueues; |
1436 | 1469 | ||
@@ -1488,7 +1521,7 @@ xfs_fs_fill_super( | |||
1488 | out_free_sb: | 1521 | out_free_sb: |
1489 | xfs_freesb(mp); | 1522 | xfs_freesb(mp); |
1490 | out_destroy_counters: | 1523 | out_destroy_counters: |
1491 | xfs_icsb_destroy_counters(mp); | 1524 | xfs_destroy_percpu_counters(mp); |
1492 | out_destroy_workqueues: | 1525 | out_destroy_workqueues: |
1493 | xfs_destroy_mount_workqueues(mp); | 1526 | xfs_destroy_mount_workqueues(mp); |
1494 | out_close_devices: | 1527 | out_close_devices: |
@@ -1505,6 +1538,24 @@ out_destroy_workqueues: | |||
1505 | goto out_free_sb; | 1538 | goto out_free_sb; |
1506 | } | 1539 | } |
1507 | 1540 | ||
1541 | STATIC void | ||
1542 | xfs_fs_put_super( | ||
1543 | struct super_block *sb) | ||
1544 | { | ||
1545 | struct xfs_mount *mp = XFS_M(sb); | ||
1546 | |||
1547 | xfs_notice(mp, "Unmounting Filesystem"); | ||
1548 | xfs_filestream_unmount(mp); | ||
1549 | xfs_unmountfs(mp); | ||
1550 | |||
1551 | xfs_freesb(mp); | ||
1552 | xfs_destroy_percpu_counters(mp); | ||
1553 | xfs_destroy_mount_workqueues(mp); | ||
1554 | xfs_close_devices(mp); | ||
1555 | xfs_free_fsname(mp); | ||
1556 | kfree(mp); | ||
1557 | } | ||
1558 | |||
1508 | STATIC struct dentry * | 1559 | STATIC struct dentry * |
1509 | xfs_fs_mount( | 1560 | xfs_fs_mount( |
1510 | struct file_system_type *fs_type, | 1561 | struct file_system_type *fs_type, |
diff --git a/fs/xfs/xfs_super.h b/fs/xfs/xfs_super.h index 2b830c2f322e..499058fea303 100644 --- a/fs/xfs/xfs_super.h +++ b/fs/xfs/xfs_super.h | |||
@@ -72,6 +72,8 @@ extern const struct export_operations xfs_export_operations; | |||
72 | extern const struct xattr_handler *xfs_xattr_handlers[]; | 72 | extern const struct xattr_handler *xfs_xattr_handlers[]; |
73 | extern const struct quotactl_ops xfs_quotactl_operations; | 73 | extern const struct quotactl_ops xfs_quotactl_operations; |
74 | 74 | ||
75 | extern void xfs_reinit_percpu_counters(struct xfs_mount *mp); | ||
76 | |||
75 | #define XFS_M(sb) ((struct xfs_mount *)((sb)->s_fs_info)) | 77 | #define XFS_M(sb) ((struct xfs_mount *)((sb)->s_fs_info)) |
76 | 78 | ||
77 | #endif /* __XFS_SUPER_H__ */ | 79 | #endif /* __XFS_SUPER_H__ */ |
diff --git a/fs/xfs/xfs_trans.c b/fs/xfs/xfs_trans.c index eb90cd59a0ec..220ef2c906b2 100644 --- a/fs/xfs/xfs_trans.c +++ b/fs/xfs/xfs_trans.c | |||
@@ -173,7 +173,7 @@ xfs_trans_reserve( | |||
173 | uint rtextents) | 173 | uint rtextents) |
174 | { | 174 | { |
175 | int error = 0; | 175 | int error = 0; |
176 | int rsvd = (tp->t_flags & XFS_TRANS_RESERVE) != 0; | 176 | bool rsvd = (tp->t_flags & XFS_TRANS_RESERVE) != 0; |
177 | 177 | ||
178 | /* Mark this thread as being in a transaction */ | 178 | /* Mark this thread as being in a transaction */ |
179 | current_set_flags_nested(&tp->t_pflags, PF_FSTRANS); | 179 | current_set_flags_nested(&tp->t_pflags, PF_FSTRANS); |
@@ -184,8 +184,7 @@ xfs_trans_reserve( | |||
184 | * fail if the count would go below zero. | 184 | * fail if the count would go below zero. |
185 | */ | 185 | */ |
186 | if (blocks > 0) { | 186 | if (blocks > 0) { |
187 | error = xfs_icsb_modify_counters(tp->t_mountp, XFS_SBS_FDBLOCKS, | 187 | error = xfs_mod_fdblocks(tp->t_mountp, -((int64_t)blocks), rsvd); |
188 | -((int64_t)blocks), rsvd); | ||
189 | if (error != 0) { | 188 | if (error != 0) { |
190 | current_restore_flags_nested(&tp->t_pflags, PF_FSTRANS); | 189 | current_restore_flags_nested(&tp->t_pflags, PF_FSTRANS); |
191 | return -ENOSPC; | 190 | return -ENOSPC; |
@@ -236,8 +235,7 @@ xfs_trans_reserve( | |||
236 | * fail if the count would go below zero. | 235 | * fail if the count would go below zero. |
237 | */ | 236 | */ |
238 | if (rtextents > 0) { | 237 | if (rtextents > 0) { |
239 | error = xfs_mod_incore_sb(tp->t_mountp, XFS_SBS_FREXTENTS, | 238 | error = xfs_mod_frextents(tp->t_mountp, -((int64_t)rtextents)); |
240 | -((int64_t)rtextents), rsvd); | ||
241 | if (error) { | 239 | if (error) { |
242 | error = -ENOSPC; | 240 | error = -ENOSPC; |
243 | goto undo_log; | 241 | goto undo_log; |
@@ -268,8 +266,7 @@ undo_log: | |||
268 | 266 | ||
269 | undo_blocks: | 267 | undo_blocks: |
270 | if (blocks > 0) { | 268 | if (blocks > 0) { |
271 | xfs_icsb_modify_counters(tp->t_mountp, XFS_SBS_FDBLOCKS, | 269 | xfs_mod_fdblocks(tp->t_mountp, -((int64_t)blocks), rsvd); |
272 | (int64_t)blocks, rsvd); | ||
273 | tp->t_blk_res = 0; | 270 | tp->t_blk_res = 0; |
274 | } | 271 | } |
275 | 272 | ||
@@ -488,6 +485,54 @@ xfs_trans_apply_sb_deltas( | |||
488 | sizeof(sbp->sb_frextents) - 1); | 485 | sizeof(sbp->sb_frextents) - 1); |
489 | } | 486 | } |
490 | 487 | ||
488 | STATIC int | ||
489 | xfs_sb_mod8( | ||
490 | uint8_t *field, | ||
491 | int8_t delta) | ||
492 | { | ||
493 | int8_t counter = *field; | ||
494 | |||
495 | counter += delta; | ||
496 | if (counter < 0) { | ||
497 | ASSERT(0); | ||
498 | return -EINVAL; | ||
499 | } | ||
500 | *field = counter; | ||
501 | return 0; | ||
502 | } | ||
503 | |||
504 | STATIC int | ||
505 | xfs_sb_mod32( | ||
506 | uint32_t *field, | ||
507 | int32_t delta) | ||
508 | { | ||
509 | int32_t counter = *field; | ||
510 | |||
511 | counter += delta; | ||
512 | if (counter < 0) { | ||
513 | ASSERT(0); | ||
514 | return -EINVAL; | ||
515 | } | ||
516 | *field = counter; | ||
517 | return 0; | ||
518 | } | ||
519 | |||
520 | STATIC int | ||
521 | xfs_sb_mod64( | ||
522 | uint64_t *field, | ||
523 | int64_t delta) | ||
524 | { | ||
525 | int64_t counter = *field; | ||
526 | |||
527 | counter += delta; | ||
528 | if (counter < 0) { | ||
529 | ASSERT(0); | ||
530 | return -EINVAL; | ||
531 | } | ||
532 | *field = counter; | ||
533 | return 0; | ||
534 | } | ||
535 | |||
491 | /* | 536 | /* |
492 | * xfs_trans_unreserve_and_mod_sb() is called to release unused reservations | 537 | * xfs_trans_unreserve_and_mod_sb() is called to release unused reservations |
493 | * and apply superblock counter changes to the in-core superblock. The | 538 | * and apply superblock counter changes to the in-core superblock. The |
@@ -495,13 +540,6 @@ xfs_trans_apply_sb_deltas( | |||
495 | * applied to the in-core superblock. The idea is that that has already been | 540 | * applied to the in-core superblock. The idea is that that has already been |
496 | * done. | 541 | * done. |
497 | * | 542 | * |
498 | * This is done efficiently with a single call to xfs_mod_incore_sb_batch(). | ||
499 | * However, we have to ensure that we only modify each superblock field only | ||
500 | * once because the application of the delta values may not be atomic. That can | ||
501 | * lead to ENOSPC races occurring if we have two separate modifcations of the | ||
502 | * free space counter to put back the entire reservation and then take away | ||
503 | * what we used. | ||
504 | * | ||
505 | * If we are not logging superblock counters, then the inode allocated/free and | 543 | * If we are not logging superblock counters, then the inode allocated/free and |
506 | * used block counts are not updated in the on disk superblock. In this case, | 544 | * used block counts are not updated in the on disk superblock. In this case, |
507 | * XFS_TRANS_SB_DIRTY will not be set when the transaction is updated but we | 545 | * XFS_TRANS_SB_DIRTY will not be set when the transaction is updated but we |
@@ -509,21 +547,15 @@ xfs_trans_apply_sb_deltas( | |||
509 | */ | 547 | */ |
510 | void | 548 | void |
511 | xfs_trans_unreserve_and_mod_sb( | 549 | xfs_trans_unreserve_and_mod_sb( |
512 | xfs_trans_t *tp) | 550 | struct xfs_trans *tp) |
513 | { | 551 | { |
514 | xfs_mod_sb_t msb[9]; /* If you add cases, add entries */ | 552 | struct xfs_mount *mp = tp->t_mountp; |
515 | xfs_mod_sb_t *msbp; | 553 | bool rsvd = (tp->t_flags & XFS_TRANS_RESERVE) != 0; |
516 | xfs_mount_t *mp = tp->t_mountp; | 554 | int64_t blkdelta = 0; |
517 | /* REFERENCED */ | 555 | int64_t rtxdelta = 0; |
518 | int error; | 556 | int64_t idelta = 0; |
519 | int rsvd; | 557 | int64_t ifreedelta = 0; |
520 | int64_t blkdelta = 0; | 558 | int error; |
521 | int64_t rtxdelta = 0; | ||
522 | int64_t idelta = 0; | ||
523 | int64_t ifreedelta = 0; | ||
524 | |||
525 | msbp = msb; | ||
526 | rsvd = (tp->t_flags & XFS_TRANS_RESERVE) != 0; | ||
527 | 559 | ||
528 | /* calculate deltas */ | 560 | /* calculate deltas */ |
529 | if (tp->t_blk_res > 0) | 561 | if (tp->t_blk_res > 0) |
@@ -547,97 +579,115 @@ xfs_trans_unreserve_and_mod_sb( | |||
547 | 579 | ||
548 | /* apply the per-cpu counters */ | 580 | /* apply the per-cpu counters */ |
549 | if (blkdelta) { | 581 | if (blkdelta) { |
550 | error = xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, | 582 | error = xfs_mod_fdblocks(mp, blkdelta, rsvd); |
551 | blkdelta, rsvd); | ||
552 | if (error) | 583 | if (error) |
553 | goto out; | 584 | goto out; |
554 | } | 585 | } |
555 | 586 | ||
556 | if (idelta) { | 587 | if (idelta) { |
557 | error = xfs_icsb_modify_counters(mp, XFS_SBS_ICOUNT, | 588 | error = xfs_mod_icount(mp, idelta); |
558 | idelta, rsvd); | ||
559 | if (error) | 589 | if (error) |
560 | goto out_undo_fdblocks; | 590 | goto out_undo_fdblocks; |
561 | } | 591 | } |
562 | 592 | ||
563 | if (ifreedelta) { | 593 | if (ifreedelta) { |
564 | error = xfs_icsb_modify_counters(mp, XFS_SBS_IFREE, | 594 | error = xfs_mod_ifree(mp, ifreedelta); |
565 | ifreedelta, rsvd); | ||
566 | if (error) | 595 | if (error) |
567 | goto out_undo_icount; | 596 | goto out_undo_icount; |
568 | } | 597 | } |
569 | 598 | ||
599 | if (rtxdelta == 0 && !(tp->t_flags & XFS_TRANS_SB_DIRTY)) | ||
600 | return; | ||
601 | |||
570 | /* apply remaining deltas */ | 602 | /* apply remaining deltas */ |
571 | if (rtxdelta != 0) { | 603 | spin_lock(&mp->m_sb_lock); |
572 | msbp->msb_field = XFS_SBS_FREXTENTS; | 604 | if (rtxdelta) { |
573 | msbp->msb_delta = rtxdelta; | 605 | error = xfs_sb_mod64(&mp->m_sb.sb_frextents, rtxdelta); |
574 | msbp++; | 606 | if (error) |
607 | goto out_undo_ifree; | ||
575 | } | 608 | } |
576 | 609 | ||
577 | if (tp->t_flags & XFS_TRANS_SB_DIRTY) { | 610 | if (tp->t_dblocks_delta != 0) { |
578 | if (tp->t_dblocks_delta != 0) { | 611 | error = xfs_sb_mod64(&mp->m_sb.sb_dblocks, tp->t_dblocks_delta); |
579 | msbp->msb_field = XFS_SBS_DBLOCKS; | 612 | if (error) |
580 | msbp->msb_delta = tp->t_dblocks_delta; | 613 | goto out_undo_frextents; |
581 | msbp++; | ||
582 | } | ||
583 | if (tp->t_agcount_delta != 0) { | ||
584 | msbp->msb_field = XFS_SBS_AGCOUNT; | ||
585 | msbp->msb_delta = tp->t_agcount_delta; | ||
586 | msbp++; | ||
587 | } | ||
588 | if (tp->t_imaxpct_delta != 0) { | ||
589 | msbp->msb_field = XFS_SBS_IMAX_PCT; | ||
590 | msbp->msb_delta = tp->t_imaxpct_delta; | ||
591 | msbp++; | ||
592 | } | ||
593 | if (tp->t_rextsize_delta != 0) { | ||
594 | msbp->msb_field = XFS_SBS_REXTSIZE; | ||
595 | msbp->msb_delta = tp->t_rextsize_delta; | ||
596 | msbp++; | ||
597 | } | ||
598 | if (tp->t_rbmblocks_delta != 0) { | ||
599 | msbp->msb_field = XFS_SBS_RBMBLOCKS; | ||
600 | msbp->msb_delta = tp->t_rbmblocks_delta; | ||
601 | msbp++; | ||
602 | } | ||
603 | if (tp->t_rblocks_delta != 0) { | ||
604 | msbp->msb_field = XFS_SBS_RBLOCKS; | ||
605 | msbp->msb_delta = tp->t_rblocks_delta; | ||
606 | msbp++; | ||
607 | } | ||
608 | if (tp->t_rextents_delta != 0) { | ||
609 | msbp->msb_field = XFS_SBS_REXTENTS; | ||
610 | msbp->msb_delta = tp->t_rextents_delta; | ||
611 | msbp++; | ||
612 | } | ||
613 | if (tp->t_rextslog_delta != 0) { | ||
614 | msbp->msb_field = XFS_SBS_REXTSLOG; | ||
615 | msbp->msb_delta = tp->t_rextslog_delta; | ||
616 | msbp++; | ||
617 | } | ||
618 | } | 614 | } |
619 | 615 | if (tp->t_agcount_delta != 0) { | |
620 | /* | 616 | error = xfs_sb_mod32(&mp->m_sb.sb_agcount, tp->t_agcount_delta); |
621 | * If we need to change anything, do it. | ||
622 | */ | ||
623 | if (msbp > msb) { | ||
624 | error = xfs_mod_incore_sb_batch(tp->t_mountp, msb, | ||
625 | (uint)(msbp - msb), rsvd); | ||
626 | if (error) | 617 | if (error) |
627 | goto out_undo_ifreecount; | 618 | goto out_undo_dblocks; |
628 | } | 619 | } |
629 | 620 | if (tp->t_imaxpct_delta != 0) { | |
621 | error = xfs_sb_mod8(&mp->m_sb.sb_imax_pct, tp->t_imaxpct_delta); | ||
622 | if (error) | ||
623 | goto out_undo_agcount; | ||
624 | } | ||
625 | if (tp->t_rextsize_delta != 0) { | ||
626 | error = xfs_sb_mod32(&mp->m_sb.sb_rextsize, | ||
627 | tp->t_rextsize_delta); | ||
628 | if (error) | ||
629 | goto out_undo_imaxpct; | ||
630 | } | ||
631 | if (tp->t_rbmblocks_delta != 0) { | ||
632 | error = xfs_sb_mod32(&mp->m_sb.sb_rbmblocks, | ||
633 | tp->t_rbmblocks_delta); | ||
634 | if (error) | ||
635 | goto out_undo_rextsize; | ||
636 | } | ||
637 | if (tp->t_rblocks_delta != 0) { | ||
638 | error = xfs_sb_mod64(&mp->m_sb.sb_rblocks, tp->t_rblocks_delta); | ||
639 | if (error) | ||
640 | goto out_undo_rbmblocks; | ||
641 | } | ||
642 | if (tp->t_rextents_delta != 0) { | ||
643 | error = xfs_sb_mod64(&mp->m_sb.sb_rextents, | ||
644 | tp->t_rextents_delta); | ||
645 | if (error) | ||
646 | goto out_undo_rblocks; | ||
647 | } | ||
648 | if (tp->t_rextslog_delta != 0) { | ||
649 | error = xfs_sb_mod8(&mp->m_sb.sb_rextslog, | ||
650 | tp->t_rextslog_delta); | ||
651 | if (error) | ||
652 | goto out_undo_rextents; | ||
653 | } | ||
654 | spin_unlock(&mp->m_sb_lock); | ||
630 | return; | 655 | return; |
631 | 656 | ||
632 | out_undo_ifreecount: | 657 | out_undo_rextents: |
658 | if (tp->t_rextents_delta) | ||
659 | xfs_sb_mod64(&mp->m_sb.sb_rextents, -tp->t_rextents_delta); | ||
660 | out_undo_rblocks: | ||
661 | if (tp->t_rblocks_delta) | ||
662 | xfs_sb_mod64(&mp->m_sb.sb_rblocks, -tp->t_rblocks_delta); | ||
663 | out_undo_rbmblocks: | ||
664 | if (tp->t_rbmblocks_delta) | ||
665 | xfs_sb_mod32(&mp->m_sb.sb_rbmblocks, -tp->t_rbmblocks_delta); | ||
666 | out_undo_rextsize: | ||
667 | if (tp->t_rextsize_delta) | ||
668 | xfs_sb_mod32(&mp->m_sb.sb_rextsize, -tp->t_rextsize_delta); | ||
669 | out_undo_imaxpct: | ||
670 | if (tp->t_rextsize_delta) | ||
671 | xfs_sb_mod8(&mp->m_sb.sb_imax_pct, -tp->t_imaxpct_delta); | ||
672 | out_undo_agcount: | ||
673 | if (tp->t_agcount_delta) | ||
674 | xfs_sb_mod32(&mp->m_sb.sb_agcount, -tp->t_agcount_delta); | ||
675 | out_undo_dblocks: | ||
676 | if (tp->t_dblocks_delta) | ||
677 | xfs_sb_mod64(&mp->m_sb.sb_dblocks, -tp->t_dblocks_delta); | ||
678 | out_undo_frextents: | ||
679 | if (rtxdelta) | ||
680 | xfs_sb_mod64(&mp->m_sb.sb_frextents, -rtxdelta); | ||
681 | out_undo_ifree: | ||
682 | spin_unlock(&mp->m_sb_lock); | ||
633 | if (ifreedelta) | 683 | if (ifreedelta) |
634 | xfs_icsb_modify_counters(mp, XFS_SBS_IFREE, -ifreedelta, rsvd); | 684 | xfs_mod_ifree(mp, -ifreedelta); |
635 | out_undo_icount: | 685 | out_undo_icount: |
636 | if (idelta) | 686 | if (idelta) |
637 | xfs_icsb_modify_counters(mp, XFS_SBS_ICOUNT, -idelta, rsvd); | 687 | xfs_mod_icount(mp, -idelta); |
638 | out_undo_fdblocks: | 688 | out_undo_fdblocks: |
639 | if (blkdelta) | 689 | if (blkdelta) |
640 | xfs_icsb_modify_counters(mp, XFS_SBS_FDBLOCKS, -blkdelta, rsvd); | 690 | xfs_mod_fdblocks(mp, -blkdelta, rsvd); |
641 | out: | 691 | out: |
642 | ASSERT(error == 0); | 692 | ASSERT(error == 0); |
643 | return; | 693 | return; |