diff options
-rw-r--r-- | fs/xfs/xfs_iget.c | 49 | ||||
-rw-r--r-- | fs/xfs/xfs_inode.c | 268 | ||||
-rw-r--r-- | fs/xfs/xfs_inode.h | 16 | ||||
-rw-r--r-- | fs/xfs/xfs_vfsops.c | 5 |
4 files changed, 156 insertions, 182 deletions
diff --git a/fs/xfs/xfs_iget.c b/fs/xfs/xfs_iget.c index 8e09b71f4104..e657c5128460 100644 --- a/fs/xfs/xfs_iget.c +++ b/fs/xfs/xfs_iget.c | |||
@@ -78,7 +78,6 @@ xfs_iget_core( | |||
78 | xfs_inode_t *ip; | 78 | xfs_inode_t *ip; |
79 | xfs_inode_t *iq; | 79 | xfs_inode_t *iq; |
80 | int error; | 80 | int error; |
81 | xfs_icluster_t *icl, *new_icl = NULL; | ||
82 | unsigned long first_index, mask; | 81 | unsigned long first_index, mask; |
83 | xfs_perag_t *pag; | 82 | xfs_perag_t *pag; |
84 | xfs_agino_t agino; | 83 | xfs_agino_t agino; |
@@ -229,11 +228,9 @@ finish_inode: | |||
229 | } | 228 | } |
230 | 229 | ||
231 | /* | 230 | /* |
232 | * This is a bit messy - we preallocate everything we _might_ | 231 | * Preload the radix tree so we can insert safely under the |
233 | * need before we pick up the ici lock. That way we don't have to | 232 | * write spinlock. |
234 | * juggle locks and go all the way back to the start. | ||
235 | */ | 233 | */ |
236 | new_icl = kmem_zone_alloc(xfs_icluster_zone, KM_SLEEP); | ||
237 | if (radix_tree_preload(GFP_KERNEL)) { | 234 | if (radix_tree_preload(GFP_KERNEL)) { |
238 | xfs_idestroy(ip); | 235 | xfs_idestroy(ip); |
239 | delay(1); | 236 | delay(1); |
@@ -242,17 +239,6 @@ finish_inode: | |||
242 | mask = ~(((XFS_INODE_CLUSTER_SIZE(mp) >> mp->m_sb.sb_inodelog)) - 1); | 239 | mask = ~(((XFS_INODE_CLUSTER_SIZE(mp) >> mp->m_sb.sb_inodelog)) - 1); |
243 | first_index = agino & mask; | 240 | first_index = agino & mask; |
244 | write_lock(&pag->pag_ici_lock); | 241 | write_lock(&pag->pag_ici_lock); |
245 | |||
246 | /* | ||
247 | * Find the cluster if it exists | ||
248 | */ | ||
249 | icl = NULL; | ||
250 | if (radix_tree_gang_lookup(&pag->pag_ici_root, (void**)&iq, | ||
251 | first_index, 1)) { | ||
252 | if ((XFS_INO_TO_AGINO(mp, iq->i_ino) & mask) == first_index) | ||
253 | icl = iq->i_cluster; | ||
254 | } | ||
255 | |||
256 | /* | 242 | /* |
257 | * insert the new inode | 243 | * insert the new inode |
258 | */ | 244 | */ |
@@ -267,30 +253,13 @@ finish_inode: | |||
267 | } | 253 | } |
268 | 254 | ||
269 | /* | 255 | /* |
270 | * These values _must_ be set before releasing ihlock! | 256 | * These values _must_ be set before releasing the radix tree lock! |
271 | */ | 257 | */ |
272 | ip->i_udquot = ip->i_gdquot = NULL; | 258 | ip->i_udquot = ip->i_gdquot = NULL; |
273 | xfs_iflags_set(ip, XFS_INEW); | 259 | xfs_iflags_set(ip, XFS_INEW); |
274 | 260 | ||
275 | ASSERT(ip->i_cluster == NULL); | ||
276 | |||
277 | if (!icl) { | ||
278 | spin_lock_init(&new_icl->icl_lock); | ||
279 | INIT_HLIST_HEAD(&new_icl->icl_inodes); | ||
280 | icl = new_icl; | ||
281 | new_icl = NULL; | ||
282 | } else { | ||
283 | ASSERT(!hlist_empty(&icl->icl_inodes)); | ||
284 | } | ||
285 | spin_lock(&icl->icl_lock); | ||
286 | hlist_add_head(&ip->i_cnode, &icl->icl_inodes); | ||
287 | ip->i_cluster = icl; | ||
288 | spin_unlock(&icl->icl_lock); | ||
289 | |||
290 | write_unlock(&pag->pag_ici_lock); | 261 | write_unlock(&pag->pag_ici_lock); |
291 | radix_tree_preload_end(); | 262 | radix_tree_preload_end(); |
292 | if (new_icl) | ||
293 | kmem_zone_free(xfs_icluster_zone, new_icl); | ||
294 | 263 | ||
295 | /* | 264 | /* |
296 | * Link ip to its mount and thread it on the mount's inode list. | 265 | * Link ip to its mount and thread it on the mount's inode list. |
@@ -529,18 +498,6 @@ xfs_iextract( | |||
529 | xfs_put_perag(mp, pag); | 498 | xfs_put_perag(mp, pag); |
530 | 499 | ||
531 | /* | 500 | /* |
532 | * Remove from cluster list | ||
533 | */ | ||
534 | mp = ip->i_mount; | ||
535 | spin_lock(&ip->i_cluster->icl_lock); | ||
536 | hlist_del(&ip->i_cnode); | ||
537 | spin_unlock(&ip->i_cluster->icl_lock); | ||
538 | |||
539 | /* was last inode in cluster? */ | ||
540 | if (hlist_empty(&ip->i_cluster->icl_inodes)) | ||
541 | kmem_zone_free(xfs_icluster_zone, ip->i_cluster); | ||
542 | |||
543 | /* | ||
544 | * Remove from mount's inode list. | 501 | * Remove from mount's inode list. |
545 | */ | 502 | */ |
546 | XFS_MOUNT_ILOCK(mp); | 503 | XFS_MOUNT_ILOCK(mp); |
diff --git a/fs/xfs/xfs_inode.c b/fs/xfs/xfs_inode.c index 3c3e9e3c1da8..040c0e41729b 100644 --- a/fs/xfs/xfs_inode.c +++ b/fs/xfs/xfs_inode.c | |||
@@ -55,7 +55,6 @@ | |||
55 | 55 | ||
56 | kmem_zone_t *xfs_ifork_zone; | 56 | kmem_zone_t *xfs_ifork_zone; |
57 | kmem_zone_t *xfs_inode_zone; | 57 | kmem_zone_t *xfs_inode_zone; |
58 | kmem_zone_t *xfs_icluster_zone; | ||
59 | 58 | ||
60 | /* | 59 | /* |
61 | * Used in xfs_itruncate(). This is the maximum number of extents | 60 | * Used in xfs_itruncate(). This is the maximum number of extents |
@@ -2994,6 +2993,153 @@ xfs_iflush_fork( | |||
2994 | return 0; | 2993 | return 0; |
2995 | } | 2994 | } |
2996 | 2995 | ||
2996 | STATIC int | ||
2997 | xfs_iflush_cluster( | ||
2998 | xfs_inode_t *ip, | ||
2999 | xfs_buf_t *bp) | ||
3000 | { | ||
3001 | xfs_mount_t *mp = ip->i_mount; | ||
3002 | xfs_perag_t *pag = xfs_get_perag(mp, ip->i_ino); | ||
3003 | unsigned long first_index, mask; | ||
3004 | int ilist_size; | ||
3005 | xfs_inode_t **ilist; | ||
3006 | xfs_inode_t *iq; | ||
3007 | xfs_inode_log_item_t *iip; | ||
3008 | int nr_found; | ||
3009 | int clcount = 0; | ||
3010 | int bufwasdelwri; | ||
3011 | int i; | ||
3012 | |||
3013 | ASSERT(pag->pagi_inodeok); | ||
3014 | ASSERT(pag->pag_ici_init); | ||
3015 | |||
3016 | ilist_size = XFS_INODE_CLUSTER_SIZE(mp) * sizeof(xfs_inode_t *); | ||
3017 | ilist = kmem_alloc(ilist_size, KM_MAYFAIL); | ||
3018 | if (!ilist) | ||
3019 | return 0; | ||
3020 | |||
3021 | mask = ~(((XFS_INODE_CLUSTER_SIZE(mp) >> mp->m_sb.sb_inodelog)) - 1); | ||
3022 | first_index = XFS_INO_TO_AGINO(mp, ip->i_ino) & mask; | ||
3023 | read_lock(&pag->pag_ici_lock); | ||
3024 | /* really need a gang lookup range call here */ | ||
3025 | nr_found = radix_tree_gang_lookup(&pag->pag_ici_root, (void**)ilist, | ||
3026 | first_index, | ||
3027 | XFS_INODE_CLUSTER_SIZE(mp)); | ||
3028 | if (nr_found == 0) | ||
3029 | goto out_free; | ||
3030 | |||
3031 | for (i = 0; i < nr_found; i++) { | ||
3032 | iq = ilist[i]; | ||
3033 | if (iq == ip) | ||
3034 | continue; | ||
3035 | /* if the inode lies outside this cluster, we're done. */ | ||
3036 | if ((XFS_INO_TO_AGINO(mp, iq->i_ino) & mask) != first_index) | ||
3037 | break; | ||
3038 | /* | ||
3039 | * Do an un-protected check to see if the inode is dirty and | ||
3040 | * is a candidate for flushing. These checks will be repeated | ||
3041 | * later after the appropriate locks are acquired. | ||
3042 | */ | ||
3043 | iip = iq->i_itemp; | ||
3044 | if ((iq->i_update_core == 0) && | ||
3045 | ((iip == NULL) || | ||
3046 | !(iip->ili_format.ilf_fields & XFS_ILOG_ALL)) && | ||
3047 | xfs_ipincount(iq) == 0) { | ||
3048 | continue; | ||
3049 | } | ||
3050 | |||
3051 | /* | ||
3052 | * Try to get locks. If any are unavailable or it is pinned, | ||
3053 | * then this inode cannot be flushed and is skipped. | ||
3054 | */ | ||
3055 | |||
3056 | if (!xfs_ilock_nowait(iq, XFS_ILOCK_SHARED)) | ||
3057 | continue; | ||
3058 | if (!xfs_iflock_nowait(iq)) { | ||
3059 | xfs_iunlock(iq, XFS_ILOCK_SHARED); | ||
3060 | continue; | ||
3061 | } | ||
3062 | if (xfs_ipincount(iq)) { | ||
3063 | xfs_ifunlock(iq); | ||
3064 | xfs_iunlock(iq, XFS_ILOCK_SHARED); | ||
3065 | continue; | ||
3066 | } | ||
3067 | |||
3068 | /* | ||
3069 | * arriving here means that this inode can be flushed. First | ||
3070 | * re-check that it's dirty before flushing. | ||
3071 | */ | ||
3072 | iip = iq->i_itemp; | ||
3073 | if ((iq->i_update_core != 0) || ((iip != NULL) && | ||
3074 | (iip->ili_format.ilf_fields & XFS_ILOG_ALL))) { | ||
3075 | int error; | ||
3076 | error = xfs_iflush_int(iq, bp); | ||
3077 | if (error) { | ||
3078 | xfs_iunlock(iq, XFS_ILOCK_SHARED); | ||
3079 | goto cluster_corrupt_out; | ||
3080 | } | ||
3081 | clcount++; | ||
3082 | } else { | ||
3083 | xfs_ifunlock(iq); | ||
3084 | } | ||
3085 | xfs_iunlock(iq, XFS_ILOCK_SHARED); | ||
3086 | } | ||
3087 | |||
3088 | if (clcount) { | ||
3089 | XFS_STATS_INC(xs_icluster_flushcnt); | ||
3090 | XFS_STATS_ADD(xs_icluster_flushinode, clcount); | ||
3091 | } | ||
3092 | |||
3093 | out_free: | ||
3094 | read_unlock(&pag->pag_ici_lock); | ||
3095 | kmem_free(ilist, ilist_size); | ||
3096 | return 0; | ||
3097 | |||
3098 | |||
3099 | cluster_corrupt_out: | ||
3100 | /* | ||
3101 | * Corruption detected in the clustering loop. Invalidate the | ||
3102 | * inode buffer and shut down the filesystem. | ||
3103 | */ | ||
3104 | read_unlock(&pag->pag_ici_lock); | ||
3105 | /* | ||
3106 | * Clean up the buffer. If it was B_DELWRI, just release it -- | ||
3107 | * brelse can handle it with no problems. If not, shut down the | ||
3108 | * filesystem before releasing the buffer. | ||
3109 | */ | ||
3110 | bufwasdelwri = XFS_BUF_ISDELAYWRITE(bp); | ||
3111 | if (bufwasdelwri) | ||
3112 | xfs_buf_relse(bp); | ||
3113 | |||
3114 | xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE); | ||
3115 | |||
3116 | if (!bufwasdelwri) { | ||
3117 | /* | ||
3118 | * Just like incore_relse: if we have b_iodone functions, | ||
3119 | * mark the buffer as an error and call them. Otherwise | ||
3120 | * mark it as stale and brelse. | ||
3121 | */ | ||
3122 | if (XFS_BUF_IODONE_FUNC(bp)) { | ||
3123 | XFS_BUF_CLR_BDSTRAT_FUNC(bp); | ||
3124 | XFS_BUF_UNDONE(bp); | ||
3125 | XFS_BUF_STALE(bp); | ||
3126 | XFS_BUF_SHUT(bp); | ||
3127 | XFS_BUF_ERROR(bp,EIO); | ||
3128 | xfs_biodone(bp); | ||
3129 | } else { | ||
3130 | XFS_BUF_STALE(bp); | ||
3131 | xfs_buf_relse(bp); | ||
3132 | } | ||
3133 | } | ||
3134 | |||
3135 | /* | ||
3136 | * Unlocks the flush lock | ||
3137 | */ | ||
3138 | xfs_iflush_abort(iq); | ||
3139 | kmem_free(ilist, ilist_size); | ||
3140 | return XFS_ERROR(EFSCORRUPTED); | ||
3141 | } | ||
3142 | |||
2997 | /* | 3143 | /* |
2998 | * xfs_iflush() will write a modified inode's changes out to the | 3144 | * xfs_iflush() will write a modified inode's changes out to the |
2999 | * inode's on disk home. The caller must have the inode lock held | 3145 | * inode's on disk home. The caller must have the inode lock held |
@@ -3013,13 +3159,8 @@ xfs_iflush( | |||
3013 | xfs_dinode_t *dip; | 3159 | xfs_dinode_t *dip; |
3014 | xfs_mount_t *mp; | 3160 | xfs_mount_t *mp; |
3015 | int error; | 3161 | int error; |
3016 | /* REFERENCED */ | ||
3017 | xfs_inode_t *iq; | ||
3018 | int clcount; /* count of inodes clustered */ | ||
3019 | int bufwasdelwri; | ||
3020 | struct hlist_node *entry; | ||
3021 | enum { INT_DELWRI = (1 << 0), INT_ASYNC = (1 << 1) }; | ||
3022 | int noblock = (flags == XFS_IFLUSH_ASYNC_NOBLOCK); | 3162 | int noblock = (flags == XFS_IFLUSH_ASYNC_NOBLOCK); |
3163 | enum { INT_DELWRI = (1 << 0), INT_ASYNC = (1 << 1) }; | ||
3023 | 3164 | ||
3024 | XFS_STATS_INC(xs_iflush_count); | 3165 | XFS_STATS_INC(xs_iflush_count); |
3025 | 3166 | ||
@@ -3138,9 +3279,8 @@ xfs_iflush( | |||
3138 | * First flush out the inode that xfs_iflush was called with. | 3279 | * First flush out the inode that xfs_iflush was called with. |
3139 | */ | 3280 | */ |
3140 | error = xfs_iflush_int(ip, bp); | 3281 | error = xfs_iflush_int(ip, bp); |
3141 | if (error) { | 3282 | if (error) |
3142 | goto corrupt_out; | 3283 | goto corrupt_out; |
3143 | } | ||
3144 | 3284 | ||
3145 | /* | 3285 | /* |
3146 | * If the buffer is pinned then push on the log now so we won't | 3286 | * If the buffer is pinned then push on the log now so we won't |
@@ -3153,70 +3293,9 @@ xfs_iflush( | |||
3153 | * inode clustering: | 3293 | * inode clustering: |
3154 | * see if other inodes can be gathered into this write | 3294 | * see if other inodes can be gathered into this write |
3155 | */ | 3295 | */ |
3156 | spin_lock(&ip->i_cluster->icl_lock); | 3296 | error = xfs_iflush_cluster(ip, bp); |
3157 | ip->i_cluster->icl_buf = bp; | 3297 | if (error) |
3158 | 3298 | goto cluster_corrupt_out; | |
3159 | clcount = 0; | ||
3160 | hlist_for_each_entry(iq, entry, &ip->i_cluster->icl_inodes, i_cnode) { | ||
3161 | if (iq == ip) | ||
3162 | continue; | ||
3163 | |||
3164 | /* | ||
3165 | * Do an un-protected check to see if the inode is dirty and | ||
3166 | * is a candidate for flushing. These checks will be repeated | ||
3167 | * later after the appropriate locks are acquired. | ||
3168 | */ | ||
3169 | iip = iq->i_itemp; | ||
3170 | if ((iq->i_update_core == 0) && | ||
3171 | ((iip == NULL) || | ||
3172 | !(iip->ili_format.ilf_fields & XFS_ILOG_ALL)) && | ||
3173 | xfs_ipincount(iq) == 0) { | ||
3174 | continue; | ||
3175 | } | ||
3176 | |||
3177 | /* | ||
3178 | * Try to get locks. If any are unavailable, | ||
3179 | * then this inode cannot be flushed and is skipped. | ||
3180 | */ | ||
3181 | |||
3182 | /* get inode locks (just i_lock) */ | ||
3183 | if (xfs_ilock_nowait(iq, XFS_ILOCK_SHARED)) { | ||
3184 | /* get inode flush lock */ | ||
3185 | if (xfs_iflock_nowait(iq)) { | ||
3186 | /* check if pinned */ | ||
3187 | if (xfs_ipincount(iq) == 0) { | ||
3188 | /* arriving here means that | ||
3189 | * this inode can be flushed. | ||
3190 | * first re-check that it's | ||
3191 | * dirty | ||
3192 | */ | ||
3193 | iip = iq->i_itemp; | ||
3194 | if ((iq->i_update_core != 0)|| | ||
3195 | ((iip != NULL) && | ||
3196 | (iip->ili_format.ilf_fields & XFS_ILOG_ALL))) { | ||
3197 | clcount++; | ||
3198 | error = xfs_iflush_int(iq, bp); | ||
3199 | if (error) { | ||
3200 | xfs_iunlock(iq, | ||
3201 | XFS_ILOCK_SHARED); | ||
3202 | goto cluster_corrupt_out; | ||
3203 | } | ||
3204 | } else { | ||
3205 | xfs_ifunlock(iq); | ||
3206 | } | ||
3207 | } else { | ||
3208 | xfs_ifunlock(iq); | ||
3209 | } | ||
3210 | } | ||
3211 | xfs_iunlock(iq, XFS_ILOCK_SHARED); | ||
3212 | } | ||
3213 | } | ||
3214 | spin_unlock(&ip->i_cluster->icl_lock); | ||
3215 | |||
3216 | if (clcount) { | ||
3217 | XFS_STATS_INC(xs_icluster_flushcnt); | ||
3218 | XFS_STATS_ADD(xs_icluster_flushinode, clcount); | ||
3219 | } | ||
3220 | 3299 | ||
3221 | if (flags & INT_DELWRI) { | 3300 | if (flags & INT_DELWRI) { |
3222 | xfs_bdwrite(mp, bp); | 3301 | xfs_bdwrite(mp, bp); |
@@ -3230,52 +3309,11 @@ xfs_iflush( | |||
3230 | corrupt_out: | 3309 | corrupt_out: |
3231 | xfs_buf_relse(bp); | 3310 | xfs_buf_relse(bp); |
3232 | xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE); | 3311 | xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE); |
3233 | xfs_iflush_abort(ip); | ||
3234 | /* | ||
3235 | * Unlocks the flush lock | ||
3236 | */ | ||
3237 | return XFS_ERROR(EFSCORRUPTED); | ||
3238 | |||
3239 | cluster_corrupt_out: | 3312 | cluster_corrupt_out: |
3240 | /* Corruption detected in the clustering loop. Invalidate the | ||
3241 | * inode buffer and shut down the filesystem. | ||
3242 | */ | ||
3243 | spin_unlock(&ip->i_cluster->icl_lock); | ||
3244 | |||
3245 | /* | ||
3246 | * Clean up the buffer. If it was B_DELWRI, just release it -- | ||
3247 | * brelse can handle it with no problems. If not, shut down the | ||
3248 | * filesystem before releasing the buffer. | ||
3249 | */ | ||
3250 | if ((bufwasdelwri= XFS_BUF_ISDELAYWRITE(bp))) { | ||
3251 | xfs_buf_relse(bp); | ||
3252 | } | ||
3253 | |||
3254 | xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE); | ||
3255 | |||
3256 | if(!bufwasdelwri) { | ||
3257 | /* | ||
3258 | * Just like incore_relse: if we have b_iodone functions, | ||
3259 | * mark the buffer as an error and call them. Otherwise | ||
3260 | * mark it as stale and brelse. | ||
3261 | */ | ||
3262 | if (XFS_BUF_IODONE_FUNC(bp)) { | ||
3263 | XFS_BUF_CLR_BDSTRAT_FUNC(bp); | ||
3264 | XFS_BUF_UNDONE(bp); | ||
3265 | XFS_BUF_STALE(bp); | ||
3266 | XFS_BUF_SHUT(bp); | ||
3267 | XFS_BUF_ERROR(bp,EIO); | ||
3268 | xfs_biodone(bp); | ||
3269 | } else { | ||
3270 | XFS_BUF_STALE(bp); | ||
3271 | xfs_buf_relse(bp); | ||
3272 | } | ||
3273 | } | ||
3274 | |||
3275 | xfs_iflush_abort(iq); | ||
3276 | /* | 3313 | /* |
3277 | * Unlocks the flush lock | 3314 | * Unlocks the flush lock |
3278 | */ | 3315 | */ |
3316 | xfs_iflush_abort(ip); | ||
3279 | return XFS_ERROR(EFSCORRUPTED); | 3317 | return XFS_ERROR(EFSCORRUPTED); |
3280 | } | 3318 | } |
3281 | 3319 | ||
diff --git a/fs/xfs/xfs_inode.h b/fs/xfs/xfs_inode.h index c3bfffca9214..93c37697a72c 100644 --- a/fs/xfs/xfs_inode.h +++ b/fs/xfs/xfs_inode.h | |||
@@ -133,19 +133,6 @@ typedef struct dm_attrs_s { | |||
133 | } dm_attrs_t; | 133 | } dm_attrs_t; |
134 | 134 | ||
135 | /* | 135 | /* |
136 | * This is the xfs inode cluster structure. This structure is used by | ||
137 | * xfs_iflush to find inodes that share a cluster and can be flushed to disk at | ||
138 | * the same time. | ||
139 | */ | ||
140 | typedef struct xfs_icluster { | ||
141 | struct hlist_head icl_inodes; /* list of inodes on cluster */ | ||
142 | xfs_daddr_t icl_blkno; /* starting block number of | ||
143 | * the cluster */ | ||
144 | struct xfs_buf *icl_buf; /* the inode buffer */ | ||
145 | spinlock_t icl_lock; /* inode list lock */ | ||
146 | } xfs_icluster_t; | ||
147 | |||
148 | /* | ||
149 | * This is the xfs in-core inode structure. | 136 | * This is the xfs in-core inode structure. |
150 | * Most of the on-disk inode is embedded in the i_d field. | 137 | * Most of the on-disk inode is embedded in the i_d field. |
151 | * | 138 | * |
@@ -248,8 +235,6 @@ typedef struct xfs_inode { | |||
248 | unsigned int i_delayed_blks; /* count of delay alloc blks */ | 235 | unsigned int i_delayed_blks; /* count of delay alloc blks */ |
249 | 236 | ||
250 | xfs_icdinode_t i_d; /* most of ondisk inode */ | 237 | xfs_icdinode_t i_d; /* most of ondisk inode */ |
251 | xfs_icluster_t *i_cluster; /* cluster list header */ | ||
252 | struct hlist_node i_cnode; /* cluster link node */ | ||
253 | 238 | ||
254 | xfs_fsize_t i_size; /* in-memory size */ | 239 | xfs_fsize_t i_size; /* in-memory size */ |
255 | xfs_fsize_t i_new_size; /* size when write completes */ | 240 | xfs_fsize_t i_new_size; /* size when write completes */ |
@@ -594,7 +579,6 @@ void xfs_inobp_check(struct xfs_mount *, struct xfs_buf *); | |||
594 | #define xfs_inobp_check(mp, bp) | 579 | #define xfs_inobp_check(mp, bp) |
595 | #endif /* DEBUG */ | 580 | #endif /* DEBUG */ |
596 | 581 | ||
597 | extern struct kmem_zone *xfs_icluster_zone; | ||
598 | extern struct kmem_zone *xfs_ifork_zone; | 582 | extern struct kmem_zone *xfs_ifork_zone; |
599 | extern struct kmem_zone *xfs_inode_zone; | 583 | extern struct kmem_zone *xfs_inode_zone; |
600 | extern struct kmem_zone *xfs_ili_zone; | 584 | extern struct kmem_zone *xfs_ili_zone; |
diff --git a/fs/xfs/xfs_vfsops.c b/fs/xfs/xfs_vfsops.c index 79bdfb3d6081..3ec27bf8531c 100644 --- a/fs/xfs/xfs_vfsops.c +++ b/fs/xfs/xfs_vfsops.c | |||
@@ -112,9 +112,6 @@ xfs_init(void) | |||
112 | xfs_ili_zone = | 112 | xfs_ili_zone = |
113 | kmem_zone_init_flags(sizeof(xfs_inode_log_item_t), "xfs_ili", | 113 | kmem_zone_init_flags(sizeof(xfs_inode_log_item_t), "xfs_ili", |
114 | KM_ZONE_SPREAD, NULL); | 114 | KM_ZONE_SPREAD, NULL); |
115 | xfs_icluster_zone = | ||
116 | kmem_zone_init_flags(sizeof(xfs_icluster_t), "xfs_icluster", | ||
117 | KM_ZONE_SPREAD, NULL); | ||
118 | 115 | ||
119 | /* | 116 | /* |
120 | * Allocate global trace buffers. | 117 | * Allocate global trace buffers. |
@@ -152,7 +149,6 @@ xfs_cleanup(void) | |||
152 | extern kmem_zone_t *xfs_inode_zone; | 149 | extern kmem_zone_t *xfs_inode_zone; |
153 | extern kmem_zone_t *xfs_efd_zone; | 150 | extern kmem_zone_t *xfs_efd_zone; |
154 | extern kmem_zone_t *xfs_efi_zone; | 151 | extern kmem_zone_t *xfs_efi_zone; |
155 | extern kmem_zone_t *xfs_icluster_zone; | ||
156 | 152 | ||
157 | xfs_cleanup_procfs(); | 153 | xfs_cleanup_procfs(); |
158 | xfs_sysctl_unregister(); | 154 | xfs_sysctl_unregister(); |
@@ -187,7 +183,6 @@ xfs_cleanup(void) | |||
187 | kmem_zone_destroy(xfs_efi_zone); | 183 | kmem_zone_destroy(xfs_efi_zone); |
188 | kmem_zone_destroy(xfs_ifork_zone); | 184 | kmem_zone_destroy(xfs_ifork_zone); |
189 | kmem_zone_destroy(xfs_ili_zone); | 185 | kmem_zone_destroy(xfs_ili_zone); |
190 | kmem_zone_destroy(xfs_icluster_zone); | ||
191 | } | 186 | } |
192 | 187 | ||
193 | /* | 188 | /* |