diff options
| author | Linus Torvalds <torvalds@linux-foundation.org> | 2008-01-25 11:39:18 -0500 |
|---|---|---|
| committer | Linus Torvalds <torvalds@linux-foundation.org> | 2008-01-25 11:39:18 -0500 |
| commit | e07dd2ad305f6b29b47d713600aa8b722ef2a9f7 (patch) | |
| tree | 4815808e538ec625bf2766b1eb9d91c7b3beaead | |
| parent | eba0e319c12fb098d66316a8eafbaaa9174a07c3 (diff) | |
| parent | 7bc5c414fe6627ec518c82d154c796f0981f5b02 (diff) | |
Merge git://git.kernel.org/pub/scm/linux/kernel/git/steve/gfs2-2.6-nmw
* git://git.kernel.org/pub/scm/linux/kernel/git/steve/gfs2-2.6-nmw: (56 commits)
[GFS2] Allow journal recovery on read-only mount
[GFS2] Lockup on error
[GFS2] Fix page_mkwrite truncation race path
[GFS2] Fix typo
[GFS2] Fix write alloc required shortcut calculation
[GFS2] gfs2_alloc_required performance
[GFS2] Remove unneeded i_spin
[GFS2] Reduce inode size by moving i_alloc out of line
[GFS2] Fix assert in log code
[GFS2] Fix problems relating to execution of files on GFS2
[GFS2] Initialize extent_list earlier
[GFS2] Allow page migration for writeback and ordered pages
[GFS2] Remove unused variable
[GFS2] Fix log block mapper
[GFS2] Minor correction
[GFS2] Eliminate the no longer needed sd_statfs_mutex
[GFS2] Incremental patch to fix compiler warning
[GFS2] Function meta_read optimization
[GFS2] Only fetch the dinode once in block_map
[GFS2] Reorganize function gfs2_glmutex_lock
...
40 files changed, 1092 insertions, 1109 deletions
diff --git a/fs/gfs2/Makefile b/fs/gfs2/Makefile index 04ad0caebedb..8fff11058cee 100644 --- a/fs/gfs2/Makefile +++ b/fs/gfs2/Makefile | |||
| @@ -2,7 +2,7 @@ obj-$(CONFIG_GFS2_FS) += gfs2.o | |||
| 2 | gfs2-y := acl.o bmap.o daemon.o dir.o eaops.o eattr.o glock.o \ | 2 | gfs2-y := acl.o bmap.o daemon.o dir.o eaops.o eattr.o glock.o \ |
| 3 | glops.o inode.o lm.o log.o lops.o locking.o main.o meta_io.o \ | 3 | glops.o inode.o lm.o log.o lops.o locking.o main.o meta_io.o \ |
| 4 | mount.o ops_address.o ops_dentry.o ops_export.o ops_file.o \ | 4 | mount.o ops_address.o ops_dentry.o ops_export.o ops_file.o \ |
| 5 | ops_fstype.o ops_inode.o ops_super.o ops_vm.o quota.o \ | 5 | ops_fstype.o ops_inode.o ops_super.o quota.o \ |
| 6 | recovery.o rgrp.o super.o sys.o trans.o util.o | 6 | recovery.o rgrp.o super.o sys.o trans.o util.o |
| 7 | 7 | ||
| 8 | obj-$(CONFIG_GFS2_FS_LOCKING_NOLOCK) += locking/nolock/ | 8 | obj-$(CONFIG_GFS2_FS_LOCKING_NOLOCK) += locking/nolock/ |
diff --git a/fs/gfs2/bmap.c b/fs/gfs2/bmap.c index 93fa427bb5f5..e4effc47abfc 100644 --- a/fs/gfs2/bmap.c +++ b/fs/gfs2/bmap.c | |||
| @@ -59,7 +59,6 @@ struct strip_mine { | |||
| 59 | static int gfs2_unstuffer_page(struct gfs2_inode *ip, struct buffer_head *dibh, | 59 | static int gfs2_unstuffer_page(struct gfs2_inode *ip, struct buffer_head *dibh, |
| 60 | u64 block, struct page *page) | 60 | u64 block, struct page *page) |
| 61 | { | 61 | { |
| 62 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | ||
| 63 | struct inode *inode = &ip->i_inode; | 62 | struct inode *inode = &ip->i_inode; |
| 64 | struct buffer_head *bh; | 63 | struct buffer_head *bh; |
| 65 | int release = 0; | 64 | int release = 0; |
| @@ -95,7 +94,7 @@ static int gfs2_unstuffer_page(struct gfs2_inode *ip, struct buffer_head *dibh, | |||
| 95 | set_buffer_uptodate(bh); | 94 | set_buffer_uptodate(bh); |
| 96 | if (!gfs2_is_jdata(ip)) | 95 | if (!gfs2_is_jdata(ip)) |
| 97 | mark_buffer_dirty(bh); | 96 | mark_buffer_dirty(bh); |
| 98 | if (sdp->sd_args.ar_data == GFS2_DATA_ORDERED || gfs2_is_jdata(ip)) | 97 | if (!gfs2_is_writeback(ip)) |
| 99 | gfs2_trans_add_bh(ip->i_gl, bh, 0); | 98 | gfs2_trans_add_bh(ip->i_gl, bh, 0); |
| 100 | 99 | ||
| 101 | if (release) { | 100 | if (release) { |
| @@ -453,8 +452,8 @@ static inline void bmap_unlock(struct inode *inode, int create) | |||
| 453 | * Returns: errno | 452 | * Returns: errno |
| 454 | */ | 453 | */ |
| 455 | 454 | ||
| 456 | int gfs2_block_map(struct inode *inode, u64 lblock, int create, | 455 | int gfs2_block_map(struct inode *inode, sector_t lblock, |
| 457 | struct buffer_head *bh_map) | 456 | struct buffer_head *bh_map, int create) |
| 458 | { | 457 | { |
| 459 | struct gfs2_inode *ip = GFS2_I(inode); | 458 | struct gfs2_inode *ip = GFS2_I(inode); |
| 460 | struct gfs2_sbd *sdp = GFS2_SB(inode); | 459 | struct gfs2_sbd *sdp = GFS2_SB(inode); |
| @@ -470,6 +469,7 @@ int gfs2_block_map(struct inode *inode, u64 lblock, int create, | |||
| 470 | unsigned int maxlen = bh_map->b_size >> inode->i_blkbits; | 469 | unsigned int maxlen = bh_map->b_size >> inode->i_blkbits; |
| 471 | struct metapath mp; | 470 | struct metapath mp; |
| 472 | u64 size; | 471 | u64 size; |
| 472 | struct buffer_head *dibh = NULL; | ||
| 473 | 473 | ||
| 474 | BUG_ON(maxlen == 0); | 474 | BUG_ON(maxlen == 0); |
| 475 | 475 | ||
| @@ -500,6 +500,8 @@ int gfs2_block_map(struct inode *inode, u64 lblock, int create, | |||
| 500 | error = gfs2_meta_inode_buffer(ip, &bh); | 500 | error = gfs2_meta_inode_buffer(ip, &bh); |
| 501 | if (error) | 501 | if (error) |
| 502 | goto out_fail; | 502 | goto out_fail; |
| 503 | dibh = bh; | ||
| 504 | get_bh(dibh); | ||
| 503 | 505 | ||
| 504 | for (x = 0; x < end_of_metadata; x++) { | 506 | for (x = 0; x < end_of_metadata; x++) { |
| 505 | lookup_block(ip, bh, x, &mp, create, &new, &dblock); | 507 | lookup_block(ip, bh, x, &mp, create, &new, &dblock); |
| @@ -518,13 +520,8 @@ int gfs2_block_map(struct inode *inode, u64 lblock, int create, | |||
| 518 | if (boundary) | 520 | if (boundary) |
| 519 | set_buffer_boundary(bh_map); | 521 | set_buffer_boundary(bh_map); |
| 520 | if (new) { | 522 | if (new) { |
| 521 | struct buffer_head *dibh; | 523 | gfs2_trans_add_bh(ip->i_gl, dibh, 1); |
| 522 | error = gfs2_meta_inode_buffer(ip, &dibh); | 524 | gfs2_dinode_out(ip, dibh->b_data); |
| 523 | if (!error) { | ||
| 524 | gfs2_trans_add_bh(ip->i_gl, dibh, 1); | ||
| 525 | gfs2_dinode_out(ip, dibh->b_data); | ||
| 526 | brelse(dibh); | ||
| 527 | } | ||
| 528 | set_buffer_new(bh_map); | 525 | set_buffer_new(bh_map); |
| 529 | goto out_brelse; | 526 | goto out_brelse; |
| 530 | } | 527 | } |
| @@ -545,6 +542,8 @@ out_brelse: | |||
| 545 | out_ok: | 542 | out_ok: |
| 546 | error = 0; | 543 | error = 0; |
| 547 | out_fail: | 544 | out_fail: |
| 545 | if (dibh) | ||
| 546 | brelse(dibh); | ||
| 548 | bmap_unlock(inode, create); | 547 | bmap_unlock(inode, create); |
| 549 | return error; | 548 | return error; |
| 550 | } | 549 | } |
| @@ -560,7 +559,7 @@ int gfs2_extent_map(struct inode *inode, u64 lblock, int *new, u64 *dblock, unsi | |||
| 560 | BUG_ON(!new); | 559 | BUG_ON(!new); |
| 561 | 560 | ||
| 562 | bh.b_size = 1 << (inode->i_blkbits + 5); | 561 | bh.b_size = 1 << (inode->i_blkbits + 5); |
| 563 | ret = gfs2_block_map(inode, lblock, create, &bh); | 562 | ret = gfs2_block_map(inode, lblock, &bh, create); |
| 564 | *extlen = bh.b_size >> inode->i_blkbits; | 563 | *extlen = bh.b_size >> inode->i_blkbits; |
| 565 | *dblock = bh.b_blocknr; | 564 | *dblock = bh.b_blocknr; |
| 566 | if (buffer_new(&bh)) | 565 | if (buffer_new(&bh)) |
| @@ -684,7 +683,7 @@ static int do_strip(struct gfs2_inode *ip, struct buffer_head *dibh, | |||
| 684 | if (metadata) | 683 | if (metadata) |
| 685 | revokes = (height) ? sdp->sd_inptrs : sdp->sd_diptrs; | 684 | revokes = (height) ? sdp->sd_inptrs : sdp->sd_diptrs; |
| 686 | 685 | ||
| 687 | error = gfs2_rindex_hold(sdp, &ip->i_alloc.al_ri_gh); | 686 | error = gfs2_rindex_hold(sdp, &ip->i_alloc->al_ri_gh); |
| 688 | if (error) | 687 | if (error) |
| 689 | return error; | 688 | return error; |
| 690 | 689 | ||
| @@ -786,7 +785,7 @@ out_rg_gunlock: | |||
| 786 | out_rlist: | 785 | out_rlist: |
| 787 | gfs2_rlist_free(&rlist); | 786 | gfs2_rlist_free(&rlist); |
| 788 | out: | 787 | out: |
| 789 | gfs2_glock_dq_uninit(&ip->i_alloc.al_ri_gh); | 788 | gfs2_glock_dq_uninit(&ip->i_alloc->al_ri_gh); |
| 790 | return error; | 789 | return error; |
| 791 | } | 790 | } |
| 792 | 791 | ||
| @@ -879,7 +878,6 @@ static int gfs2_block_truncate_page(struct address_space *mapping) | |||
| 879 | { | 878 | { |
| 880 | struct inode *inode = mapping->host; | 879 | struct inode *inode = mapping->host; |
| 881 | struct gfs2_inode *ip = GFS2_I(inode); | 880 | struct gfs2_inode *ip = GFS2_I(inode); |
| 882 | struct gfs2_sbd *sdp = GFS2_SB(inode); | ||
| 883 | loff_t from = inode->i_size; | 881 | loff_t from = inode->i_size; |
| 884 | unsigned long index = from >> PAGE_CACHE_SHIFT; | 882 | unsigned long index = from >> PAGE_CACHE_SHIFT; |
| 885 | unsigned offset = from & (PAGE_CACHE_SIZE-1); | 883 | unsigned offset = from & (PAGE_CACHE_SIZE-1); |
| @@ -911,7 +909,7 @@ static int gfs2_block_truncate_page(struct address_space *mapping) | |||
| 911 | err = 0; | 909 | err = 0; |
| 912 | 910 | ||
| 913 | if (!buffer_mapped(bh)) { | 911 | if (!buffer_mapped(bh)) { |
| 914 | gfs2_get_block(inode, iblock, bh, 0); | 912 | gfs2_block_map(inode, iblock, bh, 0); |
| 915 | /* unmapped? It's a hole - nothing to do */ | 913 | /* unmapped? It's a hole - nothing to do */ |
| 916 | if (!buffer_mapped(bh)) | 914 | if (!buffer_mapped(bh)) |
| 917 | goto unlock; | 915 | goto unlock; |
| @@ -931,7 +929,7 @@ static int gfs2_block_truncate_page(struct address_space *mapping) | |||
| 931 | err = 0; | 929 | err = 0; |
| 932 | } | 930 | } |
| 933 | 931 | ||
| 934 | if (sdp->sd_args.ar_data == GFS2_DATA_ORDERED || gfs2_is_jdata(ip)) | 932 | if (!gfs2_is_writeback(ip)) |
| 935 | gfs2_trans_add_bh(ip->i_gl, bh, 0); | 933 | gfs2_trans_add_bh(ip->i_gl, bh, 0); |
| 936 | 934 | ||
| 937 | zero_user_page(page, offset, length, KM_USER0); | 935 | zero_user_page(page, offset, length, KM_USER0); |
| @@ -1224,8 +1222,13 @@ int gfs2_write_alloc_required(struct gfs2_inode *ip, u64 offset, | |||
| 1224 | do_div(lblock_stop, bsize); | 1222 | do_div(lblock_stop, bsize); |
| 1225 | } else { | 1223 | } else { |
| 1226 | unsigned int shift = sdp->sd_sb.sb_bsize_shift; | 1224 | unsigned int shift = sdp->sd_sb.sb_bsize_shift; |
| 1225 | u64 end_of_file = (ip->i_di.di_size + sdp->sd_sb.sb_bsize - 1) >> shift; | ||
| 1227 | lblock = offset >> shift; | 1226 | lblock = offset >> shift; |
| 1228 | lblock_stop = (offset + len + sdp->sd_sb.sb_bsize - 1) >> shift; | 1227 | lblock_stop = (offset + len + sdp->sd_sb.sb_bsize - 1) >> shift; |
| 1228 | if (lblock_stop > end_of_file) { | ||
| 1229 | *alloc_required = 1; | ||
| 1230 | return 0; | ||
| 1231 | } | ||
| 1229 | } | 1232 | } |
| 1230 | 1233 | ||
| 1231 | for (; lblock < lblock_stop; lblock += extlen) { | 1234 | for (; lblock < lblock_stop; lblock += extlen) { |
diff --git a/fs/gfs2/bmap.h b/fs/gfs2/bmap.h index ac2fd04370dc..4e6cde2943bd 100644 --- a/fs/gfs2/bmap.h +++ b/fs/gfs2/bmap.h | |||
| @@ -15,7 +15,7 @@ struct gfs2_inode; | |||
| 15 | struct page; | 15 | struct page; |
| 16 | 16 | ||
| 17 | int gfs2_unstuff_dinode(struct gfs2_inode *ip, struct page *page); | 17 | int gfs2_unstuff_dinode(struct gfs2_inode *ip, struct page *page); |
| 18 | int gfs2_block_map(struct inode *inode, u64 lblock, int create, struct buffer_head *bh); | 18 | int gfs2_block_map(struct inode *inode, sector_t lblock, struct buffer_head *bh, int create); |
| 19 | int gfs2_extent_map(struct inode *inode, u64 lblock, int *new, u64 *dblock, unsigned *extlen); | 19 | int gfs2_extent_map(struct inode *inode, u64 lblock, int *new, u64 *dblock, unsigned *extlen); |
| 20 | 20 | ||
| 21 | int gfs2_truncatei(struct gfs2_inode *ip, u64 size); | 21 | int gfs2_truncatei(struct gfs2_inode *ip, u64 size); |
diff --git a/fs/gfs2/daemon.c b/fs/gfs2/daemon.c index 3731ab0771d5..e51991947d2c 100644 --- a/fs/gfs2/daemon.c +++ b/fs/gfs2/daemon.c | |||
| @@ -83,56 +83,6 @@ int gfs2_recoverd(void *data) | |||
| 83 | } | 83 | } |
| 84 | 84 | ||
| 85 | /** | 85 | /** |
| 86 | * gfs2_logd - Update log tail as Active Items get flushed to in-place blocks | ||
| 87 | * @sdp: Pointer to GFS2 superblock | ||
| 88 | * | ||
| 89 | * Also, periodically check to make sure that we're using the most recent | ||
| 90 | * journal index. | ||
| 91 | */ | ||
| 92 | |||
| 93 | int gfs2_logd(void *data) | ||
| 94 | { | ||
| 95 | struct gfs2_sbd *sdp = data; | ||
| 96 | struct gfs2_holder ji_gh; | ||
| 97 | unsigned long t; | ||
| 98 | int need_flush; | ||
| 99 | |||
| 100 | while (!kthread_should_stop()) { | ||
| 101 | /* Advance the log tail */ | ||
| 102 | |||
| 103 | t = sdp->sd_log_flush_time + | ||
| 104 | gfs2_tune_get(sdp, gt_log_flush_secs) * HZ; | ||
| 105 | |||
| 106 | gfs2_ail1_empty(sdp, DIO_ALL); | ||
| 107 | gfs2_log_lock(sdp); | ||
| 108 | need_flush = sdp->sd_log_num_buf > gfs2_tune_get(sdp, gt_incore_log_blocks); | ||
| 109 | gfs2_log_unlock(sdp); | ||
| 110 | if (need_flush || time_after_eq(jiffies, t)) { | ||
| 111 | gfs2_log_flush(sdp, NULL); | ||
| 112 | sdp->sd_log_flush_time = jiffies; | ||
| 113 | } | ||
| 114 | |||
| 115 | /* Check for latest journal index */ | ||
| 116 | |||
| 117 | t = sdp->sd_jindex_refresh_time + | ||
| 118 | gfs2_tune_get(sdp, gt_jindex_refresh_secs) * HZ; | ||
| 119 | |||
| 120 | if (time_after_eq(jiffies, t)) { | ||
| 121 | if (!gfs2_jindex_hold(sdp, &ji_gh)) | ||
| 122 | gfs2_glock_dq_uninit(&ji_gh); | ||
| 123 | sdp->sd_jindex_refresh_time = jiffies; | ||
| 124 | } | ||
| 125 | |||
| 126 | t = gfs2_tune_get(sdp, gt_logd_secs) * HZ; | ||
| 127 | if (freezing(current)) | ||
| 128 | refrigerator(); | ||
| 129 | schedule_timeout_interruptible(t); | ||
| 130 | } | ||
| 131 | |||
| 132 | return 0; | ||
| 133 | } | ||
| 134 | |||
| 135 | /** | ||
| 136 | * gfs2_quotad - Write cached quota changes into the quota file | 86 | * gfs2_quotad - Write cached quota changes into the quota file |
| 137 | * @sdp: Pointer to GFS2 superblock | 87 | * @sdp: Pointer to GFS2 superblock |
| 138 | * | 88 | * |
diff --git a/fs/gfs2/daemon.h b/fs/gfs2/daemon.h index 0de9b3557955..4be084fb6a62 100644 --- a/fs/gfs2/daemon.h +++ b/fs/gfs2/daemon.h | |||
| @@ -12,7 +12,6 @@ | |||
| 12 | 12 | ||
| 13 | int gfs2_glockd(void *data); | 13 | int gfs2_glockd(void *data); |
| 14 | int gfs2_recoverd(void *data); | 14 | int gfs2_recoverd(void *data); |
| 15 | int gfs2_logd(void *data); | ||
| 16 | int gfs2_quotad(void *data); | 15 | int gfs2_quotad(void *data); |
| 17 | 16 | ||
| 18 | #endif /* __DAEMON_DOT_H__ */ | 17 | #endif /* __DAEMON_DOT_H__ */ |
diff --git a/fs/gfs2/dir.c b/fs/gfs2/dir.c index 9949bb746a52..57e2ed932adc 100644 --- a/fs/gfs2/dir.c +++ b/fs/gfs2/dir.c | |||
| @@ -1876,7 +1876,7 @@ static int leaf_dealloc(struct gfs2_inode *dip, u32 index, u32 len, | |||
| 1876 | if (error) | 1876 | if (error) |
| 1877 | goto out; | 1877 | goto out; |
| 1878 | 1878 | ||
| 1879 | error = gfs2_rindex_hold(sdp, &dip->i_alloc.al_ri_gh); | 1879 | error = gfs2_rindex_hold(sdp, &dip->i_alloc->al_ri_gh); |
| 1880 | if (error) | 1880 | if (error) |
| 1881 | goto out_qs; | 1881 | goto out_qs; |
| 1882 | 1882 | ||
| @@ -1949,7 +1949,7 @@ out_rg_gunlock: | |||
| 1949 | gfs2_glock_dq_m(rlist.rl_rgrps, rlist.rl_ghs); | 1949 | gfs2_glock_dq_m(rlist.rl_rgrps, rlist.rl_ghs); |
| 1950 | out_rlist: | 1950 | out_rlist: |
| 1951 | gfs2_rlist_free(&rlist); | 1951 | gfs2_rlist_free(&rlist); |
| 1952 | gfs2_glock_dq_uninit(&dip->i_alloc.al_ri_gh); | 1952 | gfs2_glock_dq_uninit(&dip->i_alloc->al_ri_gh); |
| 1953 | out_qs: | 1953 | out_qs: |
| 1954 | gfs2_quota_unhold(dip); | 1954 | gfs2_quota_unhold(dip); |
| 1955 | out: | 1955 | out: |
diff --git a/fs/gfs2/eaops.c b/fs/gfs2/eaops.c index aa8dbf303f6d..f114ba2b3557 100644 --- a/fs/gfs2/eaops.c +++ b/fs/gfs2/eaops.c | |||
| @@ -56,46 +56,6 @@ unsigned int gfs2_ea_name2type(const char *name, const char **truncated_name) | |||
| 56 | return type; | 56 | return type; |
| 57 | } | 57 | } |
| 58 | 58 | ||
| 59 | static int user_eo_get(struct gfs2_inode *ip, struct gfs2_ea_request *er) | ||
| 60 | { | ||
| 61 | struct inode *inode = &ip->i_inode; | ||
| 62 | int error = permission(inode, MAY_READ, NULL); | ||
| 63 | if (error) | ||
| 64 | return error; | ||
| 65 | |||
| 66 | return gfs2_ea_get_i(ip, er); | ||
| 67 | } | ||
| 68 | |||
| 69 | static int user_eo_set(struct gfs2_inode *ip, struct gfs2_ea_request *er) | ||
| 70 | { | ||
| 71 | struct inode *inode = &ip->i_inode; | ||
| 72 | |||
| 73 | if (S_ISREG(inode->i_mode) || | ||
| 74 | (S_ISDIR(inode->i_mode) && !(inode->i_mode & S_ISVTX))) { | ||
| 75 | int error = permission(inode, MAY_WRITE, NULL); | ||
| 76 | if (error) | ||
| 77 | return error; | ||
| 78 | } else | ||
| 79 | return -EPERM; | ||
| 80 | |||
| 81 | return gfs2_ea_set_i(ip, er); | ||
| 82 | } | ||
| 83 | |||
| 84 | static int user_eo_remove(struct gfs2_inode *ip, struct gfs2_ea_request *er) | ||
| 85 | { | ||
| 86 | struct inode *inode = &ip->i_inode; | ||
| 87 | |||
| 88 | if (S_ISREG(inode->i_mode) || | ||
| 89 | (S_ISDIR(inode->i_mode) && !(inode->i_mode & S_ISVTX))) { | ||
| 90 | int error = permission(inode, MAY_WRITE, NULL); | ||
| 91 | if (error) | ||
| 92 | return error; | ||
| 93 | } else | ||
| 94 | return -EPERM; | ||
| 95 | |||
| 96 | return gfs2_ea_remove_i(ip, er); | ||
| 97 | } | ||
| 98 | |||
| 99 | static int system_eo_get(struct gfs2_inode *ip, struct gfs2_ea_request *er) | 59 | static int system_eo_get(struct gfs2_inode *ip, struct gfs2_ea_request *er) |
| 100 | { | 60 | { |
| 101 | if (!GFS2_ACL_IS_ACCESS(er->er_name, er->er_name_len) && | 61 | if (!GFS2_ACL_IS_ACCESS(er->er_name, er->er_name_len) && |
| @@ -108,8 +68,6 @@ static int system_eo_get(struct gfs2_inode *ip, struct gfs2_ea_request *er) | |||
| 108 | GFS2_ACL_IS_DEFAULT(er->er_name, er->er_name_len))) | 68 | GFS2_ACL_IS_DEFAULT(er->er_name, er->er_name_len))) |
| 109 | return -EOPNOTSUPP; | 69 | return -EOPNOTSUPP; |
| 110 | 70 | ||
| 111 | |||
| 112 | |||
| 113 | return gfs2_ea_get_i(ip, er); | 71 | return gfs2_ea_get_i(ip, er); |
| 114 | } | 72 | } |
| 115 | 73 | ||
| @@ -170,40 +128,10 @@ static int system_eo_remove(struct gfs2_inode *ip, struct gfs2_ea_request *er) | |||
| 170 | return gfs2_ea_remove_i(ip, er); | 128 | return gfs2_ea_remove_i(ip, er); |
| 171 | } | 129 | } |
| 172 | 130 | ||
| 173 | static int security_eo_get(struct gfs2_inode *ip, struct gfs2_ea_request *er) | ||
| 174 | { | ||
| 175 | struct inode *inode = &ip->i_inode; | ||
| 176 | int error = permission(inode, MAY_READ, NULL); | ||
| 177 | if (error) | ||
| 178 | return error; | ||
| 179 | |||
| 180 | return gfs2_ea_get_i(ip, er); | ||
| 181 | } | ||
| 182 | |||
| 183 | static int security_eo_set(struct gfs2_inode *ip, struct gfs2_ea_request *er) | ||
| 184 | { | ||
| 185 | struct inode *inode = &ip->i_inode; | ||
| 186 | int error = permission(inode, MAY_WRITE, NULL); | ||
| 187 | if (error) | ||
| 188 | return error; | ||
| 189 | |||
| 190 | return gfs2_ea_set_i(ip, er); | ||
| 191 | } | ||
| 192 | |||
| 193 | static int security_eo_remove(struct gfs2_inode *ip, struct gfs2_ea_request *er) | ||
| 194 | { | ||
| 195 | struct inode *inode = &ip->i_inode; | ||
| 196 | int error = permission(inode, MAY_WRITE, NULL); | ||
| 197 | if (error) | ||
| 198 | return error; | ||
| 199 | |||
| 200 | return gfs2_ea_remove_i(ip, er); | ||
| 201 | } | ||
| 202 | |||
| 203 | static const struct gfs2_eattr_operations gfs2_user_eaops = { | 131 | static const struct gfs2_eattr_operations gfs2_user_eaops = { |
| 204 | .eo_get = user_eo_get, | 132 | .eo_get = gfs2_ea_get_i, |
| 205 | .eo_set = user_eo_set, | 133 | .eo_set = gfs2_ea_set_i, |
| 206 | .eo_remove = user_eo_remove, | 134 | .eo_remove = gfs2_ea_remove_i, |
| 207 | .eo_name = "user", | 135 | .eo_name = "user", |
| 208 | }; | 136 | }; |
| 209 | 137 | ||
| @@ -215,9 +143,9 @@ const struct gfs2_eattr_operations gfs2_system_eaops = { | |||
| 215 | }; | 143 | }; |
| 216 | 144 | ||
| 217 | static const struct gfs2_eattr_operations gfs2_security_eaops = { | 145 | static const struct gfs2_eattr_operations gfs2_security_eaops = { |
| 218 | .eo_get = security_eo_get, | 146 | .eo_get = gfs2_ea_get_i, |
| 219 | .eo_set = security_eo_set, | 147 | .eo_set = gfs2_ea_set_i, |
| 220 | .eo_remove = security_eo_remove, | 148 | .eo_remove = gfs2_ea_remove_i, |
| 221 | .eo_name = "security", | 149 | .eo_name = "security", |
| 222 | }; | 150 | }; |
| 223 | 151 | ||
diff --git a/fs/gfs2/eattr.c b/fs/gfs2/eattr.c index 2a7435b5c4dc..bee99704ea10 100644 --- a/fs/gfs2/eattr.c +++ b/fs/gfs2/eattr.c | |||
| @@ -1418,7 +1418,7 @@ out: | |||
| 1418 | static int ea_dealloc_block(struct gfs2_inode *ip) | 1418 | static int ea_dealloc_block(struct gfs2_inode *ip) |
| 1419 | { | 1419 | { |
| 1420 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 1420 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
| 1421 | struct gfs2_alloc *al = &ip->i_alloc; | 1421 | struct gfs2_alloc *al = ip->i_alloc; |
| 1422 | struct gfs2_rgrpd *rgd; | 1422 | struct gfs2_rgrpd *rgd; |
| 1423 | struct buffer_head *dibh; | 1423 | struct buffer_head *dibh; |
| 1424 | int error; | 1424 | int error; |
diff --git a/fs/gfs2/glock.c b/fs/gfs2/glock.c index a37efe4aae6f..80e09c50590a 100644 --- a/fs/gfs2/glock.c +++ b/fs/gfs2/glock.c | |||
| @@ -1,6 +1,6 @@ | |||
| 1 | /* | 1 | /* |
| 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. |
| 3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | 3 | * Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved. |
| 4 | * | 4 | * |
| 5 | * This copyrighted material is made available to anyone wishing to use, | 5 | * This copyrighted material is made available to anyone wishing to use, |
| 6 | * modify, copy, or redistribute it subject to the terms and conditions | 6 | * modify, copy, or redistribute it subject to the terms and conditions |
| @@ -217,7 +217,6 @@ int gfs2_glock_put(struct gfs2_glock *gl) | |||
| 217 | if (atomic_dec_and_test(&gl->gl_ref)) { | 217 | if (atomic_dec_and_test(&gl->gl_ref)) { |
| 218 | hlist_del(&gl->gl_list); | 218 | hlist_del(&gl->gl_list); |
| 219 | write_unlock(gl_lock_addr(gl->gl_hash)); | 219 | write_unlock(gl_lock_addr(gl->gl_hash)); |
| 220 | BUG_ON(spin_is_locked(&gl->gl_spin)); | ||
| 221 | gfs2_assert(sdp, gl->gl_state == LM_ST_UNLOCKED); | 220 | gfs2_assert(sdp, gl->gl_state == LM_ST_UNLOCKED); |
| 222 | gfs2_assert(sdp, list_empty(&gl->gl_reclaim)); | 221 | gfs2_assert(sdp, list_empty(&gl->gl_reclaim)); |
| 223 | gfs2_assert(sdp, list_empty(&gl->gl_holders)); | 222 | gfs2_assert(sdp, list_empty(&gl->gl_holders)); |
| @@ -346,7 +345,6 @@ int gfs2_glock_get(struct gfs2_sbd *sdp, u64 number, | |||
| 346 | gl->gl_object = NULL; | 345 | gl->gl_object = NULL; |
| 347 | gl->gl_sbd = sdp; | 346 | gl->gl_sbd = sdp; |
| 348 | gl->gl_aspace = NULL; | 347 | gl->gl_aspace = NULL; |
| 349 | lops_init_le(&gl->gl_le, &gfs2_glock_lops); | ||
| 350 | INIT_DELAYED_WORK(&gl->gl_work, glock_work_func); | 348 | INIT_DELAYED_WORK(&gl->gl_work, glock_work_func); |
| 351 | 349 | ||
| 352 | /* If this glock protects actual on-disk data or metadata blocks, | 350 | /* If this glock protects actual on-disk data or metadata blocks, |
| @@ -461,7 +459,6 @@ static void wait_on_holder(struct gfs2_holder *gh) | |||
| 461 | 459 | ||
| 462 | static void gfs2_demote_wake(struct gfs2_glock *gl) | 460 | static void gfs2_demote_wake(struct gfs2_glock *gl) |
| 463 | { | 461 | { |
| 464 | BUG_ON(!spin_is_locked(&gl->gl_spin)); | ||
| 465 | gl->gl_demote_state = LM_ST_EXCLUSIVE; | 462 | gl->gl_demote_state = LM_ST_EXCLUSIVE; |
| 466 | clear_bit(GLF_DEMOTE, &gl->gl_flags); | 463 | clear_bit(GLF_DEMOTE, &gl->gl_flags); |
| 467 | smp_mb__after_clear_bit(); | 464 | smp_mb__after_clear_bit(); |
| @@ -507,21 +504,12 @@ static int rq_mutex(struct gfs2_holder *gh) | |||
| 507 | static int rq_promote(struct gfs2_holder *gh) | 504 | static int rq_promote(struct gfs2_holder *gh) |
| 508 | { | 505 | { |
| 509 | struct gfs2_glock *gl = gh->gh_gl; | 506 | struct gfs2_glock *gl = gh->gh_gl; |
| 510 | struct gfs2_sbd *sdp = gl->gl_sbd; | ||
| 511 | 507 | ||
| 512 | if (!relaxed_state_ok(gl->gl_state, gh->gh_state, gh->gh_flags)) { | 508 | if (!relaxed_state_ok(gl->gl_state, gh->gh_state, gh->gh_flags)) { |
| 513 | if (list_empty(&gl->gl_holders)) { | 509 | if (list_empty(&gl->gl_holders)) { |
| 514 | gl->gl_req_gh = gh; | 510 | gl->gl_req_gh = gh; |
| 515 | set_bit(GLF_LOCK, &gl->gl_flags); | 511 | set_bit(GLF_LOCK, &gl->gl_flags); |
| 516 | spin_unlock(&gl->gl_spin); | 512 | spin_unlock(&gl->gl_spin); |
| 517 | |||
| 518 | if (atomic_read(&sdp->sd_reclaim_count) > | ||
| 519 | gfs2_tune_get(sdp, gt_reclaim_limit) && | ||
| 520 | !(gh->gh_flags & LM_FLAG_PRIORITY)) { | ||
| 521 | gfs2_reclaim_glock(sdp); | ||
| 522 | gfs2_reclaim_glock(sdp); | ||
| 523 | } | ||
| 524 | |||
| 525 | gfs2_glock_xmote_th(gh->gh_gl, gh); | 513 | gfs2_glock_xmote_th(gh->gh_gl, gh); |
| 526 | spin_lock(&gl->gl_spin); | 514 | spin_lock(&gl->gl_spin); |
| 527 | } | 515 | } |
| @@ -567,7 +555,10 @@ static int rq_demote(struct gfs2_glock *gl) | |||
| 567 | gfs2_demote_wake(gl); | 555 | gfs2_demote_wake(gl); |
| 568 | return 0; | 556 | return 0; |
| 569 | } | 557 | } |
| 558 | |||
| 570 | set_bit(GLF_LOCK, &gl->gl_flags); | 559 | set_bit(GLF_LOCK, &gl->gl_flags); |
| 560 | set_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags); | ||
| 561 | |||
| 571 | if (gl->gl_demote_state == LM_ST_UNLOCKED || | 562 | if (gl->gl_demote_state == LM_ST_UNLOCKED || |
| 572 | gl->gl_state != LM_ST_EXCLUSIVE) { | 563 | gl->gl_state != LM_ST_EXCLUSIVE) { |
| 573 | spin_unlock(&gl->gl_spin); | 564 | spin_unlock(&gl->gl_spin); |
| @@ -576,7 +567,9 @@ static int rq_demote(struct gfs2_glock *gl) | |||
| 576 | spin_unlock(&gl->gl_spin); | 567 | spin_unlock(&gl->gl_spin); |
| 577 | gfs2_glock_xmote_th(gl, NULL); | 568 | gfs2_glock_xmote_th(gl, NULL); |
| 578 | } | 569 | } |
| 570 | |||
| 579 | spin_lock(&gl->gl_spin); | 571 | spin_lock(&gl->gl_spin); |
| 572 | clear_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags); | ||
| 580 | 573 | ||
| 581 | return 0; | 574 | return 0; |
| 582 | } | 575 | } |
| @@ -598,23 +591,18 @@ static void run_queue(struct gfs2_glock *gl) | |||
| 598 | if (!list_empty(&gl->gl_waiters1)) { | 591 | if (!list_empty(&gl->gl_waiters1)) { |
| 599 | gh = list_entry(gl->gl_waiters1.next, | 592 | gh = list_entry(gl->gl_waiters1.next, |
| 600 | struct gfs2_holder, gh_list); | 593 | struct gfs2_holder, gh_list); |
| 601 | 594 | blocked = rq_mutex(gh); | |
| 602 | if (test_bit(HIF_MUTEX, &gh->gh_iflags)) | ||
| 603 | blocked = rq_mutex(gh); | ||
| 604 | else | ||
| 605 | gfs2_assert_warn(gl->gl_sbd, 0); | ||
| 606 | |||
| 607 | } else if (test_bit(GLF_DEMOTE, &gl->gl_flags)) { | 595 | } else if (test_bit(GLF_DEMOTE, &gl->gl_flags)) { |
| 608 | blocked = rq_demote(gl); | 596 | blocked = rq_demote(gl); |
| 597 | if (gl->gl_waiters2 && !blocked) { | ||
| 598 | set_bit(GLF_DEMOTE, &gl->gl_flags); | ||
| 599 | gl->gl_demote_state = LM_ST_UNLOCKED; | ||
| 600 | } | ||
| 601 | gl->gl_waiters2 = 0; | ||
| 609 | } else if (!list_empty(&gl->gl_waiters3)) { | 602 | } else if (!list_empty(&gl->gl_waiters3)) { |
| 610 | gh = list_entry(gl->gl_waiters3.next, | 603 | gh = list_entry(gl->gl_waiters3.next, |
| 611 | struct gfs2_holder, gh_list); | 604 | struct gfs2_holder, gh_list); |
| 612 | 605 | blocked = rq_promote(gh); | |
| 613 | if (test_bit(HIF_PROMOTE, &gh->gh_iflags)) | ||
| 614 | blocked = rq_promote(gh); | ||
| 615 | else | ||
| 616 | gfs2_assert_warn(gl->gl_sbd, 0); | ||
| 617 | |||
| 618 | } else | 606 | } else |
| 619 | break; | 607 | break; |
| 620 | 608 | ||
| @@ -632,27 +620,21 @@ static void run_queue(struct gfs2_glock *gl) | |||
| 632 | 620 | ||
| 633 | static void gfs2_glmutex_lock(struct gfs2_glock *gl) | 621 | static void gfs2_glmutex_lock(struct gfs2_glock *gl) |
| 634 | { | 622 | { |
| 635 | struct gfs2_holder gh; | ||
| 636 | |||
| 637 | gfs2_holder_init(gl, 0, 0, &gh); | ||
| 638 | set_bit(HIF_MUTEX, &gh.gh_iflags); | ||
| 639 | if (test_and_set_bit(HIF_WAIT, &gh.gh_iflags)) | ||
| 640 | BUG(); | ||
| 641 | |||
| 642 | spin_lock(&gl->gl_spin); | 623 | spin_lock(&gl->gl_spin); |
| 643 | if (test_and_set_bit(GLF_LOCK, &gl->gl_flags)) { | 624 | if (test_and_set_bit(GLF_LOCK, &gl->gl_flags)) { |
| 625 | struct gfs2_holder gh; | ||
| 626 | |||
| 627 | gfs2_holder_init(gl, 0, 0, &gh); | ||
| 628 | set_bit(HIF_WAIT, &gh.gh_iflags); | ||
| 644 | list_add_tail(&gh.gh_list, &gl->gl_waiters1); | 629 | list_add_tail(&gh.gh_list, &gl->gl_waiters1); |
| 630 | spin_unlock(&gl->gl_spin); | ||
| 631 | wait_on_holder(&gh); | ||
| 632 | gfs2_holder_uninit(&gh); | ||
| 645 | } else { | 633 | } else { |
| 646 | gl->gl_owner_pid = current->pid; | 634 | gl->gl_owner_pid = current->pid; |
| 647 | gl->gl_ip = (unsigned long)__builtin_return_address(0); | 635 | gl->gl_ip = (unsigned long)__builtin_return_address(0); |
| 648 | clear_bit(HIF_WAIT, &gh.gh_iflags); | 636 | spin_unlock(&gl->gl_spin); |
| 649 | smp_mb(); | ||
| 650 | wake_up_bit(&gh.gh_iflags, HIF_WAIT); | ||
| 651 | } | 637 | } |
| 652 | spin_unlock(&gl->gl_spin); | ||
| 653 | |||
| 654 | wait_on_holder(&gh); | ||
| 655 | gfs2_holder_uninit(&gh); | ||
| 656 | } | 638 | } |
| 657 | 639 | ||
| 658 | /** | 640 | /** |
| @@ -691,7 +673,6 @@ static void gfs2_glmutex_unlock(struct gfs2_glock *gl) | |||
| 691 | gl->gl_owner_pid = 0; | 673 | gl->gl_owner_pid = 0; |
| 692 | gl->gl_ip = 0; | 674 | gl->gl_ip = 0; |
| 693 | run_queue(gl); | 675 | run_queue(gl); |
| 694 | BUG_ON(!spin_is_locked(&gl->gl_spin)); | ||
| 695 | spin_unlock(&gl->gl_spin); | 676 | spin_unlock(&gl->gl_spin); |
| 696 | } | 677 | } |
| 697 | 678 | ||
| @@ -722,7 +703,10 @@ static void handle_callback(struct gfs2_glock *gl, unsigned int state, | |||
| 722 | } | 703 | } |
| 723 | } else if (gl->gl_demote_state != LM_ST_UNLOCKED && | 704 | } else if (gl->gl_demote_state != LM_ST_UNLOCKED && |
| 724 | gl->gl_demote_state != state) { | 705 | gl->gl_demote_state != state) { |
| 725 | gl->gl_demote_state = LM_ST_UNLOCKED; | 706 | if (test_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags)) |
| 707 | gl->gl_waiters2 = 1; | ||
| 708 | else | ||
| 709 | gl->gl_demote_state = LM_ST_UNLOCKED; | ||
| 726 | } | 710 | } |
| 727 | spin_unlock(&gl->gl_spin); | 711 | spin_unlock(&gl->gl_spin); |
| 728 | } | 712 | } |
| @@ -943,8 +927,8 @@ static void gfs2_glock_drop_th(struct gfs2_glock *gl) | |||
| 943 | const struct gfs2_glock_operations *glops = gl->gl_ops; | 927 | const struct gfs2_glock_operations *glops = gl->gl_ops; |
| 944 | unsigned int ret; | 928 | unsigned int ret; |
| 945 | 929 | ||
| 946 | if (glops->go_drop_th) | 930 | if (glops->go_xmote_th) |
| 947 | glops->go_drop_th(gl); | 931 | glops->go_xmote_th(gl); |
| 948 | 932 | ||
| 949 | gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags)); | 933 | gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags)); |
| 950 | gfs2_assert_warn(sdp, list_empty(&gl->gl_holders)); | 934 | gfs2_assert_warn(sdp, list_empty(&gl->gl_holders)); |
| @@ -1156,8 +1140,6 @@ restart: | |||
| 1156 | return -EIO; | 1140 | return -EIO; |
| 1157 | } | 1141 | } |
| 1158 | 1142 | ||
| 1159 | set_bit(HIF_PROMOTE, &gh->gh_iflags); | ||
| 1160 | |||
| 1161 | spin_lock(&gl->gl_spin); | 1143 | spin_lock(&gl->gl_spin); |
| 1162 | add_to_queue(gh); | 1144 | add_to_queue(gh); |
| 1163 | run_queue(gl); | 1145 | run_queue(gl); |
| @@ -1248,12 +1230,11 @@ void gfs2_glock_dq(struct gfs2_holder *gh) | |||
| 1248 | list_del_init(&gh->gh_list); | 1230 | list_del_init(&gh->gh_list); |
| 1249 | 1231 | ||
| 1250 | if (list_empty(&gl->gl_holders)) { | 1232 | if (list_empty(&gl->gl_holders)) { |
| 1251 | spin_unlock(&gl->gl_spin); | 1233 | if (glops->go_unlock) { |
| 1252 | 1234 | spin_unlock(&gl->gl_spin); | |
| 1253 | if (glops->go_unlock) | ||
| 1254 | glops->go_unlock(gh); | 1235 | glops->go_unlock(gh); |
| 1255 | 1236 | spin_lock(&gl->gl_spin); | |
| 1256 | spin_lock(&gl->gl_spin); | 1237 | } |
| 1257 | gl->gl_stamp = jiffies; | 1238 | gl->gl_stamp = jiffies; |
| 1258 | } | 1239 | } |
| 1259 | 1240 | ||
| @@ -1910,8 +1891,6 @@ static int dump_glock(struct glock_iter *gi, struct gfs2_glock *gl) | |||
| 1910 | print_dbg(gi, " req_bh = %s\n", (gl->gl_req_bh) ? "yes" : "no"); | 1891 | print_dbg(gi, " req_bh = %s\n", (gl->gl_req_bh) ? "yes" : "no"); |
| 1911 | print_dbg(gi, " lvb_count = %d\n", atomic_read(&gl->gl_lvb_count)); | 1892 | print_dbg(gi, " lvb_count = %d\n", atomic_read(&gl->gl_lvb_count)); |
| 1912 | print_dbg(gi, " object = %s\n", (gl->gl_object) ? "yes" : "no"); | 1893 | print_dbg(gi, " object = %s\n", (gl->gl_object) ? "yes" : "no"); |
| 1913 | print_dbg(gi, " le = %s\n", | ||
| 1914 | (list_empty(&gl->gl_le.le_list)) ? "no" : "yes"); | ||
| 1915 | print_dbg(gi, " reclaim = %s\n", | 1894 | print_dbg(gi, " reclaim = %s\n", |
| 1916 | (list_empty(&gl->gl_reclaim)) ? "no" : "yes"); | 1895 | (list_empty(&gl->gl_reclaim)) ? "no" : "yes"); |
| 1917 | if (gl->gl_aspace) | 1896 | if (gl->gl_aspace) |
diff --git a/fs/gfs2/glops.c b/fs/gfs2/glops.c index 4670dcb2a877..c663b7a0f410 100644 --- a/fs/gfs2/glops.c +++ b/fs/gfs2/glops.c | |||
| @@ -56,7 +56,7 @@ static void gfs2_ail_empty_gl(struct gfs2_glock *gl) | |||
| 56 | bd = list_entry(head->next, struct gfs2_bufdata, | 56 | bd = list_entry(head->next, struct gfs2_bufdata, |
| 57 | bd_ail_gl_list); | 57 | bd_ail_gl_list); |
| 58 | bh = bd->bd_bh; | 58 | bh = bd->bd_bh; |
| 59 | gfs2_remove_from_ail(NULL, bd); | 59 | gfs2_remove_from_ail(bd); |
| 60 | bd->bd_bh = NULL; | 60 | bd->bd_bh = NULL; |
| 61 | bh->b_private = NULL; | 61 | bh->b_private = NULL; |
| 62 | bd->bd_blkno = bh->b_blocknr; | 62 | bd->bd_blkno = bh->b_blocknr; |
| @@ -86,15 +86,10 @@ static void gfs2_pte_inval(struct gfs2_glock *gl) | |||
| 86 | if (!ip || !S_ISREG(inode->i_mode)) | 86 | if (!ip || !S_ISREG(inode->i_mode)) |
| 87 | return; | 87 | return; |
| 88 | 88 | ||
| 89 | if (!test_bit(GIF_PAGED, &ip->i_flags)) | ||
| 90 | return; | ||
| 91 | |||
| 92 | unmap_shared_mapping_range(inode->i_mapping, 0, 0); | 89 | unmap_shared_mapping_range(inode->i_mapping, 0, 0); |
| 93 | |||
| 94 | if (test_bit(GIF_SW_PAGED, &ip->i_flags)) | 90 | if (test_bit(GIF_SW_PAGED, &ip->i_flags)) |
| 95 | set_bit(GLF_DIRTY, &gl->gl_flags); | 91 | set_bit(GLF_DIRTY, &gl->gl_flags); |
| 96 | 92 | ||
| 97 | clear_bit(GIF_SW_PAGED, &ip->i_flags); | ||
| 98 | } | 93 | } |
| 99 | 94 | ||
| 100 | /** | 95 | /** |
| @@ -143,44 +138,34 @@ static void meta_go_inval(struct gfs2_glock *gl, int flags) | |||
| 143 | static void inode_go_sync(struct gfs2_glock *gl) | 138 | static void inode_go_sync(struct gfs2_glock *gl) |
| 144 | { | 139 | { |
| 145 | struct gfs2_inode *ip = gl->gl_object; | 140 | struct gfs2_inode *ip = gl->gl_object; |
| 141 | struct address_space *metamapping = gl->gl_aspace->i_mapping; | ||
| 142 | int error; | ||
| 143 | |||
| 144 | if (gl->gl_state != LM_ST_UNLOCKED) | ||
| 145 | gfs2_pte_inval(gl); | ||
| 146 | if (gl->gl_state != LM_ST_EXCLUSIVE) | ||
| 147 | return; | ||
| 146 | 148 | ||
| 147 | if (ip && !S_ISREG(ip->i_inode.i_mode)) | 149 | if (ip && !S_ISREG(ip->i_inode.i_mode)) |
| 148 | ip = NULL; | 150 | ip = NULL; |
| 149 | 151 | ||
| 150 | if (test_bit(GLF_DIRTY, &gl->gl_flags)) { | 152 | if (test_bit(GLF_DIRTY, &gl->gl_flags)) { |
| 151 | if (ip && !gfs2_is_jdata(ip)) | ||
| 152 | filemap_fdatawrite(ip->i_inode.i_mapping); | ||
| 153 | gfs2_log_flush(gl->gl_sbd, gl); | 153 | gfs2_log_flush(gl->gl_sbd, gl); |
| 154 | if (ip && gfs2_is_jdata(ip)) | 154 | filemap_fdatawrite(metamapping); |
| 155 | filemap_fdatawrite(ip->i_inode.i_mapping); | ||
| 156 | gfs2_meta_sync(gl); | ||
| 157 | if (ip) { | 155 | if (ip) { |
| 158 | struct address_space *mapping = ip->i_inode.i_mapping; | 156 | struct address_space *mapping = ip->i_inode.i_mapping; |
| 159 | int error = filemap_fdatawait(mapping); | 157 | filemap_fdatawrite(mapping); |
| 158 | error = filemap_fdatawait(mapping); | ||
| 160 | mapping_set_error(mapping, error); | 159 | mapping_set_error(mapping, error); |
| 161 | } | 160 | } |
| 161 | error = filemap_fdatawait(metamapping); | ||
| 162 | mapping_set_error(metamapping, error); | ||
| 162 | clear_bit(GLF_DIRTY, &gl->gl_flags); | 163 | clear_bit(GLF_DIRTY, &gl->gl_flags); |
| 163 | gfs2_ail_empty_gl(gl); | 164 | gfs2_ail_empty_gl(gl); |
| 164 | } | 165 | } |
| 165 | } | 166 | } |
| 166 | 167 | ||
| 167 | /** | 168 | /** |
| 168 | * inode_go_xmote_th - promote/demote a glock | ||
| 169 | * @gl: the glock | ||
| 170 | * @state: the requested state | ||
| 171 | * @flags: | ||
| 172 | * | ||
| 173 | */ | ||
| 174 | |||
| 175 | static void inode_go_xmote_th(struct gfs2_glock *gl) | ||
| 176 | { | ||
| 177 | if (gl->gl_state != LM_ST_UNLOCKED) | ||
| 178 | gfs2_pte_inval(gl); | ||
| 179 | if (gl->gl_state == LM_ST_EXCLUSIVE) | ||
| 180 | inode_go_sync(gl); | ||
| 181 | } | ||
| 182 | |||
| 183 | /** | ||
| 184 | * inode_go_xmote_bh - After promoting/demoting a glock | 169 | * inode_go_xmote_bh - After promoting/demoting a glock |
| 185 | * @gl: the glock | 170 | * @gl: the glock |
| 186 | * | 171 | * |
| @@ -201,22 +186,6 @@ static void inode_go_xmote_bh(struct gfs2_glock *gl) | |||
| 201 | } | 186 | } |
| 202 | 187 | ||
| 203 | /** | 188 | /** |
| 204 | * inode_go_drop_th - unlock a glock | ||
| 205 | * @gl: the glock | ||
| 206 | * | ||
| 207 | * Invoked from rq_demote(). | ||
| 208 | * Another node needs the lock in EXCLUSIVE mode, or lock (unused for too long) | ||
| 209 | * is being purged from our node's glock cache; we're dropping lock. | ||
| 210 | */ | ||
| 211 | |||
| 212 | static void inode_go_drop_th(struct gfs2_glock *gl) | ||
| 213 | { | ||
| 214 | gfs2_pte_inval(gl); | ||
| 215 | if (gl->gl_state == LM_ST_EXCLUSIVE) | ||
| 216 | inode_go_sync(gl); | ||
| 217 | } | ||
| 218 | |||
| 219 | /** | ||
| 220 | * inode_go_inval - prepare a inode glock to be released | 189 | * inode_go_inval - prepare a inode glock to be released |
| 221 | * @gl: the glock | 190 | * @gl: the glock |
| 222 | * @flags: | 191 | * @flags: |
| @@ -234,10 +203,8 @@ static void inode_go_inval(struct gfs2_glock *gl, int flags) | |||
| 234 | set_bit(GIF_INVALID, &ip->i_flags); | 203 | set_bit(GIF_INVALID, &ip->i_flags); |
| 235 | } | 204 | } |
| 236 | 205 | ||
| 237 | if (ip && S_ISREG(ip->i_inode.i_mode)) { | 206 | if (ip && S_ISREG(ip->i_inode.i_mode)) |
| 238 | truncate_inode_pages(ip->i_inode.i_mapping, 0); | 207 | truncate_inode_pages(ip->i_inode.i_mapping, 0); |
| 239 | clear_bit(GIF_PAGED, &ip->i_flags); | ||
| 240 | } | ||
| 241 | } | 208 | } |
| 242 | 209 | ||
| 243 | /** | 210 | /** |
| @@ -294,23 +261,6 @@ static int inode_go_lock(struct gfs2_holder *gh) | |||
| 294 | } | 261 | } |
| 295 | 262 | ||
| 296 | /** | 263 | /** |
| 297 | * inode_go_unlock - operation done before an inode lock is unlocked by a | ||
| 298 | * process | ||
| 299 | * @gl: the glock | ||
| 300 | * @flags: | ||
| 301 | * | ||
| 302 | */ | ||
| 303 | |||
| 304 | static void inode_go_unlock(struct gfs2_holder *gh) | ||
| 305 | { | ||
| 306 | struct gfs2_glock *gl = gh->gh_gl; | ||
| 307 | struct gfs2_inode *ip = gl->gl_object; | ||
| 308 | |||
| 309 | if (ip) | ||
| 310 | gfs2_meta_cache_flush(ip); | ||
| 311 | } | ||
| 312 | |||
| 313 | /** | ||
| 314 | * rgrp_go_demote_ok - Check to see if it's ok to unlock a RG's glock | 264 | * rgrp_go_demote_ok - Check to see if it's ok to unlock a RG's glock |
| 315 | * @gl: the glock | 265 | * @gl: the glock |
| 316 | * | 266 | * |
| @@ -350,14 +300,14 @@ static void rgrp_go_unlock(struct gfs2_holder *gh) | |||
| 350 | } | 300 | } |
| 351 | 301 | ||
| 352 | /** | 302 | /** |
| 353 | * trans_go_xmote_th - promote/demote the transaction glock | 303 | * trans_go_sync - promote/demote the transaction glock |
| 354 | * @gl: the glock | 304 | * @gl: the glock |
| 355 | * @state: the requested state | 305 | * @state: the requested state |
| 356 | * @flags: | 306 | * @flags: |
| 357 | * | 307 | * |
| 358 | */ | 308 | */ |
| 359 | 309 | ||
| 360 | static void trans_go_xmote_th(struct gfs2_glock *gl) | 310 | static void trans_go_sync(struct gfs2_glock *gl) |
| 361 | { | 311 | { |
| 362 | struct gfs2_sbd *sdp = gl->gl_sbd; | 312 | struct gfs2_sbd *sdp = gl->gl_sbd; |
| 363 | 313 | ||
| @@ -384,7 +334,6 @@ static void trans_go_xmote_bh(struct gfs2_glock *gl) | |||
| 384 | 334 | ||
| 385 | if (gl->gl_state != LM_ST_UNLOCKED && | 335 | if (gl->gl_state != LM_ST_UNLOCKED && |
| 386 | test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags)) { | 336 | test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags)) { |
| 387 | gfs2_meta_cache_flush(GFS2_I(sdp->sd_jdesc->jd_inode)); | ||
| 388 | j_gl->gl_ops->go_inval(j_gl, DIO_METADATA); | 337 | j_gl->gl_ops->go_inval(j_gl, DIO_METADATA); |
| 389 | 338 | ||
| 390 | error = gfs2_find_jhead(sdp->sd_jdesc, &head); | 339 | error = gfs2_find_jhead(sdp->sd_jdesc, &head); |
| @@ -402,24 +351,6 @@ static void trans_go_xmote_bh(struct gfs2_glock *gl) | |||
| 402 | } | 351 | } |
| 403 | 352 | ||
| 404 | /** | 353 | /** |
| 405 | * trans_go_drop_th - unlock the transaction glock | ||
| 406 | * @gl: the glock | ||
| 407 | * | ||
| 408 | * We want to sync the device even with localcaching. Remember | ||
| 409 | * that localcaching journal replay only marks buffers dirty. | ||
| 410 | */ | ||
| 411 | |||
| 412 | static void trans_go_drop_th(struct gfs2_glock *gl) | ||
| 413 | { | ||
| 414 | struct gfs2_sbd *sdp = gl->gl_sbd; | ||
| 415 | |||
| 416 | if (test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags)) { | ||
| 417 | gfs2_meta_syncfs(sdp); | ||
| 418 | gfs2_log_shutdown(sdp); | ||
| 419 | } | ||
| 420 | } | ||
| 421 | |||
| 422 | /** | ||
| 423 | * quota_go_demote_ok - Check to see if it's ok to unlock a quota glock | 354 | * quota_go_demote_ok - Check to see if it's ok to unlock a quota glock |
| 424 | * @gl: the glock | 355 | * @gl: the glock |
| 425 | * | 356 | * |
| @@ -433,25 +364,21 @@ static int quota_go_demote_ok(struct gfs2_glock *gl) | |||
| 433 | 364 | ||
| 434 | const struct gfs2_glock_operations gfs2_meta_glops = { | 365 | const struct gfs2_glock_operations gfs2_meta_glops = { |
| 435 | .go_xmote_th = meta_go_sync, | 366 | .go_xmote_th = meta_go_sync, |
| 436 | .go_drop_th = meta_go_sync, | ||
| 437 | .go_type = LM_TYPE_META, | 367 | .go_type = LM_TYPE_META, |
| 438 | }; | 368 | }; |
| 439 | 369 | ||
| 440 | const struct gfs2_glock_operations gfs2_inode_glops = { | 370 | const struct gfs2_glock_operations gfs2_inode_glops = { |
| 441 | .go_xmote_th = inode_go_xmote_th, | 371 | .go_xmote_th = inode_go_sync, |
| 442 | .go_xmote_bh = inode_go_xmote_bh, | 372 | .go_xmote_bh = inode_go_xmote_bh, |
| 443 | .go_drop_th = inode_go_drop_th, | ||
| 444 | .go_inval = inode_go_inval, | 373 | .go_inval = inode_go_inval, |
| 445 | .go_demote_ok = inode_go_demote_ok, | 374 | .go_demote_ok = inode_go_demote_ok, |
| 446 | .go_lock = inode_go_lock, | 375 | .go_lock = inode_go_lock, |
| 447 | .go_unlock = inode_go_unlock, | ||
| 448 | .go_type = LM_TYPE_INODE, | 376 | .go_type = LM_TYPE_INODE, |
| 449 | .go_min_hold_time = HZ / 10, | 377 | .go_min_hold_time = HZ / 10, |
| 450 | }; | 378 | }; |
| 451 | 379 | ||
| 452 | const struct gfs2_glock_operations gfs2_rgrp_glops = { | 380 | const struct gfs2_glock_operations gfs2_rgrp_glops = { |
| 453 | .go_xmote_th = meta_go_sync, | 381 | .go_xmote_th = meta_go_sync, |
| 454 | .go_drop_th = meta_go_sync, | ||
| 455 | .go_inval = meta_go_inval, | 382 | .go_inval = meta_go_inval, |
| 456 | .go_demote_ok = rgrp_go_demote_ok, | 383 | .go_demote_ok = rgrp_go_demote_ok, |
| 457 | .go_lock = rgrp_go_lock, | 384 | .go_lock = rgrp_go_lock, |
| @@ -461,9 +388,8 @@ const struct gfs2_glock_operations gfs2_rgrp_glops = { | |||
| 461 | }; | 388 | }; |
| 462 | 389 | ||
| 463 | const struct gfs2_glock_operations gfs2_trans_glops = { | 390 | const struct gfs2_glock_operations gfs2_trans_glops = { |
| 464 | .go_xmote_th = trans_go_xmote_th, | 391 | .go_xmote_th = trans_go_sync, |
| 465 | .go_xmote_bh = trans_go_xmote_bh, | 392 | .go_xmote_bh = trans_go_xmote_bh, |
| 466 | .go_drop_th = trans_go_drop_th, | ||
| 467 | .go_type = LM_TYPE_NONDISK, | 393 | .go_type = LM_TYPE_NONDISK, |
| 468 | }; | 394 | }; |
| 469 | 395 | ||
diff --git a/fs/gfs2/incore.h b/fs/gfs2/incore.h index eaddfb5a8e6f..513aaf0dc0ab 100644 --- a/fs/gfs2/incore.h +++ b/fs/gfs2/incore.h | |||
| @@ -1,6 +1,6 @@ | |||
| 1 | /* | 1 | /* |
| 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. |
| 3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | 3 | * Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved. |
| 4 | * | 4 | * |
| 5 | * This copyrighted material is made available to anyone wishing to use, | 5 | * This copyrighted material is made available to anyone wishing to use, |
| 6 | * modify, copy, or redistribute it subject to the terms and conditions | 6 | * modify, copy, or redistribute it subject to the terms and conditions |
| @@ -131,7 +131,6 @@ struct gfs2_bufdata { | |||
| 131 | struct gfs2_glock_operations { | 131 | struct gfs2_glock_operations { |
| 132 | void (*go_xmote_th) (struct gfs2_glock *gl); | 132 | void (*go_xmote_th) (struct gfs2_glock *gl); |
| 133 | void (*go_xmote_bh) (struct gfs2_glock *gl); | 133 | void (*go_xmote_bh) (struct gfs2_glock *gl); |
| 134 | void (*go_drop_th) (struct gfs2_glock *gl); | ||
| 135 | void (*go_inval) (struct gfs2_glock *gl, int flags); | 134 | void (*go_inval) (struct gfs2_glock *gl, int flags); |
| 136 | int (*go_demote_ok) (struct gfs2_glock *gl); | 135 | int (*go_demote_ok) (struct gfs2_glock *gl); |
| 137 | int (*go_lock) (struct gfs2_holder *gh); | 136 | int (*go_lock) (struct gfs2_holder *gh); |
| @@ -141,10 +140,6 @@ struct gfs2_glock_operations { | |||
| 141 | }; | 140 | }; |
| 142 | 141 | ||
| 143 | enum { | 142 | enum { |
| 144 | /* Actions */ | ||
| 145 | HIF_MUTEX = 0, | ||
| 146 | HIF_PROMOTE = 1, | ||
| 147 | |||
| 148 | /* States */ | 143 | /* States */ |
| 149 | HIF_HOLDER = 6, | 144 | HIF_HOLDER = 6, |
| 150 | HIF_FIRST = 7, | 145 | HIF_FIRST = 7, |
| @@ -171,6 +166,8 @@ enum { | |||
| 171 | GLF_DEMOTE = 3, | 166 | GLF_DEMOTE = 3, |
| 172 | GLF_PENDING_DEMOTE = 4, | 167 | GLF_PENDING_DEMOTE = 4, |
| 173 | GLF_DIRTY = 5, | 168 | GLF_DIRTY = 5, |
| 169 | GLF_DEMOTE_IN_PROGRESS = 6, | ||
| 170 | GLF_LFLUSH = 7, | ||
| 174 | }; | 171 | }; |
| 175 | 172 | ||
| 176 | struct gfs2_glock { | 173 | struct gfs2_glock { |
| @@ -190,6 +187,7 @@ struct gfs2_glock { | |||
| 190 | struct list_head gl_holders; | 187 | struct list_head gl_holders; |
| 191 | struct list_head gl_waiters1; /* HIF_MUTEX */ | 188 | struct list_head gl_waiters1; /* HIF_MUTEX */ |
| 192 | struct list_head gl_waiters3; /* HIF_PROMOTE */ | 189 | struct list_head gl_waiters3; /* HIF_PROMOTE */ |
| 190 | int gl_waiters2; /* GIF_DEMOTE */ | ||
| 193 | 191 | ||
| 194 | const struct gfs2_glock_operations *gl_ops; | 192 | const struct gfs2_glock_operations *gl_ops; |
| 195 | 193 | ||
| @@ -210,7 +208,6 @@ struct gfs2_glock { | |||
| 210 | struct gfs2_sbd *gl_sbd; | 208 | struct gfs2_sbd *gl_sbd; |
| 211 | 209 | ||
| 212 | struct inode *gl_aspace; | 210 | struct inode *gl_aspace; |
| 213 | struct gfs2_log_element gl_le; | ||
| 214 | struct list_head gl_ail_list; | 211 | struct list_head gl_ail_list; |
| 215 | atomic_t gl_ail_count; | 212 | atomic_t gl_ail_count; |
| 216 | struct delayed_work gl_work; | 213 | struct delayed_work gl_work; |
| @@ -239,7 +236,6 @@ struct gfs2_alloc { | |||
| 239 | enum { | 236 | enum { |
| 240 | GIF_INVALID = 0, | 237 | GIF_INVALID = 0, |
| 241 | GIF_QD_LOCKED = 1, | 238 | GIF_QD_LOCKED = 1, |
| 242 | GIF_PAGED = 2, | ||
| 243 | GIF_SW_PAGED = 3, | 239 | GIF_SW_PAGED = 3, |
| 244 | }; | 240 | }; |
| 245 | 241 | ||
| @@ -268,14 +264,10 @@ struct gfs2_inode { | |||
| 268 | struct gfs2_glock *i_gl; /* Move into i_gh? */ | 264 | struct gfs2_glock *i_gl; /* Move into i_gh? */ |
| 269 | struct gfs2_holder i_iopen_gh; | 265 | struct gfs2_holder i_iopen_gh; |
| 270 | struct gfs2_holder i_gh; /* for prepare/commit_write only */ | 266 | struct gfs2_holder i_gh; /* for prepare/commit_write only */ |
| 271 | struct gfs2_alloc i_alloc; | 267 | struct gfs2_alloc *i_alloc; |
| 272 | u64 i_last_rg_alloc; | 268 | u64 i_last_rg_alloc; |
| 273 | 269 | ||
| 274 | spinlock_t i_spin; | ||
| 275 | struct rw_semaphore i_rw_mutex; | 270 | struct rw_semaphore i_rw_mutex; |
| 276 | unsigned long i_last_pfault; | ||
| 277 | |||
| 278 | struct buffer_head *i_cache[GFS2_MAX_META_HEIGHT]; | ||
| 279 | }; | 271 | }; |
| 280 | 272 | ||
| 281 | /* | 273 | /* |
| @@ -287,19 +279,12 @@ static inline struct gfs2_inode *GFS2_I(struct inode *inode) | |||
| 287 | return container_of(inode, struct gfs2_inode, i_inode); | 279 | return container_of(inode, struct gfs2_inode, i_inode); |
| 288 | } | 280 | } |
| 289 | 281 | ||
| 290 | /* To be removed? */ | 282 | static inline struct gfs2_sbd *GFS2_SB(const struct inode *inode) |
| 291 | static inline struct gfs2_sbd *GFS2_SB(struct inode *inode) | ||
| 292 | { | 283 | { |
| 293 | return inode->i_sb->s_fs_info; | 284 | return inode->i_sb->s_fs_info; |
| 294 | } | 285 | } |
| 295 | 286 | ||
| 296 | enum { | ||
| 297 | GFF_DID_DIRECT_ALLOC = 0, | ||
| 298 | GFF_EXLOCK = 1, | ||
| 299 | }; | ||
| 300 | |||
| 301 | struct gfs2_file { | 287 | struct gfs2_file { |
| 302 | unsigned long f_flags; /* GFF_... */ | ||
| 303 | struct mutex f_fl_mutex; | 288 | struct mutex f_fl_mutex; |
| 304 | struct gfs2_holder f_fl_gh; | 289 | struct gfs2_holder f_fl_gh; |
| 305 | }; | 290 | }; |
| @@ -373,8 +358,17 @@ struct gfs2_ail { | |||
| 373 | u64 ai_sync_gen; | 358 | u64 ai_sync_gen; |
| 374 | }; | 359 | }; |
| 375 | 360 | ||
| 361 | struct gfs2_journal_extent { | ||
| 362 | struct list_head extent_list; | ||
| 363 | |||
| 364 | unsigned int lblock; /* First logical block */ | ||
| 365 | u64 dblock; /* First disk block */ | ||
| 366 | u64 blocks; | ||
| 367 | }; | ||
| 368 | |||
| 376 | struct gfs2_jdesc { | 369 | struct gfs2_jdesc { |
| 377 | struct list_head jd_list; | 370 | struct list_head jd_list; |
| 371 | struct list_head extent_list; | ||
| 378 | 372 | ||
| 379 | struct inode *jd_inode; | 373 | struct inode *jd_inode; |
| 380 | unsigned int jd_jid; | 374 | unsigned int jd_jid; |
| @@ -421,13 +415,9 @@ struct gfs2_args { | |||
| 421 | struct gfs2_tune { | 415 | struct gfs2_tune { |
| 422 | spinlock_t gt_spin; | 416 | spinlock_t gt_spin; |
| 423 | 417 | ||
| 424 | unsigned int gt_ilimit; | ||
| 425 | unsigned int gt_ilimit_tries; | ||
| 426 | unsigned int gt_ilimit_min; | ||
| 427 | unsigned int gt_demote_secs; /* Cache retention for unheld glock */ | 418 | unsigned int gt_demote_secs; /* Cache retention for unheld glock */ |
| 428 | unsigned int gt_incore_log_blocks; | 419 | unsigned int gt_incore_log_blocks; |
| 429 | unsigned int gt_log_flush_secs; | 420 | unsigned int gt_log_flush_secs; |
| 430 | unsigned int gt_jindex_refresh_secs; /* Check for new journal index */ | ||
| 431 | 421 | ||
| 432 | unsigned int gt_recoverd_secs; | 422 | unsigned int gt_recoverd_secs; |
| 433 | unsigned int gt_logd_secs; | 423 | unsigned int gt_logd_secs; |
| @@ -443,10 +433,8 @@ struct gfs2_tune { | |||
| 443 | unsigned int gt_new_files_jdata; | 433 | unsigned int gt_new_files_jdata; |
| 444 | unsigned int gt_new_files_directio; | 434 | unsigned int gt_new_files_directio; |
| 445 | unsigned int gt_max_readahead; /* Max bytes to read-ahead from disk */ | 435 | unsigned int gt_max_readahead; /* Max bytes to read-ahead from disk */ |
| 446 | unsigned int gt_lockdump_size; | ||
| 447 | unsigned int gt_stall_secs; /* Detects trouble! */ | 436 | unsigned int gt_stall_secs; /* Detects trouble! */ |
| 448 | unsigned int gt_complain_secs; | 437 | unsigned int gt_complain_secs; |
| 449 | unsigned int gt_reclaim_limit; /* Max num of glocks in reclaim list */ | ||
| 450 | unsigned int gt_statfs_quantum; | 438 | unsigned int gt_statfs_quantum; |
| 451 | unsigned int gt_statfs_slow; | 439 | unsigned int gt_statfs_slow; |
| 452 | }; | 440 | }; |
| @@ -539,7 +527,6 @@ struct gfs2_sbd { | |||
| 539 | /* StatFS stuff */ | 527 | /* StatFS stuff */ |
| 540 | 528 | ||
| 541 | spinlock_t sd_statfs_spin; | 529 | spinlock_t sd_statfs_spin; |
| 542 | struct mutex sd_statfs_mutex; | ||
| 543 | struct gfs2_statfs_change_host sd_statfs_master; | 530 | struct gfs2_statfs_change_host sd_statfs_master; |
| 544 | struct gfs2_statfs_change_host sd_statfs_local; | 531 | struct gfs2_statfs_change_host sd_statfs_local; |
| 545 | unsigned long sd_statfs_sync_time; | 532 | unsigned long sd_statfs_sync_time; |
| @@ -602,20 +589,18 @@ struct gfs2_sbd { | |||
| 602 | unsigned int sd_log_commited_databuf; | 589 | unsigned int sd_log_commited_databuf; |
| 603 | unsigned int sd_log_commited_revoke; | 590 | unsigned int sd_log_commited_revoke; |
| 604 | 591 | ||
| 605 | unsigned int sd_log_num_gl; | ||
| 606 | unsigned int sd_log_num_buf; | 592 | unsigned int sd_log_num_buf; |
| 607 | unsigned int sd_log_num_revoke; | 593 | unsigned int sd_log_num_revoke; |
| 608 | unsigned int sd_log_num_rg; | 594 | unsigned int sd_log_num_rg; |
| 609 | unsigned int sd_log_num_databuf; | 595 | unsigned int sd_log_num_databuf; |
| 610 | 596 | ||
| 611 | struct list_head sd_log_le_gl; | ||
| 612 | struct list_head sd_log_le_buf; | 597 | struct list_head sd_log_le_buf; |
| 613 | struct list_head sd_log_le_revoke; | 598 | struct list_head sd_log_le_revoke; |
| 614 | struct list_head sd_log_le_rg; | 599 | struct list_head sd_log_le_rg; |
| 615 | struct list_head sd_log_le_databuf; | 600 | struct list_head sd_log_le_databuf; |
| 616 | struct list_head sd_log_le_ordered; | 601 | struct list_head sd_log_le_ordered; |
| 617 | 602 | ||
| 618 | unsigned int sd_log_blks_free; | 603 | atomic_t sd_log_blks_free; |
| 619 | struct mutex sd_log_reserve_mutex; | 604 | struct mutex sd_log_reserve_mutex; |
| 620 | 605 | ||
| 621 | u64 sd_log_sequence; | 606 | u64 sd_log_sequence; |
diff --git a/fs/gfs2/inode.c b/fs/gfs2/inode.c index 5f6dc32946cd..728d3169e7bd 100644 --- a/fs/gfs2/inode.c +++ b/fs/gfs2/inode.c | |||
| @@ -31,7 +31,6 @@ | |||
| 31 | #include "log.h" | 31 | #include "log.h" |
| 32 | #include "meta_io.h" | 32 | #include "meta_io.h" |
| 33 | #include "ops_address.h" | 33 | #include "ops_address.h" |
| 34 | #include "ops_file.h" | ||
| 35 | #include "ops_inode.h" | 34 | #include "ops_inode.h" |
| 36 | #include "quota.h" | 35 | #include "quota.h" |
| 37 | #include "rgrp.h" | 36 | #include "rgrp.h" |
| @@ -132,15 +131,21 @@ static struct inode *gfs2_iget_skip(struct super_block *sb, | |||
| 132 | 131 | ||
| 133 | void gfs2_set_iop(struct inode *inode) | 132 | void gfs2_set_iop(struct inode *inode) |
| 134 | { | 133 | { |
| 134 | struct gfs2_sbd *sdp = GFS2_SB(inode); | ||
| 135 | umode_t mode = inode->i_mode; | 135 | umode_t mode = inode->i_mode; |
| 136 | 136 | ||
| 137 | if (S_ISREG(mode)) { | 137 | if (S_ISREG(mode)) { |
| 138 | inode->i_op = &gfs2_file_iops; | 138 | inode->i_op = &gfs2_file_iops; |
| 139 | inode->i_fop = &gfs2_file_fops; | 139 | if (sdp->sd_args.ar_localflocks) |
| 140 | inode->i_mapping->a_ops = &gfs2_file_aops; | 140 | inode->i_fop = &gfs2_file_fops_nolock; |
| 141 | else | ||
| 142 | inode->i_fop = &gfs2_file_fops; | ||
| 141 | } else if (S_ISDIR(mode)) { | 143 | } else if (S_ISDIR(mode)) { |
| 142 | inode->i_op = &gfs2_dir_iops; | 144 | inode->i_op = &gfs2_dir_iops; |
| 143 | inode->i_fop = &gfs2_dir_fops; | 145 | if (sdp->sd_args.ar_localflocks) |
| 146 | inode->i_fop = &gfs2_dir_fops_nolock; | ||
| 147 | else | ||
| 148 | inode->i_fop = &gfs2_dir_fops; | ||
| 144 | } else if (S_ISLNK(mode)) { | 149 | } else if (S_ISLNK(mode)) { |
| 145 | inode->i_op = &gfs2_symlink_iops; | 150 | inode->i_op = &gfs2_symlink_iops; |
| 146 | } else { | 151 | } else { |
| @@ -291,12 +296,10 @@ static int gfs2_dinode_in(struct gfs2_inode *ip, const void *buf) | |||
| 291 | di->di_entries = be32_to_cpu(str->di_entries); | 296 | di->di_entries = be32_to_cpu(str->di_entries); |
| 292 | 297 | ||
| 293 | di->di_eattr = be64_to_cpu(str->di_eattr); | 298 | di->di_eattr = be64_to_cpu(str->di_eattr); |
| 294 | return 0; | 299 | if (S_ISREG(ip->i_inode.i_mode)) |
| 295 | } | 300 | gfs2_set_aops(&ip->i_inode); |
| 296 | 301 | ||
| 297 | static void gfs2_inode_bh(struct gfs2_inode *ip, struct buffer_head *bh) | 302 | return 0; |
| 298 | { | ||
| 299 | ip->i_cache[0] = bh; | ||
| 300 | } | 303 | } |
| 301 | 304 | ||
| 302 | /** | 305 | /** |
| @@ -366,7 +369,8 @@ int gfs2_dinode_dealloc(struct gfs2_inode *ip) | |||
| 366 | if (error) | 369 | if (error) |
| 367 | goto out_rg_gunlock; | 370 | goto out_rg_gunlock; |
| 368 | 371 | ||
| 369 | gfs2_trans_add_gl(ip->i_gl); | 372 | set_bit(GLF_DIRTY, &ip->i_gl->gl_flags); |
| 373 | set_bit(GLF_LFLUSH, &ip->i_gl->gl_flags); | ||
| 370 | 374 | ||
| 371 | gfs2_free_di(rgd, ip); | 375 | gfs2_free_di(rgd, ip); |
| 372 | 376 | ||
| @@ -707,9 +711,10 @@ static int alloc_dinode(struct gfs2_inode *dip, u64 *no_addr, u64 *generation) | |||
| 707 | struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); | 711 | struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); |
| 708 | int error; | 712 | int error; |
| 709 | 713 | ||
| 710 | gfs2_alloc_get(dip); | 714 | if (gfs2_alloc_get(dip) == NULL) |
| 715 | return -ENOMEM; | ||
| 711 | 716 | ||
| 712 | dip->i_alloc.al_requested = RES_DINODE; | 717 | dip->i_alloc->al_requested = RES_DINODE; |
| 713 | error = gfs2_inplace_reserve(dip); | 718 | error = gfs2_inplace_reserve(dip); |
| 714 | if (error) | 719 | if (error) |
| 715 | goto out; | 720 | goto out; |
| @@ -855,7 +860,7 @@ static int link_dinode(struct gfs2_inode *dip, const struct qstr *name, | |||
| 855 | 860 | ||
| 856 | error = alloc_required = gfs2_diradd_alloc_required(&dip->i_inode, name); | 861 | error = alloc_required = gfs2_diradd_alloc_required(&dip->i_inode, name); |
| 857 | if (alloc_required < 0) | 862 | if (alloc_required < 0) |
| 858 | goto fail; | 863 | goto fail_quota_locks; |
| 859 | if (alloc_required) { | 864 | if (alloc_required) { |
| 860 | error = gfs2_quota_check(dip, dip->i_inode.i_uid, dip->i_inode.i_gid); | 865 | error = gfs2_quota_check(dip, dip->i_inode.i_uid, dip->i_inode.i_gid); |
| 861 | if (error) | 866 | if (error) |
| @@ -896,7 +901,7 @@ fail_end_trans: | |||
| 896 | gfs2_trans_end(sdp); | 901 | gfs2_trans_end(sdp); |
| 897 | 902 | ||
| 898 | fail_ipreserv: | 903 | fail_ipreserv: |
| 899 | if (dip->i_alloc.al_rgd) | 904 | if (dip->i_alloc->al_rgd) |
| 900 | gfs2_inplace_release(dip); | 905 | gfs2_inplace_release(dip); |
| 901 | 906 | ||
| 902 | fail_quota_locks: | 907 | fail_quota_locks: |
| @@ -966,7 +971,7 @@ struct inode *gfs2_createi(struct gfs2_holder *ghs, const struct qstr *name, | |||
| 966 | struct gfs2_inum_host inum = { .no_addr = 0, .no_formal_ino = 0 }; | 971 | struct gfs2_inum_host inum = { .no_addr = 0, .no_formal_ino = 0 }; |
| 967 | int error; | 972 | int error; |
| 968 | u64 generation; | 973 | u64 generation; |
| 969 | struct buffer_head *bh=NULL; | 974 | struct buffer_head *bh = NULL; |
| 970 | 975 | ||
| 971 | if (!name->len || name->len > GFS2_FNAMESIZE) | 976 | if (!name->len || name->len > GFS2_FNAMESIZE) |
| 972 | return ERR_PTR(-ENAMETOOLONG); | 977 | return ERR_PTR(-ENAMETOOLONG); |
| @@ -1003,8 +1008,6 @@ struct inode *gfs2_createi(struct gfs2_holder *ghs, const struct qstr *name, | |||
| 1003 | if (IS_ERR(inode)) | 1008 | if (IS_ERR(inode)) |
| 1004 | goto fail_gunlock2; | 1009 | goto fail_gunlock2; |
| 1005 | 1010 | ||
| 1006 | gfs2_inode_bh(GFS2_I(inode), bh); | ||
| 1007 | |||
| 1008 | error = gfs2_inode_refresh(GFS2_I(inode)); | 1011 | error = gfs2_inode_refresh(GFS2_I(inode)); |
| 1009 | if (error) | 1012 | if (error) |
| 1010 | goto fail_gunlock2; | 1013 | goto fail_gunlock2; |
| @@ -1021,6 +1024,8 @@ struct inode *gfs2_createi(struct gfs2_holder *ghs, const struct qstr *name, | |||
| 1021 | if (error) | 1024 | if (error) |
| 1022 | goto fail_gunlock2; | 1025 | goto fail_gunlock2; |
| 1023 | 1026 | ||
| 1027 | if (bh) | ||
| 1028 | brelse(bh); | ||
| 1024 | if (!inode) | 1029 | if (!inode) |
| 1025 | return ERR_PTR(-ENOMEM); | 1030 | return ERR_PTR(-ENOMEM); |
| 1026 | return inode; | 1031 | return inode; |
| @@ -1032,6 +1037,8 @@ fail_gunlock2: | |||
| 1032 | fail_gunlock: | 1037 | fail_gunlock: |
| 1033 | gfs2_glock_dq(ghs); | 1038 | gfs2_glock_dq(ghs); |
| 1034 | fail: | 1039 | fail: |
| 1040 | if (bh) | ||
| 1041 | brelse(bh); | ||
| 1035 | return ERR_PTR(error); | 1042 | return ERR_PTR(error); |
| 1036 | } | 1043 | } |
| 1037 | 1044 | ||
diff --git a/fs/gfs2/inode.h b/fs/gfs2/inode.h index 351ac87ab384..d44650662615 100644 --- a/fs/gfs2/inode.h +++ b/fs/gfs2/inode.h | |||
| @@ -20,6 +20,18 @@ static inline int gfs2_is_jdata(const struct gfs2_inode *ip) | |||
| 20 | return ip->i_di.di_flags & GFS2_DIF_JDATA; | 20 | return ip->i_di.di_flags & GFS2_DIF_JDATA; |
| 21 | } | 21 | } |
| 22 | 22 | ||
| 23 | static inline int gfs2_is_writeback(const struct gfs2_inode *ip) | ||
| 24 | { | ||
| 25 | const struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | ||
| 26 | return (sdp->sd_args.ar_data == GFS2_DATA_WRITEBACK) && !gfs2_is_jdata(ip); | ||
| 27 | } | ||
| 28 | |||
| 29 | static inline int gfs2_is_ordered(const struct gfs2_inode *ip) | ||
| 30 | { | ||
| 31 | const struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | ||
| 32 | return (sdp->sd_args.ar_data == GFS2_DATA_ORDERED) && !gfs2_is_jdata(ip); | ||
| 33 | } | ||
| 34 | |||
| 23 | static inline int gfs2_is_dir(const struct gfs2_inode *ip) | 35 | static inline int gfs2_is_dir(const struct gfs2_inode *ip) |
| 24 | { | 36 | { |
| 25 | return S_ISDIR(ip->i_inode.i_mode); | 37 | return S_ISDIR(ip->i_inode.i_mode); |
diff --git a/fs/gfs2/locking/dlm/mount.c b/fs/gfs2/locking/dlm/mount.c index 41c5b04caaba..f2efff424224 100644 --- a/fs/gfs2/locking/dlm/mount.c +++ b/fs/gfs2/locking/dlm/mount.c | |||
| @@ -67,6 +67,11 @@ static int make_args(struct gdlm_ls *ls, char *data_arg, int *nodir) | |||
| 67 | memset(data, 0, 256); | 67 | memset(data, 0, 256); |
| 68 | strncpy(data, data_arg, 255); | 68 | strncpy(data, data_arg, 255); |
| 69 | 69 | ||
| 70 | if (!strlen(data)) { | ||
| 71 | log_error("no mount options, (u)mount helpers not installed"); | ||
| 72 | return -EINVAL; | ||
| 73 | } | ||
| 74 | |||
| 70 | for (options = data; (x = strsep(&options, ":")); ) { | 75 | for (options = data; (x = strsep(&options, ":")); ) { |
| 71 | if (!*x) | 76 | if (!*x) |
| 72 | continue; | 77 | continue; |
diff --git a/fs/gfs2/locking/dlm/plock.c b/fs/gfs2/locking/dlm/plock.c index 1f7b038530b4..2ebd374b3143 100644 --- a/fs/gfs2/locking/dlm/plock.c +++ b/fs/gfs2/locking/dlm/plock.c | |||
| @@ -89,15 +89,19 @@ int gdlm_plock(void *lockspace, struct lm_lockname *name, | |||
| 89 | op->info.number = name->ln_number; | 89 | op->info.number = name->ln_number; |
| 90 | op->info.start = fl->fl_start; | 90 | op->info.start = fl->fl_start; |
| 91 | op->info.end = fl->fl_end; | 91 | op->info.end = fl->fl_end; |
| 92 | op->info.owner = (__u64)(long) fl->fl_owner; | ||
| 93 | if (fl->fl_lmops && fl->fl_lmops->fl_grant) { | 92 | if (fl->fl_lmops && fl->fl_lmops->fl_grant) { |
| 93 | /* fl_owner is lockd which doesn't distinguish | ||
| 94 | processes on the nfs client */ | ||
| 95 | op->info.owner = (__u64) fl->fl_pid; | ||
| 94 | xop->callback = fl->fl_lmops->fl_grant; | 96 | xop->callback = fl->fl_lmops->fl_grant; |
| 95 | locks_init_lock(&xop->flc); | 97 | locks_init_lock(&xop->flc); |
| 96 | locks_copy_lock(&xop->flc, fl); | 98 | locks_copy_lock(&xop->flc, fl); |
| 97 | xop->fl = fl; | 99 | xop->fl = fl; |
| 98 | xop->file = file; | 100 | xop->file = file; |
| 99 | } else | 101 | } else { |
| 102 | op->info.owner = (__u64)(long) fl->fl_owner; | ||
| 100 | xop->callback = NULL; | 103 | xop->callback = NULL; |
| 104 | } | ||
| 101 | 105 | ||
| 102 | send_op(op); | 106 | send_op(op); |
| 103 | 107 | ||
| @@ -203,7 +207,10 @@ int gdlm_punlock(void *lockspace, struct lm_lockname *name, | |||
| 203 | op->info.number = name->ln_number; | 207 | op->info.number = name->ln_number; |
| 204 | op->info.start = fl->fl_start; | 208 | op->info.start = fl->fl_start; |
| 205 | op->info.end = fl->fl_end; | 209 | op->info.end = fl->fl_end; |
| 206 | op->info.owner = (__u64)(long) fl->fl_owner; | 210 | if (fl->fl_lmops && fl->fl_lmops->fl_grant) |
| 211 | op->info.owner = (__u64) fl->fl_pid; | ||
| 212 | else | ||
| 213 | op->info.owner = (__u64)(long) fl->fl_owner; | ||
| 207 | 214 | ||
| 208 | send_op(op); | 215 | send_op(op); |
| 209 | wait_event(recv_wq, (op->done != 0)); | 216 | wait_event(recv_wq, (op->done != 0)); |
| @@ -242,7 +249,10 @@ int gdlm_plock_get(void *lockspace, struct lm_lockname *name, | |||
| 242 | op->info.number = name->ln_number; | 249 | op->info.number = name->ln_number; |
| 243 | op->info.start = fl->fl_start; | 250 | op->info.start = fl->fl_start; |
| 244 | op->info.end = fl->fl_end; | 251 | op->info.end = fl->fl_end; |
| 245 | op->info.owner = (__u64)(long) fl->fl_owner; | 252 | if (fl->fl_lmops && fl->fl_lmops->fl_grant) |
| 253 | op->info.owner = (__u64) fl->fl_pid; | ||
| 254 | else | ||
| 255 | op->info.owner = (__u64)(long) fl->fl_owner; | ||
| 246 | 256 | ||
| 247 | send_op(op); | 257 | send_op(op); |
| 248 | wait_event(recv_wq, (op->done != 0)); | 258 | wait_event(recv_wq, (op->done != 0)); |
diff --git a/fs/gfs2/locking/dlm/thread.c b/fs/gfs2/locking/dlm/thread.c index bd938f06481d..521694fc19d6 100644 --- a/fs/gfs2/locking/dlm/thread.c +++ b/fs/gfs2/locking/dlm/thread.c | |||
| @@ -273,18 +273,13 @@ static int gdlm_thread(void *data, int blist) | |||
| 273 | struct gdlm_ls *ls = (struct gdlm_ls *) data; | 273 | struct gdlm_ls *ls = (struct gdlm_ls *) data; |
| 274 | struct gdlm_lock *lp = NULL; | 274 | struct gdlm_lock *lp = NULL; |
| 275 | uint8_t complete, blocking, submit, drop; | 275 | uint8_t complete, blocking, submit, drop; |
| 276 | DECLARE_WAITQUEUE(wait, current); | ||
| 277 | 276 | ||
| 278 | /* Only thread1 is allowed to do blocking callbacks since gfs | 277 | /* Only thread1 is allowed to do blocking callbacks since gfs |
| 279 | may wait for a completion callback within a blocking cb. */ | 278 | may wait for a completion callback within a blocking cb. */ |
| 280 | 279 | ||
| 281 | while (!kthread_should_stop()) { | 280 | while (!kthread_should_stop()) { |
| 282 | set_current_state(TASK_INTERRUPTIBLE); | 281 | wait_event_interruptible(ls->thread_wait, |
| 283 | add_wait_queue(&ls->thread_wait, &wait); | 282 | !no_work(ls, blist) || kthread_should_stop()); |
| 284 | if (no_work(ls, blist)) | ||
| 285 | schedule(); | ||
| 286 | remove_wait_queue(&ls->thread_wait, &wait); | ||
| 287 | set_current_state(TASK_RUNNING); | ||
| 288 | 283 | ||
| 289 | complete = blocking = submit = drop = 0; | 284 | complete = blocking = submit = drop = 0; |
| 290 | 285 | ||
diff --git a/fs/gfs2/log.c b/fs/gfs2/log.c index 7df702473252..161ab6f2058e 100644 --- a/fs/gfs2/log.c +++ b/fs/gfs2/log.c | |||
| @@ -1,6 +1,6 @@ | |||
| 1 | /* | 1 | /* |
| 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. |
| 3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | 3 | * Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved. |
| 4 | * | 4 | * |
| 5 | * This copyrighted material is made available to anyone wishing to use, | 5 | * This copyrighted material is made available to anyone wishing to use, |
| 6 | * modify, copy, or redistribute it subject to the terms and conditions | 6 | * modify, copy, or redistribute it subject to the terms and conditions |
| @@ -16,6 +16,8 @@ | |||
| 16 | #include <linux/crc32.h> | 16 | #include <linux/crc32.h> |
| 17 | #include <linux/lm_interface.h> | 17 | #include <linux/lm_interface.h> |
| 18 | #include <linux/delay.h> | 18 | #include <linux/delay.h> |
| 19 | #include <linux/kthread.h> | ||
| 20 | #include <linux/freezer.h> | ||
| 19 | 21 | ||
| 20 | #include "gfs2.h" | 22 | #include "gfs2.h" |
| 21 | #include "incore.h" | 23 | #include "incore.h" |
| @@ -68,14 +70,12 @@ unsigned int gfs2_struct2blk(struct gfs2_sbd *sdp, unsigned int nstruct, | |||
| 68 | * | 70 | * |
| 69 | */ | 71 | */ |
| 70 | 72 | ||
| 71 | void gfs2_remove_from_ail(struct address_space *mapping, struct gfs2_bufdata *bd) | 73 | void gfs2_remove_from_ail(struct gfs2_bufdata *bd) |
| 72 | { | 74 | { |
| 73 | bd->bd_ail = NULL; | 75 | bd->bd_ail = NULL; |
| 74 | list_del_init(&bd->bd_ail_st_list); | 76 | list_del_init(&bd->bd_ail_st_list); |
| 75 | list_del_init(&bd->bd_ail_gl_list); | 77 | list_del_init(&bd->bd_ail_gl_list); |
| 76 | atomic_dec(&bd->bd_gl->gl_ail_count); | 78 | atomic_dec(&bd->bd_gl->gl_ail_count); |
| 77 | if (mapping) | ||
| 78 | gfs2_meta_cache_flush(GFS2_I(mapping->host)); | ||
| 79 | brelse(bd->bd_bh); | 79 | brelse(bd->bd_bh); |
| 80 | } | 80 | } |
| 81 | 81 | ||
| @@ -92,8 +92,6 @@ static void gfs2_ail1_start_one(struct gfs2_sbd *sdp, struct gfs2_ail *ai) | |||
| 92 | struct buffer_head *bh; | 92 | struct buffer_head *bh; |
| 93 | int retry; | 93 | int retry; |
| 94 | 94 | ||
| 95 | BUG_ON(!spin_is_locked(&sdp->sd_log_lock)); | ||
| 96 | |||
| 97 | do { | 95 | do { |
| 98 | retry = 0; | 96 | retry = 0; |
| 99 | 97 | ||
| @@ -210,7 +208,7 @@ static void gfs2_ail1_start(struct gfs2_sbd *sdp, int flags) | |||
| 210 | gfs2_log_unlock(sdp); | 208 | gfs2_log_unlock(sdp); |
| 211 | } | 209 | } |
| 212 | 210 | ||
| 213 | int gfs2_ail1_empty(struct gfs2_sbd *sdp, int flags) | 211 | static int gfs2_ail1_empty(struct gfs2_sbd *sdp, int flags) |
| 214 | { | 212 | { |
| 215 | struct gfs2_ail *ai, *s; | 213 | struct gfs2_ail *ai, *s; |
| 216 | int ret; | 214 | int ret; |
| @@ -248,7 +246,7 @@ static void gfs2_ail2_empty_one(struct gfs2_sbd *sdp, struct gfs2_ail *ai) | |||
| 248 | bd = list_entry(head->prev, struct gfs2_bufdata, | 246 | bd = list_entry(head->prev, struct gfs2_bufdata, |
| 249 | bd_ail_st_list); | 247 | bd_ail_st_list); |
| 250 | gfs2_assert(sdp, bd->bd_ail == ai); | 248 | gfs2_assert(sdp, bd->bd_ail == ai); |
| 251 | gfs2_remove_from_ail(bd->bd_bh->b_page->mapping, bd); | 249 | gfs2_remove_from_ail(bd); |
| 252 | } | 250 | } |
| 253 | } | 251 | } |
| 254 | 252 | ||
| @@ -303,7 +301,7 @@ int gfs2_log_reserve(struct gfs2_sbd *sdp, unsigned int blks) | |||
| 303 | 301 | ||
| 304 | mutex_lock(&sdp->sd_log_reserve_mutex); | 302 | mutex_lock(&sdp->sd_log_reserve_mutex); |
| 305 | gfs2_log_lock(sdp); | 303 | gfs2_log_lock(sdp); |
| 306 | while(sdp->sd_log_blks_free <= (blks + reserved_blks)) { | 304 | while(atomic_read(&sdp->sd_log_blks_free) <= (blks + reserved_blks)) { |
| 307 | gfs2_log_unlock(sdp); | 305 | gfs2_log_unlock(sdp); |
| 308 | gfs2_ail1_empty(sdp, 0); | 306 | gfs2_ail1_empty(sdp, 0); |
| 309 | gfs2_log_flush(sdp, NULL); | 307 | gfs2_log_flush(sdp, NULL); |
| @@ -312,7 +310,7 @@ int gfs2_log_reserve(struct gfs2_sbd *sdp, unsigned int blks) | |||
| 312 | gfs2_ail1_start(sdp, 0); | 310 | gfs2_ail1_start(sdp, 0); |
| 313 | gfs2_log_lock(sdp); | 311 | gfs2_log_lock(sdp); |
| 314 | } | 312 | } |
| 315 | sdp->sd_log_blks_free -= blks; | 313 | atomic_sub(blks, &sdp->sd_log_blks_free); |
| 316 | gfs2_log_unlock(sdp); | 314 | gfs2_log_unlock(sdp); |
| 317 | mutex_unlock(&sdp->sd_log_reserve_mutex); | 315 | mutex_unlock(&sdp->sd_log_reserve_mutex); |
| 318 | 316 | ||
| @@ -332,27 +330,23 @@ void gfs2_log_release(struct gfs2_sbd *sdp, unsigned int blks) | |||
| 332 | { | 330 | { |
| 333 | 331 | ||
| 334 | gfs2_log_lock(sdp); | 332 | gfs2_log_lock(sdp); |
| 335 | sdp->sd_log_blks_free += blks; | 333 | atomic_add(blks, &sdp->sd_log_blks_free); |
| 336 | gfs2_assert_withdraw(sdp, | 334 | gfs2_assert_withdraw(sdp, |
| 337 | sdp->sd_log_blks_free <= sdp->sd_jdesc->jd_blocks); | 335 | atomic_read(&sdp->sd_log_blks_free) <= sdp->sd_jdesc->jd_blocks); |
| 338 | gfs2_log_unlock(sdp); | 336 | gfs2_log_unlock(sdp); |
| 339 | up_read(&sdp->sd_log_flush_lock); | 337 | up_read(&sdp->sd_log_flush_lock); |
| 340 | } | 338 | } |
| 341 | 339 | ||
| 342 | static u64 log_bmap(struct gfs2_sbd *sdp, unsigned int lbn) | 340 | static u64 log_bmap(struct gfs2_sbd *sdp, unsigned int lbn) |
| 343 | { | 341 | { |
| 344 | struct inode *inode = sdp->sd_jdesc->jd_inode; | 342 | struct gfs2_journal_extent *je; |
| 345 | int error; | 343 | |
| 346 | struct buffer_head bh_map = { .b_state = 0, .b_blocknr = 0 }; | 344 | list_for_each_entry(je, &sdp->sd_jdesc->extent_list, extent_list) { |
| 347 | 345 | if (lbn >= je->lblock && lbn < je->lblock + je->blocks) | |
| 348 | bh_map.b_size = 1 << inode->i_blkbits; | 346 | return je->dblock + lbn - je->lblock; |
| 349 | error = gfs2_block_map(inode, lbn, 0, &bh_map); | 347 | } |
| 350 | if (error || !bh_map.b_blocknr) | 348 | |
| 351 | printk(KERN_INFO "error=%d, dbn=%llu lbn=%u", error, | 349 | return -1; |
| 352 | (unsigned long long)bh_map.b_blocknr, lbn); | ||
| 353 | gfs2_assert_withdraw(sdp, !error && bh_map.b_blocknr); | ||
| 354 | |||
| 355 | return bh_map.b_blocknr; | ||
| 356 | } | 350 | } |
| 357 | 351 | ||
| 358 | /** | 352 | /** |
| @@ -561,8 +555,8 @@ static void log_pull_tail(struct gfs2_sbd *sdp, unsigned int new_tail) | |||
| 561 | ail2_empty(sdp, new_tail); | 555 | ail2_empty(sdp, new_tail); |
| 562 | 556 | ||
| 563 | gfs2_log_lock(sdp); | 557 | gfs2_log_lock(sdp); |
| 564 | sdp->sd_log_blks_free += dist; | 558 | atomic_add(dist, &sdp->sd_log_blks_free); |
| 565 | gfs2_assert_withdraw(sdp, sdp->sd_log_blks_free <= sdp->sd_jdesc->jd_blocks); | 559 | gfs2_assert_withdraw(sdp, atomic_read(&sdp->sd_log_blks_free) <= sdp->sd_jdesc->jd_blocks); |
| 566 | gfs2_log_unlock(sdp); | 560 | gfs2_log_unlock(sdp); |
| 567 | 561 | ||
| 568 | sdp->sd_log_tail = new_tail; | 562 | sdp->sd_log_tail = new_tail; |
| @@ -652,7 +646,7 @@ static void gfs2_ordered_write(struct gfs2_sbd *sdp) | |||
| 652 | get_bh(bh); | 646 | get_bh(bh); |
| 653 | gfs2_log_unlock(sdp); | 647 | gfs2_log_unlock(sdp); |
| 654 | lock_buffer(bh); | 648 | lock_buffer(bh); |
| 655 | if (test_clear_buffer_dirty(bh)) { | 649 | if (buffer_mapped(bh) && test_clear_buffer_dirty(bh)) { |
| 656 | bh->b_end_io = end_buffer_write_sync; | 650 | bh->b_end_io = end_buffer_write_sync; |
| 657 | submit_bh(WRITE, bh); | 651 | submit_bh(WRITE, bh); |
| 658 | } else { | 652 | } else { |
| @@ -694,20 +688,16 @@ static void gfs2_ordered_wait(struct gfs2_sbd *sdp) | |||
| 694 | * | 688 | * |
| 695 | */ | 689 | */ |
| 696 | 690 | ||
| 697 | void gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl) | 691 | void __gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl) |
| 698 | { | 692 | { |
| 699 | struct gfs2_ail *ai; | 693 | struct gfs2_ail *ai; |
| 700 | 694 | ||
| 701 | down_write(&sdp->sd_log_flush_lock); | 695 | down_write(&sdp->sd_log_flush_lock); |
| 702 | 696 | ||
| 703 | if (gl) { | 697 | /* Log might have been flushed while we waited for the flush lock */ |
| 704 | gfs2_log_lock(sdp); | 698 | if (gl && !test_bit(GLF_LFLUSH, &gl->gl_flags)) { |
| 705 | if (list_empty(&gl->gl_le.le_list)) { | 699 | up_write(&sdp->sd_log_flush_lock); |
| 706 | gfs2_log_unlock(sdp); | 700 | return; |
| 707 | up_write(&sdp->sd_log_flush_lock); | ||
| 708 | return; | ||
| 709 | } | ||
| 710 | gfs2_log_unlock(sdp); | ||
| 711 | } | 701 | } |
| 712 | 702 | ||
| 713 | ai = kzalloc(sizeof(struct gfs2_ail), GFP_NOFS | __GFP_NOFAIL); | 703 | ai = kzalloc(sizeof(struct gfs2_ail), GFP_NOFS | __GFP_NOFAIL); |
| @@ -739,7 +729,7 @@ void gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl) | |||
| 739 | log_flush_commit(sdp); | 729 | log_flush_commit(sdp); |
| 740 | else if (sdp->sd_log_tail != current_tail(sdp) && !sdp->sd_log_idle){ | 730 | else if (sdp->sd_log_tail != current_tail(sdp) && !sdp->sd_log_idle){ |
| 741 | gfs2_log_lock(sdp); | 731 | gfs2_log_lock(sdp); |
| 742 | sdp->sd_log_blks_free--; /* Adjust for unreserved buffer */ | 732 | atomic_dec(&sdp->sd_log_blks_free); /* Adjust for unreserved buffer */ |
| 743 | gfs2_log_unlock(sdp); | 733 | gfs2_log_unlock(sdp); |
| 744 | log_write_header(sdp, 0, PULL); | 734 | log_write_header(sdp, 0, PULL); |
| 745 | } | 735 | } |
| @@ -767,7 +757,7 @@ void gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl) | |||
| 767 | static void log_refund(struct gfs2_sbd *sdp, struct gfs2_trans *tr) | 757 | static void log_refund(struct gfs2_sbd *sdp, struct gfs2_trans *tr) |
| 768 | { | 758 | { |
| 769 | unsigned int reserved; | 759 | unsigned int reserved; |
| 770 | unsigned int old; | 760 | unsigned int unused; |
| 771 | 761 | ||
| 772 | gfs2_log_lock(sdp); | 762 | gfs2_log_lock(sdp); |
| 773 | 763 | ||
| @@ -779,14 +769,11 @@ static void log_refund(struct gfs2_sbd *sdp, struct gfs2_trans *tr) | |||
| 779 | sdp->sd_log_commited_revoke += tr->tr_num_revoke - tr->tr_num_revoke_rm; | 769 | sdp->sd_log_commited_revoke += tr->tr_num_revoke - tr->tr_num_revoke_rm; |
| 780 | gfs2_assert_withdraw(sdp, ((int)sdp->sd_log_commited_revoke) >= 0); | 770 | gfs2_assert_withdraw(sdp, ((int)sdp->sd_log_commited_revoke) >= 0); |
| 781 | reserved = calc_reserved(sdp); | 771 | reserved = calc_reserved(sdp); |
| 782 | old = sdp->sd_log_blks_free; | 772 | unused = sdp->sd_log_blks_reserved - reserved + tr->tr_reserved; |
| 783 | sdp->sd_log_blks_free += tr->tr_reserved - | 773 | gfs2_assert_withdraw(sdp, unused >= 0); |
| 784 | (reserved - sdp->sd_log_blks_reserved); | 774 | atomic_add(unused, &sdp->sd_log_blks_free); |
| 785 | 775 | gfs2_assert_withdraw(sdp, atomic_read(&sdp->sd_log_blks_free) <= | |
| 786 | gfs2_assert_withdraw(sdp, sdp->sd_log_blks_free >= old); | ||
| 787 | gfs2_assert_withdraw(sdp, sdp->sd_log_blks_free <= | ||
| 788 | sdp->sd_jdesc->jd_blocks); | 776 | sdp->sd_jdesc->jd_blocks); |
| 789 | |||
| 790 | sdp->sd_log_blks_reserved = reserved; | 777 | sdp->sd_log_blks_reserved = reserved; |
| 791 | 778 | ||
| 792 | gfs2_log_unlock(sdp); | 779 | gfs2_log_unlock(sdp); |
| @@ -825,7 +812,6 @@ void gfs2_log_shutdown(struct gfs2_sbd *sdp) | |||
| 825 | down_write(&sdp->sd_log_flush_lock); | 812 | down_write(&sdp->sd_log_flush_lock); |
| 826 | 813 | ||
| 827 | gfs2_assert_withdraw(sdp, !sdp->sd_log_blks_reserved); | 814 | gfs2_assert_withdraw(sdp, !sdp->sd_log_blks_reserved); |
| 828 | gfs2_assert_withdraw(sdp, !sdp->sd_log_num_gl); | ||
| 829 | gfs2_assert_withdraw(sdp, !sdp->sd_log_num_buf); | 815 | gfs2_assert_withdraw(sdp, !sdp->sd_log_num_buf); |
| 830 | gfs2_assert_withdraw(sdp, !sdp->sd_log_num_revoke); | 816 | gfs2_assert_withdraw(sdp, !sdp->sd_log_num_revoke); |
| 831 | gfs2_assert_withdraw(sdp, !sdp->sd_log_num_rg); | 817 | gfs2_assert_withdraw(sdp, !sdp->sd_log_num_rg); |
| @@ -838,7 +824,7 @@ void gfs2_log_shutdown(struct gfs2_sbd *sdp) | |||
| 838 | log_write_header(sdp, GFS2_LOG_HEAD_UNMOUNT, | 824 | log_write_header(sdp, GFS2_LOG_HEAD_UNMOUNT, |
| 839 | (sdp->sd_log_tail == current_tail(sdp)) ? 0 : PULL); | 825 | (sdp->sd_log_tail == current_tail(sdp)) ? 0 : PULL); |
| 840 | 826 | ||
| 841 | gfs2_assert_warn(sdp, sdp->sd_log_blks_free == sdp->sd_jdesc->jd_blocks); | 827 | gfs2_assert_warn(sdp, atomic_read(&sdp->sd_log_blks_free) == sdp->sd_jdesc->jd_blocks); |
| 842 | gfs2_assert_warn(sdp, sdp->sd_log_head == sdp->sd_log_tail); | 828 | gfs2_assert_warn(sdp, sdp->sd_log_head == sdp->sd_log_tail); |
| 843 | gfs2_assert_warn(sdp, list_empty(&sdp->sd_ail2_list)); | 829 | gfs2_assert_warn(sdp, list_empty(&sdp->sd_ail2_list)); |
| 844 | 830 | ||
| @@ -866,3 +852,42 @@ void gfs2_meta_syncfs(struct gfs2_sbd *sdp) | |||
| 866 | } | 852 | } |
| 867 | } | 853 | } |
| 868 | 854 | ||
| 855 | |||
| 856 | /** | ||
| 857 | * gfs2_logd - Update log tail as Active Items get flushed to in-place blocks | ||
| 858 | * @sdp: Pointer to GFS2 superblock | ||
| 859 | * | ||
| 860 | * Also, periodically check to make sure that we're using the most recent | ||
| 861 | * journal index. | ||
| 862 | */ | ||
| 863 | |||
| 864 | int gfs2_logd(void *data) | ||
| 865 | { | ||
| 866 | struct gfs2_sbd *sdp = data; | ||
| 867 | unsigned long t; | ||
| 868 | int need_flush; | ||
| 869 | |||
| 870 | while (!kthread_should_stop()) { | ||
| 871 | /* Advance the log tail */ | ||
| 872 | |||
| 873 | t = sdp->sd_log_flush_time + | ||
| 874 | gfs2_tune_get(sdp, gt_log_flush_secs) * HZ; | ||
| 875 | |||
| 876 | gfs2_ail1_empty(sdp, DIO_ALL); | ||
| 877 | gfs2_log_lock(sdp); | ||
| 878 | need_flush = sdp->sd_log_num_buf > gfs2_tune_get(sdp, gt_incore_log_blocks); | ||
| 879 | gfs2_log_unlock(sdp); | ||
| 880 | if (need_flush || time_after_eq(jiffies, t)) { | ||
| 881 | gfs2_log_flush(sdp, NULL); | ||
| 882 | sdp->sd_log_flush_time = jiffies; | ||
| 883 | } | ||
| 884 | |||
| 885 | t = gfs2_tune_get(sdp, gt_logd_secs) * HZ; | ||
| 886 | if (freezing(current)) | ||
| 887 | refrigerator(); | ||
| 888 | schedule_timeout_interruptible(t); | ||
| 889 | } | ||
| 890 | |||
| 891 | return 0; | ||
| 892 | } | ||
| 893 | |||
diff --git a/fs/gfs2/log.h b/fs/gfs2/log.h index dae282400627..771152816508 100644 --- a/fs/gfs2/log.h +++ b/fs/gfs2/log.h | |||
| @@ -48,8 +48,6 @@ static inline void gfs2_log_pointers_init(struct gfs2_sbd *sdp, | |||
| 48 | unsigned int gfs2_struct2blk(struct gfs2_sbd *sdp, unsigned int nstruct, | 48 | unsigned int gfs2_struct2blk(struct gfs2_sbd *sdp, unsigned int nstruct, |
| 49 | unsigned int ssize); | 49 | unsigned int ssize); |
| 50 | 50 | ||
| 51 | int gfs2_ail1_empty(struct gfs2_sbd *sdp, int flags); | ||
| 52 | |||
| 53 | int gfs2_log_reserve(struct gfs2_sbd *sdp, unsigned int blks); | 51 | int gfs2_log_reserve(struct gfs2_sbd *sdp, unsigned int blks); |
| 54 | void gfs2_log_release(struct gfs2_sbd *sdp, unsigned int blks); | 52 | void gfs2_log_release(struct gfs2_sbd *sdp, unsigned int blks); |
| 55 | void gfs2_log_incr_head(struct gfs2_sbd *sdp); | 53 | void gfs2_log_incr_head(struct gfs2_sbd *sdp); |
| @@ -57,11 +55,19 @@ void gfs2_log_incr_head(struct gfs2_sbd *sdp); | |||
| 57 | struct buffer_head *gfs2_log_get_buf(struct gfs2_sbd *sdp); | 55 | struct buffer_head *gfs2_log_get_buf(struct gfs2_sbd *sdp); |
| 58 | struct buffer_head *gfs2_log_fake_buf(struct gfs2_sbd *sdp, | 56 | struct buffer_head *gfs2_log_fake_buf(struct gfs2_sbd *sdp, |
| 59 | struct buffer_head *real); | 57 | struct buffer_head *real); |
| 60 | void gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl); | 58 | void __gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl); |
| 59 | |||
| 60 | static inline void gfs2_log_flush(struct gfs2_sbd *sbd, struct gfs2_glock *gl) | ||
| 61 | { | ||
| 62 | if (!gl || test_bit(GLF_LFLUSH, &gl->gl_flags)) | ||
| 63 | __gfs2_log_flush(sbd, gl); | ||
| 64 | } | ||
| 65 | |||
| 61 | void gfs2_log_commit(struct gfs2_sbd *sdp, struct gfs2_trans *trans); | 66 | void gfs2_log_commit(struct gfs2_sbd *sdp, struct gfs2_trans *trans); |
| 62 | void gfs2_remove_from_ail(struct address_space *mapping, struct gfs2_bufdata *bd); | 67 | void gfs2_remove_from_ail(struct gfs2_bufdata *bd); |
| 63 | 68 | ||
| 64 | void gfs2_log_shutdown(struct gfs2_sbd *sdp); | 69 | void gfs2_log_shutdown(struct gfs2_sbd *sdp); |
| 65 | void gfs2_meta_syncfs(struct gfs2_sbd *sdp); | 70 | void gfs2_meta_syncfs(struct gfs2_sbd *sdp); |
| 71 | int gfs2_logd(void *data); | ||
| 66 | 72 | ||
| 67 | #endif /* __LOG_DOT_H__ */ | 73 | #endif /* __LOG_DOT_H__ */ |
diff --git a/fs/gfs2/lops.c b/fs/gfs2/lops.c index 6c27cea761c6..fae59d69d01a 100644 --- a/fs/gfs2/lops.c +++ b/fs/gfs2/lops.c | |||
| @@ -87,6 +87,7 @@ static void gfs2_unpin(struct gfs2_sbd *sdp, struct buffer_head *bh, | |||
| 87 | } | 87 | } |
| 88 | bd->bd_ail = ai; | 88 | bd->bd_ail = ai; |
| 89 | list_add(&bd->bd_ail_st_list, &ai->ai_ail1_list); | 89 | list_add(&bd->bd_ail_st_list, &ai->ai_ail1_list); |
| 90 | clear_bit(GLF_LFLUSH, &bd->bd_gl->gl_flags); | ||
| 90 | gfs2_log_unlock(sdp); | 91 | gfs2_log_unlock(sdp); |
| 91 | unlock_buffer(bh); | 92 | unlock_buffer(bh); |
| 92 | } | 93 | } |
| @@ -124,49 +125,6 @@ static struct buffer_head *gfs2_get_log_desc(struct gfs2_sbd *sdp, u32 ld_type) | |||
| 124 | return bh; | 125 | return bh; |
| 125 | } | 126 | } |
| 126 | 127 | ||
| 127 | static void __glock_lo_add(struct gfs2_sbd *sdp, struct gfs2_log_element *le) | ||
| 128 | { | ||
| 129 | struct gfs2_glock *gl; | ||
| 130 | struct gfs2_trans *tr = current->journal_info; | ||
| 131 | |||
| 132 | tr->tr_touched = 1; | ||
| 133 | |||
| 134 | gl = container_of(le, struct gfs2_glock, gl_le); | ||
| 135 | if (gfs2_assert_withdraw(sdp, gfs2_glock_is_held_excl(gl))) | ||
| 136 | return; | ||
| 137 | |||
| 138 | if (!list_empty(&le->le_list)) | ||
| 139 | return; | ||
| 140 | |||
| 141 | gfs2_glock_hold(gl); | ||
| 142 | set_bit(GLF_DIRTY, &gl->gl_flags); | ||
| 143 | sdp->sd_log_num_gl++; | ||
| 144 | list_add(&le->le_list, &sdp->sd_log_le_gl); | ||
| 145 | } | ||
| 146 | |||
| 147 | static void glock_lo_add(struct gfs2_sbd *sdp, struct gfs2_log_element *le) | ||
| 148 | { | ||
| 149 | gfs2_log_lock(sdp); | ||
| 150 | __glock_lo_add(sdp, le); | ||
| 151 | gfs2_log_unlock(sdp); | ||
| 152 | } | ||
| 153 | |||
| 154 | static void glock_lo_after_commit(struct gfs2_sbd *sdp, struct gfs2_ail *ai) | ||
| 155 | { | ||
| 156 | struct list_head *head = &sdp->sd_log_le_gl; | ||
| 157 | struct gfs2_glock *gl; | ||
| 158 | |||
| 159 | while (!list_empty(head)) { | ||
| 160 | gl = list_entry(head->next, struct gfs2_glock, gl_le.le_list); | ||
| 161 | list_del_init(&gl->gl_le.le_list); | ||
| 162 | sdp->sd_log_num_gl--; | ||
| 163 | |||
| 164 | gfs2_assert_withdraw(sdp, gfs2_glock_is_held_excl(gl)); | ||
| 165 | gfs2_glock_put(gl); | ||
| 166 | } | ||
| 167 | gfs2_assert_warn(sdp, !sdp->sd_log_num_gl); | ||
| 168 | } | ||
| 169 | |||
| 170 | static void buf_lo_add(struct gfs2_sbd *sdp, struct gfs2_log_element *le) | 128 | static void buf_lo_add(struct gfs2_sbd *sdp, struct gfs2_log_element *le) |
| 171 | { | 129 | { |
| 172 | struct gfs2_bufdata *bd = container_of(le, struct gfs2_bufdata, bd_le); | 130 | struct gfs2_bufdata *bd = container_of(le, struct gfs2_bufdata, bd_le); |
| @@ -182,7 +140,8 @@ static void buf_lo_add(struct gfs2_sbd *sdp, struct gfs2_log_element *le) | |||
| 182 | list_add(&bd->bd_list_tr, &tr->tr_list_buf); | 140 | list_add(&bd->bd_list_tr, &tr->tr_list_buf); |
| 183 | if (!list_empty(&le->le_list)) | 141 | if (!list_empty(&le->le_list)) |
| 184 | goto out; | 142 | goto out; |
| 185 | __glock_lo_add(sdp, &bd->bd_gl->gl_le); | 143 | set_bit(GLF_LFLUSH, &bd->bd_gl->gl_flags); |
| 144 | set_bit(GLF_DIRTY, &bd->bd_gl->gl_flags); | ||
| 186 | gfs2_meta_check(sdp, bd->bd_bh); | 145 | gfs2_meta_check(sdp, bd->bd_bh); |
| 187 | gfs2_pin(sdp, bd->bd_bh); | 146 | gfs2_pin(sdp, bd->bd_bh); |
| 188 | sdp->sd_log_num_buf++; | 147 | sdp->sd_log_num_buf++; |
| @@ -556,17 +515,20 @@ static void databuf_lo_add(struct gfs2_sbd *sdp, struct gfs2_log_element *le) | |||
| 556 | 515 | ||
| 557 | lock_buffer(bd->bd_bh); | 516 | lock_buffer(bd->bd_bh); |
| 558 | gfs2_log_lock(sdp); | 517 | gfs2_log_lock(sdp); |
| 559 | if (!list_empty(&bd->bd_list_tr)) | 518 | if (tr) { |
| 560 | goto out; | 519 | if (!list_empty(&bd->bd_list_tr)) |
| 561 | tr->tr_touched = 1; | 520 | goto out; |
| 562 | if (gfs2_is_jdata(ip)) { | 521 | tr->tr_touched = 1; |
| 563 | tr->tr_num_buf++; | 522 | if (gfs2_is_jdata(ip)) { |
| 564 | list_add(&bd->bd_list_tr, &tr->tr_list_buf); | 523 | tr->tr_num_buf++; |
| 524 | list_add(&bd->bd_list_tr, &tr->tr_list_buf); | ||
| 525 | } | ||
| 565 | } | 526 | } |
| 566 | if (!list_empty(&le->le_list)) | 527 | if (!list_empty(&le->le_list)) |
| 567 | goto out; | 528 | goto out; |
| 568 | 529 | ||
| 569 | __glock_lo_add(sdp, &bd->bd_gl->gl_le); | 530 | set_bit(GLF_LFLUSH, &bd->bd_gl->gl_flags); |
| 531 | set_bit(GLF_DIRTY, &bd->bd_gl->gl_flags); | ||
| 570 | if (gfs2_is_jdata(ip)) { | 532 | if (gfs2_is_jdata(ip)) { |
| 571 | gfs2_pin(sdp, bd->bd_bh); | 533 | gfs2_pin(sdp, bd->bd_bh); |
| 572 | tr->tr_num_databuf_new++; | 534 | tr->tr_num_databuf_new++; |
| @@ -773,12 +735,6 @@ static void databuf_lo_after_commit(struct gfs2_sbd *sdp, struct gfs2_ail *ai) | |||
| 773 | } | 735 | } |
| 774 | 736 | ||
| 775 | 737 | ||
| 776 | const struct gfs2_log_operations gfs2_glock_lops = { | ||
| 777 | .lo_add = glock_lo_add, | ||
| 778 | .lo_after_commit = glock_lo_after_commit, | ||
| 779 | .lo_name = "glock", | ||
| 780 | }; | ||
| 781 | |||
| 782 | const struct gfs2_log_operations gfs2_buf_lops = { | 738 | const struct gfs2_log_operations gfs2_buf_lops = { |
| 783 | .lo_add = buf_lo_add, | 739 | .lo_add = buf_lo_add, |
| 784 | .lo_incore_commit = buf_lo_incore_commit, | 740 | .lo_incore_commit = buf_lo_incore_commit, |
| @@ -816,7 +772,6 @@ const struct gfs2_log_operations gfs2_databuf_lops = { | |||
| 816 | }; | 772 | }; |
| 817 | 773 | ||
| 818 | const struct gfs2_log_operations *gfs2_log_ops[] = { | 774 | const struct gfs2_log_operations *gfs2_log_ops[] = { |
| 819 | &gfs2_glock_lops, | ||
| 820 | &gfs2_databuf_lops, | 775 | &gfs2_databuf_lops, |
| 821 | &gfs2_buf_lops, | 776 | &gfs2_buf_lops, |
| 822 | &gfs2_rg_lops, | 777 | &gfs2_rg_lops, |
diff --git a/fs/gfs2/main.c b/fs/gfs2/main.c index 7ecfe0d3a491..9c7765c12d62 100644 --- a/fs/gfs2/main.c +++ b/fs/gfs2/main.c | |||
| @@ -29,9 +29,8 @@ static void gfs2_init_inode_once(struct kmem_cache *cachep, void *foo) | |||
| 29 | struct gfs2_inode *ip = foo; | 29 | struct gfs2_inode *ip = foo; |
| 30 | 30 | ||
| 31 | inode_init_once(&ip->i_inode); | 31 | inode_init_once(&ip->i_inode); |
| 32 | spin_lock_init(&ip->i_spin); | ||
| 33 | init_rwsem(&ip->i_rw_mutex); | 32 | init_rwsem(&ip->i_rw_mutex); |
| 34 | memset(ip->i_cache, 0, sizeof(ip->i_cache)); | 33 | ip->i_alloc = NULL; |
| 35 | } | 34 | } |
| 36 | 35 | ||
| 37 | static void gfs2_init_glock_once(struct kmem_cache *cachep, void *foo) | 36 | static void gfs2_init_glock_once(struct kmem_cache *cachep, void *foo) |
diff --git a/fs/gfs2/meta_io.c b/fs/gfs2/meta_io.c index 4da423985e4f..85aea27b4a86 100644 --- a/fs/gfs2/meta_io.c +++ b/fs/gfs2/meta_io.c | |||
| @@ -50,6 +50,7 @@ static int gfs2_aspace_writepage(struct page *page, | |||
| 50 | static const struct address_space_operations aspace_aops = { | 50 | static const struct address_space_operations aspace_aops = { |
| 51 | .writepage = gfs2_aspace_writepage, | 51 | .writepage = gfs2_aspace_writepage, |
| 52 | .releasepage = gfs2_releasepage, | 52 | .releasepage = gfs2_releasepage, |
| 53 | .sync_page = block_sync_page, | ||
| 53 | }; | 54 | }; |
| 54 | 55 | ||
| 55 | /** | 56 | /** |
| @@ -221,13 +222,14 @@ int gfs2_meta_read(struct gfs2_glock *gl, u64 blkno, int flags, | |||
| 221 | struct buffer_head **bhp) | 222 | struct buffer_head **bhp) |
| 222 | { | 223 | { |
| 223 | *bhp = getbuf(gl, blkno, CREATE); | 224 | *bhp = getbuf(gl, blkno, CREATE); |
| 224 | if (!buffer_uptodate(*bhp)) | 225 | if (!buffer_uptodate(*bhp)) { |
| 225 | ll_rw_block(READ_META, 1, bhp); | 226 | ll_rw_block(READ_META, 1, bhp); |
| 226 | if (flags & DIO_WAIT) { | 227 | if (flags & DIO_WAIT) { |
| 227 | int error = gfs2_meta_wait(gl->gl_sbd, *bhp); | 228 | int error = gfs2_meta_wait(gl->gl_sbd, *bhp); |
| 228 | if (error) { | 229 | if (error) { |
| 229 | brelse(*bhp); | 230 | brelse(*bhp); |
| 230 | return error; | 231 | return error; |
| 232 | } | ||
| 231 | } | 233 | } |
| 232 | } | 234 | } |
| 233 | 235 | ||
| @@ -282,7 +284,7 @@ void gfs2_attach_bufdata(struct gfs2_glock *gl, struct buffer_head *bh, | |||
| 282 | return; | 284 | return; |
| 283 | } | 285 | } |
| 284 | 286 | ||
| 285 | bd = kmem_cache_zalloc(gfs2_bufdata_cachep, GFP_NOFS | __GFP_NOFAIL), | 287 | bd = kmem_cache_zalloc(gfs2_bufdata_cachep, GFP_NOFS | __GFP_NOFAIL); |
| 286 | bd->bd_bh = bh; | 288 | bd->bd_bh = bh; |
| 287 | bd->bd_gl = gl; | 289 | bd->bd_gl = gl; |
| 288 | 290 | ||
| @@ -317,7 +319,7 @@ void gfs2_remove_from_journal(struct buffer_head *bh, struct gfs2_trans *tr, int | |||
| 317 | } | 319 | } |
| 318 | if (bd) { | 320 | if (bd) { |
| 319 | if (bd->bd_ail) { | 321 | if (bd->bd_ail) { |
| 320 | gfs2_remove_from_ail(NULL, bd); | 322 | gfs2_remove_from_ail(bd); |
| 321 | bh->b_private = NULL; | 323 | bh->b_private = NULL; |
| 322 | bd->bd_bh = NULL; | 324 | bd->bd_bh = NULL; |
| 323 | bd->bd_blkno = bh->b_blocknr; | 325 | bd->bd_blkno = bh->b_blocknr; |
| @@ -358,32 +360,6 @@ void gfs2_meta_wipe(struct gfs2_inode *ip, u64 bstart, u32 blen) | |||
| 358 | } | 360 | } |
| 359 | 361 | ||
| 360 | /** | 362 | /** |
| 361 | * gfs2_meta_cache_flush - get rid of any references on buffers for this inode | ||
| 362 | * @ip: The GFS2 inode | ||
| 363 | * | ||
| 364 | * This releases buffers that are in the most-recently-used array of | ||
| 365 | * blocks used for indirect block addressing for this inode. | ||
| 366 | */ | ||
| 367 | |||
| 368 | void gfs2_meta_cache_flush(struct gfs2_inode *ip) | ||
| 369 | { | ||
| 370 | struct buffer_head **bh_slot; | ||
| 371 | unsigned int x; | ||
| 372 | |||
| 373 | spin_lock(&ip->i_spin); | ||
| 374 | |||
| 375 | for (x = 0; x < GFS2_MAX_META_HEIGHT; x++) { | ||
| 376 | bh_slot = &ip->i_cache[x]; | ||
| 377 | if (*bh_slot) { | ||
| 378 | brelse(*bh_slot); | ||
| 379 | *bh_slot = NULL; | ||
| 380 | } | ||
| 381 | } | ||
| 382 | |||
| 383 | spin_unlock(&ip->i_spin); | ||
| 384 | } | ||
| 385 | |||
| 386 | /** | ||
| 387 | * gfs2_meta_indirect_buffer - Get a metadata buffer | 363 | * gfs2_meta_indirect_buffer - Get a metadata buffer |
| 388 | * @ip: The GFS2 inode | 364 | * @ip: The GFS2 inode |
| 389 | * @height: The level of this buf in the metadata (indir addr) tree (if any) | 365 | * @height: The level of this buf in the metadata (indir addr) tree (if any) |
| @@ -391,8 +367,6 @@ void gfs2_meta_cache_flush(struct gfs2_inode *ip) | |||
| 391 | * @new: Non-zero if we may create a new buffer | 367 | * @new: Non-zero if we may create a new buffer |
| 392 | * @bhp: the buffer is returned here | 368 | * @bhp: the buffer is returned here |
| 393 | * | 369 | * |
| 394 | * Try to use the gfs2_inode's MRU metadata tree cache. | ||
| 395 | * | ||
| 396 | * Returns: errno | 370 | * Returns: errno |
| 397 | */ | 371 | */ |
| 398 | 372 | ||
| @@ -401,58 +375,25 @@ int gfs2_meta_indirect_buffer(struct gfs2_inode *ip, int height, u64 num, | |||
| 401 | { | 375 | { |
| 402 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 376 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
| 403 | struct gfs2_glock *gl = ip->i_gl; | 377 | struct gfs2_glock *gl = ip->i_gl; |
| 404 | struct buffer_head *bh = NULL, **bh_slot = ip->i_cache + height; | 378 | struct buffer_head *bh; |
| 405 | int in_cache = 0; | 379 | int ret = 0; |
| 406 | |||
| 407 | BUG_ON(!gl); | ||
| 408 | BUG_ON(!sdp); | ||
| 409 | |||
| 410 | spin_lock(&ip->i_spin); | ||
| 411 | if (*bh_slot && (*bh_slot)->b_blocknr == num) { | ||
| 412 | bh = *bh_slot; | ||
| 413 | get_bh(bh); | ||
| 414 | in_cache = 1; | ||
| 415 | } | ||
| 416 | spin_unlock(&ip->i_spin); | ||
| 417 | |||
| 418 | if (!bh) | ||
| 419 | bh = getbuf(gl, num, CREATE); | ||
| 420 | |||
| 421 | if (!bh) | ||
| 422 | return -ENOBUFS; | ||
| 423 | 380 | ||
| 424 | if (new) { | 381 | if (new) { |
| 425 | if (gfs2_assert_warn(sdp, height)) | 382 | BUG_ON(height == 0); |
| 426 | goto err; | 383 | bh = gfs2_meta_new(gl, num); |
| 427 | meta_prep_new(bh); | ||
| 428 | gfs2_trans_add_bh(ip->i_gl, bh, 1); | 384 | gfs2_trans_add_bh(ip->i_gl, bh, 1); |
| 429 | gfs2_metatype_set(bh, GFS2_METATYPE_IN, GFS2_FORMAT_IN); | 385 | gfs2_metatype_set(bh, GFS2_METATYPE_IN, GFS2_FORMAT_IN); |
| 430 | gfs2_buffer_clear_tail(bh, sizeof(struct gfs2_meta_header)); | 386 | gfs2_buffer_clear_tail(bh, sizeof(struct gfs2_meta_header)); |
| 431 | } else { | 387 | } else { |
| 432 | u32 mtype = height ? GFS2_METATYPE_IN : GFS2_METATYPE_DI; | 388 | u32 mtype = height ? GFS2_METATYPE_IN : GFS2_METATYPE_DI; |
| 433 | if (!buffer_uptodate(bh)) { | 389 | ret = gfs2_meta_read(gl, num, DIO_WAIT, &bh); |
| 434 | ll_rw_block(READ_META, 1, &bh); | 390 | if (ret == 0 && gfs2_metatype_check(sdp, bh, mtype)) { |
| 435 | if (gfs2_meta_wait(sdp, bh)) | 391 | brelse(bh); |
| 436 | goto err; | 392 | ret = -EIO; |
| 437 | } | 393 | } |
| 438 | if (gfs2_metatype_check(sdp, bh, mtype)) | ||
| 439 | goto err; | ||
| 440 | } | ||
| 441 | |||
| 442 | if (!in_cache) { | ||
| 443 | spin_lock(&ip->i_spin); | ||
| 444 | if (*bh_slot) | ||
| 445 | brelse(*bh_slot); | ||
| 446 | *bh_slot = bh; | ||
| 447 | get_bh(bh); | ||
| 448 | spin_unlock(&ip->i_spin); | ||
| 449 | } | 394 | } |
| 450 | |||
| 451 | *bhp = bh; | 395 | *bhp = bh; |
| 452 | return 0; | 396 | return ret; |
| 453 | err: | ||
| 454 | brelse(bh); | ||
| 455 | return -EIO; | ||
| 456 | } | 397 | } |
| 457 | 398 | ||
| 458 | /** | 399 | /** |
diff --git a/fs/gfs2/meta_io.h b/fs/gfs2/meta_io.h index b7048222ebb4..73e3b1c76fe1 100644 --- a/fs/gfs2/meta_io.h +++ b/fs/gfs2/meta_io.h | |||
| @@ -56,7 +56,6 @@ void gfs2_remove_from_journal(struct buffer_head *bh, struct gfs2_trans *tr, | |||
| 56 | 56 | ||
| 57 | void gfs2_meta_wipe(struct gfs2_inode *ip, u64 bstart, u32 blen); | 57 | void gfs2_meta_wipe(struct gfs2_inode *ip, u64 bstart, u32 blen); |
| 58 | 58 | ||
| 59 | void gfs2_meta_cache_flush(struct gfs2_inode *ip); | ||
| 60 | int gfs2_meta_indirect_buffer(struct gfs2_inode *ip, int height, u64 num, | 59 | int gfs2_meta_indirect_buffer(struct gfs2_inode *ip, int height, u64 num, |
| 61 | int new, struct buffer_head **bhp); | 60 | int new, struct buffer_head **bhp); |
| 62 | 61 | ||
diff --git a/fs/gfs2/ops_address.c b/fs/gfs2/ops_address.c index 9679f8b9870d..38dbe99a30ed 100644 --- a/fs/gfs2/ops_address.c +++ b/fs/gfs2/ops_address.c | |||
| @@ -20,6 +20,8 @@ | |||
| 20 | #include <linux/swap.h> | 20 | #include <linux/swap.h> |
| 21 | #include <linux/gfs2_ondisk.h> | 21 | #include <linux/gfs2_ondisk.h> |
| 22 | #include <linux/lm_interface.h> | 22 | #include <linux/lm_interface.h> |
| 23 | #include <linux/backing-dev.h> | ||
| 24 | #include <linux/pagevec.h> | ||
| 23 | 25 | ||
| 24 | #include "gfs2.h" | 26 | #include "gfs2.h" |
| 25 | #include "incore.h" | 27 | #include "incore.h" |
| @@ -32,7 +34,6 @@ | |||
| 32 | #include "quota.h" | 34 | #include "quota.h" |
| 33 | #include "trans.h" | 35 | #include "trans.h" |
| 34 | #include "rgrp.h" | 36 | #include "rgrp.h" |
| 35 | #include "ops_file.h" | ||
| 36 | #include "super.h" | 37 | #include "super.h" |
| 37 | #include "util.h" | 38 | #include "util.h" |
| 38 | #include "glops.h" | 39 | #include "glops.h" |
| @@ -58,22 +59,6 @@ static void gfs2_page_add_databufs(struct gfs2_inode *ip, struct page *page, | |||
| 58 | } | 59 | } |
| 59 | 60 | ||
| 60 | /** | 61 | /** |
| 61 | * gfs2_get_block - Fills in a buffer head with details about a block | ||
| 62 | * @inode: The inode | ||
| 63 | * @lblock: The block number to look up | ||
| 64 | * @bh_result: The buffer head to return the result in | ||
| 65 | * @create: Non-zero if we may add block to the file | ||
| 66 | * | ||
| 67 | * Returns: errno | ||
| 68 | */ | ||
| 69 | |||
| 70 | int gfs2_get_block(struct inode *inode, sector_t lblock, | ||
| 71 | struct buffer_head *bh_result, int create) | ||
| 72 | { | ||
| 73 | return gfs2_block_map(inode, lblock, create, bh_result); | ||
| 74 | } | ||
| 75 | |||
| 76 | /** | ||
| 77 | * gfs2_get_block_noalloc - Fills in a buffer head with details about a block | 62 | * gfs2_get_block_noalloc - Fills in a buffer head with details about a block |
| 78 | * @inode: The inode | 63 | * @inode: The inode |
| 79 | * @lblock: The block number to look up | 64 | * @lblock: The block number to look up |
| @@ -88,7 +73,7 @@ static int gfs2_get_block_noalloc(struct inode *inode, sector_t lblock, | |||
| 88 | { | 73 | { |
| 89 | int error; | 74 | int error; |
| 90 | 75 | ||
| 91 | error = gfs2_block_map(inode, lblock, 0, bh_result); | 76 | error = gfs2_block_map(inode, lblock, bh_result, 0); |
| 92 | if (error) | 77 | if (error) |
| 93 | return error; | 78 | return error; |
| 94 | if (!buffer_mapped(bh_result)) | 79 | if (!buffer_mapped(bh_result)) |
| @@ -99,20 +84,19 @@ static int gfs2_get_block_noalloc(struct inode *inode, sector_t lblock, | |||
| 99 | static int gfs2_get_block_direct(struct inode *inode, sector_t lblock, | 84 | static int gfs2_get_block_direct(struct inode *inode, sector_t lblock, |
| 100 | struct buffer_head *bh_result, int create) | 85 | struct buffer_head *bh_result, int create) |
| 101 | { | 86 | { |
| 102 | return gfs2_block_map(inode, lblock, 0, bh_result); | 87 | return gfs2_block_map(inode, lblock, bh_result, 0); |
| 103 | } | 88 | } |
| 104 | 89 | ||
| 105 | /** | 90 | /** |
| 106 | * gfs2_writepage - Write complete page | 91 | * gfs2_writepage_common - Common bits of writepage |
| 107 | * @page: Page to write | 92 | * @page: The page to be written |
| 93 | * @wbc: The writeback control | ||
| 108 | * | 94 | * |
| 109 | * Returns: errno | 95 | * Returns: 1 if writepage is ok, otherwise an error code or zero if no error. |
| 110 | * | ||
| 111 | * Some of this is copied from block_write_full_page() although we still | ||
| 112 | * call it to do most of the work. | ||
| 113 | */ | 96 | */ |
| 114 | 97 | ||
| 115 | static int gfs2_writepage(struct page *page, struct writeback_control *wbc) | 98 | static int gfs2_writepage_common(struct page *page, |
| 99 | struct writeback_control *wbc) | ||
| 116 | { | 100 | { |
| 117 | struct inode *inode = page->mapping->host; | 101 | struct inode *inode = page->mapping->host; |
| 118 | struct gfs2_inode *ip = GFS2_I(inode); | 102 | struct gfs2_inode *ip = GFS2_I(inode); |
| @@ -120,41 +104,133 @@ static int gfs2_writepage(struct page *page, struct writeback_control *wbc) | |||
| 120 | loff_t i_size = i_size_read(inode); | 104 | loff_t i_size = i_size_read(inode); |
| 121 | pgoff_t end_index = i_size >> PAGE_CACHE_SHIFT; | 105 | pgoff_t end_index = i_size >> PAGE_CACHE_SHIFT; |
| 122 | unsigned offset; | 106 | unsigned offset; |
| 123 | int error; | 107 | int ret = -EIO; |
| 124 | int done_trans = 0; | ||
| 125 | 108 | ||
| 126 | if (gfs2_assert_withdraw(sdp, gfs2_glock_is_held_excl(ip->i_gl))) { | 109 | if (gfs2_assert_withdraw(sdp, gfs2_glock_is_held_excl(ip->i_gl))) |
| 127 | unlock_page(page); | 110 | goto out; |
| 128 | return -EIO; | 111 | ret = 0; |
| 129 | } | ||
| 130 | if (current->journal_info) | 112 | if (current->journal_info) |
| 131 | goto out_ignore; | 113 | goto redirty; |
| 132 | |||
| 133 | /* Is the page fully outside i_size? (truncate in progress) */ | 114 | /* Is the page fully outside i_size? (truncate in progress) */ |
| 134 | offset = i_size & (PAGE_CACHE_SIZE-1); | 115 | offset = i_size & (PAGE_CACHE_SIZE-1); |
| 135 | if (page->index > end_index || (page->index == end_index && !offset)) { | 116 | if (page->index > end_index || (page->index == end_index && !offset)) { |
| 136 | page->mapping->a_ops->invalidatepage(page, 0); | 117 | page->mapping->a_ops->invalidatepage(page, 0); |
| 137 | unlock_page(page); | 118 | goto out; |
| 138 | return 0; /* don't care */ | 119 | } |
| 120 | return 1; | ||
| 121 | redirty: | ||
| 122 | redirty_page_for_writepage(wbc, page); | ||
| 123 | out: | ||
| 124 | unlock_page(page); | ||
| 125 | return 0; | ||
| 126 | } | ||
| 127 | |||
| 128 | /** | ||
| 129 | * gfs2_writeback_writepage - Write page for writeback mappings | ||
| 130 | * @page: The page | ||
| 131 | * @wbc: The writeback control | ||
| 132 | * | ||
| 133 | */ | ||
| 134 | |||
| 135 | static int gfs2_writeback_writepage(struct page *page, | ||
| 136 | struct writeback_control *wbc) | ||
| 137 | { | ||
| 138 | int ret; | ||
| 139 | |||
| 140 | ret = gfs2_writepage_common(page, wbc); | ||
| 141 | if (ret <= 0) | ||
| 142 | return ret; | ||
| 143 | |||
| 144 | ret = mpage_writepage(page, gfs2_get_block_noalloc, wbc); | ||
| 145 | if (ret == -EAGAIN) | ||
| 146 | ret = block_write_full_page(page, gfs2_get_block_noalloc, wbc); | ||
| 147 | return ret; | ||
| 148 | } | ||
| 149 | |||
| 150 | /** | ||
| 151 | * gfs2_ordered_writepage - Write page for ordered data files | ||
| 152 | * @page: The page to write | ||
| 153 | * @wbc: The writeback control | ||
| 154 | * | ||
| 155 | */ | ||
| 156 | |||
| 157 | static int gfs2_ordered_writepage(struct page *page, | ||
| 158 | struct writeback_control *wbc) | ||
| 159 | { | ||
| 160 | struct inode *inode = page->mapping->host; | ||
| 161 | struct gfs2_inode *ip = GFS2_I(inode); | ||
| 162 | int ret; | ||
| 163 | |||
| 164 | ret = gfs2_writepage_common(page, wbc); | ||
| 165 | if (ret <= 0) | ||
| 166 | return ret; | ||
| 167 | |||
| 168 | if (!page_has_buffers(page)) { | ||
| 169 | create_empty_buffers(page, inode->i_sb->s_blocksize, | ||
| 170 | (1 << BH_Dirty)|(1 << BH_Uptodate)); | ||
| 139 | } | 171 | } |
| 172 | gfs2_page_add_databufs(ip, page, 0, inode->i_sb->s_blocksize-1); | ||
| 173 | return block_write_full_page(page, gfs2_get_block_noalloc, wbc); | ||
| 174 | } | ||
| 140 | 175 | ||
| 141 | if ((sdp->sd_args.ar_data == GFS2_DATA_ORDERED || gfs2_is_jdata(ip)) && | 176 | /** |
| 142 | PageChecked(page)) { | 177 | * __gfs2_jdata_writepage - The core of jdata writepage |
| 178 | * @page: The page to write | ||
| 179 | * @wbc: The writeback control | ||
| 180 | * | ||
| 181 | * This is shared between writepage and writepages and implements the | ||
| 182 | * core of the writepage operation. If a transaction is required then | ||
| 183 | * PageChecked will have been set and the transaction will have | ||
| 184 | * already been started before this is called. | ||
| 185 | */ | ||
| 186 | |||
| 187 | static int __gfs2_jdata_writepage(struct page *page, struct writeback_control *wbc) | ||
| 188 | { | ||
| 189 | struct inode *inode = page->mapping->host; | ||
| 190 | struct gfs2_inode *ip = GFS2_I(inode); | ||
| 191 | struct gfs2_sbd *sdp = GFS2_SB(inode); | ||
| 192 | |||
| 193 | if (PageChecked(page)) { | ||
| 143 | ClearPageChecked(page); | 194 | ClearPageChecked(page); |
| 144 | error = gfs2_trans_begin(sdp, RES_DINODE + 1, 0); | ||
| 145 | if (error) | ||
| 146 | goto out_ignore; | ||
| 147 | if (!page_has_buffers(page)) { | 195 | if (!page_has_buffers(page)) { |
| 148 | create_empty_buffers(page, inode->i_sb->s_blocksize, | 196 | create_empty_buffers(page, inode->i_sb->s_blocksize, |
| 149 | (1 << BH_Dirty)|(1 << BH_Uptodate)); | 197 | (1 << BH_Dirty)|(1 << BH_Uptodate)); |
| 150 | } | 198 | } |
| 151 | gfs2_page_add_databufs(ip, page, 0, sdp->sd_vfs->s_blocksize-1); | 199 | gfs2_page_add_databufs(ip, page, 0, sdp->sd_vfs->s_blocksize-1); |
| 200 | } | ||
| 201 | return block_write_full_page(page, gfs2_get_block_noalloc, wbc); | ||
| 202 | } | ||
| 203 | |||
| 204 | /** | ||
| 205 | * gfs2_jdata_writepage - Write complete page | ||
| 206 | * @page: Page to write | ||
| 207 | * | ||
| 208 | * Returns: errno | ||
| 209 | * | ||
| 210 | */ | ||
| 211 | |||
| 212 | static int gfs2_jdata_writepage(struct page *page, struct writeback_control *wbc) | ||
| 213 | { | ||
| 214 | struct inode *inode = page->mapping->host; | ||
| 215 | struct gfs2_sbd *sdp = GFS2_SB(inode); | ||
| 216 | int error; | ||
| 217 | int done_trans = 0; | ||
| 218 | |||
| 219 | error = gfs2_writepage_common(page, wbc); | ||
| 220 | if (error <= 0) | ||
| 221 | return error; | ||
| 222 | |||
| 223 | if (PageChecked(page)) { | ||
| 224 | if (wbc->sync_mode != WB_SYNC_ALL) | ||
| 225 | goto out_ignore; | ||
| 226 | error = gfs2_trans_begin(sdp, RES_DINODE + 1, 0); | ||
| 227 | if (error) | ||
| 228 | goto out_ignore; | ||
| 152 | done_trans = 1; | 229 | done_trans = 1; |
| 153 | } | 230 | } |
| 154 | error = block_write_full_page(page, gfs2_get_block_noalloc, wbc); | 231 | error = __gfs2_jdata_writepage(page, wbc); |
| 155 | if (done_trans) | 232 | if (done_trans) |
| 156 | gfs2_trans_end(sdp); | 233 | gfs2_trans_end(sdp); |
| 157 | gfs2_meta_cache_flush(ip); | ||
| 158 | return error; | 234 | return error; |
| 159 | 235 | ||
| 160 | out_ignore: | 236 | out_ignore: |
| @@ -164,29 +240,190 @@ out_ignore: | |||
| 164 | } | 240 | } |
| 165 | 241 | ||
| 166 | /** | 242 | /** |
| 167 | * gfs2_writepages - Write a bunch of dirty pages back to disk | 243 | * gfs2_writeback_writepages - Write a bunch of dirty pages back to disk |
| 168 | * @mapping: The mapping to write | 244 | * @mapping: The mapping to write |
| 169 | * @wbc: Write-back control | 245 | * @wbc: Write-back control |
| 170 | * | 246 | * |
| 171 | * For journaled files and/or ordered writes this just falls back to the | 247 | * For the data=writeback case we can already ignore buffer heads |
| 172 | * kernel's default writepages path for now. We will probably want to change | ||
| 173 | * that eventually (i.e. when we look at allocate on flush). | ||
| 174 | * | ||
| 175 | * For the data=writeback case though we can already ignore buffer heads | ||
| 176 | * and write whole extents at once. This is a big reduction in the | 248 | * and write whole extents at once. This is a big reduction in the |
| 177 | * number of I/O requests we send and the bmap calls we make in this case. | 249 | * number of I/O requests we send and the bmap calls we make in this case. |
| 178 | */ | 250 | */ |
| 179 | static int gfs2_writepages(struct address_space *mapping, | 251 | static int gfs2_writeback_writepages(struct address_space *mapping, |
| 180 | struct writeback_control *wbc) | 252 | struct writeback_control *wbc) |
| 253 | { | ||
| 254 | return mpage_writepages(mapping, wbc, gfs2_get_block_noalloc); | ||
| 255 | } | ||
| 256 | |||
| 257 | /** | ||
| 258 | * gfs2_write_jdata_pagevec - Write back a pagevec's worth of pages | ||
| 259 | * @mapping: The mapping | ||
| 260 | * @wbc: The writeback control | ||
| 261 | * @writepage: The writepage function to call for each page | ||
| 262 | * @pvec: The vector of pages | ||
| 263 | * @nr_pages: The number of pages to write | ||
| 264 | * | ||
| 265 | * Returns: non-zero if loop should terminate, zero otherwise | ||
| 266 | */ | ||
| 267 | |||
| 268 | static int gfs2_write_jdata_pagevec(struct address_space *mapping, | ||
| 269 | struct writeback_control *wbc, | ||
| 270 | struct pagevec *pvec, | ||
| 271 | int nr_pages, pgoff_t end) | ||
| 181 | { | 272 | { |
| 182 | struct inode *inode = mapping->host; | 273 | struct inode *inode = mapping->host; |
| 183 | struct gfs2_inode *ip = GFS2_I(inode); | ||
| 184 | struct gfs2_sbd *sdp = GFS2_SB(inode); | 274 | struct gfs2_sbd *sdp = GFS2_SB(inode); |
| 275 | loff_t i_size = i_size_read(inode); | ||
| 276 | pgoff_t end_index = i_size >> PAGE_CACHE_SHIFT; | ||
| 277 | unsigned offset = i_size & (PAGE_CACHE_SIZE-1); | ||
| 278 | unsigned nrblocks = nr_pages * (PAGE_CACHE_SIZE/inode->i_sb->s_blocksize); | ||
| 279 | struct backing_dev_info *bdi = mapping->backing_dev_info; | ||
| 280 | int i; | ||
| 281 | int ret; | ||
| 282 | |||
| 283 | ret = gfs2_trans_begin(sdp, nrblocks, 0); | ||
| 284 | if (ret < 0) | ||
| 285 | return ret; | ||
| 286 | |||
| 287 | for(i = 0; i < nr_pages; i++) { | ||
| 288 | struct page *page = pvec->pages[i]; | ||
| 289 | |||
| 290 | lock_page(page); | ||
| 291 | |||
| 292 | if (unlikely(page->mapping != mapping)) { | ||
| 293 | unlock_page(page); | ||
| 294 | continue; | ||
| 295 | } | ||
| 296 | |||
| 297 | if (!wbc->range_cyclic && page->index > end) { | ||
| 298 | ret = 1; | ||
| 299 | unlock_page(page); | ||
| 300 | continue; | ||
| 301 | } | ||
| 302 | |||
| 303 | if (wbc->sync_mode != WB_SYNC_NONE) | ||
| 304 | wait_on_page_writeback(page); | ||
| 305 | |||
| 306 | if (PageWriteback(page) || | ||
| 307 | !clear_page_dirty_for_io(page)) { | ||
| 308 | unlock_page(page); | ||
| 309 | continue; | ||
| 310 | } | ||
| 311 | |||
| 312 | /* Is the page fully outside i_size? (truncate in progress) */ | ||
| 313 | if (page->index > end_index || (page->index == end_index && !offset)) { | ||
| 314 | page->mapping->a_ops->invalidatepage(page, 0); | ||
| 315 | unlock_page(page); | ||
| 316 | continue; | ||
| 317 | } | ||
| 318 | |||
| 319 | ret = __gfs2_jdata_writepage(page, wbc); | ||
| 320 | |||
| 321 | if (ret || (--(wbc->nr_to_write) <= 0)) | ||
| 322 | ret = 1; | ||
| 323 | if (wbc->nonblocking && bdi_write_congested(bdi)) { | ||
| 324 | wbc->encountered_congestion = 1; | ||
| 325 | ret = 1; | ||
| 326 | } | ||
| 327 | |||
| 328 | } | ||
| 329 | gfs2_trans_end(sdp); | ||
| 330 | return ret; | ||
| 331 | } | ||
| 332 | |||
| 333 | /** | ||
| 334 | * gfs2_write_cache_jdata - Like write_cache_pages but different | ||
| 335 | * @mapping: The mapping to write | ||
| 336 | * @wbc: The writeback control | ||
| 337 | * @writepage: The writepage function to call | ||
| 338 | * @data: The data to pass to writepage | ||
| 339 | * | ||
| 340 | * The reason that we use our own function here is that we need to | ||
| 341 | * start transactions before we grab page locks. This allows us | ||
| 342 | * to get the ordering right. | ||
| 343 | */ | ||
| 344 | |||
| 345 | static int gfs2_write_cache_jdata(struct address_space *mapping, | ||
| 346 | struct writeback_control *wbc) | ||
| 347 | { | ||
| 348 | struct backing_dev_info *bdi = mapping->backing_dev_info; | ||
| 349 | int ret = 0; | ||
| 350 | int done = 0; | ||
| 351 | struct pagevec pvec; | ||
| 352 | int nr_pages; | ||
| 353 | pgoff_t index; | ||
| 354 | pgoff_t end; | ||
| 355 | int scanned = 0; | ||
| 356 | int range_whole = 0; | ||
| 357 | |||
| 358 | if (wbc->nonblocking && bdi_write_congested(bdi)) { | ||
| 359 | wbc->encountered_congestion = 1; | ||
| 360 | return 0; | ||
| 361 | } | ||
| 362 | |||
| 363 | pagevec_init(&pvec, 0); | ||
| 364 | if (wbc->range_cyclic) { | ||
| 365 | index = mapping->writeback_index; /* Start from prev offset */ | ||
| 366 | end = -1; | ||
| 367 | } else { | ||
| 368 | index = wbc->range_start >> PAGE_CACHE_SHIFT; | ||
| 369 | end = wbc->range_end >> PAGE_CACHE_SHIFT; | ||
| 370 | if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) | ||
| 371 | range_whole = 1; | ||
| 372 | scanned = 1; | ||
| 373 | } | ||
| 185 | 374 | ||
| 186 | if (sdp->sd_args.ar_data == GFS2_DATA_WRITEBACK && !gfs2_is_jdata(ip)) | 375 | retry: |
| 187 | return mpage_writepages(mapping, wbc, gfs2_get_block_noalloc); | 376 | while (!done && (index <= end) && |
| 377 | (nr_pages = pagevec_lookup_tag(&pvec, mapping, &index, | ||
| 378 | PAGECACHE_TAG_DIRTY, | ||
| 379 | min(end - index, (pgoff_t)PAGEVEC_SIZE-1) + 1))) { | ||
| 380 | scanned = 1; | ||
| 381 | ret = gfs2_write_jdata_pagevec(mapping, wbc, &pvec, nr_pages, end); | ||
| 382 | if (ret) | ||
| 383 | done = 1; | ||
| 384 | if (ret > 0) | ||
| 385 | ret = 0; | ||
| 386 | |||
| 387 | pagevec_release(&pvec); | ||
| 388 | cond_resched(); | ||
| 389 | } | ||
| 390 | |||
| 391 | if (!scanned && !done) { | ||
| 392 | /* | ||
| 393 | * We hit the last page and there is more work to be done: wrap | ||
| 394 | * back to the start of the file | ||
| 395 | */ | ||
| 396 | scanned = 1; | ||
| 397 | index = 0; | ||
| 398 | goto retry; | ||
| 399 | } | ||
| 400 | |||
| 401 | if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) | ||
| 402 | mapping->writeback_index = index; | ||
| 403 | return ret; | ||
| 404 | } | ||
| 405 | |||
| 406 | |||
| 407 | /** | ||
| 408 | * gfs2_jdata_writepages - Write a bunch of dirty pages back to disk | ||
| 409 | * @mapping: The mapping to write | ||
| 410 | * @wbc: The writeback control | ||
| 411 | * | ||
| 412 | */ | ||
| 188 | 413 | ||
| 189 | return generic_writepages(mapping, wbc); | 414 | static int gfs2_jdata_writepages(struct address_space *mapping, |
| 415 | struct writeback_control *wbc) | ||
| 416 | { | ||
| 417 | struct gfs2_inode *ip = GFS2_I(mapping->host); | ||
| 418 | struct gfs2_sbd *sdp = GFS2_SB(mapping->host); | ||
| 419 | int ret; | ||
| 420 | |||
| 421 | ret = gfs2_write_cache_jdata(mapping, wbc); | ||
| 422 | if (ret == 0 && wbc->sync_mode == WB_SYNC_ALL) { | ||
| 423 | gfs2_log_flush(sdp, ip->i_gl); | ||
| 424 | ret = gfs2_write_cache_jdata(mapping, wbc); | ||
| 425 | } | ||
| 426 | return ret; | ||
| 190 | } | 427 | } |
| 191 | 428 | ||
| 192 | /** | 429 | /** |
| @@ -231,62 +468,107 @@ static int stuffed_readpage(struct gfs2_inode *ip, struct page *page) | |||
| 231 | 468 | ||
| 232 | 469 | ||
| 233 | /** | 470 | /** |
| 234 | * gfs2_readpage - readpage with locking | 471 | * __gfs2_readpage - readpage |
| 235 | * @file: The file to read a page for. N.B. This may be NULL if we are | 472 | * @file: The file to read a page for |
| 236 | * reading an internal file. | ||
| 237 | * @page: The page to read | 473 | * @page: The page to read |
| 238 | * | 474 | * |
| 239 | * Returns: errno | 475 | * This is the core of gfs2's readpage. Its used by the internal file |
| 476 | * reading code as in that case we already hold the glock. Also its | ||
| 477 | * called by gfs2_readpage() once the required lock has been granted. | ||
| 478 | * | ||
| 240 | */ | 479 | */ |
| 241 | 480 | ||
| 242 | static int gfs2_readpage(struct file *file, struct page *page) | 481 | static int __gfs2_readpage(void *file, struct page *page) |
| 243 | { | 482 | { |
| 244 | struct gfs2_inode *ip = GFS2_I(page->mapping->host); | 483 | struct gfs2_inode *ip = GFS2_I(page->mapping->host); |
| 245 | struct gfs2_sbd *sdp = GFS2_SB(page->mapping->host); | 484 | struct gfs2_sbd *sdp = GFS2_SB(page->mapping->host); |
| 246 | struct gfs2_file *gf = NULL; | ||
| 247 | struct gfs2_holder gh; | ||
| 248 | int error; | 485 | int error; |
| 249 | int do_unlock = 0; | ||
| 250 | |||
| 251 | if (likely(file != &gfs2_internal_file_sentinel)) { | ||
| 252 | if (file) { | ||
| 253 | gf = file->private_data; | ||
| 254 | if (test_bit(GFF_EXLOCK, &gf->f_flags)) | ||
| 255 | /* gfs2_sharewrite_fault has grabbed the ip->i_gl already */ | ||
| 256 | goto skip_lock; | ||
| 257 | } | ||
| 258 | gfs2_holder_init(ip->i_gl, LM_ST_SHARED, GL_ATIME|LM_FLAG_TRY_1CB, &gh); | ||
| 259 | do_unlock = 1; | ||
| 260 | error = gfs2_glock_nq_atime(&gh); | ||
| 261 | if (unlikely(error)) | ||
| 262 | goto out_unlock; | ||
| 263 | } | ||
| 264 | 486 | ||
| 265 | skip_lock: | ||
| 266 | if (gfs2_is_stuffed(ip)) { | 487 | if (gfs2_is_stuffed(ip)) { |
| 267 | error = stuffed_readpage(ip, page); | 488 | error = stuffed_readpage(ip, page); |
| 268 | unlock_page(page); | 489 | unlock_page(page); |
| 269 | } else | 490 | } else { |
| 270 | error = mpage_readpage(page, gfs2_get_block); | 491 | error = mpage_readpage(page, gfs2_block_map); |
| 492 | } | ||
| 271 | 493 | ||
| 272 | if (unlikely(test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) | 494 | if (unlikely(test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) |
| 273 | error = -EIO; | 495 | return -EIO; |
| 496 | |||
| 497 | return error; | ||
| 498 | } | ||
| 499 | |||
| 500 | /** | ||
| 501 | * gfs2_readpage - read a page of a file | ||
| 502 | * @file: The file to read | ||
| 503 | * @page: The page of the file | ||
| 504 | * | ||
| 505 | * This deals with the locking required. We use a trylock in order to | ||
| 506 | * avoid the page lock / glock ordering problems returning AOP_TRUNCATED_PAGE | ||
| 507 | * in the event that we are unable to get the lock. | ||
| 508 | */ | ||
| 509 | |||
| 510 | static int gfs2_readpage(struct file *file, struct page *page) | ||
| 511 | { | ||
| 512 | struct gfs2_inode *ip = GFS2_I(page->mapping->host); | ||
| 513 | struct gfs2_holder gh; | ||
| 514 | int error; | ||
| 274 | 515 | ||
| 275 | if (do_unlock) { | 516 | gfs2_holder_init(ip->i_gl, LM_ST_SHARED, GL_ATIME|LM_FLAG_TRY_1CB, &gh); |
| 276 | gfs2_glock_dq_m(1, &gh); | 517 | error = gfs2_glock_nq_atime(&gh); |
| 277 | gfs2_holder_uninit(&gh); | 518 | if (unlikely(error)) { |
| 519 | unlock_page(page); | ||
| 520 | goto out; | ||
| 278 | } | 521 | } |
| 522 | error = __gfs2_readpage(file, page); | ||
| 523 | gfs2_glock_dq(&gh); | ||
| 279 | out: | 524 | out: |
| 280 | return error; | 525 | gfs2_holder_uninit(&gh); |
| 281 | out_unlock: | ||
| 282 | unlock_page(page); | ||
| 283 | if (error == GLR_TRYFAILED) { | 526 | if (error == GLR_TRYFAILED) { |
| 284 | error = AOP_TRUNCATED_PAGE; | ||
| 285 | yield(); | 527 | yield(); |
| 528 | return AOP_TRUNCATED_PAGE; | ||
| 286 | } | 529 | } |
| 287 | if (do_unlock) | 530 | return error; |
| 288 | gfs2_holder_uninit(&gh); | 531 | } |
| 289 | goto out; | 532 | |
| 533 | /** | ||
| 534 | * gfs2_internal_read - read an internal file | ||
| 535 | * @ip: The gfs2 inode | ||
| 536 | * @ra_state: The readahead state (or NULL for no readahead) | ||
| 537 | * @buf: The buffer to fill | ||
| 538 | * @pos: The file position | ||
| 539 | * @size: The amount to read | ||
| 540 | * | ||
| 541 | */ | ||
| 542 | |||
| 543 | int gfs2_internal_read(struct gfs2_inode *ip, struct file_ra_state *ra_state, | ||
| 544 | char *buf, loff_t *pos, unsigned size) | ||
| 545 | { | ||
| 546 | struct address_space *mapping = ip->i_inode.i_mapping; | ||
| 547 | unsigned long index = *pos / PAGE_CACHE_SIZE; | ||
| 548 | unsigned offset = *pos & (PAGE_CACHE_SIZE - 1); | ||
| 549 | unsigned copied = 0; | ||
| 550 | unsigned amt; | ||
| 551 | struct page *page; | ||
| 552 | void *p; | ||
| 553 | |||
| 554 | do { | ||
| 555 | amt = size - copied; | ||
| 556 | if (offset + size > PAGE_CACHE_SIZE) | ||
| 557 | amt = PAGE_CACHE_SIZE - offset; | ||
| 558 | page = read_cache_page(mapping, index, __gfs2_readpage, NULL); | ||
| 559 | if (IS_ERR(page)) | ||
| 560 | return PTR_ERR(page); | ||
| 561 | p = kmap_atomic(page, KM_USER0); | ||
| 562 | memcpy(buf + copied, p + offset, amt); | ||
| 563 | kunmap_atomic(p, KM_USER0); | ||
| 564 | mark_page_accessed(page); | ||
| 565 | page_cache_release(page); | ||
| 566 | copied += amt; | ||
| 567 | index++; | ||
| 568 | offset = 0; | ||
| 569 | } while(copied < size); | ||
| 570 | (*pos) += size; | ||
| 571 | return size; | ||
| 290 | } | 572 | } |
| 291 | 573 | ||
| 292 | /** | 574 | /** |
| @@ -300,10 +582,9 @@ out_unlock: | |||
| 300 | * Any I/O we ignore at this time will be done via readpage later. | 582 | * Any I/O we ignore at this time will be done via readpage later. |
| 301 | * 2. We don't handle stuffed files here we let readpage do the honours. | 583 | * 2. We don't handle stuffed files here we let readpage do the honours. |
| 302 | * 3. mpage_readpages() does most of the heavy lifting in the common case. | 584 | * 3. mpage_readpages() does most of the heavy lifting in the common case. |
| 303 | * 4. gfs2_get_block() is relied upon to set BH_Boundary in the right places. | 585 | * 4. gfs2_block_map() is relied upon to set BH_Boundary in the right places. |
| 304 | * 5. We use LM_FLAG_TRY_1CB here, effectively we then have lock-ahead as | ||
| 305 | * well as read-ahead. | ||
| 306 | */ | 586 | */ |
| 587 | |||
| 307 | static int gfs2_readpages(struct file *file, struct address_space *mapping, | 588 | static int gfs2_readpages(struct file *file, struct address_space *mapping, |
| 308 | struct list_head *pages, unsigned nr_pages) | 589 | struct list_head *pages, unsigned nr_pages) |
| 309 | { | 590 | { |
| @@ -311,42 +592,20 @@ static int gfs2_readpages(struct file *file, struct address_space *mapping, | |||
| 311 | struct gfs2_inode *ip = GFS2_I(inode); | 592 | struct gfs2_inode *ip = GFS2_I(inode); |
| 312 | struct gfs2_sbd *sdp = GFS2_SB(inode); | 593 | struct gfs2_sbd *sdp = GFS2_SB(inode); |
| 313 | struct gfs2_holder gh; | 594 | struct gfs2_holder gh; |
| 314 | int ret = 0; | 595 | int ret; |
| 315 | int do_unlock = 0; | ||
| 316 | 596 | ||
| 317 | if (likely(file != &gfs2_internal_file_sentinel)) { | 597 | gfs2_holder_init(ip->i_gl, LM_ST_SHARED, GL_ATIME, &gh); |
| 318 | if (file) { | 598 | ret = gfs2_glock_nq_atime(&gh); |
| 319 | struct gfs2_file *gf = file->private_data; | 599 | if (unlikely(ret)) |
| 320 | if (test_bit(GFF_EXLOCK, &gf->f_flags)) | 600 | goto out_uninit; |
| 321 | goto skip_lock; | ||
| 322 | } | ||
| 323 | gfs2_holder_init(ip->i_gl, LM_ST_SHARED, | ||
| 324 | LM_FLAG_TRY_1CB|GL_ATIME, &gh); | ||
| 325 | do_unlock = 1; | ||
| 326 | ret = gfs2_glock_nq_atime(&gh); | ||
| 327 | if (ret == GLR_TRYFAILED) | ||
| 328 | goto out_noerror; | ||
| 329 | if (unlikely(ret)) | ||
| 330 | goto out_unlock; | ||
| 331 | } | ||
| 332 | skip_lock: | ||
| 333 | if (!gfs2_is_stuffed(ip)) | 601 | if (!gfs2_is_stuffed(ip)) |
| 334 | ret = mpage_readpages(mapping, pages, nr_pages, gfs2_get_block); | 602 | ret = mpage_readpages(mapping, pages, nr_pages, gfs2_block_map); |
| 335 | 603 | gfs2_glock_dq(&gh); | |
| 336 | if (do_unlock) { | 604 | out_uninit: |
| 337 | gfs2_glock_dq_m(1, &gh); | 605 | gfs2_holder_uninit(&gh); |
| 338 | gfs2_holder_uninit(&gh); | ||
| 339 | } | ||
| 340 | out: | ||
| 341 | if (unlikely(test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) | 606 | if (unlikely(test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) |
| 342 | ret = -EIO; | 607 | ret = -EIO; |
| 343 | return ret; | 608 | return ret; |
| 344 | out_noerror: | ||
| 345 | ret = 0; | ||
| 346 | out_unlock: | ||
| 347 | if (do_unlock) | ||
| 348 | gfs2_holder_uninit(&gh); | ||
| 349 | goto out; | ||
| 350 | } | 609 | } |
| 351 | 610 | ||
| 352 | /** | 611 | /** |
| @@ -382,20 +641,11 @@ static int gfs2_write_begin(struct file *file, struct address_space *mapping, | |||
| 382 | if (unlikely(error)) | 641 | if (unlikely(error)) |
| 383 | goto out_uninit; | 642 | goto out_uninit; |
| 384 | 643 | ||
| 385 | error = -ENOMEM; | ||
| 386 | page = __grab_cache_page(mapping, index); | ||
| 387 | *pagep = page; | ||
| 388 | if (!page) | ||
| 389 | goto out_unlock; | ||
| 390 | |||
| 391 | gfs2_write_calc_reserv(ip, len, &data_blocks, &ind_blocks); | 644 | gfs2_write_calc_reserv(ip, len, &data_blocks, &ind_blocks); |
| 392 | |||
| 393 | error = gfs2_write_alloc_required(ip, pos, len, &alloc_required); | 645 | error = gfs2_write_alloc_required(ip, pos, len, &alloc_required); |
| 394 | if (error) | 646 | if (error) |
| 395 | goto out_putpage; | 647 | goto out_unlock; |
| 396 | |||
| 397 | 648 | ||
| 398 | ip->i_alloc.al_requested = 0; | ||
| 399 | if (alloc_required) { | 649 | if (alloc_required) { |
| 400 | al = gfs2_alloc_get(ip); | 650 | al = gfs2_alloc_get(ip); |
| 401 | 651 | ||
| @@ -424,40 +674,47 @@ static int gfs2_write_begin(struct file *file, struct address_space *mapping, | |||
| 424 | if (error) | 674 | if (error) |
| 425 | goto out_trans_fail; | 675 | goto out_trans_fail; |
| 426 | 676 | ||
| 677 | error = -ENOMEM; | ||
| 678 | page = __grab_cache_page(mapping, index); | ||
| 679 | *pagep = page; | ||
| 680 | if (unlikely(!page)) | ||
| 681 | goto out_endtrans; | ||
| 682 | |||
| 427 | if (gfs2_is_stuffed(ip)) { | 683 | if (gfs2_is_stuffed(ip)) { |
| 684 | error = 0; | ||
| 428 | if (pos + len > sdp->sd_sb.sb_bsize - sizeof(struct gfs2_dinode)) { | 685 | if (pos + len > sdp->sd_sb.sb_bsize - sizeof(struct gfs2_dinode)) { |
| 429 | error = gfs2_unstuff_dinode(ip, page); | 686 | error = gfs2_unstuff_dinode(ip, page); |
| 430 | if (error == 0) | 687 | if (error == 0) |
| 431 | goto prepare_write; | 688 | goto prepare_write; |
| 432 | } else if (!PageUptodate(page)) | 689 | } else if (!PageUptodate(page)) { |
| 433 | error = stuffed_readpage(ip, page); | 690 | error = stuffed_readpage(ip, page); |
| 691 | } | ||
| 434 | goto out; | 692 | goto out; |
| 435 | } | 693 | } |
| 436 | 694 | ||
| 437 | prepare_write: | 695 | prepare_write: |
| 438 | error = block_prepare_write(page, from, to, gfs2_get_block); | 696 | error = block_prepare_write(page, from, to, gfs2_block_map); |
| 439 | |||
| 440 | out: | 697 | out: |
| 441 | if (error) { | 698 | if (error == 0) |
| 442 | gfs2_trans_end(sdp); | 699 | return 0; |
| 700 | |||
| 701 | page_cache_release(page); | ||
| 702 | if (pos + len > ip->i_inode.i_size) | ||
| 703 | vmtruncate(&ip->i_inode, ip->i_inode.i_size); | ||
| 704 | out_endtrans: | ||
| 705 | gfs2_trans_end(sdp); | ||
| 443 | out_trans_fail: | 706 | out_trans_fail: |
| 444 | if (alloc_required) { | 707 | if (alloc_required) { |
| 445 | gfs2_inplace_release(ip); | 708 | gfs2_inplace_release(ip); |
| 446 | out_qunlock: | 709 | out_qunlock: |
| 447 | gfs2_quota_unlock(ip); | 710 | gfs2_quota_unlock(ip); |
| 448 | out_alloc_put: | 711 | out_alloc_put: |
| 449 | gfs2_alloc_put(ip); | 712 | gfs2_alloc_put(ip); |
| 450 | } | 713 | } |
| 451 | out_putpage: | ||
| 452 | page_cache_release(page); | ||
| 453 | if (pos + len > ip->i_inode.i_size) | ||
| 454 | vmtruncate(&ip->i_inode, ip->i_inode.i_size); | ||
| 455 | out_unlock: | 714 | out_unlock: |
| 456 | gfs2_glock_dq_m(1, &ip->i_gh); | 715 | gfs2_glock_dq(&ip->i_gh); |
| 457 | out_uninit: | 716 | out_uninit: |
| 458 | gfs2_holder_uninit(&ip->i_gh); | 717 | gfs2_holder_uninit(&ip->i_gh); |
| 459 | } | ||
| 460 | |||
| 461 | return error; | 718 | return error; |
| 462 | } | 719 | } |
| 463 | 720 | ||
| @@ -565,7 +822,7 @@ static int gfs2_write_end(struct file *file, struct address_space *mapping, | |||
| 565 | struct gfs2_inode *ip = GFS2_I(inode); | 822 | struct gfs2_inode *ip = GFS2_I(inode); |
| 566 | struct gfs2_sbd *sdp = GFS2_SB(inode); | 823 | struct gfs2_sbd *sdp = GFS2_SB(inode); |
| 567 | struct buffer_head *dibh; | 824 | struct buffer_head *dibh; |
| 568 | struct gfs2_alloc *al = &ip->i_alloc; | 825 | struct gfs2_alloc *al = ip->i_alloc; |
| 569 | struct gfs2_dinode *di; | 826 | struct gfs2_dinode *di; |
| 570 | unsigned int from = pos & (PAGE_CACHE_SIZE - 1); | 827 | unsigned int from = pos & (PAGE_CACHE_SIZE - 1); |
| 571 | unsigned int to = from + len; | 828 | unsigned int to = from + len; |
| @@ -585,19 +842,16 @@ static int gfs2_write_end(struct file *file, struct address_space *mapping, | |||
| 585 | if (gfs2_is_stuffed(ip)) | 842 | if (gfs2_is_stuffed(ip)) |
| 586 | return gfs2_stuffed_write_end(inode, dibh, pos, len, copied, page); | 843 | return gfs2_stuffed_write_end(inode, dibh, pos, len, copied, page); |
| 587 | 844 | ||
| 588 | if (sdp->sd_args.ar_data == GFS2_DATA_ORDERED || gfs2_is_jdata(ip)) | 845 | if (!gfs2_is_writeback(ip)) |
| 589 | gfs2_page_add_databufs(ip, page, from, to); | 846 | gfs2_page_add_databufs(ip, page, from, to); |
| 590 | 847 | ||
| 591 | ret = generic_write_end(file, mapping, pos, len, copied, page, fsdata); | 848 | ret = generic_write_end(file, mapping, pos, len, copied, page, fsdata); |
| 592 | 849 | ||
| 593 | if (likely(ret >= 0)) { | 850 | if (likely(ret >= 0) && (inode->i_size > ip->i_di.di_size)) { |
| 594 | copied = ret; | 851 | di = (struct gfs2_dinode *)dibh->b_data; |
| 595 | if ((pos + copied) > inode->i_size) { | 852 | ip->i_di.di_size = inode->i_size; |
| 596 | di = (struct gfs2_dinode *)dibh->b_data; | 853 | di->di_size = cpu_to_be64(inode->i_size); |
| 597 | ip->i_di.di_size = inode->i_size; | 854 | mark_inode_dirty(inode); |
| 598 | di->di_size = cpu_to_be64(inode->i_size); | ||
| 599 | mark_inode_dirty(inode); | ||
| 600 | } | ||
| 601 | } | 855 | } |
| 602 | 856 | ||
| 603 | if (inode == sdp->sd_rindex) | 857 | if (inode == sdp->sd_rindex) |
| @@ -606,7 +860,7 @@ static int gfs2_write_end(struct file *file, struct address_space *mapping, | |||
| 606 | brelse(dibh); | 860 | brelse(dibh); |
| 607 | gfs2_trans_end(sdp); | 861 | gfs2_trans_end(sdp); |
| 608 | failed: | 862 | failed: |
| 609 | if (al->al_requested) { | 863 | if (al) { |
| 610 | gfs2_inplace_release(ip); | 864 | gfs2_inplace_release(ip); |
| 611 | gfs2_quota_unlock(ip); | 865 | gfs2_quota_unlock(ip); |
| 612 | gfs2_alloc_put(ip); | 866 | gfs2_alloc_put(ip); |
| @@ -625,11 +879,7 @@ failed: | |||
| 625 | 879 | ||
| 626 | static int gfs2_set_page_dirty(struct page *page) | 880 | static int gfs2_set_page_dirty(struct page *page) |
| 627 | { | 881 | { |
| 628 | struct gfs2_inode *ip = GFS2_I(page->mapping->host); | 882 | SetPageChecked(page); |
| 629 | struct gfs2_sbd *sdp = GFS2_SB(page->mapping->host); | ||
| 630 | |||
| 631 | if (sdp->sd_args.ar_data == GFS2_DATA_ORDERED || gfs2_is_jdata(ip)) | ||
| 632 | SetPageChecked(page); | ||
| 633 | return __set_page_dirty_buffers(page); | 883 | return __set_page_dirty_buffers(page); |
| 634 | } | 884 | } |
| 635 | 885 | ||
| @@ -653,7 +903,7 @@ static sector_t gfs2_bmap(struct address_space *mapping, sector_t lblock) | |||
| 653 | return 0; | 903 | return 0; |
| 654 | 904 | ||
| 655 | if (!gfs2_is_stuffed(ip)) | 905 | if (!gfs2_is_stuffed(ip)) |
| 656 | dblock = generic_block_bmap(mapping, lblock, gfs2_get_block); | 906 | dblock = generic_block_bmap(mapping, lblock, gfs2_block_map); |
| 657 | 907 | ||
| 658 | gfs2_glock_dq_uninit(&i_gh); | 908 | gfs2_glock_dq_uninit(&i_gh); |
| 659 | 909 | ||
| @@ -719,13 +969,9 @@ static int gfs2_ok_for_dio(struct gfs2_inode *ip, int rw, loff_t offset) | |||
| 719 | { | 969 | { |
| 720 | /* | 970 | /* |
| 721 | * Should we return an error here? I can't see that O_DIRECT for | 971 | * Should we return an error here? I can't see that O_DIRECT for |
| 722 | * a journaled file makes any sense. For now we'll silently fall | 972 | * a stuffed file makes any sense. For now we'll silently fall |
| 723 | * back to buffered I/O, likewise we do the same for stuffed | 973 | * back to buffered I/O |
| 724 | * files since they are (a) small and (b) unaligned. | ||
| 725 | */ | 974 | */ |
| 726 | if (gfs2_is_jdata(ip)) | ||
| 727 | return 0; | ||
| 728 | |||
| 729 | if (gfs2_is_stuffed(ip)) | 975 | if (gfs2_is_stuffed(ip)) |
| 730 | return 0; | 976 | return 0; |
| 731 | 977 | ||
| @@ -836,9 +1082,23 @@ cannot_release: | |||
| 836 | return 0; | 1082 | return 0; |
| 837 | } | 1083 | } |
| 838 | 1084 | ||
| 839 | const struct address_space_operations gfs2_file_aops = { | 1085 | static const struct address_space_operations gfs2_writeback_aops = { |
| 840 | .writepage = gfs2_writepage, | 1086 | .writepage = gfs2_writeback_writepage, |
| 841 | .writepages = gfs2_writepages, | 1087 | .writepages = gfs2_writeback_writepages, |
| 1088 | .readpage = gfs2_readpage, | ||
| 1089 | .readpages = gfs2_readpages, | ||
| 1090 | .sync_page = block_sync_page, | ||
| 1091 | .write_begin = gfs2_write_begin, | ||
| 1092 | .write_end = gfs2_write_end, | ||
| 1093 | .bmap = gfs2_bmap, | ||
| 1094 | .invalidatepage = gfs2_invalidatepage, | ||
| 1095 | .releasepage = gfs2_releasepage, | ||
| 1096 | .direct_IO = gfs2_direct_IO, | ||
| 1097 | .migratepage = buffer_migrate_page, | ||
| 1098 | }; | ||
| 1099 | |||
| 1100 | static const struct address_space_operations gfs2_ordered_aops = { | ||
| 1101 | .writepage = gfs2_ordered_writepage, | ||
| 842 | .readpage = gfs2_readpage, | 1102 | .readpage = gfs2_readpage, |
| 843 | .readpages = gfs2_readpages, | 1103 | .readpages = gfs2_readpages, |
| 844 | .sync_page = block_sync_page, | 1104 | .sync_page = block_sync_page, |
| @@ -849,5 +1109,34 @@ const struct address_space_operations gfs2_file_aops = { | |||
| 849 | .invalidatepage = gfs2_invalidatepage, | 1109 | .invalidatepage = gfs2_invalidatepage, |
| 850 | .releasepage = gfs2_releasepage, | 1110 | .releasepage = gfs2_releasepage, |
| 851 | .direct_IO = gfs2_direct_IO, | 1111 | .direct_IO = gfs2_direct_IO, |
| 1112 | .migratepage = buffer_migrate_page, | ||
| 852 | }; | 1113 | }; |
| 853 | 1114 | ||
| 1115 | static const struct address_space_operations gfs2_jdata_aops = { | ||
| 1116 | .writepage = gfs2_jdata_writepage, | ||
| 1117 | .writepages = gfs2_jdata_writepages, | ||
| 1118 | .readpage = gfs2_readpage, | ||
| 1119 | .readpages = gfs2_readpages, | ||
| 1120 | .sync_page = block_sync_page, | ||
| 1121 | .write_begin = gfs2_write_begin, | ||
| 1122 | .write_end = gfs2_write_end, | ||
| 1123 | .set_page_dirty = gfs2_set_page_dirty, | ||
| 1124 | .bmap = gfs2_bmap, | ||
| 1125 | .invalidatepage = gfs2_invalidatepage, | ||
| 1126 | .releasepage = gfs2_releasepage, | ||
| 1127 | }; | ||
| 1128 | |||
| 1129 | void gfs2_set_aops(struct inode *inode) | ||
| 1130 | { | ||
| 1131 | struct gfs2_inode *ip = GFS2_I(inode); | ||
| 1132 | |||
| 1133 | if (gfs2_is_writeback(ip)) | ||
| 1134 | inode->i_mapping->a_ops = &gfs2_writeback_aops; | ||
| 1135 | else if (gfs2_is_ordered(ip)) | ||
| 1136 | inode->i_mapping->a_ops = &gfs2_ordered_aops; | ||
| 1137 | else if (gfs2_is_jdata(ip)) | ||
| 1138 | inode->i_mapping->a_ops = &gfs2_jdata_aops; | ||
| 1139 | else | ||
| 1140 | BUG(); | ||
| 1141 | } | ||
| 1142 | |||
diff --git a/fs/gfs2/ops_address.h b/fs/gfs2/ops_address.h index fa1b5b3d28b9..5da21285bba4 100644 --- a/fs/gfs2/ops_address.h +++ b/fs/gfs2/ops_address.h | |||
| @@ -14,9 +14,10 @@ | |||
| 14 | #include <linux/buffer_head.h> | 14 | #include <linux/buffer_head.h> |
| 15 | #include <linux/mm.h> | 15 | #include <linux/mm.h> |
| 16 | 16 | ||
| 17 | extern const struct address_space_operations gfs2_file_aops; | ||
| 18 | extern int gfs2_get_block(struct inode *inode, sector_t lblock, | ||
| 19 | struct buffer_head *bh_result, int create); | ||
| 20 | extern int gfs2_releasepage(struct page *page, gfp_t gfp_mask); | 17 | extern int gfs2_releasepage(struct page *page, gfp_t gfp_mask); |
| 18 | extern int gfs2_internal_read(struct gfs2_inode *ip, | ||
| 19 | struct file_ra_state *ra_state, | ||
| 20 | char *buf, loff_t *pos, unsigned size); | ||
| 21 | extern void gfs2_set_aops(struct inode *inode); | ||
| 21 | 22 | ||
| 22 | #endif /* __OPS_ADDRESS_DOT_H__ */ | 23 | #endif /* __OPS_ADDRESS_DOT_H__ */ |
diff --git a/fs/gfs2/ops_file.c b/fs/gfs2/ops_file.c index bb11fd6752d3..f4842f2548cd 100644 --- a/fs/gfs2/ops_file.c +++ b/fs/gfs2/ops_file.c | |||
| @@ -33,57 +33,12 @@ | |||
| 33 | #include "lm.h" | 33 | #include "lm.h" |
| 34 | #include "log.h" | 34 | #include "log.h" |
| 35 | #include "meta_io.h" | 35 | #include "meta_io.h" |
| 36 | #include "ops_file.h" | ||
| 37 | #include "ops_vm.h" | ||
| 38 | #include "quota.h" | 36 | #include "quota.h" |
| 39 | #include "rgrp.h" | 37 | #include "rgrp.h" |
| 40 | #include "trans.h" | 38 | #include "trans.h" |
| 41 | #include "util.h" | 39 | #include "util.h" |
| 42 | #include "eaops.h" | 40 | #include "eaops.h" |
| 43 | 41 | #include "ops_address.h" | |
| 44 | /* | ||
| 45 | * Most fields left uninitialised to catch anybody who tries to | ||
| 46 | * use them. f_flags set to prevent file_accessed() from touching | ||
| 47 | * any other part of this. Its use is purely as a flag so that we | ||
| 48 | * know (in readpage()) whether or not do to locking. | ||
| 49 | */ | ||
| 50 | struct file gfs2_internal_file_sentinel = { | ||
| 51 | .f_flags = O_NOATIME|O_RDONLY, | ||
| 52 | }; | ||
| 53 | |||
| 54 | static int gfs2_read_actor(read_descriptor_t *desc, struct page *page, | ||
| 55 | unsigned long offset, unsigned long size) | ||
| 56 | { | ||
| 57 | char *kaddr; | ||
| 58 | unsigned long count = desc->count; | ||
| 59 | |||
| 60 | if (size > count) | ||
| 61 | size = count; | ||
| 62 | |||
| 63 | kaddr = kmap(page); | ||
| 64 | memcpy(desc->arg.data, kaddr + offset, size); | ||
| 65 | kunmap(page); | ||
| 66 | |||
| 67 | desc->count = count - size; | ||
| 68 | desc->written += size; | ||
| 69 | desc->arg.buf += size; | ||
| 70 | return size; | ||
| 71 | } | ||
| 72 | |||
| 73 | int gfs2_internal_read(struct gfs2_inode *ip, struct file_ra_state *ra_state, | ||
| 74 | char *buf, loff_t *pos, unsigned size) | ||
| 75 | { | ||
| 76 | struct inode *inode = &ip->i_inode; | ||
| 77 | read_descriptor_t desc; | ||
| 78 | desc.written = 0; | ||
| 79 | desc.arg.data = buf; | ||
| 80 | desc.count = size; | ||
| 81 | desc.error = 0; | ||
| 82 | do_generic_mapping_read(inode->i_mapping, ra_state, | ||
| 83 | &gfs2_internal_file_sentinel, pos, &desc, | ||
| 84 | gfs2_read_actor); | ||
| 85 | return desc.written ? desc.written : desc.error; | ||
| 86 | } | ||
| 87 | 42 | ||
| 88 | /** | 43 | /** |
| 89 | * gfs2_llseek - seek to a location in a file | 44 | * gfs2_llseek - seek to a location in a file |
| @@ -214,7 +169,7 @@ static int gfs2_get_flags(struct file *filp, u32 __user *ptr) | |||
| 214 | if (put_user(fsflags, ptr)) | 169 | if (put_user(fsflags, ptr)) |
| 215 | error = -EFAULT; | 170 | error = -EFAULT; |
| 216 | 171 | ||
| 217 | gfs2_glock_dq_m(1, &gh); | 172 | gfs2_glock_dq(&gh); |
| 218 | gfs2_holder_uninit(&gh); | 173 | gfs2_holder_uninit(&gh); |
| 219 | return error; | 174 | return error; |
| 220 | } | 175 | } |
| @@ -291,7 +246,16 @@ static int do_gfs2_set_flags(struct file *filp, u32 reqflags, u32 mask) | |||
| 291 | if (error) | 246 | if (error) |
| 292 | goto out; | 247 | goto out; |
| 293 | } | 248 | } |
| 294 | 249 | if ((flags ^ new_flags) & GFS2_DIF_JDATA) { | |
| 250 | if (flags & GFS2_DIF_JDATA) | ||
| 251 | gfs2_log_flush(sdp, ip->i_gl); | ||
| 252 | error = filemap_fdatawrite(inode->i_mapping); | ||
| 253 | if (error) | ||
| 254 | goto out; | ||
| 255 | error = filemap_fdatawait(inode->i_mapping); | ||
| 256 | if (error) | ||
| 257 | goto out; | ||
| 258 | } | ||
| 295 | error = gfs2_trans_begin(sdp, RES_DINODE, 0); | 259 | error = gfs2_trans_begin(sdp, RES_DINODE, 0); |
| 296 | if (error) | 260 | if (error) |
| 297 | goto out; | 261 | goto out; |
| @@ -303,6 +267,7 @@ static int do_gfs2_set_flags(struct file *filp, u32 reqflags, u32 mask) | |||
| 303 | gfs2_dinode_out(ip, bh->b_data); | 267 | gfs2_dinode_out(ip, bh->b_data); |
| 304 | brelse(bh); | 268 | brelse(bh); |
| 305 | gfs2_set_inode_flags(inode); | 269 | gfs2_set_inode_flags(inode); |
| 270 | gfs2_set_aops(inode); | ||
| 306 | out_trans_end: | 271 | out_trans_end: |
| 307 | gfs2_trans_end(sdp); | 272 | gfs2_trans_end(sdp); |
| 308 | out: | 273 | out: |
| @@ -338,6 +303,128 @@ static long gfs2_ioctl(struct file *filp, unsigned int cmd, unsigned long arg) | |||
| 338 | return -ENOTTY; | 303 | return -ENOTTY; |
| 339 | } | 304 | } |
| 340 | 305 | ||
| 306 | /** | ||
| 307 | * gfs2_allocate_page_backing - Use bmap to allocate blocks | ||
| 308 | * @page: The (locked) page to allocate backing for | ||
| 309 | * | ||
| 310 | * We try to allocate all the blocks required for the page in | ||
| 311 | * one go. This might fail for various reasons, so we keep | ||
| 312 | * trying until all the blocks to back this page are allocated. | ||
| 313 | * If some of the blocks are already allocated, thats ok too. | ||
| 314 | */ | ||
| 315 | |||
| 316 | static int gfs2_allocate_page_backing(struct page *page) | ||
| 317 | { | ||
| 318 | struct inode *inode = page->mapping->host; | ||
| 319 | struct buffer_head bh; | ||
| 320 | unsigned long size = PAGE_CACHE_SIZE; | ||
| 321 | u64 lblock = page->index << (PAGE_CACHE_SHIFT - inode->i_blkbits); | ||
| 322 | |||
| 323 | do { | ||
| 324 | bh.b_state = 0; | ||
| 325 | bh.b_size = size; | ||
| 326 | gfs2_block_map(inode, lblock, &bh, 1); | ||
| 327 | if (!buffer_mapped(&bh)) | ||
| 328 | return -EIO; | ||
| 329 | size -= bh.b_size; | ||
| 330 | lblock += (bh.b_size >> inode->i_blkbits); | ||
| 331 | } while(size > 0); | ||
| 332 | return 0; | ||
| 333 | } | ||
| 334 | |||
| 335 | /** | ||
| 336 | * gfs2_page_mkwrite - Make a shared, mmap()ed, page writable | ||
| 337 | * @vma: The virtual memory area | ||
| 338 | * @page: The page which is about to become writable | ||
| 339 | * | ||
| 340 | * When the page becomes writable, we need to ensure that we have | ||
| 341 | * blocks allocated on disk to back that page. | ||
| 342 | */ | ||
| 343 | |||
| 344 | static int gfs2_page_mkwrite(struct vm_area_struct *vma, struct page *page) | ||
| 345 | { | ||
| 346 | struct inode *inode = vma->vm_file->f_path.dentry->d_inode; | ||
| 347 | struct gfs2_inode *ip = GFS2_I(inode); | ||
| 348 | struct gfs2_sbd *sdp = GFS2_SB(inode); | ||
| 349 | unsigned long last_index; | ||
| 350 | u64 pos = page->index << (PAGE_CACHE_SIZE - inode->i_blkbits); | ||
| 351 | unsigned int data_blocks, ind_blocks, rblocks; | ||
| 352 | int alloc_required = 0; | ||
| 353 | struct gfs2_holder gh; | ||
| 354 | struct gfs2_alloc *al; | ||
| 355 | int ret; | ||
| 356 | |||
| 357 | gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, GL_ATIME, &gh); | ||
| 358 | ret = gfs2_glock_nq_atime(&gh); | ||
| 359 | if (ret) | ||
| 360 | goto out; | ||
| 361 | |||
| 362 | set_bit(GIF_SW_PAGED, &ip->i_flags); | ||
| 363 | gfs2_write_calc_reserv(ip, PAGE_CACHE_SIZE, &data_blocks, &ind_blocks); | ||
| 364 | ret = gfs2_write_alloc_required(ip, pos, PAGE_CACHE_SIZE, &alloc_required); | ||
| 365 | if (ret || !alloc_required) | ||
| 366 | goto out_unlock; | ||
| 367 | ret = -ENOMEM; | ||
| 368 | al = gfs2_alloc_get(ip); | ||
| 369 | if (al == NULL) | ||
| 370 | goto out_unlock; | ||
| 371 | |||
| 372 | ret = gfs2_quota_lock(ip, NO_QUOTA_CHANGE, NO_QUOTA_CHANGE); | ||
| 373 | if (ret) | ||
| 374 | goto out_alloc_put; | ||
| 375 | ret = gfs2_quota_check(ip, ip->i_inode.i_uid, ip->i_inode.i_gid); | ||
| 376 | if (ret) | ||
| 377 | goto out_quota_unlock; | ||
| 378 | al->al_requested = data_blocks + ind_blocks; | ||
| 379 | ret = gfs2_inplace_reserve(ip); | ||
| 380 | if (ret) | ||
| 381 | goto out_quota_unlock; | ||
| 382 | |||
| 383 | rblocks = RES_DINODE + ind_blocks; | ||
| 384 | if (gfs2_is_jdata(ip)) | ||
| 385 | rblocks += data_blocks ? data_blocks : 1; | ||
| 386 | if (ind_blocks || data_blocks) | ||
| 387 | rblocks += RES_STATFS + RES_QUOTA; | ||
| 388 | ret = gfs2_trans_begin(sdp, rblocks, 0); | ||
| 389 | if (ret) | ||
| 390 | goto out_trans_fail; | ||
| 391 | |||
| 392 | lock_page(page); | ||
| 393 | ret = -EINVAL; | ||
| 394 | last_index = ip->i_inode.i_size >> PAGE_CACHE_SHIFT; | ||
| 395 | if (page->index > last_index) | ||
| 396 | goto out_unlock_page; | ||
| 397 | ret = 0; | ||
| 398 | if (!PageUptodate(page) || page->mapping != ip->i_inode.i_mapping) | ||
| 399 | goto out_unlock_page; | ||
| 400 | if (gfs2_is_stuffed(ip)) { | ||
| 401 | ret = gfs2_unstuff_dinode(ip, page); | ||
| 402 | if (ret) | ||
| 403 | goto out_unlock_page; | ||
| 404 | } | ||
| 405 | ret = gfs2_allocate_page_backing(page); | ||
| 406 | |||
| 407 | out_unlock_page: | ||
| 408 | unlock_page(page); | ||
| 409 | gfs2_trans_end(sdp); | ||
| 410 | out_trans_fail: | ||
| 411 | gfs2_inplace_release(ip); | ||
| 412 | out_quota_unlock: | ||
| 413 | gfs2_quota_unlock(ip); | ||
| 414 | out_alloc_put: | ||
| 415 | gfs2_alloc_put(ip); | ||
| 416 | out_unlock: | ||
| 417 | gfs2_glock_dq(&gh); | ||
| 418 | out: | ||
| 419 | gfs2_holder_uninit(&gh); | ||
| 420 | return ret; | ||
| 421 | } | ||
| 422 | |||
| 423 | static struct vm_operations_struct gfs2_vm_ops = { | ||
| 424 | .fault = filemap_fault, | ||
| 425 | .page_mkwrite = gfs2_page_mkwrite, | ||
| 426 | }; | ||
| 427 | |||
| 341 | 428 | ||
| 342 | /** | 429 | /** |
| 343 | * gfs2_mmap - | 430 | * gfs2_mmap - |
| @@ -360,14 +447,7 @@ static int gfs2_mmap(struct file *file, struct vm_area_struct *vma) | |||
| 360 | return error; | 447 | return error; |
| 361 | } | 448 | } |
| 362 | 449 | ||
| 363 | /* This is VM_MAYWRITE instead of VM_WRITE because a call | 450 | vma->vm_ops = &gfs2_vm_ops; |
| 364 | to mprotect() can turn on VM_WRITE later. */ | ||
| 365 | |||
| 366 | if ((vma->vm_flags & (VM_MAYSHARE | VM_MAYWRITE)) == | ||
| 367 | (VM_MAYSHARE | VM_MAYWRITE)) | ||
| 368 | vma->vm_ops = &gfs2_vm_ops_sharewrite; | ||
| 369 | else | ||
| 370 | vma->vm_ops = &gfs2_vm_ops_private; | ||
| 371 | 451 | ||
| 372 | gfs2_glock_dq_uninit(&i_gh); | 452 | gfs2_glock_dq_uninit(&i_gh); |
| 373 | 453 | ||
| @@ -538,15 +618,6 @@ static int gfs2_lock(struct file *file, int cmd, struct file_lock *fl) | |||
| 538 | if (__mandatory_lock(&ip->i_inode)) | 618 | if (__mandatory_lock(&ip->i_inode)) |
| 539 | return -ENOLCK; | 619 | return -ENOLCK; |
| 540 | 620 | ||
| 541 | if (sdp->sd_args.ar_localflocks) { | ||
| 542 | if (IS_GETLK(cmd)) { | ||
| 543 | posix_test_lock(file, fl); | ||
| 544 | return 0; | ||
| 545 | } else { | ||
| 546 | return posix_lock_file_wait(file, fl); | ||
| 547 | } | ||
| 548 | } | ||
| 549 | |||
| 550 | if (cmd == F_CANCELLK) { | 621 | if (cmd == F_CANCELLK) { |
| 551 | /* Hack: */ | 622 | /* Hack: */ |
| 552 | cmd = F_SETLK; | 623 | cmd = F_SETLK; |
| @@ -632,16 +703,12 @@ static void do_unflock(struct file *file, struct file_lock *fl) | |||
| 632 | static int gfs2_flock(struct file *file, int cmd, struct file_lock *fl) | 703 | static int gfs2_flock(struct file *file, int cmd, struct file_lock *fl) |
| 633 | { | 704 | { |
| 634 | struct gfs2_inode *ip = GFS2_I(file->f_mapping->host); | 705 | struct gfs2_inode *ip = GFS2_I(file->f_mapping->host); |
| 635 | struct gfs2_sbd *sdp = GFS2_SB(file->f_mapping->host); | ||
| 636 | 706 | ||
| 637 | if (!(fl->fl_flags & FL_FLOCK)) | 707 | if (!(fl->fl_flags & FL_FLOCK)) |
| 638 | return -ENOLCK; | 708 | return -ENOLCK; |
| 639 | if (__mandatory_lock(&ip->i_inode)) | 709 | if (__mandatory_lock(&ip->i_inode)) |
| 640 | return -ENOLCK; | 710 | return -ENOLCK; |
| 641 | 711 | ||
| 642 | if (sdp->sd_args.ar_localflocks) | ||
| 643 | return flock_lock_file_wait(file, fl); | ||
| 644 | |||
| 645 | if (fl->fl_type == F_UNLCK) { | 712 | if (fl->fl_type == F_UNLCK) { |
| 646 | do_unflock(file, fl); | 713 | do_unflock(file, fl); |
| 647 | return 0; | 714 | return 0; |
| @@ -678,3 +745,27 @@ const struct file_operations gfs2_dir_fops = { | |||
| 678 | .flock = gfs2_flock, | 745 | .flock = gfs2_flock, |
| 679 | }; | 746 | }; |
| 680 | 747 | ||
| 748 | const struct file_operations gfs2_file_fops_nolock = { | ||
| 749 | .llseek = gfs2_llseek, | ||
| 750 | .read = do_sync_read, | ||
| 751 | .aio_read = generic_file_aio_read, | ||
| 752 | .write = do_sync_write, | ||
| 753 | .aio_write = generic_file_aio_write, | ||
| 754 | .unlocked_ioctl = gfs2_ioctl, | ||
| 755 | .mmap = gfs2_mmap, | ||
| 756 | .open = gfs2_open, | ||
| 757 | .release = gfs2_close, | ||
| 758 | .fsync = gfs2_fsync, | ||
| 759 | .splice_read = generic_file_splice_read, | ||
| 760 | .splice_write = generic_file_splice_write, | ||
| 761 | .setlease = gfs2_setlease, | ||
| 762 | }; | ||
| 763 | |||
| 764 | const struct file_operations gfs2_dir_fops_nolock = { | ||
| 765 | .readdir = gfs2_readdir, | ||
| 766 | .unlocked_ioctl = gfs2_ioctl, | ||
| 767 | .open = gfs2_open, | ||
| 768 | .release = gfs2_close, | ||
| 769 | .fsync = gfs2_fsync, | ||
| 770 | }; | ||
| 771 | |||
diff --git a/fs/gfs2/ops_file.h b/fs/gfs2/ops_file.h deleted file mode 100644 index 7e5d8ec9c846..000000000000 --- a/fs/gfs2/ops_file.h +++ /dev/null | |||
| @@ -1,24 +0,0 @@ | |||
| 1 | /* | ||
| 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | ||
| 3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | ||
| 4 | * | ||
| 5 | * This copyrighted material is made available to anyone wishing to use, | ||
| 6 | * modify, copy, or redistribute it subject to the terms and conditions | ||
| 7 | * of the GNU General Public License version 2. | ||
| 8 | */ | ||
| 9 | |||
| 10 | #ifndef __OPS_FILE_DOT_H__ | ||
| 11 | #define __OPS_FILE_DOT_H__ | ||
| 12 | |||
| 13 | #include <linux/fs.h> | ||
| 14 | struct gfs2_inode; | ||
| 15 | |||
| 16 | extern struct file gfs2_internal_file_sentinel; | ||
| 17 | extern int gfs2_internal_read(struct gfs2_inode *ip, | ||
| 18 | struct file_ra_state *ra_state, | ||
| 19 | char *buf, loff_t *pos, unsigned size); | ||
| 20 | extern void gfs2_set_inode_flags(struct inode *inode); | ||
| 21 | extern const struct file_operations gfs2_file_fops; | ||
| 22 | extern const struct file_operations gfs2_dir_fops; | ||
| 23 | |||
| 24 | #endif /* __OPS_FILE_DOT_H__ */ | ||
diff --git a/fs/gfs2/ops_fstype.c b/fs/gfs2/ops_fstype.c index 17de58e83d92..43d511bba52d 100644 --- a/fs/gfs2/ops_fstype.c +++ b/fs/gfs2/ops_fstype.c | |||
| @@ -1,6 +1,6 @@ | |||
| 1 | /* | 1 | /* |
| 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. |
| 3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | 3 | * Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved. |
| 4 | * | 4 | * |
| 5 | * This copyrighted material is made available to anyone wishing to use, | 5 | * This copyrighted material is made available to anyone wishing to use, |
| 6 | * modify, copy, or redistribute it subject to the terms and conditions | 6 | * modify, copy, or redistribute it subject to the terms and conditions |
| @@ -21,6 +21,7 @@ | |||
| 21 | 21 | ||
| 22 | #include "gfs2.h" | 22 | #include "gfs2.h" |
| 23 | #include "incore.h" | 23 | #include "incore.h" |
| 24 | #include "bmap.h" | ||
| 24 | #include "daemon.h" | 25 | #include "daemon.h" |
| 25 | #include "glock.h" | 26 | #include "glock.h" |
| 26 | #include "glops.h" | 27 | #include "glops.h" |
| @@ -59,7 +60,6 @@ static struct gfs2_sbd *init_sbd(struct super_block *sb) | |||
| 59 | 60 | ||
| 60 | mutex_init(&sdp->sd_inum_mutex); | 61 | mutex_init(&sdp->sd_inum_mutex); |
| 61 | spin_lock_init(&sdp->sd_statfs_spin); | 62 | spin_lock_init(&sdp->sd_statfs_spin); |
| 62 | mutex_init(&sdp->sd_statfs_mutex); | ||
| 63 | 63 | ||
| 64 | spin_lock_init(&sdp->sd_rindex_spin); | 64 | spin_lock_init(&sdp->sd_rindex_spin); |
| 65 | mutex_init(&sdp->sd_rindex_mutex); | 65 | mutex_init(&sdp->sd_rindex_mutex); |
| @@ -77,7 +77,6 @@ static struct gfs2_sbd *init_sbd(struct super_block *sb) | |||
| 77 | 77 | ||
| 78 | spin_lock_init(&sdp->sd_log_lock); | 78 | spin_lock_init(&sdp->sd_log_lock); |
| 79 | 79 | ||
| 80 | INIT_LIST_HEAD(&sdp->sd_log_le_gl); | ||
| 81 | INIT_LIST_HEAD(&sdp->sd_log_le_buf); | 80 | INIT_LIST_HEAD(&sdp->sd_log_le_buf); |
| 82 | INIT_LIST_HEAD(&sdp->sd_log_le_revoke); | 81 | INIT_LIST_HEAD(&sdp->sd_log_le_revoke); |
| 83 | INIT_LIST_HEAD(&sdp->sd_log_le_rg); | 82 | INIT_LIST_HEAD(&sdp->sd_log_le_rg); |
| @@ -303,6 +302,67 @@ out: | |||
| 303 | return error; | 302 | return error; |
| 304 | } | 303 | } |
| 305 | 304 | ||
| 305 | /** | ||
| 306 | * map_journal_extents - create a reusable "extent" mapping from all logical | ||
| 307 | * blocks to all physical blocks for the given journal. This will save | ||
| 308 | * us time when writing journal blocks. Most journals will have only one | ||
| 309 | * extent that maps all their logical blocks. That's because gfs2.mkfs | ||
| 310 | * arranges the journal blocks sequentially to maximize performance. | ||
| 311 | * So the extent would map the first block for the entire file length. | ||
| 312 | * However, gfs2_jadd can happen while file activity is happening, so | ||
| 313 | * those journals may not be sequential. Less likely is the case where | ||
| 314 | * the users created their own journals by mounting the metafs and | ||
| 315 | * laying it out. But it's still possible. These journals might have | ||
| 316 | * several extents. | ||
| 317 | * | ||
| 318 | * TODO: This should be done in bigger chunks rather than one block at a time, | ||
| 319 | * but since it's only done at mount time, I'm not worried about the | ||
| 320 | * time it takes. | ||
| 321 | */ | ||
| 322 | static int map_journal_extents(struct gfs2_sbd *sdp) | ||
| 323 | { | ||
| 324 | struct gfs2_jdesc *jd = sdp->sd_jdesc; | ||
| 325 | unsigned int lb; | ||
| 326 | u64 db, prev_db; /* logical block, disk block, prev disk block */ | ||
| 327 | struct gfs2_inode *ip = GFS2_I(jd->jd_inode); | ||
| 328 | struct gfs2_journal_extent *jext = NULL; | ||
| 329 | struct buffer_head bh; | ||
| 330 | int rc = 0; | ||
| 331 | |||
| 332 | prev_db = 0; | ||
| 333 | |||
| 334 | for (lb = 0; lb < ip->i_di.di_size >> sdp->sd_sb.sb_bsize_shift; lb++) { | ||
| 335 | bh.b_state = 0; | ||
| 336 | bh.b_blocknr = 0; | ||
| 337 | bh.b_size = 1 << ip->i_inode.i_blkbits; | ||
| 338 | rc = gfs2_block_map(jd->jd_inode, lb, &bh, 0); | ||
| 339 | db = bh.b_blocknr; | ||
| 340 | if (rc || !db) { | ||
| 341 | printk(KERN_INFO "GFS2 journal mapping error %d: lb=" | ||
| 342 | "%u db=%llu\n", rc, lb, (unsigned long long)db); | ||
| 343 | break; | ||
| 344 | } | ||
| 345 | if (!prev_db || db != prev_db + 1) { | ||
| 346 | jext = kzalloc(sizeof(struct gfs2_journal_extent), | ||
| 347 | GFP_KERNEL); | ||
| 348 | if (!jext) { | ||
| 349 | printk(KERN_INFO "GFS2 error: out of memory " | ||
| 350 | "mapping journal extents.\n"); | ||
| 351 | rc = -ENOMEM; | ||
| 352 | break; | ||
| 353 | } | ||
| 354 | jext->dblock = db; | ||
| 355 | jext->lblock = lb; | ||
| 356 | jext->blocks = 1; | ||
| 357 | list_add_tail(&jext->extent_list, &jd->extent_list); | ||
| 358 | } else { | ||
| 359 | jext->blocks++; | ||
| 360 | } | ||
| 361 | prev_db = db; | ||
| 362 | } | ||
| 363 | return rc; | ||
| 364 | } | ||
| 365 | |||
| 306 | static int init_journal(struct gfs2_sbd *sdp, int undo) | 366 | static int init_journal(struct gfs2_sbd *sdp, int undo) |
| 307 | { | 367 | { |
| 308 | struct gfs2_holder ji_gh; | 368 | struct gfs2_holder ji_gh; |
| @@ -340,7 +400,7 @@ static int init_journal(struct gfs2_sbd *sdp, int undo) | |||
| 340 | 400 | ||
| 341 | if (sdp->sd_args.ar_spectator) { | 401 | if (sdp->sd_args.ar_spectator) { |
| 342 | sdp->sd_jdesc = gfs2_jdesc_find(sdp, 0); | 402 | sdp->sd_jdesc = gfs2_jdesc_find(sdp, 0); |
| 343 | sdp->sd_log_blks_free = sdp->sd_jdesc->jd_blocks; | 403 | atomic_set(&sdp->sd_log_blks_free, sdp->sd_jdesc->jd_blocks); |
| 344 | } else { | 404 | } else { |
| 345 | if (sdp->sd_lockstruct.ls_jid >= gfs2_jindex_size(sdp)) { | 405 | if (sdp->sd_lockstruct.ls_jid >= gfs2_jindex_size(sdp)) { |
| 346 | fs_err(sdp, "can't mount journal #%u\n", | 406 | fs_err(sdp, "can't mount journal #%u\n", |
| @@ -377,7 +437,10 @@ static int init_journal(struct gfs2_sbd *sdp, int undo) | |||
| 377 | sdp->sd_jdesc->jd_jid, error); | 437 | sdp->sd_jdesc->jd_jid, error); |
| 378 | goto fail_jinode_gh; | 438 | goto fail_jinode_gh; |
| 379 | } | 439 | } |
| 380 | sdp->sd_log_blks_free = sdp->sd_jdesc->jd_blocks; | 440 | atomic_set(&sdp->sd_log_blks_free, sdp->sd_jdesc->jd_blocks); |
| 441 | |||
| 442 | /* Map the extents for this journal's blocks */ | ||
| 443 | map_journal_extents(sdp); | ||
| 381 | } | 444 | } |
| 382 | 445 | ||
| 383 | if (sdp->sd_lockstruct.ls_first) { | 446 | if (sdp->sd_lockstruct.ls_first) { |
diff --git a/fs/gfs2/ops_inode.c b/fs/gfs2/ops_inode.c index 291f0c7eaa3b..9f71372c1757 100644 --- a/fs/gfs2/ops_inode.c +++ b/fs/gfs2/ops_inode.c | |||
| @@ -61,7 +61,7 @@ static int gfs2_create(struct inode *dir, struct dentry *dentry, | |||
| 61 | inode = gfs2_createi(ghs, &dentry->d_name, S_IFREG | mode, 0); | 61 | inode = gfs2_createi(ghs, &dentry->d_name, S_IFREG | mode, 0); |
| 62 | if (!IS_ERR(inode)) { | 62 | if (!IS_ERR(inode)) { |
| 63 | gfs2_trans_end(sdp); | 63 | gfs2_trans_end(sdp); |
| 64 | if (dip->i_alloc.al_rgd) | 64 | if (dip->i_alloc->al_rgd) |
| 65 | gfs2_inplace_release(dip); | 65 | gfs2_inplace_release(dip); |
| 66 | gfs2_quota_unlock(dip); | 66 | gfs2_quota_unlock(dip); |
| 67 | gfs2_alloc_put(dip); | 67 | gfs2_alloc_put(dip); |
| @@ -113,8 +113,18 @@ static struct dentry *gfs2_lookup(struct inode *dir, struct dentry *dentry, | |||
| 113 | if (inode && IS_ERR(inode)) | 113 | if (inode && IS_ERR(inode)) |
| 114 | return ERR_PTR(PTR_ERR(inode)); | 114 | return ERR_PTR(PTR_ERR(inode)); |
| 115 | 115 | ||
| 116 | if (inode) | 116 | if (inode) { |
| 117 | struct gfs2_glock *gl = GFS2_I(inode)->i_gl; | ||
| 118 | struct gfs2_holder gh; | ||
| 119 | int error; | ||
| 120 | error = gfs2_glock_nq_init(gl, LM_ST_SHARED, LM_FLAG_ANY, &gh); | ||
| 121 | if (error) { | ||
| 122 | iput(inode); | ||
| 123 | return ERR_PTR(error); | ||
| 124 | } | ||
| 125 | gfs2_glock_dq_uninit(&gh); | ||
| 117 | return d_splice_alias(inode, dentry); | 126 | return d_splice_alias(inode, dentry); |
| 127 | } | ||
| 118 | d_add(dentry, inode); | 128 | d_add(dentry, inode); |
| 119 | 129 | ||
| 120 | return NULL; | 130 | return NULL; |
| @@ -366,7 +376,7 @@ static int gfs2_symlink(struct inode *dir, struct dentry *dentry, | |||
| 366 | } | 376 | } |
| 367 | 377 | ||
| 368 | gfs2_trans_end(sdp); | 378 | gfs2_trans_end(sdp); |
| 369 | if (dip->i_alloc.al_rgd) | 379 | if (dip->i_alloc->al_rgd) |
| 370 | gfs2_inplace_release(dip); | 380 | gfs2_inplace_release(dip); |
| 371 | gfs2_quota_unlock(dip); | 381 | gfs2_quota_unlock(dip); |
| 372 | gfs2_alloc_put(dip); | 382 | gfs2_alloc_put(dip); |
| @@ -442,7 +452,7 @@ static int gfs2_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
| 442 | gfs2_assert_withdraw(sdp, !error); /* dip already pinned */ | 452 | gfs2_assert_withdraw(sdp, !error); /* dip already pinned */ |
| 443 | 453 | ||
| 444 | gfs2_trans_end(sdp); | 454 | gfs2_trans_end(sdp); |
| 445 | if (dip->i_alloc.al_rgd) | 455 | if (dip->i_alloc->al_rgd) |
| 446 | gfs2_inplace_release(dip); | 456 | gfs2_inplace_release(dip); |
| 447 | gfs2_quota_unlock(dip); | 457 | gfs2_quota_unlock(dip); |
| 448 | gfs2_alloc_put(dip); | 458 | gfs2_alloc_put(dip); |
| @@ -548,7 +558,7 @@ static int gfs2_mknod(struct inode *dir, struct dentry *dentry, int mode, | |||
| 548 | } | 558 | } |
| 549 | 559 | ||
| 550 | gfs2_trans_end(sdp); | 560 | gfs2_trans_end(sdp); |
| 551 | if (dip->i_alloc.al_rgd) | 561 | if (dip->i_alloc->al_rgd) |
| 552 | gfs2_inplace_release(dip); | 562 | gfs2_inplace_release(dip); |
| 553 | gfs2_quota_unlock(dip); | 563 | gfs2_quota_unlock(dip); |
| 554 | gfs2_alloc_put(dip); | 564 | gfs2_alloc_put(dip); |
diff --git a/fs/gfs2/ops_inode.h b/fs/gfs2/ops_inode.h index 34f0caac1a03..fd8cee231e1d 100644 --- a/fs/gfs2/ops_inode.h +++ b/fs/gfs2/ops_inode.h | |||
| @@ -16,5 +16,11 @@ extern const struct inode_operations gfs2_file_iops; | |||
| 16 | extern const struct inode_operations gfs2_dir_iops; | 16 | extern const struct inode_operations gfs2_dir_iops; |
| 17 | extern const struct inode_operations gfs2_symlink_iops; | 17 | extern const struct inode_operations gfs2_symlink_iops; |
| 18 | extern const struct inode_operations gfs2_dev_iops; | 18 | extern const struct inode_operations gfs2_dev_iops; |
| 19 | extern const struct file_operations gfs2_file_fops; | ||
| 20 | extern const struct file_operations gfs2_dir_fops; | ||
| 21 | extern const struct file_operations gfs2_file_fops_nolock; | ||
| 22 | extern const struct file_operations gfs2_dir_fops_nolock; | ||
| 23 | |||
| 24 | extern void gfs2_set_inode_flags(struct inode *inode); | ||
| 19 | 25 | ||
| 20 | #endif /* __OPS_INODE_DOT_H__ */ | 26 | #endif /* __OPS_INODE_DOT_H__ */ |
diff --git a/fs/gfs2/ops_super.c b/fs/gfs2/ops_super.c index 950f31460e8b..5e524217944a 100644 --- a/fs/gfs2/ops_super.c +++ b/fs/gfs2/ops_super.c | |||
| @@ -487,7 +487,6 @@ static struct inode *gfs2_alloc_inode(struct super_block *sb) | |||
| 487 | if (ip) { | 487 | if (ip) { |
| 488 | ip->i_flags = 0; | 488 | ip->i_flags = 0; |
| 489 | ip->i_gl = NULL; | 489 | ip->i_gl = NULL; |
| 490 | ip->i_last_pfault = jiffies; | ||
| 491 | } | 490 | } |
| 492 | return &ip->i_inode; | 491 | return &ip->i_inode; |
| 493 | } | 492 | } |
diff --git a/fs/gfs2/ops_vm.c b/fs/gfs2/ops_vm.c deleted file mode 100644 index 927d739d4685..000000000000 --- a/fs/gfs2/ops_vm.c +++ /dev/null | |||
| @@ -1,169 +0,0 @@ | |||
| 1 | /* | ||
| 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | ||
| 3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | ||
| 4 | * | ||
| 5 | * This copyrighted material is made available to anyone wishing to use, | ||
| 6 | * modify, copy, or redistribute it subject to the terms and conditions | ||
| 7 | * of the GNU General Public License version 2. | ||
| 8 | */ | ||
| 9 | |||
| 10 | #include <linux/slab.h> | ||
| 11 | #include <linux/spinlock.h> | ||
| 12 | #include <linux/completion.h> | ||
| 13 | #include <linux/buffer_head.h> | ||
| 14 | #include <linux/mm.h> | ||
| 15 | #include <linux/pagemap.h> | ||
| 16 | #include <linux/gfs2_ondisk.h> | ||
| 17 | #include <linux/lm_interface.h> | ||
| 18 | |||
| 19 | #include "gfs2.h" | ||
| 20 | #include "incore.h" | ||
| 21 | #include "bmap.h" | ||
| 22 | #include "glock.h" | ||
| 23 | #include "inode.h" | ||
| 24 | #include "ops_vm.h" | ||
| 25 | #include "quota.h" | ||
| 26 | #include "rgrp.h" | ||
| 27 | #include "trans.h" | ||
| 28 | #include "util.h" | ||
| 29 | |||
| 30 | static int gfs2_private_fault(struct vm_area_struct *vma, struct vm_fault *vmf) | ||
| 31 | { | ||
| 32 | struct gfs2_inode *ip = GFS2_I(vma->vm_file->f_mapping->host); | ||
| 33 | |||
| 34 | set_bit(GIF_PAGED, &ip->i_flags); | ||
| 35 | return filemap_fault(vma, vmf); | ||
| 36 | } | ||
| 37 | |||
| 38 | static int alloc_page_backing(struct gfs2_inode *ip, struct page *page) | ||
| 39 | { | ||
| 40 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | ||
| 41 | unsigned long index = page->index; | ||
| 42 | u64 lblock = index << (PAGE_CACHE_SHIFT - | ||
| 43 | sdp->sd_sb.sb_bsize_shift); | ||
| 44 | unsigned int blocks = PAGE_CACHE_SIZE >> sdp->sd_sb.sb_bsize_shift; | ||
| 45 | struct gfs2_alloc *al; | ||
| 46 | unsigned int data_blocks, ind_blocks; | ||
| 47 | unsigned int x; | ||
| 48 | int error; | ||
| 49 | |||
| 50 | al = gfs2_alloc_get(ip); | ||
| 51 | |||
| 52 | error = gfs2_quota_lock(ip, NO_QUOTA_CHANGE, NO_QUOTA_CHANGE); | ||
| 53 | if (error) | ||
| 54 | goto out; | ||
| 55 | |||
| 56 | error = gfs2_quota_check(ip, ip->i_inode.i_uid, ip->i_inode.i_gid); | ||
| 57 | if (error) | ||
| 58 | goto out_gunlock_q; | ||
| 59 | |||
| 60 | gfs2_write_calc_reserv(ip, PAGE_CACHE_SIZE, &data_blocks, &ind_blocks); | ||
| 61 | |||
| 62 | al->al_requested = data_blocks + ind_blocks; | ||
| 63 | |||
| 64 | error = gfs2_inplace_reserve(ip); | ||
| 65 | if (error) | ||
| 66 | goto out_gunlock_q; | ||
| 67 | |||
| 68 | error = gfs2_trans_begin(sdp, al->al_rgd->rd_length + | ||
| 69 | ind_blocks + RES_DINODE + | ||
| 70 | RES_STATFS + RES_QUOTA, 0); | ||
| 71 | if (error) | ||
| 72 | goto out_ipres; | ||
| 73 | |||
| 74 | if (gfs2_is_stuffed(ip)) { | ||
| 75 | error = gfs2_unstuff_dinode(ip, NULL); | ||
| 76 | if (error) | ||
| 77 | goto out_trans; | ||
| 78 | } | ||
| 79 | |||
| 80 | for (x = 0; x < blocks; ) { | ||
| 81 | u64 dblock; | ||
| 82 | unsigned int extlen; | ||
| 83 | int new = 1; | ||
| 84 | |||
| 85 | error = gfs2_extent_map(&ip->i_inode, lblock, &new, &dblock, &extlen); | ||
| 86 | if (error) | ||
| 87 | goto out_trans; | ||
| 88 | |||
| 89 | lblock += extlen; | ||
| 90 | x += extlen; | ||
| 91 | } | ||
| 92 | |||
| 93 | gfs2_assert_warn(sdp, al->al_alloced); | ||
| 94 | |||
| 95 | out_trans: | ||
| 96 | gfs2_trans_end(sdp); | ||
| 97 | out_ipres: | ||
| 98 | gfs2_inplace_release(ip); | ||
| 99 | out_gunlock_q: | ||
| 100 | gfs2_quota_unlock(ip); | ||
| 101 | out: | ||
| 102 | gfs2_alloc_put(ip); | ||
| 103 | return error; | ||
| 104 | } | ||
| 105 | |||
| 106 | static int gfs2_sharewrite_fault(struct vm_area_struct *vma, | ||
| 107 | struct vm_fault *vmf) | ||
| 108 | { | ||
| 109 | struct file *file = vma->vm_file; | ||
| 110 | struct gfs2_file *gf = file->private_data; | ||
| 111 | struct gfs2_inode *ip = GFS2_I(file->f_mapping->host); | ||
| 112 | struct gfs2_holder i_gh; | ||
| 113 | int alloc_required; | ||
| 114 | int error; | ||
| 115 | int ret = 0; | ||
| 116 | |||
| 117 | error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh); | ||
| 118 | if (error) | ||
| 119 | goto out; | ||
| 120 | |||
| 121 | set_bit(GIF_PAGED, &ip->i_flags); | ||
| 122 | set_bit(GIF_SW_PAGED, &ip->i_flags); | ||
| 123 | |||
| 124 | error = gfs2_write_alloc_required(ip, | ||
| 125 | (u64)vmf->pgoff << PAGE_CACHE_SHIFT, | ||
| 126 | PAGE_CACHE_SIZE, &alloc_required); | ||
| 127 | if (error) { | ||
| 128 | ret = VM_FAULT_OOM; /* XXX: are these right? */ | ||
| 129 | goto out_unlock; | ||
| 130 | } | ||
| 131 | |||
| 132 | set_bit(GFF_EXLOCK, &gf->f_flags); | ||
| 133 | ret = filemap_fault(vma, vmf); | ||
| 134 | clear_bit(GFF_EXLOCK, &gf->f_flags); | ||
| 135 | if (ret & VM_FAULT_ERROR) | ||
| 136 | goto out_unlock; | ||
| 137 | |||
| 138 | if (alloc_required) { | ||
| 139 | /* XXX: do we need to drop page lock around alloc_page_backing?*/ | ||
| 140 | error = alloc_page_backing(ip, vmf->page); | ||
| 141 | if (error) { | ||
| 142 | /* | ||
| 143 | * VM_FAULT_LOCKED should always be the case for | ||
| 144 | * filemap_fault, but it may not be in a future | ||
| 145 | * implementation. | ||
| 146 | */ | ||
| 147 | if (ret & VM_FAULT_LOCKED) | ||
| 148 | unlock_page(vmf->page); | ||
| 149 | page_cache_release(vmf->page); | ||
| 150 | ret = VM_FAULT_OOM; | ||
| 151 | goto out_unlock; | ||
| 152 | } | ||
| 153 | set_page_dirty(vmf->page); | ||
| 154 | } | ||
| 155 | |||
| 156 | out_unlock: | ||
| 157 | gfs2_glock_dq_uninit(&i_gh); | ||
| 158 | out: | ||
| 159 | return ret; | ||
| 160 | } | ||
| 161 | |||
| 162 | struct vm_operations_struct gfs2_vm_ops_private = { | ||
| 163 | .fault = gfs2_private_fault, | ||
| 164 | }; | ||
| 165 | |||
| 166 | struct vm_operations_struct gfs2_vm_ops_sharewrite = { | ||
| 167 | .fault = gfs2_sharewrite_fault, | ||
| 168 | }; | ||
| 169 | |||
diff --git a/fs/gfs2/ops_vm.h b/fs/gfs2/ops_vm.h deleted file mode 100644 index 4ae8f43ed5e3..000000000000 --- a/fs/gfs2/ops_vm.h +++ /dev/null | |||
| @@ -1,18 +0,0 @@ | |||
| 1 | /* | ||
| 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | ||
| 3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | ||
| 4 | * | ||
| 5 | * This copyrighted material is made available to anyone wishing to use, | ||
| 6 | * modify, copy, or redistribute it subject to the terms and conditions | ||
| 7 | * of the GNU General Public License version 2. | ||
| 8 | */ | ||
| 9 | |||
| 10 | #ifndef __OPS_VM_DOT_H__ | ||
| 11 | #define __OPS_VM_DOT_H__ | ||
| 12 | |||
| 13 | #include <linux/mm.h> | ||
| 14 | |||
| 15 | extern struct vm_operations_struct gfs2_vm_ops_private; | ||
| 16 | extern struct vm_operations_struct gfs2_vm_ops_sharewrite; | ||
| 17 | |||
| 18 | #endif /* __OPS_VM_DOT_H__ */ | ||
diff --git a/fs/gfs2/quota.c b/fs/gfs2/quota.c index addb51e0f135..a08dabd6ce90 100644 --- a/fs/gfs2/quota.c +++ b/fs/gfs2/quota.c | |||
| @@ -1,6 +1,6 @@ | |||
| 1 | /* | 1 | /* |
| 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. |
| 3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | 3 | * Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved. |
| 4 | * | 4 | * |
| 5 | * This copyrighted material is made available to anyone wishing to use, | 5 | * This copyrighted material is made available to anyone wishing to use, |
| 6 | * modify, copy, or redistribute it subject to the terms and conditions | 6 | * modify, copy, or redistribute it subject to the terms and conditions |
| @@ -59,7 +59,6 @@ | |||
| 59 | #include "super.h" | 59 | #include "super.h" |
| 60 | #include "trans.h" | 60 | #include "trans.h" |
| 61 | #include "inode.h" | 61 | #include "inode.h" |
| 62 | #include "ops_file.h" | ||
| 63 | #include "ops_address.h" | 62 | #include "ops_address.h" |
| 64 | #include "util.h" | 63 | #include "util.h" |
| 65 | 64 | ||
| @@ -274,10 +273,10 @@ static int bh_get(struct gfs2_quota_data *qd) | |||
| 274 | } | 273 | } |
| 275 | 274 | ||
| 276 | block = qd->qd_slot / sdp->sd_qc_per_block; | 275 | block = qd->qd_slot / sdp->sd_qc_per_block; |
| 277 | offset = qd->qd_slot % sdp->sd_qc_per_block;; | 276 | offset = qd->qd_slot % sdp->sd_qc_per_block; |
| 278 | 277 | ||
| 279 | bh_map.b_size = 1 << ip->i_inode.i_blkbits; | 278 | bh_map.b_size = 1 << ip->i_inode.i_blkbits; |
| 280 | error = gfs2_block_map(&ip->i_inode, block, 0, &bh_map); | 279 | error = gfs2_block_map(&ip->i_inode, block, &bh_map, 0); |
| 281 | if (error) | 280 | if (error) |
| 282 | goto fail; | 281 | goto fail; |
| 283 | error = gfs2_meta_read(ip->i_gl, bh_map.b_blocknr, DIO_WAIT, &bh); | 282 | error = gfs2_meta_read(ip->i_gl, bh_map.b_blocknr, DIO_WAIT, &bh); |
| @@ -454,7 +453,7 @@ static void qdsb_put(struct gfs2_quota_data *qd) | |||
| 454 | int gfs2_quota_hold(struct gfs2_inode *ip, u32 uid, u32 gid) | 453 | int gfs2_quota_hold(struct gfs2_inode *ip, u32 uid, u32 gid) |
| 455 | { | 454 | { |
| 456 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 455 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
| 457 | struct gfs2_alloc *al = &ip->i_alloc; | 456 | struct gfs2_alloc *al = ip->i_alloc; |
| 458 | struct gfs2_quota_data **qd = al->al_qd; | 457 | struct gfs2_quota_data **qd = al->al_qd; |
| 459 | int error; | 458 | int error; |
| 460 | 459 | ||
| @@ -502,7 +501,7 @@ out: | |||
| 502 | void gfs2_quota_unhold(struct gfs2_inode *ip) | 501 | void gfs2_quota_unhold(struct gfs2_inode *ip) |
| 503 | { | 502 | { |
| 504 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 503 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
| 505 | struct gfs2_alloc *al = &ip->i_alloc; | 504 | struct gfs2_alloc *al = ip->i_alloc; |
| 506 | unsigned int x; | 505 | unsigned int x; |
| 507 | 506 | ||
| 508 | gfs2_assert_warn(sdp, !test_bit(GIF_QD_LOCKED, &ip->i_flags)); | 507 | gfs2_assert_warn(sdp, !test_bit(GIF_QD_LOCKED, &ip->i_flags)); |
| @@ -646,7 +645,7 @@ static int gfs2_adjust_quota(struct gfs2_inode *ip, loff_t loc, | |||
| 646 | } | 645 | } |
| 647 | 646 | ||
| 648 | if (!buffer_mapped(bh)) { | 647 | if (!buffer_mapped(bh)) { |
| 649 | gfs2_get_block(inode, iblock, bh, 1); | 648 | gfs2_block_map(inode, iblock, bh, 1); |
| 650 | if (!buffer_mapped(bh)) | 649 | if (!buffer_mapped(bh)) |
| 651 | goto unlock; | 650 | goto unlock; |
| 652 | } | 651 | } |
| @@ -793,11 +792,9 @@ static int do_glock(struct gfs2_quota_data *qd, int force_refresh, | |||
| 793 | struct gfs2_holder i_gh; | 792 | struct gfs2_holder i_gh; |
| 794 | struct gfs2_quota_host q; | 793 | struct gfs2_quota_host q; |
| 795 | char buf[sizeof(struct gfs2_quota)]; | 794 | char buf[sizeof(struct gfs2_quota)]; |
| 796 | struct file_ra_state ra_state; | ||
| 797 | int error; | 795 | int error; |
| 798 | struct gfs2_quota_lvb *qlvb; | 796 | struct gfs2_quota_lvb *qlvb; |
| 799 | 797 | ||
| 800 | file_ra_state_init(&ra_state, sdp->sd_quota_inode->i_mapping); | ||
| 801 | restart: | 798 | restart: |
| 802 | error = gfs2_glock_nq_init(qd->qd_gl, LM_ST_SHARED, 0, q_gh); | 799 | error = gfs2_glock_nq_init(qd->qd_gl, LM_ST_SHARED, 0, q_gh); |
| 803 | if (error) | 800 | if (error) |
| @@ -820,8 +817,8 @@ restart: | |||
| 820 | 817 | ||
| 821 | memset(buf, 0, sizeof(struct gfs2_quota)); | 818 | memset(buf, 0, sizeof(struct gfs2_quota)); |
| 822 | pos = qd2offset(qd); | 819 | pos = qd2offset(qd); |
| 823 | error = gfs2_internal_read(ip, &ra_state, buf, | 820 | error = gfs2_internal_read(ip, NULL, buf, &pos, |
| 824 | &pos, sizeof(struct gfs2_quota)); | 821 | sizeof(struct gfs2_quota)); |
| 825 | if (error < 0) | 822 | if (error < 0) |
| 826 | goto fail_gunlock; | 823 | goto fail_gunlock; |
| 827 | 824 | ||
| @@ -856,7 +853,7 @@ fail: | |||
| 856 | int gfs2_quota_lock(struct gfs2_inode *ip, u32 uid, u32 gid) | 853 | int gfs2_quota_lock(struct gfs2_inode *ip, u32 uid, u32 gid) |
| 857 | { | 854 | { |
| 858 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 855 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
| 859 | struct gfs2_alloc *al = &ip->i_alloc; | 856 | struct gfs2_alloc *al = ip->i_alloc; |
| 860 | unsigned int x; | 857 | unsigned int x; |
| 861 | int error = 0; | 858 | int error = 0; |
| 862 | 859 | ||
| @@ -924,7 +921,7 @@ static int need_sync(struct gfs2_quota_data *qd) | |||
| 924 | 921 | ||
| 925 | void gfs2_quota_unlock(struct gfs2_inode *ip) | 922 | void gfs2_quota_unlock(struct gfs2_inode *ip) |
| 926 | { | 923 | { |
| 927 | struct gfs2_alloc *al = &ip->i_alloc; | 924 | struct gfs2_alloc *al = ip->i_alloc; |
| 928 | struct gfs2_quota_data *qda[4]; | 925 | struct gfs2_quota_data *qda[4]; |
| 929 | unsigned int count = 0; | 926 | unsigned int count = 0; |
| 930 | unsigned int x; | 927 | unsigned int x; |
| @@ -972,7 +969,7 @@ static int print_message(struct gfs2_quota_data *qd, char *type) | |||
| 972 | int gfs2_quota_check(struct gfs2_inode *ip, u32 uid, u32 gid) | 969 | int gfs2_quota_check(struct gfs2_inode *ip, u32 uid, u32 gid) |
| 973 | { | 970 | { |
| 974 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 971 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
| 975 | struct gfs2_alloc *al = &ip->i_alloc; | 972 | struct gfs2_alloc *al = ip->i_alloc; |
| 976 | struct gfs2_quota_data *qd; | 973 | struct gfs2_quota_data *qd; |
| 977 | s64 value; | 974 | s64 value; |
| 978 | unsigned int x; | 975 | unsigned int x; |
| @@ -1016,10 +1013,9 @@ int gfs2_quota_check(struct gfs2_inode *ip, u32 uid, u32 gid) | |||
| 1016 | void gfs2_quota_change(struct gfs2_inode *ip, s64 change, | 1013 | void gfs2_quota_change(struct gfs2_inode *ip, s64 change, |
| 1017 | u32 uid, u32 gid) | 1014 | u32 uid, u32 gid) |
| 1018 | { | 1015 | { |
| 1019 | struct gfs2_alloc *al = &ip->i_alloc; | 1016 | struct gfs2_alloc *al = ip->i_alloc; |
| 1020 | struct gfs2_quota_data *qd; | 1017 | struct gfs2_quota_data *qd; |
| 1021 | unsigned int x; | 1018 | unsigned int x; |
| 1022 | unsigned int found = 0; | ||
| 1023 | 1019 | ||
| 1024 | if (gfs2_assert_warn(GFS2_SB(&ip->i_inode), change)) | 1020 | if (gfs2_assert_warn(GFS2_SB(&ip->i_inode), change)) |
| 1025 | return; | 1021 | return; |
| @@ -1032,7 +1028,6 @@ void gfs2_quota_change(struct gfs2_inode *ip, s64 change, | |||
| 1032 | if ((qd->qd_id == uid && test_bit(QDF_USER, &qd->qd_flags)) || | 1028 | if ((qd->qd_id == uid && test_bit(QDF_USER, &qd->qd_flags)) || |
| 1033 | (qd->qd_id == gid && !test_bit(QDF_USER, &qd->qd_flags))) { | 1029 | (qd->qd_id == gid && !test_bit(QDF_USER, &qd->qd_flags))) { |
| 1034 | do_qc(qd, change); | 1030 | do_qc(qd, change); |
| 1035 | found++; | ||
| 1036 | } | 1031 | } |
| 1037 | } | 1032 | } |
| 1038 | } | 1033 | } |
diff --git a/fs/gfs2/recovery.c b/fs/gfs2/recovery.c index beb6c7ac0086..b249e294a95b 100644 --- a/fs/gfs2/recovery.c +++ b/fs/gfs2/recovery.c | |||
| @@ -391,7 +391,7 @@ static int clean_journal(struct gfs2_jdesc *jd, struct gfs2_log_header_host *hea | |||
| 391 | lblock = head->lh_blkno; | 391 | lblock = head->lh_blkno; |
| 392 | gfs2_replay_incr_blk(sdp, &lblock); | 392 | gfs2_replay_incr_blk(sdp, &lblock); |
| 393 | bh_map.b_size = 1 << ip->i_inode.i_blkbits; | 393 | bh_map.b_size = 1 << ip->i_inode.i_blkbits; |
| 394 | error = gfs2_block_map(&ip->i_inode, lblock, 0, &bh_map); | 394 | error = gfs2_block_map(&ip->i_inode, lblock, &bh_map, 0); |
| 395 | if (error) | 395 | if (error) |
| 396 | return error; | 396 | return error; |
| 397 | if (!bh_map.b_blocknr) { | 397 | if (!bh_map.b_blocknr) { |
| @@ -504,13 +504,21 @@ int gfs2_recover_journal(struct gfs2_jdesc *jd) | |||
| 504 | if (!test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags)) | 504 | if (!test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags)) |
| 505 | ro = 1; | 505 | ro = 1; |
| 506 | } else { | 506 | } else { |
| 507 | if (sdp->sd_vfs->s_flags & MS_RDONLY) | 507 | if (sdp->sd_vfs->s_flags & MS_RDONLY) { |
| 508 | ro = 1; | 508 | /* check if device itself is read-only */ |
| 509 | ro = bdev_read_only(sdp->sd_vfs->s_bdev); | ||
| 510 | if (!ro) { | ||
| 511 | fs_info(sdp, "recovery required on " | ||
| 512 | "read-only filesystem.\n"); | ||
| 513 | fs_info(sdp, "write access will be " | ||
| 514 | "enabled during recovery.\n"); | ||
| 515 | } | ||
| 516 | } | ||
| 509 | } | 517 | } |
| 510 | 518 | ||
| 511 | if (ro) { | 519 | if (ro) { |
| 512 | fs_warn(sdp, "jid=%u: Can't replay: read-only FS\n", | 520 | fs_warn(sdp, "jid=%u: Can't replay: read-only block " |
| 513 | jd->jd_jid); | 521 | "device\n", jd->jd_jid); |
| 514 | error = -EROFS; | 522 | error = -EROFS; |
| 515 | goto fail_gunlock_tr; | 523 | goto fail_gunlock_tr; |
| 516 | } | 524 | } |
diff --git a/fs/gfs2/rgrp.c b/fs/gfs2/rgrp.c index 708c287e1d0e..3552110b2e5f 100644 --- a/fs/gfs2/rgrp.c +++ b/fs/gfs2/rgrp.c | |||
| @@ -25,10 +25,10 @@ | |||
| 25 | #include "rgrp.h" | 25 | #include "rgrp.h" |
| 26 | #include "super.h" | 26 | #include "super.h" |
| 27 | #include "trans.h" | 27 | #include "trans.h" |
| 28 | #include "ops_file.h" | ||
| 29 | #include "util.h" | 28 | #include "util.h" |
| 30 | #include "log.h" | 29 | #include "log.h" |
| 31 | #include "inode.h" | 30 | #include "inode.h" |
| 31 | #include "ops_address.h" | ||
| 32 | 32 | ||
| 33 | #define BFITNOENT ((u32)~0) | 33 | #define BFITNOENT ((u32)~0) |
| 34 | #define NO_BLOCK ((u64)~0) | 34 | #define NO_BLOCK ((u64)~0) |
| @@ -126,41 +126,43 @@ static unsigned char gfs2_testbit(struct gfs2_rgrpd *rgd, unsigned char *buffer, | |||
| 126 | * Return: the block number (bitmap buffer scope) that was found | 126 | * Return: the block number (bitmap buffer scope) that was found |
| 127 | */ | 127 | */ |
| 128 | 128 | ||
| 129 | static u32 gfs2_bitfit(struct gfs2_rgrpd *rgd, unsigned char *buffer, | 129 | static u32 gfs2_bitfit(unsigned char *buffer, unsigned int buflen, u32 goal, |
| 130 | unsigned int buflen, u32 goal, | 130 | unsigned char old_state) |
| 131 | unsigned char old_state) | ||
| 132 | { | 131 | { |
| 133 | unsigned char *byte, *end, alloc; | 132 | unsigned char *byte; |
| 134 | u32 blk = goal; | 133 | u32 blk = goal; |
| 135 | unsigned int bit; | 134 | unsigned int bit, bitlong; |
| 135 | unsigned long *plong, plong55; | ||
| 136 | 136 | ||
| 137 | byte = buffer + (goal / GFS2_NBBY); | 137 | byte = buffer + (goal / GFS2_NBBY); |
| 138 | plong = (unsigned long *)(buffer + (goal / GFS2_NBBY)); | ||
| 138 | bit = (goal % GFS2_NBBY) * GFS2_BIT_SIZE; | 139 | bit = (goal % GFS2_NBBY) * GFS2_BIT_SIZE; |
| 139 | end = buffer + buflen; | 140 | bitlong = bit; |
| 140 | alloc = (old_state == GFS2_BLKST_FREE) ? 0x55 : 0; | 141 | #if BITS_PER_LONG == 32 |
| 141 | 142 | plong55 = 0x55555555; | |
| 142 | while (byte < end) { | 143 | #else |
| 143 | /* If we're looking for a free block we can eliminate all | 144 | plong55 = 0x5555555555555555; |
| 144 | bitmap settings with 0x55, which represents four data | 145 | #endif |
| 145 | blocks in a row. If we're looking for a data block, we can | 146 | while (byte < buffer + buflen) { |
| 146 | eliminate 0x00 which corresponds to four free blocks. */ | 147 | |
| 147 | if ((*byte & 0x55) == alloc) { | 148 | if (bitlong == 0 && old_state == 0 && *plong == plong55) { |
| 148 | blk += (8 - bit) >> 1; | 149 | plong++; |
| 149 | 150 | byte += sizeof(unsigned long); | |
| 150 | bit = 0; | 151 | blk += sizeof(unsigned long) * GFS2_NBBY; |
| 151 | byte++; | ||
| 152 | |||
| 153 | continue; | 152 | continue; |
| 154 | } | 153 | } |
| 155 | |||
| 156 | if (((*byte >> bit) & GFS2_BIT_MASK) == old_state) | 154 | if (((*byte >> bit) & GFS2_BIT_MASK) == old_state) |
| 157 | return blk; | 155 | return blk; |
| 158 | |||
| 159 | bit += GFS2_BIT_SIZE; | 156 | bit += GFS2_BIT_SIZE; |
| 160 | if (bit >= 8) { | 157 | if (bit >= 8) { |
| 161 | bit = 0; | 158 | bit = 0; |
| 162 | byte++; | 159 | byte++; |
| 163 | } | 160 | } |
| 161 | bitlong += GFS2_BIT_SIZE; | ||
| 162 | if (bitlong >= sizeof(unsigned long) * 8) { | ||
| 163 | bitlong = 0; | ||
| 164 | plong++; | ||
| 165 | } | ||
| 164 | 166 | ||
| 165 | blk++; | 167 | blk++; |
| 166 | } | 168 | } |
| @@ -817,11 +819,9 @@ void gfs2_rgrp_repolish_clones(struct gfs2_rgrpd *rgd) | |||
| 817 | 819 | ||
| 818 | struct gfs2_alloc *gfs2_alloc_get(struct gfs2_inode *ip) | 820 | struct gfs2_alloc *gfs2_alloc_get(struct gfs2_inode *ip) |
| 819 | { | 821 | { |
| 820 | struct gfs2_alloc *al = &ip->i_alloc; | 822 | BUG_ON(ip->i_alloc != NULL); |
| 821 | 823 | ip->i_alloc = kzalloc(sizeof(struct gfs2_alloc), GFP_KERNEL); | |
| 822 | /* FIXME: Should assert that the correct locks are held here... */ | 824 | return ip->i_alloc; |
| 823 | memset(al, 0, sizeof(*al)); | ||
| 824 | return al; | ||
| 825 | } | 825 | } |
| 826 | 826 | ||
| 827 | /** | 827 | /** |
| @@ -1059,26 +1059,34 @@ static struct inode *get_local_rgrp(struct gfs2_inode *ip, u64 *last_unlinked) | |||
| 1059 | struct inode *inode = NULL; | 1059 | struct inode *inode = NULL; |
| 1060 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 1060 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
| 1061 | struct gfs2_rgrpd *rgd, *begin = NULL; | 1061 | struct gfs2_rgrpd *rgd, *begin = NULL; |
| 1062 | struct gfs2_alloc *al = &ip->i_alloc; | 1062 | struct gfs2_alloc *al = ip->i_alloc; |
| 1063 | int flags = LM_FLAG_TRY; | 1063 | int flags = LM_FLAG_TRY; |
| 1064 | int skipped = 0; | 1064 | int skipped = 0; |
| 1065 | int loops = 0; | 1065 | int loops = 0; |
| 1066 | int error; | 1066 | int error, rg_locked; |
| 1067 | 1067 | ||
| 1068 | /* Try recently successful rgrps */ | 1068 | /* Try recently successful rgrps */ |
| 1069 | 1069 | ||
| 1070 | rgd = recent_rgrp_first(sdp, ip->i_last_rg_alloc); | 1070 | rgd = recent_rgrp_first(sdp, ip->i_last_rg_alloc); |
| 1071 | 1071 | ||
| 1072 | while (rgd) { | 1072 | while (rgd) { |
| 1073 | error = gfs2_glock_nq_init(rgd->rd_gl, LM_ST_EXCLUSIVE, | 1073 | rg_locked = 0; |
| 1074 | LM_FLAG_TRY, &al->al_rgd_gh); | 1074 | |
| 1075 | if (gfs2_glock_is_locked_by_me(rgd->rd_gl)) { | ||
| 1076 | rg_locked = 1; | ||
| 1077 | error = 0; | ||
| 1078 | } else { | ||
| 1079 | error = gfs2_glock_nq_init(rgd->rd_gl, LM_ST_EXCLUSIVE, | ||
| 1080 | LM_FLAG_TRY, &al->al_rgd_gh); | ||
| 1081 | } | ||
| 1075 | switch (error) { | 1082 | switch (error) { |
| 1076 | case 0: | 1083 | case 0: |
| 1077 | if (try_rgrp_fit(rgd, al)) | 1084 | if (try_rgrp_fit(rgd, al)) |
| 1078 | goto out; | 1085 | goto out; |
| 1079 | if (rgd->rd_flags & GFS2_RDF_CHECK) | 1086 | if (rgd->rd_flags & GFS2_RDF_CHECK) |
| 1080 | inode = try_rgrp_unlink(rgd, last_unlinked); | 1087 | inode = try_rgrp_unlink(rgd, last_unlinked); |
| 1081 | gfs2_glock_dq_uninit(&al->al_rgd_gh); | 1088 | if (!rg_locked) |
| 1089 | gfs2_glock_dq_uninit(&al->al_rgd_gh); | ||
| 1082 | if (inode) | 1090 | if (inode) |
| 1083 | return inode; | 1091 | return inode; |
| 1084 | rgd = recent_rgrp_next(rgd, 1); | 1092 | rgd = recent_rgrp_next(rgd, 1); |
| @@ -1098,15 +1106,23 @@ static struct inode *get_local_rgrp(struct gfs2_inode *ip, u64 *last_unlinked) | |||
| 1098 | begin = rgd = forward_rgrp_get(sdp); | 1106 | begin = rgd = forward_rgrp_get(sdp); |
| 1099 | 1107 | ||
| 1100 | for (;;) { | 1108 | for (;;) { |
| 1101 | error = gfs2_glock_nq_init(rgd->rd_gl, LM_ST_EXCLUSIVE, flags, | 1109 | rg_locked = 0; |
| 1102 | &al->al_rgd_gh); | 1110 | |
| 1111 | if (gfs2_glock_is_locked_by_me(rgd->rd_gl)) { | ||
| 1112 | rg_locked = 1; | ||
| 1113 | error = 0; | ||
| 1114 | } else { | ||
| 1115 | error = gfs2_glock_nq_init(rgd->rd_gl, LM_ST_EXCLUSIVE, flags, | ||
| 1116 | &al->al_rgd_gh); | ||
| 1117 | } | ||
| 1103 | switch (error) { | 1118 | switch (error) { |
| 1104 | case 0: | 1119 | case 0: |
| 1105 | if (try_rgrp_fit(rgd, al)) | 1120 | if (try_rgrp_fit(rgd, al)) |
| 1106 | goto out; | 1121 | goto out; |
| 1107 | if (rgd->rd_flags & GFS2_RDF_CHECK) | 1122 | if (rgd->rd_flags & GFS2_RDF_CHECK) |
| 1108 | inode = try_rgrp_unlink(rgd, last_unlinked); | 1123 | inode = try_rgrp_unlink(rgd, last_unlinked); |
| 1109 | gfs2_glock_dq_uninit(&al->al_rgd_gh); | 1124 | if (!rg_locked) |
| 1125 | gfs2_glock_dq_uninit(&al->al_rgd_gh); | ||
| 1110 | if (inode) | 1126 | if (inode) |
| 1111 | return inode; | 1127 | return inode; |
| 1112 | break; | 1128 | break; |
| @@ -1158,7 +1174,7 @@ out: | |||
| 1158 | int gfs2_inplace_reserve_i(struct gfs2_inode *ip, char *file, unsigned int line) | 1174 | int gfs2_inplace_reserve_i(struct gfs2_inode *ip, char *file, unsigned int line) |
| 1159 | { | 1175 | { |
| 1160 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 1176 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
| 1161 | struct gfs2_alloc *al = &ip->i_alloc; | 1177 | struct gfs2_alloc *al = ip->i_alloc; |
| 1162 | struct inode *inode; | 1178 | struct inode *inode; |
| 1163 | int error = 0; | 1179 | int error = 0; |
| 1164 | u64 last_unlinked = NO_BLOCK; | 1180 | u64 last_unlinked = NO_BLOCK; |
| @@ -1204,7 +1220,7 @@ try_again: | |||
| 1204 | void gfs2_inplace_release(struct gfs2_inode *ip) | 1220 | void gfs2_inplace_release(struct gfs2_inode *ip) |
| 1205 | { | 1221 | { |
| 1206 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 1222 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
| 1207 | struct gfs2_alloc *al = &ip->i_alloc; | 1223 | struct gfs2_alloc *al = ip->i_alloc; |
| 1208 | 1224 | ||
| 1209 | if (gfs2_assert_warn(sdp, al->al_alloced <= al->al_requested) == -1) | 1225 | if (gfs2_assert_warn(sdp, al->al_alloced <= al->al_requested) == -1) |
| 1210 | fs_warn(sdp, "al_alloced = %u, al_requested = %u " | 1226 | fs_warn(sdp, "al_alloced = %u, al_requested = %u " |
| @@ -1213,7 +1229,8 @@ void gfs2_inplace_release(struct gfs2_inode *ip) | |||
| 1213 | al->al_line); | 1229 | al->al_line); |
| 1214 | 1230 | ||
| 1215 | al->al_rgd = NULL; | 1231 | al->al_rgd = NULL; |
| 1216 | gfs2_glock_dq_uninit(&al->al_rgd_gh); | 1232 | if (al->al_rgd_gh.gh_gl) |
| 1233 | gfs2_glock_dq_uninit(&al->al_rgd_gh); | ||
| 1217 | if (ip != GFS2_I(sdp->sd_rindex)) | 1234 | if (ip != GFS2_I(sdp->sd_rindex)) |
| 1218 | gfs2_glock_dq_uninit(&al->al_ri_gh); | 1235 | gfs2_glock_dq_uninit(&al->al_ri_gh); |
| 1219 | } | 1236 | } |
| @@ -1301,11 +1318,10 @@ static u32 rgblk_search(struct gfs2_rgrpd *rgd, u32 goal, | |||
| 1301 | /* The GFS2_BLKST_UNLINKED state doesn't apply to the clone | 1318 | /* The GFS2_BLKST_UNLINKED state doesn't apply to the clone |
| 1302 | bitmaps, so we must search the originals for that. */ | 1319 | bitmaps, so we must search the originals for that. */ |
| 1303 | if (old_state != GFS2_BLKST_UNLINKED && bi->bi_clone) | 1320 | if (old_state != GFS2_BLKST_UNLINKED && bi->bi_clone) |
| 1304 | blk = gfs2_bitfit(rgd, bi->bi_clone + bi->bi_offset, | 1321 | blk = gfs2_bitfit(bi->bi_clone + bi->bi_offset, |
| 1305 | bi->bi_len, goal, old_state); | 1322 | bi->bi_len, goal, old_state); |
| 1306 | else | 1323 | else |
| 1307 | blk = gfs2_bitfit(rgd, | 1324 | blk = gfs2_bitfit(bi->bi_bh->b_data + bi->bi_offset, |
| 1308 | bi->bi_bh->b_data + bi->bi_offset, | ||
| 1309 | bi->bi_len, goal, old_state); | 1325 | bi->bi_len, goal, old_state); |
| 1310 | if (blk != BFITNOENT) | 1326 | if (blk != BFITNOENT) |
| 1311 | break; | 1327 | break; |
| @@ -1394,7 +1410,7 @@ static struct gfs2_rgrpd *rgblk_free(struct gfs2_sbd *sdp, u64 bstart, | |||
| 1394 | u64 gfs2_alloc_data(struct gfs2_inode *ip) | 1410 | u64 gfs2_alloc_data(struct gfs2_inode *ip) |
| 1395 | { | 1411 | { |
| 1396 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 1412 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
| 1397 | struct gfs2_alloc *al = &ip->i_alloc; | 1413 | struct gfs2_alloc *al = ip->i_alloc; |
| 1398 | struct gfs2_rgrpd *rgd = al->al_rgd; | 1414 | struct gfs2_rgrpd *rgd = al->al_rgd; |
| 1399 | u32 goal, blk; | 1415 | u32 goal, blk; |
| 1400 | u64 block; | 1416 | u64 block; |
| @@ -1439,7 +1455,7 @@ u64 gfs2_alloc_data(struct gfs2_inode *ip) | |||
| 1439 | u64 gfs2_alloc_meta(struct gfs2_inode *ip) | 1455 | u64 gfs2_alloc_meta(struct gfs2_inode *ip) |
| 1440 | { | 1456 | { |
| 1441 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 1457 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
| 1442 | struct gfs2_alloc *al = &ip->i_alloc; | 1458 | struct gfs2_alloc *al = ip->i_alloc; |
| 1443 | struct gfs2_rgrpd *rgd = al->al_rgd; | 1459 | struct gfs2_rgrpd *rgd = al->al_rgd; |
| 1444 | u32 goal, blk; | 1460 | u32 goal, blk; |
| 1445 | u64 block; | 1461 | u64 block; |
| @@ -1485,7 +1501,7 @@ u64 gfs2_alloc_meta(struct gfs2_inode *ip) | |||
| 1485 | u64 gfs2_alloc_di(struct gfs2_inode *dip, u64 *generation) | 1501 | u64 gfs2_alloc_di(struct gfs2_inode *dip, u64 *generation) |
| 1486 | { | 1502 | { |
| 1487 | struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); | 1503 | struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); |
| 1488 | struct gfs2_alloc *al = &dip->i_alloc; | 1504 | struct gfs2_alloc *al = dip->i_alloc; |
| 1489 | struct gfs2_rgrpd *rgd = al->al_rgd; | 1505 | struct gfs2_rgrpd *rgd = al->al_rgd; |
| 1490 | u32 blk; | 1506 | u32 blk; |
| 1491 | u64 block; | 1507 | u64 block; |
diff --git a/fs/gfs2/rgrp.h b/fs/gfs2/rgrp.h index b4c6adfc6f2e..149bb161f4b6 100644 --- a/fs/gfs2/rgrp.h +++ b/fs/gfs2/rgrp.h | |||
| @@ -32,7 +32,9 @@ void gfs2_rgrp_repolish_clones(struct gfs2_rgrpd *rgd); | |||
| 32 | struct gfs2_alloc *gfs2_alloc_get(struct gfs2_inode *ip); | 32 | struct gfs2_alloc *gfs2_alloc_get(struct gfs2_inode *ip); |
| 33 | static inline void gfs2_alloc_put(struct gfs2_inode *ip) | 33 | static inline void gfs2_alloc_put(struct gfs2_inode *ip) |
| 34 | { | 34 | { |
| 35 | return; /* So we can see where ip->i_alloc is used */ | 35 | BUG_ON(ip->i_alloc == NULL); |
| 36 | kfree(ip->i_alloc); | ||
| 37 | ip->i_alloc = NULL; | ||
| 36 | } | 38 | } |
| 37 | 39 | ||
| 38 | int gfs2_inplace_reserve_i(struct gfs2_inode *ip, | 40 | int gfs2_inplace_reserve_i(struct gfs2_inode *ip, |
diff --git a/fs/gfs2/super.c b/fs/gfs2/super.c index dd3e737f528e..ef0562c3bc71 100644 --- a/fs/gfs2/super.c +++ b/fs/gfs2/super.c | |||
| @@ -1,6 +1,6 @@ | |||
| 1 | /* | 1 | /* |
| 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. |
| 3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | 3 | * Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved. |
| 4 | * | 4 | * |
| 5 | * This copyrighted material is made available to anyone wishing to use, | 5 | * This copyrighted material is made available to anyone wishing to use, |
| 6 | * modify, copy, or redistribute it subject to the terms and conditions | 6 | * modify, copy, or redistribute it subject to the terms and conditions |
| @@ -51,13 +51,9 @@ void gfs2_tune_init(struct gfs2_tune *gt) | |||
| 51 | { | 51 | { |
| 52 | spin_lock_init(>->gt_spin); | 52 | spin_lock_init(>->gt_spin); |
| 53 | 53 | ||
| 54 | gt->gt_ilimit = 100; | ||
| 55 | gt->gt_ilimit_tries = 3; | ||
| 56 | gt->gt_ilimit_min = 1; | ||
| 57 | gt->gt_demote_secs = 300; | 54 | gt->gt_demote_secs = 300; |
| 58 | gt->gt_incore_log_blocks = 1024; | 55 | gt->gt_incore_log_blocks = 1024; |
| 59 | gt->gt_log_flush_secs = 60; | 56 | gt->gt_log_flush_secs = 60; |
| 60 | gt->gt_jindex_refresh_secs = 60; | ||
| 61 | gt->gt_recoverd_secs = 60; | 57 | gt->gt_recoverd_secs = 60; |
| 62 | gt->gt_logd_secs = 1; | 58 | gt->gt_logd_secs = 1; |
| 63 | gt->gt_quotad_secs = 5; | 59 | gt->gt_quotad_secs = 5; |
| @@ -71,10 +67,8 @@ void gfs2_tune_init(struct gfs2_tune *gt) | |||
| 71 | gt->gt_new_files_jdata = 0; | 67 | gt->gt_new_files_jdata = 0; |
| 72 | gt->gt_new_files_directio = 0; | 68 | gt->gt_new_files_directio = 0; |
| 73 | gt->gt_max_readahead = 1 << 18; | 69 | gt->gt_max_readahead = 1 << 18; |
| 74 | gt->gt_lockdump_size = 131072; | ||
| 75 | gt->gt_stall_secs = 600; | 70 | gt->gt_stall_secs = 600; |
| 76 | gt->gt_complain_secs = 10; | 71 | gt->gt_complain_secs = 10; |
| 77 | gt->gt_reclaim_limit = 5000; | ||
| 78 | gt->gt_statfs_quantum = 30; | 72 | gt->gt_statfs_quantum = 30; |
| 79 | gt->gt_statfs_slow = 0; | 73 | gt->gt_statfs_slow = 0; |
| 80 | } | 74 | } |
| @@ -393,6 +387,7 @@ int gfs2_jindex_hold(struct gfs2_sbd *sdp, struct gfs2_holder *ji_gh) | |||
| 393 | if (!jd) | 387 | if (!jd) |
| 394 | break; | 388 | break; |
| 395 | 389 | ||
| 390 | INIT_LIST_HEAD(&jd->extent_list); | ||
| 396 | jd->jd_inode = gfs2_lookupi(sdp->sd_jindex, &name, 1, NULL); | 391 | jd->jd_inode = gfs2_lookupi(sdp->sd_jindex, &name, 1, NULL); |
| 397 | if (!jd->jd_inode || IS_ERR(jd->jd_inode)) { | 392 | if (!jd->jd_inode || IS_ERR(jd->jd_inode)) { |
| 398 | if (!jd->jd_inode) | 393 | if (!jd->jd_inode) |
| @@ -422,8 +417,9 @@ int gfs2_jindex_hold(struct gfs2_sbd *sdp, struct gfs2_holder *ji_gh) | |||
| 422 | 417 | ||
| 423 | void gfs2_jindex_free(struct gfs2_sbd *sdp) | 418 | void gfs2_jindex_free(struct gfs2_sbd *sdp) |
| 424 | { | 419 | { |
| 425 | struct list_head list; | 420 | struct list_head list, *head; |
| 426 | struct gfs2_jdesc *jd; | 421 | struct gfs2_jdesc *jd; |
| 422 | struct gfs2_journal_extent *jext; | ||
| 427 | 423 | ||
| 428 | spin_lock(&sdp->sd_jindex_spin); | 424 | spin_lock(&sdp->sd_jindex_spin); |
| 429 | list_add(&list, &sdp->sd_jindex_list); | 425 | list_add(&list, &sdp->sd_jindex_list); |
| @@ -433,6 +429,14 @@ void gfs2_jindex_free(struct gfs2_sbd *sdp) | |||
| 433 | 429 | ||
| 434 | while (!list_empty(&list)) { | 430 | while (!list_empty(&list)) { |
| 435 | jd = list_entry(list.next, struct gfs2_jdesc, jd_list); | 431 | jd = list_entry(list.next, struct gfs2_jdesc, jd_list); |
| 432 | head = &jd->extent_list; | ||
| 433 | while (!list_empty(head)) { | ||
| 434 | jext = list_entry(head->next, | ||
| 435 | struct gfs2_journal_extent, | ||
| 436 | extent_list); | ||
| 437 | list_del(&jext->extent_list); | ||
| 438 | kfree(jext); | ||
| 439 | } | ||
| 436 | list_del(&jd->jd_list); | 440 | list_del(&jd->jd_list); |
| 437 | iput(jd->jd_inode); | 441 | iput(jd->jd_inode); |
| 438 | kfree(jd); | 442 | kfree(jd); |
| @@ -543,7 +547,6 @@ int gfs2_make_fs_rw(struct gfs2_sbd *sdp) | |||
| 543 | if (error) | 547 | if (error) |
| 544 | return error; | 548 | return error; |
| 545 | 549 | ||
| 546 | gfs2_meta_cache_flush(ip); | ||
| 547 | j_gl->gl_ops->go_inval(j_gl, DIO_METADATA); | 550 | j_gl->gl_ops->go_inval(j_gl, DIO_METADATA); |
| 548 | 551 | ||
| 549 | error = gfs2_find_jhead(sdp->sd_jdesc, &head); | 552 | error = gfs2_find_jhead(sdp->sd_jdesc, &head); |
| @@ -686,9 +689,7 @@ void gfs2_statfs_change(struct gfs2_sbd *sdp, s64 total, s64 free, | |||
| 686 | if (error) | 689 | if (error) |
| 687 | return; | 690 | return; |
| 688 | 691 | ||
| 689 | mutex_lock(&sdp->sd_statfs_mutex); | ||
| 690 | gfs2_trans_add_bh(l_ip->i_gl, l_bh, 1); | 692 | gfs2_trans_add_bh(l_ip->i_gl, l_bh, 1); |
| 691 | mutex_unlock(&sdp->sd_statfs_mutex); | ||
| 692 | 693 | ||
| 693 | spin_lock(&sdp->sd_statfs_spin); | 694 | spin_lock(&sdp->sd_statfs_spin); |
| 694 | l_sc->sc_total += total; | 695 | l_sc->sc_total += total; |
| @@ -736,9 +737,7 @@ int gfs2_statfs_sync(struct gfs2_sbd *sdp) | |||
| 736 | if (error) | 737 | if (error) |
| 737 | goto out_bh2; | 738 | goto out_bh2; |
| 738 | 739 | ||
| 739 | mutex_lock(&sdp->sd_statfs_mutex); | ||
| 740 | gfs2_trans_add_bh(l_ip->i_gl, l_bh, 1); | 740 | gfs2_trans_add_bh(l_ip->i_gl, l_bh, 1); |
| 741 | mutex_unlock(&sdp->sd_statfs_mutex); | ||
| 742 | 741 | ||
| 743 | spin_lock(&sdp->sd_statfs_spin); | 742 | spin_lock(&sdp->sd_statfs_spin); |
| 744 | m_sc->sc_total += l_sc->sc_total; | 743 | m_sc->sc_total += l_sc->sc_total; |
diff --git a/fs/gfs2/sys.c b/fs/gfs2/sys.c index 3a3176b846f3..eaa3b7b2f99e 100644 --- a/fs/gfs2/sys.c +++ b/fs/gfs2/sys.c | |||
| @@ -32,7 +32,8 @@ spinlock_t gfs2_sys_margs_lock; | |||
| 32 | 32 | ||
| 33 | static ssize_t id_show(struct gfs2_sbd *sdp, char *buf) | 33 | static ssize_t id_show(struct gfs2_sbd *sdp, char *buf) |
| 34 | { | 34 | { |
| 35 | return snprintf(buf, PAGE_SIZE, "%s\n", sdp->sd_vfs->s_id); | 35 | return snprintf(buf, PAGE_SIZE, "%u:%u\n", |
| 36 | MAJOR(sdp->sd_vfs->s_dev), MINOR(sdp->sd_vfs->s_dev)); | ||
| 36 | } | 37 | } |
| 37 | 38 | ||
| 38 | static ssize_t fsname_show(struct gfs2_sbd *sdp, char *buf) | 39 | static ssize_t fsname_show(struct gfs2_sbd *sdp, char *buf) |
| @@ -425,13 +426,11 @@ TUNE_ATTR_2(name, name##_store) | |||
| 425 | TUNE_ATTR(demote_secs, 0); | 426 | TUNE_ATTR(demote_secs, 0); |
| 426 | TUNE_ATTR(incore_log_blocks, 0); | 427 | TUNE_ATTR(incore_log_blocks, 0); |
| 427 | TUNE_ATTR(log_flush_secs, 0); | 428 | TUNE_ATTR(log_flush_secs, 0); |
| 428 | TUNE_ATTR(jindex_refresh_secs, 0); | ||
| 429 | TUNE_ATTR(quota_warn_period, 0); | 429 | TUNE_ATTR(quota_warn_period, 0); |
| 430 | TUNE_ATTR(quota_quantum, 0); | 430 | TUNE_ATTR(quota_quantum, 0); |
| 431 | TUNE_ATTR(atime_quantum, 0); | 431 | TUNE_ATTR(atime_quantum, 0); |
| 432 | TUNE_ATTR(max_readahead, 0); | 432 | TUNE_ATTR(max_readahead, 0); |
| 433 | TUNE_ATTR(complain_secs, 0); | 433 | TUNE_ATTR(complain_secs, 0); |
| 434 | TUNE_ATTR(reclaim_limit, 0); | ||
| 435 | TUNE_ATTR(statfs_slow, 0); | 434 | TUNE_ATTR(statfs_slow, 0); |
| 436 | TUNE_ATTR(new_files_jdata, 0); | 435 | TUNE_ATTR(new_files_jdata, 0); |
| 437 | TUNE_ATTR(new_files_directio, 0); | 436 | TUNE_ATTR(new_files_directio, 0); |
| @@ -448,13 +447,11 @@ static struct attribute *tune_attrs[] = { | |||
| 448 | &tune_attr_demote_secs.attr, | 447 | &tune_attr_demote_secs.attr, |
| 449 | &tune_attr_incore_log_blocks.attr, | 448 | &tune_attr_incore_log_blocks.attr, |
| 450 | &tune_attr_log_flush_secs.attr, | 449 | &tune_attr_log_flush_secs.attr, |
| 451 | &tune_attr_jindex_refresh_secs.attr, | ||
| 452 | &tune_attr_quota_warn_period.attr, | 450 | &tune_attr_quota_warn_period.attr, |
| 453 | &tune_attr_quota_quantum.attr, | 451 | &tune_attr_quota_quantum.attr, |
| 454 | &tune_attr_atime_quantum.attr, | 452 | &tune_attr_atime_quantum.attr, |
| 455 | &tune_attr_max_readahead.attr, | 453 | &tune_attr_max_readahead.attr, |
| 456 | &tune_attr_complain_secs.attr, | 454 | &tune_attr_complain_secs.attr, |
| 457 | &tune_attr_reclaim_limit.attr, | ||
| 458 | &tune_attr_statfs_slow.attr, | 455 | &tune_attr_statfs_slow.attr, |
| 459 | &tune_attr_quota_simul_sync.attr, | 456 | &tune_attr_quota_simul_sync.attr, |
| 460 | &tune_attr_quota_cache_secs.attr, | 457 | &tune_attr_quota_cache_secs.attr, |
diff --git a/fs/gfs2/trans.c b/fs/gfs2/trans.c index 717983e2c2ae..73e5d92a657c 100644 --- a/fs/gfs2/trans.c +++ b/fs/gfs2/trans.c | |||
| @@ -114,11 +114,6 @@ void gfs2_trans_end(struct gfs2_sbd *sdp) | |||
| 114 | gfs2_log_flush(sdp, NULL); | 114 | gfs2_log_flush(sdp, NULL); |
| 115 | } | 115 | } |
| 116 | 116 | ||
| 117 | void gfs2_trans_add_gl(struct gfs2_glock *gl) | ||
| 118 | { | ||
| 119 | lops_add(gl->gl_sbd, &gl->gl_le); | ||
| 120 | } | ||
| 121 | |||
| 122 | /** | 117 | /** |
| 123 | * gfs2_trans_add_bh - Add a to-be-modified buffer to the current transaction | 118 | * gfs2_trans_add_bh - Add a to-be-modified buffer to the current transaction |
| 124 | * @gl: the glock the buffer belongs to | 119 | * @gl: the glock the buffer belongs to |
diff --git a/fs/gfs2/trans.h b/fs/gfs2/trans.h index 043d5f4b9c4c..e826f0dab80a 100644 --- a/fs/gfs2/trans.h +++ b/fs/gfs2/trans.h | |||
| @@ -30,7 +30,6 @@ int gfs2_trans_begin(struct gfs2_sbd *sdp, unsigned int blocks, | |||
| 30 | 30 | ||
| 31 | void gfs2_trans_end(struct gfs2_sbd *sdp); | 31 | void gfs2_trans_end(struct gfs2_sbd *sdp); |
| 32 | 32 | ||
| 33 | void gfs2_trans_add_gl(struct gfs2_glock *gl); | ||
| 34 | void gfs2_trans_add_bh(struct gfs2_glock *gl, struct buffer_head *bh, int meta); | 33 | void gfs2_trans_add_bh(struct gfs2_glock *gl, struct buffer_head *bh, int meta); |
| 35 | void gfs2_trans_add_revoke(struct gfs2_sbd *sdp, struct gfs2_bufdata *bd); | 34 | void gfs2_trans_add_revoke(struct gfs2_sbd *sdp, struct gfs2_bufdata *bd); |
| 36 | void gfs2_trans_add_unrevoke(struct gfs2_sbd *sdp, u64 blkno); | 35 | void gfs2_trans_add_unrevoke(struct gfs2_sbd *sdp, u64 blkno); |
