diff options
Diffstat (limited to 'fs/gfs2')
41 files changed, 1117 insertions, 1151 deletions
diff --git a/fs/gfs2/Makefile b/fs/gfs2/Makefile index 04ad0caebed..8fff11058ce 100644 --- a/fs/gfs2/Makefile +++ b/fs/gfs2/Makefile | |||
@@ -2,7 +2,7 @@ obj-$(CONFIG_GFS2_FS) += gfs2.o | |||
2 | gfs2-y := acl.o bmap.o daemon.o dir.o eaops.o eattr.o glock.o \ | 2 | gfs2-y := acl.o bmap.o daemon.o dir.o eaops.o eattr.o glock.o \ |
3 | glops.o inode.o lm.o log.o lops.o locking.o main.o meta_io.o \ | 3 | glops.o inode.o lm.o log.o lops.o locking.o main.o meta_io.o \ |
4 | mount.o ops_address.o ops_dentry.o ops_export.o ops_file.o \ | 4 | mount.o ops_address.o ops_dentry.o ops_export.o ops_file.o \ |
5 | ops_fstype.o ops_inode.o ops_super.o ops_vm.o quota.o \ | 5 | ops_fstype.o ops_inode.o ops_super.o quota.o \ |
6 | recovery.o rgrp.o super.o sys.o trans.o util.o | 6 | recovery.o rgrp.o super.o sys.o trans.o util.o |
7 | 7 | ||
8 | obj-$(CONFIG_GFS2_FS_LOCKING_NOLOCK) += locking/nolock/ | 8 | obj-$(CONFIG_GFS2_FS_LOCKING_NOLOCK) += locking/nolock/ |
diff --git a/fs/gfs2/bmap.c b/fs/gfs2/bmap.c index 93fa427bb5f..e4effc47abf 100644 --- a/fs/gfs2/bmap.c +++ b/fs/gfs2/bmap.c | |||
@@ -59,7 +59,6 @@ struct strip_mine { | |||
59 | static int gfs2_unstuffer_page(struct gfs2_inode *ip, struct buffer_head *dibh, | 59 | static int gfs2_unstuffer_page(struct gfs2_inode *ip, struct buffer_head *dibh, |
60 | u64 block, struct page *page) | 60 | u64 block, struct page *page) |
61 | { | 61 | { |
62 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | ||
63 | struct inode *inode = &ip->i_inode; | 62 | struct inode *inode = &ip->i_inode; |
64 | struct buffer_head *bh; | 63 | struct buffer_head *bh; |
65 | int release = 0; | 64 | int release = 0; |
@@ -95,7 +94,7 @@ static int gfs2_unstuffer_page(struct gfs2_inode *ip, struct buffer_head *dibh, | |||
95 | set_buffer_uptodate(bh); | 94 | set_buffer_uptodate(bh); |
96 | if (!gfs2_is_jdata(ip)) | 95 | if (!gfs2_is_jdata(ip)) |
97 | mark_buffer_dirty(bh); | 96 | mark_buffer_dirty(bh); |
98 | if (sdp->sd_args.ar_data == GFS2_DATA_ORDERED || gfs2_is_jdata(ip)) | 97 | if (!gfs2_is_writeback(ip)) |
99 | gfs2_trans_add_bh(ip->i_gl, bh, 0); | 98 | gfs2_trans_add_bh(ip->i_gl, bh, 0); |
100 | 99 | ||
101 | if (release) { | 100 | if (release) { |
@@ -453,8 +452,8 @@ static inline void bmap_unlock(struct inode *inode, int create) | |||
453 | * Returns: errno | 452 | * Returns: errno |
454 | */ | 453 | */ |
455 | 454 | ||
456 | int gfs2_block_map(struct inode *inode, u64 lblock, int create, | 455 | int gfs2_block_map(struct inode *inode, sector_t lblock, |
457 | struct buffer_head *bh_map) | 456 | struct buffer_head *bh_map, int create) |
458 | { | 457 | { |
459 | struct gfs2_inode *ip = GFS2_I(inode); | 458 | struct gfs2_inode *ip = GFS2_I(inode); |
460 | struct gfs2_sbd *sdp = GFS2_SB(inode); | 459 | struct gfs2_sbd *sdp = GFS2_SB(inode); |
@@ -470,6 +469,7 @@ int gfs2_block_map(struct inode *inode, u64 lblock, int create, | |||
470 | unsigned int maxlen = bh_map->b_size >> inode->i_blkbits; | 469 | unsigned int maxlen = bh_map->b_size >> inode->i_blkbits; |
471 | struct metapath mp; | 470 | struct metapath mp; |
472 | u64 size; | 471 | u64 size; |
472 | struct buffer_head *dibh = NULL; | ||
473 | 473 | ||
474 | BUG_ON(maxlen == 0); | 474 | BUG_ON(maxlen == 0); |
475 | 475 | ||
@@ -500,6 +500,8 @@ int gfs2_block_map(struct inode *inode, u64 lblock, int create, | |||
500 | error = gfs2_meta_inode_buffer(ip, &bh); | 500 | error = gfs2_meta_inode_buffer(ip, &bh); |
501 | if (error) | 501 | if (error) |
502 | goto out_fail; | 502 | goto out_fail; |
503 | dibh = bh; | ||
504 | get_bh(dibh); | ||
503 | 505 | ||
504 | for (x = 0; x < end_of_metadata; x++) { | 506 | for (x = 0; x < end_of_metadata; x++) { |
505 | lookup_block(ip, bh, x, &mp, create, &new, &dblock); | 507 | lookup_block(ip, bh, x, &mp, create, &new, &dblock); |
@@ -518,13 +520,8 @@ int gfs2_block_map(struct inode *inode, u64 lblock, int create, | |||
518 | if (boundary) | 520 | if (boundary) |
519 | set_buffer_boundary(bh_map); | 521 | set_buffer_boundary(bh_map); |
520 | if (new) { | 522 | if (new) { |
521 | struct buffer_head *dibh; | 523 | gfs2_trans_add_bh(ip->i_gl, dibh, 1); |
522 | error = gfs2_meta_inode_buffer(ip, &dibh); | 524 | gfs2_dinode_out(ip, dibh->b_data); |
523 | if (!error) { | ||
524 | gfs2_trans_add_bh(ip->i_gl, dibh, 1); | ||
525 | gfs2_dinode_out(ip, dibh->b_data); | ||
526 | brelse(dibh); | ||
527 | } | ||
528 | set_buffer_new(bh_map); | 525 | set_buffer_new(bh_map); |
529 | goto out_brelse; | 526 | goto out_brelse; |
530 | } | 527 | } |
@@ -545,6 +542,8 @@ out_brelse: | |||
545 | out_ok: | 542 | out_ok: |
546 | error = 0; | 543 | error = 0; |
547 | out_fail: | 544 | out_fail: |
545 | if (dibh) | ||
546 | brelse(dibh); | ||
548 | bmap_unlock(inode, create); | 547 | bmap_unlock(inode, create); |
549 | return error; | 548 | return error; |
550 | } | 549 | } |
@@ -560,7 +559,7 @@ int gfs2_extent_map(struct inode *inode, u64 lblock, int *new, u64 *dblock, unsi | |||
560 | BUG_ON(!new); | 559 | BUG_ON(!new); |
561 | 560 | ||
562 | bh.b_size = 1 << (inode->i_blkbits + 5); | 561 | bh.b_size = 1 << (inode->i_blkbits + 5); |
563 | ret = gfs2_block_map(inode, lblock, create, &bh); | 562 | ret = gfs2_block_map(inode, lblock, &bh, create); |
564 | *extlen = bh.b_size >> inode->i_blkbits; | 563 | *extlen = bh.b_size >> inode->i_blkbits; |
565 | *dblock = bh.b_blocknr; | 564 | *dblock = bh.b_blocknr; |
566 | if (buffer_new(&bh)) | 565 | if (buffer_new(&bh)) |
@@ -684,7 +683,7 @@ static int do_strip(struct gfs2_inode *ip, struct buffer_head *dibh, | |||
684 | if (metadata) | 683 | if (metadata) |
685 | revokes = (height) ? sdp->sd_inptrs : sdp->sd_diptrs; | 684 | revokes = (height) ? sdp->sd_inptrs : sdp->sd_diptrs; |
686 | 685 | ||
687 | error = gfs2_rindex_hold(sdp, &ip->i_alloc.al_ri_gh); | 686 | error = gfs2_rindex_hold(sdp, &ip->i_alloc->al_ri_gh); |
688 | if (error) | 687 | if (error) |
689 | return error; | 688 | return error; |
690 | 689 | ||
@@ -786,7 +785,7 @@ out_rg_gunlock: | |||
786 | out_rlist: | 785 | out_rlist: |
787 | gfs2_rlist_free(&rlist); | 786 | gfs2_rlist_free(&rlist); |
788 | out: | 787 | out: |
789 | gfs2_glock_dq_uninit(&ip->i_alloc.al_ri_gh); | 788 | gfs2_glock_dq_uninit(&ip->i_alloc->al_ri_gh); |
790 | return error; | 789 | return error; |
791 | } | 790 | } |
792 | 791 | ||
@@ -879,7 +878,6 @@ static int gfs2_block_truncate_page(struct address_space *mapping) | |||
879 | { | 878 | { |
880 | struct inode *inode = mapping->host; | 879 | struct inode *inode = mapping->host; |
881 | struct gfs2_inode *ip = GFS2_I(inode); | 880 | struct gfs2_inode *ip = GFS2_I(inode); |
882 | struct gfs2_sbd *sdp = GFS2_SB(inode); | ||
883 | loff_t from = inode->i_size; | 881 | loff_t from = inode->i_size; |
884 | unsigned long index = from >> PAGE_CACHE_SHIFT; | 882 | unsigned long index = from >> PAGE_CACHE_SHIFT; |
885 | unsigned offset = from & (PAGE_CACHE_SIZE-1); | 883 | unsigned offset = from & (PAGE_CACHE_SIZE-1); |
@@ -911,7 +909,7 @@ static int gfs2_block_truncate_page(struct address_space *mapping) | |||
911 | err = 0; | 909 | err = 0; |
912 | 910 | ||
913 | if (!buffer_mapped(bh)) { | 911 | if (!buffer_mapped(bh)) { |
914 | gfs2_get_block(inode, iblock, bh, 0); | 912 | gfs2_block_map(inode, iblock, bh, 0); |
915 | /* unmapped? It's a hole - nothing to do */ | 913 | /* unmapped? It's a hole - nothing to do */ |
916 | if (!buffer_mapped(bh)) | 914 | if (!buffer_mapped(bh)) |
917 | goto unlock; | 915 | goto unlock; |
@@ -931,7 +929,7 @@ static int gfs2_block_truncate_page(struct address_space *mapping) | |||
931 | err = 0; | 929 | err = 0; |
932 | } | 930 | } |
933 | 931 | ||
934 | if (sdp->sd_args.ar_data == GFS2_DATA_ORDERED || gfs2_is_jdata(ip)) | 932 | if (!gfs2_is_writeback(ip)) |
935 | gfs2_trans_add_bh(ip->i_gl, bh, 0); | 933 | gfs2_trans_add_bh(ip->i_gl, bh, 0); |
936 | 934 | ||
937 | zero_user_page(page, offset, length, KM_USER0); | 935 | zero_user_page(page, offset, length, KM_USER0); |
@@ -1224,8 +1222,13 @@ int gfs2_write_alloc_required(struct gfs2_inode *ip, u64 offset, | |||
1224 | do_div(lblock_stop, bsize); | 1222 | do_div(lblock_stop, bsize); |
1225 | } else { | 1223 | } else { |
1226 | unsigned int shift = sdp->sd_sb.sb_bsize_shift; | 1224 | unsigned int shift = sdp->sd_sb.sb_bsize_shift; |
1225 | u64 end_of_file = (ip->i_di.di_size + sdp->sd_sb.sb_bsize - 1) >> shift; | ||
1227 | lblock = offset >> shift; | 1226 | lblock = offset >> shift; |
1228 | lblock_stop = (offset + len + sdp->sd_sb.sb_bsize - 1) >> shift; | 1227 | lblock_stop = (offset + len + sdp->sd_sb.sb_bsize - 1) >> shift; |
1228 | if (lblock_stop > end_of_file) { | ||
1229 | *alloc_required = 1; | ||
1230 | return 0; | ||
1231 | } | ||
1229 | } | 1232 | } |
1230 | 1233 | ||
1231 | for (; lblock < lblock_stop; lblock += extlen) { | 1234 | for (; lblock < lblock_stop; lblock += extlen) { |
diff --git a/fs/gfs2/bmap.h b/fs/gfs2/bmap.h index ac2fd04370d..4e6cde2943b 100644 --- a/fs/gfs2/bmap.h +++ b/fs/gfs2/bmap.h | |||
@@ -15,7 +15,7 @@ struct gfs2_inode; | |||
15 | struct page; | 15 | struct page; |
16 | 16 | ||
17 | int gfs2_unstuff_dinode(struct gfs2_inode *ip, struct page *page); | 17 | int gfs2_unstuff_dinode(struct gfs2_inode *ip, struct page *page); |
18 | int gfs2_block_map(struct inode *inode, u64 lblock, int create, struct buffer_head *bh); | 18 | int gfs2_block_map(struct inode *inode, sector_t lblock, struct buffer_head *bh, int create); |
19 | int gfs2_extent_map(struct inode *inode, u64 lblock, int *new, u64 *dblock, unsigned *extlen); | 19 | int gfs2_extent_map(struct inode *inode, u64 lblock, int *new, u64 *dblock, unsigned *extlen); |
20 | 20 | ||
21 | int gfs2_truncatei(struct gfs2_inode *ip, u64 size); | 21 | int gfs2_truncatei(struct gfs2_inode *ip, u64 size); |
diff --git a/fs/gfs2/daemon.c b/fs/gfs2/daemon.c index 3731ab0771d..e51991947d2 100644 --- a/fs/gfs2/daemon.c +++ b/fs/gfs2/daemon.c | |||
@@ -83,56 +83,6 @@ int gfs2_recoverd(void *data) | |||
83 | } | 83 | } |
84 | 84 | ||
85 | /** | 85 | /** |
86 | * gfs2_logd - Update log tail as Active Items get flushed to in-place blocks | ||
87 | * @sdp: Pointer to GFS2 superblock | ||
88 | * | ||
89 | * Also, periodically check to make sure that we're using the most recent | ||
90 | * journal index. | ||
91 | */ | ||
92 | |||
93 | int gfs2_logd(void *data) | ||
94 | { | ||
95 | struct gfs2_sbd *sdp = data; | ||
96 | struct gfs2_holder ji_gh; | ||
97 | unsigned long t; | ||
98 | int need_flush; | ||
99 | |||
100 | while (!kthread_should_stop()) { | ||
101 | /* Advance the log tail */ | ||
102 | |||
103 | t = sdp->sd_log_flush_time + | ||
104 | gfs2_tune_get(sdp, gt_log_flush_secs) * HZ; | ||
105 | |||
106 | gfs2_ail1_empty(sdp, DIO_ALL); | ||
107 | gfs2_log_lock(sdp); | ||
108 | need_flush = sdp->sd_log_num_buf > gfs2_tune_get(sdp, gt_incore_log_blocks); | ||
109 | gfs2_log_unlock(sdp); | ||
110 | if (need_flush || time_after_eq(jiffies, t)) { | ||
111 | gfs2_log_flush(sdp, NULL); | ||
112 | sdp->sd_log_flush_time = jiffies; | ||
113 | } | ||
114 | |||
115 | /* Check for latest journal index */ | ||
116 | |||
117 | t = sdp->sd_jindex_refresh_time + | ||
118 | gfs2_tune_get(sdp, gt_jindex_refresh_secs) * HZ; | ||
119 | |||
120 | if (time_after_eq(jiffies, t)) { | ||
121 | if (!gfs2_jindex_hold(sdp, &ji_gh)) | ||
122 | gfs2_glock_dq_uninit(&ji_gh); | ||
123 | sdp->sd_jindex_refresh_time = jiffies; | ||
124 | } | ||
125 | |||
126 | t = gfs2_tune_get(sdp, gt_logd_secs) * HZ; | ||
127 | if (freezing(current)) | ||
128 | refrigerator(); | ||
129 | schedule_timeout_interruptible(t); | ||
130 | } | ||
131 | |||
132 | return 0; | ||
133 | } | ||
134 | |||
135 | /** | ||
136 | * gfs2_quotad - Write cached quota changes into the quota file | 86 | * gfs2_quotad - Write cached quota changes into the quota file |
137 | * @sdp: Pointer to GFS2 superblock | 87 | * @sdp: Pointer to GFS2 superblock |
138 | * | 88 | * |
diff --git a/fs/gfs2/daemon.h b/fs/gfs2/daemon.h index 0de9b355795..4be084fb6a6 100644 --- a/fs/gfs2/daemon.h +++ b/fs/gfs2/daemon.h | |||
@@ -12,7 +12,6 @@ | |||
12 | 12 | ||
13 | int gfs2_glockd(void *data); | 13 | int gfs2_glockd(void *data); |
14 | int gfs2_recoverd(void *data); | 14 | int gfs2_recoverd(void *data); |
15 | int gfs2_logd(void *data); | ||
16 | int gfs2_quotad(void *data); | 15 | int gfs2_quotad(void *data); |
17 | 16 | ||
18 | #endif /* __DAEMON_DOT_H__ */ | 17 | #endif /* __DAEMON_DOT_H__ */ |
diff --git a/fs/gfs2/dir.c b/fs/gfs2/dir.c index 9949bb746a5..57e2ed932ad 100644 --- a/fs/gfs2/dir.c +++ b/fs/gfs2/dir.c | |||
@@ -1876,7 +1876,7 @@ static int leaf_dealloc(struct gfs2_inode *dip, u32 index, u32 len, | |||
1876 | if (error) | 1876 | if (error) |
1877 | goto out; | 1877 | goto out; |
1878 | 1878 | ||
1879 | error = gfs2_rindex_hold(sdp, &dip->i_alloc.al_ri_gh); | 1879 | error = gfs2_rindex_hold(sdp, &dip->i_alloc->al_ri_gh); |
1880 | if (error) | 1880 | if (error) |
1881 | goto out_qs; | 1881 | goto out_qs; |
1882 | 1882 | ||
@@ -1949,7 +1949,7 @@ out_rg_gunlock: | |||
1949 | gfs2_glock_dq_m(rlist.rl_rgrps, rlist.rl_ghs); | 1949 | gfs2_glock_dq_m(rlist.rl_rgrps, rlist.rl_ghs); |
1950 | out_rlist: | 1950 | out_rlist: |
1951 | gfs2_rlist_free(&rlist); | 1951 | gfs2_rlist_free(&rlist); |
1952 | gfs2_glock_dq_uninit(&dip->i_alloc.al_ri_gh); | 1952 | gfs2_glock_dq_uninit(&dip->i_alloc->al_ri_gh); |
1953 | out_qs: | 1953 | out_qs: |
1954 | gfs2_quota_unhold(dip); | 1954 | gfs2_quota_unhold(dip); |
1955 | out: | 1955 | out: |
diff --git a/fs/gfs2/eaops.c b/fs/gfs2/eaops.c index aa8dbf303f6..f114ba2b355 100644 --- a/fs/gfs2/eaops.c +++ b/fs/gfs2/eaops.c | |||
@@ -56,46 +56,6 @@ unsigned int gfs2_ea_name2type(const char *name, const char **truncated_name) | |||
56 | return type; | 56 | return type; |
57 | } | 57 | } |
58 | 58 | ||
59 | static int user_eo_get(struct gfs2_inode *ip, struct gfs2_ea_request *er) | ||
60 | { | ||
61 | struct inode *inode = &ip->i_inode; | ||
62 | int error = permission(inode, MAY_READ, NULL); | ||
63 | if (error) | ||
64 | return error; | ||
65 | |||
66 | return gfs2_ea_get_i(ip, er); | ||
67 | } | ||
68 | |||
69 | static int user_eo_set(struct gfs2_inode *ip, struct gfs2_ea_request *er) | ||
70 | { | ||
71 | struct inode *inode = &ip->i_inode; | ||
72 | |||
73 | if (S_ISREG(inode->i_mode) || | ||
74 | (S_ISDIR(inode->i_mode) && !(inode->i_mode & S_ISVTX))) { | ||
75 | int error = permission(inode, MAY_WRITE, NULL); | ||
76 | if (error) | ||
77 | return error; | ||
78 | } else | ||
79 | return -EPERM; | ||
80 | |||
81 | return gfs2_ea_set_i(ip, er); | ||
82 | } | ||
83 | |||
84 | static int user_eo_remove(struct gfs2_inode *ip, struct gfs2_ea_request *er) | ||
85 | { | ||
86 | struct inode *inode = &ip->i_inode; | ||
87 | |||
88 | if (S_ISREG(inode->i_mode) || | ||
89 | (S_ISDIR(inode->i_mode) && !(inode->i_mode & S_ISVTX))) { | ||
90 | int error = permission(inode, MAY_WRITE, NULL); | ||
91 | if (error) | ||
92 | return error; | ||
93 | } else | ||
94 | return -EPERM; | ||
95 | |||
96 | return gfs2_ea_remove_i(ip, er); | ||
97 | } | ||
98 | |||
99 | static int system_eo_get(struct gfs2_inode *ip, struct gfs2_ea_request *er) | 59 | static int system_eo_get(struct gfs2_inode *ip, struct gfs2_ea_request *er) |
100 | { | 60 | { |
101 | if (!GFS2_ACL_IS_ACCESS(er->er_name, er->er_name_len) && | 61 | if (!GFS2_ACL_IS_ACCESS(er->er_name, er->er_name_len) && |
@@ -108,8 +68,6 @@ static int system_eo_get(struct gfs2_inode *ip, struct gfs2_ea_request *er) | |||
108 | GFS2_ACL_IS_DEFAULT(er->er_name, er->er_name_len))) | 68 | GFS2_ACL_IS_DEFAULT(er->er_name, er->er_name_len))) |
109 | return -EOPNOTSUPP; | 69 | return -EOPNOTSUPP; |
110 | 70 | ||
111 | |||
112 | |||
113 | return gfs2_ea_get_i(ip, er); | 71 | return gfs2_ea_get_i(ip, er); |
114 | } | 72 | } |
115 | 73 | ||
@@ -170,40 +128,10 @@ static int system_eo_remove(struct gfs2_inode *ip, struct gfs2_ea_request *er) | |||
170 | return gfs2_ea_remove_i(ip, er); | 128 | return gfs2_ea_remove_i(ip, er); |
171 | } | 129 | } |
172 | 130 | ||
173 | static int security_eo_get(struct gfs2_inode *ip, struct gfs2_ea_request *er) | ||
174 | { | ||
175 | struct inode *inode = &ip->i_inode; | ||
176 | int error = permission(inode, MAY_READ, NULL); | ||
177 | if (error) | ||
178 | return error; | ||
179 | |||
180 | return gfs2_ea_get_i(ip, er); | ||
181 | } | ||
182 | |||
183 | static int security_eo_set(struct gfs2_inode *ip, struct gfs2_ea_request *er) | ||
184 | { | ||
185 | struct inode *inode = &ip->i_inode; | ||
186 | int error = permission(inode, MAY_WRITE, NULL); | ||
187 | if (error) | ||
188 | return error; | ||
189 | |||
190 | return gfs2_ea_set_i(ip, er); | ||
191 | } | ||
192 | |||
193 | static int security_eo_remove(struct gfs2_inode *ip, struct gfs2_ea_request *er) | ||
194 | { | ||
195 | struct inode *inode = &ip->i_inode; | ||
196 | int error = permission(inode, MAY_WRITE, NULL); | ||
197 | if (error) | ||
198 | return error; | ||
199 | |||
200 | return gfs2_ea_remove_i(ip, er); | ||
201 | } | ||
202 | |||
203 | static const struct gfs2_eattr_operations gfs2_user_eaops = { | 131 | static const struct gfs2_eattr_operations gfs2_user_eaops = { |
204 | .eo_get = user_eo_get, | 132 | .eo_get = gfs2_ea_get_i, |
205 | .eo_set = user_eo_set, | 133 | .eo_set = gfs2_ea_set_i, |
206 | .eo_remove = user_eo_remove, | 134 | .eo_remove = gfs2_ea_remove_i, |
207 | .eo_name = "user", | 135 | .eo_name = "user", |
208 | }; | 136 | }; |
209 | 137 | ||
@@ -215,9 +143,9 @@ const struct gfs2_eattr_operations gfs2_system_eaops = { | |||
215 | }; | 143 | }; |
216 | 144 | ||
217 | static const struct gfs2_eattr_operations gfs2_security_eaops = { | 145 | static const struct gfs2_eattr_operations gfs2_security_eaops = { |
218 | .eo_get = security_eo_get, | 146 | .eo_get = gfs2_ea_get_i, |
219 | .eo_set = security_eo_set, | 147 | .eo_set = gfs2_ea_set_i, |
220 | .eo_remove = security_eo_remove, | 148 | .eo_remove = gfs2_ea_remove_i, |
221 | .eo_name = "security", | 149 | .eo_name = "security", |
222 | }; | 150 | }; |
223 | 151 | ||
diff --git a/fs/gfs2/eattr.c b/fs/gfs2/eattr.c index 2a7435b5c4d..bee99704ea1 100644 --- a/fs/gfs2/eattr.c +++ b/fs/gfs2/eattr.c | |||
@@ -1418,7 +1418,7 @@ out: | |||
1418 | static int ea_dealloc_block(struct gfs2_inode *ip) | 1418 | static int ea_dealloc_block(struct gfs2_inode *ip) |
1419 | { | 1419 | { |
1420 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 1420 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
1421 | struct gfs2_alloc *al = &ip->i_alloc; | 1421 | struct gfs2_alloc *al = ip->i_alloc; |
1422 | struct gfs2_rgrpd *rgd; | 1422 | struct gfs2_rgrpd *rgd; |
1423 | struct buffer_head *dibh; | 1423 | struct buffer_head *dibh; |
1424 | int error; | 1424 | int error; |
diff --git a/fs/gfs2/glock.c b/fs/gfs2/glock.c index a37efe4aae6..80e09c50590 100644 --- a/fs/gfs2/glock.c +++ b/fs/gfs2/glock.c | |||
@@ -1,6 +1,6 @@ | |||
1 | /* | 1 | /* |
2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. |
3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | 3 | * Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved. |
4 | * | 4 | * |
5 | * This copyrighted material is made available to anyone wishing to use, | 5 | * This copyrighted material is made available to anyone wishing to use, |
6 | * modify, copy, or redistribute it subject to the terms and conditions | 6 | * modify, copy, or redistribute it subject to the terms and conditions |
@@ -217,7 +217,6 @@ int gfs2_glock_put(struct gfs2_glock *gl) | |||
217 | if (atomic_dec_and_test(&gl->gl_ref)) { | 217 | if (atomic_dec_and_test(&gl->gl_ref)) { |
218 | hlist_del(&gl->gl_list); | 218 | hlist_del(&gl->gl_list); |
219 | write_unlock(gl_lock_addr(gl->gl_hash)); | 219 | write_unlock(gl_lock_addr(gl->gl_hash)); |
220 | BUG_ON(spin_is_locked(&gl->gl_spin)); | ||
221 | gfs2_assert(sdp, gl->gl_state == LM_ST_UNLOCKED); | 220 | gfs2_assert(sdp, gl->gl_state == LM_ST_UNLOCKED); |
222 | gfs2_assert(sdp, list_empty(&gl->gl_reclaim)); | 221 | gfs2_assert(sdp, list_empty(&gl->gl_reclaim)); |
223 | gfs2_assert(sdp, list_empty(&gl->gl_holders)); | 222 | gfs2_assert(sdp, list_empty(&gl->gl_holders)); |
@@ -346,7 +345,6 @@ int gfs2_glock_get(struct gfs2_sbd *sdp, u64 number, | |||
346 | gl->gl_object = NULL; | 345 | gl->gl_object = NULL; |
347 | gl->gl_sbd = sdp; | 346 | gl->gl_sbd = sdp; |
348 | gl->gl_aspace = NULL; | 347 | gl->gl_aspace = NULL; |
349 | lops_init_le(&gl->gl_le, &gfs2_glock_lops); | ||
350 | INIT_DELAYED_WORK(&gl->gl_work, glock_work_func); | 348 | INIT_DELAYED_WORK(&gl->gl_work, glock_work_func); |
351 | 349 | ||
352 | /* If this glock protects actual on-disk data or metadata blocks, | 350 | /* If this glock protects actual on-disk data or metadata blocks, |
@@ -461,7 +459,6 @@ static void wait_on_holder(struct gfs2_holder *gh) | |||
461 | 459 | ||
462 | static void gfs2_demote_wake(struct gfs2_glock *gl) | 460 | static void gfs2_demote_wake(struct gfs2_glock *gl) |
463 | { | 461 | { |
464 | BUG_ON(!spin_is_locked(&gl->gl_spin)); | ||
465 | gl->gl_demote_state = LM_ST_EXCLUSIVE; | 462 | gl->gl_demote_state = LM_ST_EXCLUSIVE; |
466 | clear_bit(GLF_DEMOTE, &gl->gl_flags); | 463 | clear_bit(GLF_DEMOTE, &gl->gl_flags); |
467 | smp_mb__after_clear_bit(); | 464 | smp_mb__after_clear_bit(); |
@@ -507,21 +504,12 @@ static int rq_mutex(struct gfs2_holder *gh) | |||
507 | static int rq_promote(struct gfs2_holder *gh) | 504 | static int rq_promote(struct gfs2_holder *gh) |
508 | { | 505 | { |
509 | struct gfs2_glock *gl = gh->gh_gl; | 506 | struct gfs2_glock *gl = gh->gh_gl; |
510 | struct gfs2_sbd *sdp = gl->gl_sbd; | ||
511 | 507 | ||
512 | if (!relaxed_state_ok(gl->gl_state, gh->gh_state, gh->gh_flags)) { | 508 | if (!relaxed_state_ok(gl->gl_state, gh->gh_state, gh->gh_flags)) { |
513 | if (list_empty(&gl->gl_holders)) { | 509 | if (list_empty(&gl->gl_holders)) { |
514 | gl->gl_req_gh = gh; | 510 | gl->gl_req_gh = gh; |
515 | set_bit(GLF_LOCK, &gl->gl_flags); | 511 | set_bit(GLF_LOCK, &gl->gl_flags); |
516 | spin_unlock(&gl->gl_spin); | 512 | spin_unlock(&gl->gl_spin); |
517 | |||
518 | if (atomic_read(&sdp->sd_reclaim_count) > | ||
519 | gfs2_tune_get(sdp, gt_reclaim_limit) && | ||
520 | !(gh->gh_flags & LM_FLAG_PRIORITY)) { | ||
521 | gfs2_reclaim_glock(sdp); | ||
522 | gfs2_reclaim_glock(sdp); | ||
523 | } | ||
524 | |||
525 | gfs2_glock_xmote_th(gh->gh_gl, gh); | 513 | gfs2_glock_xmote_th(gh->gh_gl, gh); |
526 | spin_lock(&gl->gl_spin); | 514 | spin_lock(&gl->gl_spin); |
527 | } | 515 | } |
@@ -567,7 +555,10 @@ static int rq_demote(struct gfs2_glock *gl) | |||
567 | gfs2_demote_wake(gl); | 555 | gfs2_demote_wake(gl); |
568 | return 0; | 556 | return 0; |
569 | } | 557 | } |
558 | |||
570 | set_bit(GLF_LOCK, &gl->gl_flags); | 559 | set_bit(GLF_LOCK, &gl->gl_flags); |
560 | set_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags); | ||
561 | |||
571 | if (gl->gl_demote_state == LM_ST_UNLOCKED || | 562 | if (gl->gl_demote_state == LM_ST_UNLOCKED || |
572 | gl->gl_state != LM_ST_EXCLUSIVE) { | 563 | gl->gl_state != LM_ST_EXCLUSIVE) { |
573 | spin_unlock(&gl->gl_spin); | 564 | spin_unlock(&gl->gl_spin); |
@@ -576,7 +567,9 @@ static int rq_demote(struct gfs2_glock *gl) | |||
576 | spin_unlock(&gl->gl_spin); | 567 | spin_unlock(&gl->gl_spin); |
577 | gfs2_glock_xmote_th(gl, NULL); | 568 | gfs2_glock_xmote_th(gl, NULL); |
578 | } | 569 | } |
570 | |||
579 | spin_lock(&gl->gl_spin); | 571 | spin_lock(&gl->gl_spin); |
572 | clear_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags); | ||
580 | 573 | ||
581 | return 0; | 574 | return 0; |
582 | } | 575 | } |
@@ -598,23 +591,18 @@ static void run_queue(struct gfs2_glock *gl) | |||
598 | if (!list_empty(&gl->gl_waiters1)) { | 591 | if (!list_empty(&gl->gl_waiters1)) { |
599 | gh = list_entry(gl->gl_waiters1.next, | 592 | gh = list_entry(gl->gl_waiters1.next, |
600 | struct gfs2_holder, gh_list); | 593 | struct gfs2_holder, gh_list); |
601 | 594 | blocked = rq_mutex(gh); | |
602 | if (test_bit(HIF_MUTEX, &gh->gh_iflags)) | ||
603 | blocked = rq_mutex(gh); | ||
604 | else | ||
605 | gfs2_assert_warn(gl->gl_sbd, 0); | ||
606 | |||
607 | } else if (test_bit(GLF_DEMOTE, &gl->gl_flags)) { | 595 | } else if (test_bit(GLF_DEMOTE, &gl->gl_flags)) { |
608 | blocked = rq_demote(gl); | 596 | blocked = rq_demote(gl); |
597 | if (gl->gl_waiters2 && !blocked) { | ||
598 | set_bit(GLF_DEMOTE, &gl->gl_flags); | ||
599 | gl->gl_demote_state = LM_ST_UNLOCKED; | ||
600 | } | ||
601 | gl->gl_waiters2 = 0; | ||
609 | } else if (!list_empty(&gl->gl_waiters3)) { | 602 | } else if (!list_empty(&gl->gl_waiters3)) { |
610 | gh = list_entry(gl->gl_waiters3.next, | 603 | gh = list_entry(gl->gl_waiters3.next, |
611 | struct gfs2_holder, gh_list); | 604 | struct gfs2_holder, gh_list); |
612 | 605 | blocked = rq_promote(gh); | |
613 | if (test_bit(HIF_PROMOTE, &gh->gh_iflags)) | ||
614 | blocked = rq_promote(gh); | ||
615 | else | ||
616 | gfs2_assert_warn(gl->gl_sbd, 0); | ||
617 | |||
618 | } else | 606 | } else |
619 | break; | 607 | break; |
620 | 608 | ||
@@ -632,27 +620,21 @@ static void run_queue(struct gfs2_glock *gl) | |||
632 | 620 | ||
633 | static void gfs2_glmutex_lock(struct gfs2_glock *gl) | 621 | static void gfs2_glmutex_lock(struct gfs2_glock *gl) |
634 | { | 622 | { |
635 | struct gfs2_holder gh; | ||
636 | |||
637 | gfs2_holder_init(gl, 0, 0, &gh); | ||
638 | set_bit(HIF_MUTEX, &gh.gh_iflags); | ||
639 | if (test_and_set_bit(HIF_WAIT, &gh.gh_iflags)) | ||
640 | BUG(); | ||
641 | |||
642 | spin_lock(&gl->gl_spin); | 623 | spin_lock(&gl->gl_spin); |
643 | if (test_and_set_bit(GLF_LOCK, &gl->gl_flags)) { | 624 | if (test_and_set_bit(GLF_LOCK, &gl->gl_flags)) { |
625 | struct gfs2_holder gh; | ||
626 | |||
627 | gfs2_holder_init(gl, 0, 0, &gh); | ||
628 | set_bit(HIF_WAIT, &gh.gh_iflags); | ||
644 | list_add_tail(&gh.gh_list, &gl->gl_waiters1); | 629 | list_add_tail(&gh.gh_list, &gl->gl_waiters1); |
630 | spin_unlock(&gl->gl_spin); | ||
631 | wait_on_holder(&gh); | ||
632 | gfs2_holder_uninit(&gh); | ||
645 | } else { | 633 | } else { |
646 | gl->gl_owner_pid = current->pid; | 634 | gl->gl_owner_pid = current->pid; |
647 | gl->gl_ip = (unsigned long)__builtin_return_address(0); | 635 | gl->gl_ip = (unsigned long)__builtin_return_address(0); |
648 | clear_bit(HIF_WAIT, &gh.gh_iflags); | 636 | spin_unlock(&gl->gl_spin); |
649 | smp_mb(); | ||
650 | wake_up_bit(&gh.gh_iflags, HIF_WAIT); | ||
651 | } | 637 | } |
652 | spin_unlock(&gl->gl_spin); | ||
653 | |||
654 | wait_on_holder(&gh); | ||
655 | gfs2_holder_uninit(&gh); | ||
656 | } | 638 | } |
657 | 639 | ||
658 | /** | 640 | /** |
@@ -691,7 +673,6 @@ static void gfs2_glmutex_unlock(struct gfs2_glock *gl) | |||
691 | gl->gl_owner_pid = 0; | 673 | gl->gl_owner_pid = 0; |
692 | gl->gl_ip = 0; | 674 | gl->gl_ip = 0; |
693 | run_queue(gl); | 675 | run_queue(gl); |
694 | BUG_ON(!spin_is_locked(&gl->gl_spin)); | ||
695 | spin_unlock(&gl->gl_spin); | 676 | spin_unlock(&gl->gl_spin); |
696 | } | 677 | } |
697 | 678 | ||
@@ -722,7 +703,10 @@ static void handle_callback(struct gfs2_glock *gl, unsigned int state, | |||
722 | } | 703 | } |
723 | } else if (gl->gl_demote_state != LM_ST_UNLOCKED && | 704 | } else if (gl->gl_demote_state != LM_ST_UNLOCKED && |
724 | gl->gl_demote_state != state) { | 705 | gl->gl_demote_state != state) { |
725 | gl->gl_demote_state = LM_ST_UNLOCKED; | 706 | if (test_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags)) |
707 | gl->gl_waiters2 = 1; | ||
708 | else | ||
709 | gl->gl_demote_state = LM_ST_UNLOCKED; | ||
726 | } | 710 | } |
727 | spin_unlock(&gl->gl_spin); | 711 | spin_unlock(&gl->gl_spin); |
728 | } | 712 | } |
@@ -943,8 +927,8 @@ static void gfs2_glock_drop_th(struct gfs2_glock *gl) | |||
943 | const struct gfs2_glock_operations *glops = gl->gl_ops; | 927 | const struct gfs2_glock_operations *glops = gl->gl_ops; |
944 | unsigned int ret; | 928 | unsigned int ret; |
945 | 929 | ||
946 | if (glops->go_drop_th) | 930 | if (glops->go_xmote_th) |
947 | glops->go_drop_th(gl); | 931 | glops->go_xmote_th(gl); |
948 | 932 | ||
949 | gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags)); | 933 | gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags)); |
950 | gfs2_assert_warn(sdp, list_empty(&gl->gl_holders)); | 934 | gfs2_assert_warn(sdp, list_empty(&gl->gl_holders)); |
@@ -1156,8 +1140,6 @@ restart: | |||
1156 | return -EIO; | 1140 | return -EIO; |
1157 | } | 1141 | } |
1158 | 1142 | ||
1159 | set_bit(HIF_PROMOTE, &gh->gh_iflags); | ||
1160 | |||
1161 | spin_lock(&gl->gl_spin); | 1143 | spin_lock(&gl->gl_spin); |
1162 | add_to_queue(gh); | 1144 | add_to_queue(gh); |
1163 | run_queue(gl); | 1145 | run_queue(gl); |
@@ -1248,12 +1230,11 @@ void gfs2_glock_dq(struct gfs2_holder *gh) | |||
1248 | list_del_init(&gh->gh_list); | 1230 | list_del_init(&gh->gh_list); |
1249 | 1231 | ||
1250 | if (list_empty(&gl->gl_holders)) { | 1232 | if (list_empty(&gl->gl_holders)) { |
1251 | spin_unlock(&gl->gl_spin); | 1233 | if (glops->go_unlock) { |
1252 | 1234 | spin_unlock(&gl->gl_spin); | |
1253 | if (glops->go_unlock) | ||
1254 | glops->go_unlock(gh); | 1235 | glops->go_unlock(gh); |
1255 | 1236 | spin_lock(&gl->gl_spin); | |
1256 | spin_lock(&gl->gl_spin); | 1237 | } |
1257 | gl->gl_stamp = jiffies; | 1238 | gl->gl_stamp = jiffies; |
1258 | } | 1239 | } |
1259 | 1240 | ||
@@ -1910,8 +1891,6 @@ static int dump_glock(struct glock_iter *gi, struct gfs2_glock *gl) | |||
1910 | print_dbg(gi, " req_bh = %s\n", (gl->gl_req_bh) ? "yes" : "no"); | 1891 | print_dbg(gi, " req_bh = %s\n", (gl->gl_req_bh) ? "yes" : "no"); |
1911 | print_dbg(gi, " lvb_count = %d\n", atomic_read(&gl->gl_lvb_count)); | 1892 | print_dbg(gi, " lvb_count = %d\n", atomic_read(&gl->gl_lvb_count)); |
1912 | print_dbg(gi, " object = %s\n", (gl->gl_object) ? "yes" : "no"); | 1893 | print_dbg(gi, " object = %s\n", (gl->gl_object) ? "yes" : "no"); |
1913 | print_dbg(gi, " le = %s\n", | ||
1914 | (list_empty(&gl->gl_le.le_list)) ? "no" : "yes"); | ||
1915 | print_dbg(gi, " reclaim = %s\n", | 1894 | print_dbg(gi, " reclaim = %s\n", |
1916 | (list_empty(&gl->gl_reclaim)) ? "no" : "yes"); | 1895 | (list_empty(&gl->gl_reclaim)) ? "no" : "yes"); |
1917 | if (gl->gl_aspace) | 1896 | if (gl->gl_aspace) |
diff --git a/fs/gfs2/glops.c b/fs/gfs2/glops.c index 4670dcb2a87..c663b7a0f41 100644 --- a/fs/gfs2/glops.c +++ b/fs/gfs2/glops.c | |||
@@ -56,7 +56,7 @@ static void gfs2_ail_empty_gl(struct gfs2_glock *gl) | |||
56 | bd = list_entry(head->next, struct gfs2_bufdata, | 56 | bd = list_entry(head->next, struct gfs2_bufdata, |
57 | bd_ail_gl_list); | 57 | bd_ail_gl_list); |
58 | bh = bd->bd_bh; | 58 | bh = bd->bd_bh; |
59 | gfs2_remove_from_ail(NULL, bd); | 59 | gfs2_remove_from_ail(bd); |
60 | bd->bd_bh = NULL; | 60 | bd->bd_bh = NULL; |
61 | bh->b_private = NULL; | 61 | bh->b_private = NULL; |
62 | bd->bd_blkno = bh->b_blocknr; | 62 | bd->bd_blkno = bh->b_blocknr; |
@@ -86,15 +86,10 @@ static void gfs2_pte_inval(struct gfs2_glock *gl) | |||
86 | if (!ip || !S_ISREG(inode->i_mode)) | 86 | if (!ip || !S_ISREG(inode->i_mode)) |
87 | return; | 87 | return; |
88 | 88 | ||
89 | if (!test_bit(GIF_PAGED, &ip->i_flags)) | ||
90 | return; | ||
91 | |||
92 | unmap_shared_mapping_range(inode->i_mapping, 0, 0); | 89 | unmap_shared_mapping_range(inode->i_mapping, 0, 0); |
93 | |||
94 | if (test_bit(GIF_SW_PAGED, &ip->i_flags)) | 90 | if (test_bit(GIF_SW_PAGED, &ip->i_flags)) |
95 | set_bit(GLF_DIRTY, &gl->gl_flags); | 91 | set_bit(GLF_DIRTY, &gl->gl_flags); |
96 | 92 | ||
97 | clear_bit(GIF_SW_PAGED, &ip->i_flags); | ||
98 | } | 93 | } |
99 | 94 | ||
100 | /** | 95 | /** |
@@ -143,44 +138,34 @@ static void meta_go_inval(struct gfs2_glock *gl, int flags) | |||
143 | static void inode_go_sync(struct gfs2_glock *gl) | 138 | static void inode_go_sync(struct gfs2_glock *gl) |
144 | { | 139 | { |
145 | struct gfs2_inode *ip = gl->gl_object; | 140 | struct gfs2_inode *ip = gl->gl_object; |
141 | struct address_space *metamapping = gl->gl_aspace->i_mapping; | ||
142 | int error; | ||
143 | |||
144 | if (gl->gl_state != LM_ST_UNLOCKED) | ||
145 | gfs2_pte_inval(gl); | ||
146 | if (gl->gl_state != LM_ST_EXCLUSIVE) | ||
147 | return; | ||
146 | 148 | ||
147 | if (ip && !S_ISREG(ip->i_inode.i_mode)) | 149 | if (ip && !S_ISREG(ip->i_inode.i_mode)) |
148 | ip = NULL; | 150 | ip = NULL; |
149 | 151 | ||
150 | if (test_bit(GLF_DIRTY, &gl->gl_flags)) { | 152 | if (test_bit(GLF_DIRTY, &gl->gl_flags)) { |
151 | if (ip && !gfs2_is_jdata(ip)) | ||
152 | filemap_fdatawrite(ip->i_inode.i_mapping); | ||
153 | gfs2_log_flush(gl->gl_sbd, gl); | 153 | gfs2_log_flush(gl->gl_sbd, gl); |
154 | if (ip && gfs2_is_jdata(ip)) | 154 | filemap_fdatawrite(metamapping); |
155 | filemap_fdatawrite(ip->i_inode.i_mapping); | ||
156 | gfs2_meta_sync(gl); | ||
157 | if (ip) { | 155 | if (ip) { |
158 | struct address_space *mapping = ip->i_inode.i_mapping; | 156 | struct address_space *mapping = ip->i_inode.i_mapping; |
159 | int error = filemap_fdatawait(mapping); | 157 | filemap_fdatawrite(mapping); |
158 | error = filemap_fdatawait(mapping); | ||
160 | mapping_set_error(mapping, error); | 159 | mapping_set_error(mapping, error); |
161 | } | 160 | } |
161 | error = filemap_fdatawait(metamapping); | ||
162 | mapping_set_error(metamapping, error); | ||
162 | clear_bit(GLF_DIRTY, &gl->gl_flags); | 163 | clear_bit(GLF_DIRTY, &gl->gl_flags); |
163 | gfs2_ail_empty_gl(gl); | 164 | gfs2_ail_empty_gl(gl); |
164 | } | 165 | } |
165 | } | 166 | } |
166 | 167 | ||
167 | /** | 168 | /** |
168 | * inode_go_xmote_th - promote/demote a glock | ||
169 | * @gl: the glock | ||
170 | * @state: the requested state | ||
171 | * @flags: | ||
172 | * | ||
173 | */ | ||
174 | |||
175 | static void inode_go_xmote_th(struct gfs2_glock *gl) | ||
176 | { | ||
177 | if (gl->gl_state != LM_ST_UNLOCKED) | ||
178 | gfs2_pte_inval(gl); | ||
179 | if (gl->gl_state == LM_ST_EXCLUSIVE) | ||
180 | inode_go_sync(gl); | ||
181 | } | ||
182 | |||
183 | /** | ||
184 | * inode_go_xmote_bh - After promoting/demoting a glock | 169 | * inode_go_xmote_bh - After promoting/demoting a glock |
185 | * @gl: the glock | 170 | * @gl: the glock |
186 | * | 171 | * |
@@ -201,22 +186,6 @@ static void inode_go_xmote_bh(struct gfs2_glock *gl) | |||
201 | } | 186 | } |
202 | 187 | ||
203 | /** | 188 | /** |
204 | * inode_go_drop_th - unlock a glock | ||
205 | * @gl: the glock | ||
206 | * | ||
207 | * Invoked from rq_demote(). | ||
208 | * Another node needs the lock in EXCLUSIVE mode, or lock (unused for too long) | ||
209 | * is being purged from our node's glock cache; we're dropping lock. | ||
210 | */ | ||
211 | |||
212 | static void inode_go_drop_th(struct gfs2_glock *gl) | ||
213 | { | ||
214 | gfs2_pte_inval(gl); | ||
215 | if (gl->gl_state == LM_ST_EXCLUSIVE) | ||
216 | inode_go_sync(gl); | ||
217 | } | ||
218 | |||
219 | /** | ||
220 | * inode_go_inval - prepare a inode glock to be released | 189 | * inode_go_inval - prepare a inode glock to be released |
221 | * @gl: the glock | 190 | * @gl: the glock |
222 | * @flags: | 191 | * @flags: |
@@ -234,10 +203,8 @@ static void inode_go_inval(struct gfs2_glock *gl, int flags) | |||
234 | set_bit(GIF_INVALID, &ip->i_flags); | 203 | set_bit(GIF_INVALID, &ip->i_flags); |
235 | } | 204 | } |
236 | 205 | ||
237 | if (ip && S_ISREG(ip->i_inode.i_mode)) { | 206 | if (ip && S_ISREG(ip->i_inode.i_mode)) |
238 | truncate_inode_pages(ip->i_inode.i_mapping, 0); | 207 | truncate_inode_pages(ip->i_inode.i_mapping, 0); |
239 | clear_bit(GIF_PAGED, &ip->i_flags); | ||
240 | } | ||
241 | } | 208 | } |
242 | 209 | ||
243 | /** | 210 | /** |
@@ -294,23 +261,6 @@ static int inode_go_lock(struct gfs2_holder *gh) | |||
294 | } | 261 | } |
295 | 262 | ||
296 | /** | 263 | /** |
297 | * inode_go_unlock - operation done before an inode lock is unlocked by a | ||
298 | * process | ||
299 | * @gl: the glock | ||
300 | * @flags: | ||
301 | * | ||
302 | */ | ||
303 | |||
304 | static void inode_go_unlock(struct gfs2_holder *gh) | ||
305 | { | ||
306 | struct gfs2_glock *gl = gh->gh_gl; | ||
307 | struct gfs2_inode *ip = gl->gl_object; | ||
308 | |||
309 | if (ip) | ||
310 | gfs2_meta_cache_flush(ip); | ||
311 | } | ||
312 | |||
313 | /** | ||
314 | * rgrp_go_demote_ok - Check to see if it's ok to unlock a RG's glock | 264 | * rgrp_go_demote_ok - Check to see if it's ok to unlock a RG's glock |
315 | * @gl: the glock | 265 | * @gl: the glock |
316 | * | 266 | * |
@@ -350,14 +300,14 @@ static void rgrp_go_unlock(struct gfs2_holder *gh) | |||
350 | } | 300 | } |
351 | 301 | ||
352 | /** | 302 | /** |
353 | * trans_go_xmote_th - promote/demote the transaction glock | 303 | * trans_go_sync - promote/demote the transaction glock |
354 | * @gl: the glock | 304 | * @gl: the glock |
355 | * @state: the requested state | 305 | * @state: the requested state |
356 | * @flags: | 306 | * @flags: |
357 | * | 307 | * |
358 | */ | 308 | */ |
359 | 309 | ||
360 | static void trans_go_xmote_th(struct gfs2_glock *gl) | 310 | static void trans_go_sync(struct gfs2_glock *gl) |
361 | { | 311 | { |
362 | struct gfs2_sbd *sdp = gl->gl_sbd; | 312 | struct gfs2_sbd *sdp = gl->gl_sbd; |
363 | 313 | ||
@@ -384,7 +334,6 @@ static void trans_go_xmote_bh(struct gfs2_glock *gl) | |||
384 | 334 | ||
385 | if (gl->gl_state != LM_ST_UNLOCKED && | 335 | if (gl->gl_state != LM_ST_UNLOCKED && |
386 | test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags)) { | 336 | test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags)) { |
387 | gfs2_meta_cache_flush(GFS2_I(sdp->sd_jdesc->jd_inode)); | ||
388 | j_gl->gl_ops->go_inval(j_gl, DIO_METADATA); | 337 | j_gl->gl_ops->go_inval(j_gl, DIO_METADATA); |
389 | 338 | ||
390 | error = gfs2_find_jhead(sdp->sd_jdesc, &head); | 339 | error = gfs2_find_jhead(sdp->sd_jdesc, &head); |
@@ -402,24 +351,6 @@ static void trans_go_xmote_bh(struct gfs2_glock *gl) | |||
402 | } | 351 | } |
403 | 352 | ||
404 | /** | 353 | /** |
405 | * trans_go_drop_th - unlock the transaction glock | ||
406 | * @gl: the glock | ||
407 | * | ||
408 | * We want to sync the device even with localcaching. Remember | ||
409 | * that localcaching journal replay only marks buffers dirty. | ||
410 | */ | ||
411 | |||
412 | static void trans_go_drop_th(struct gfs2_glock *gl) | ||
413 | { | ||
414 | struct gfs2_sbd *sdp = gl->gl_sbd; | ||
415 | |||
416 | if (test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags)) { | ||
417 | gfs2_meta_syncfs(sdp); | ||
418 | gfs2_log_shutdown(sdp); | ||
419 | } | ||
420 | } | ||
421 | |||
422 | /** | ||
423 | * quota_go_demote_ok - Check to see if it's ok to unlock a quota glock | 354 | * quota_go_demote_ok - Check to see if it's ok to unlock a quota glock |
424 | * @gl: the glock | 355 | * @gl: the glock |
425 | * | 356 | * |
@@ -433,25 +364,21 @@ static int quota_go_demote_ok(struct gfs2_glock *gl) | |||
433 | 364 | ||
434 | const struct gfs2_glock_operations gfs2_meta_glops = { | 365 | const struct gfs2_glock_operations gfs2_meta_glops = { |
435 | .go_xmote_th = meta_go_sync, | 366 | .go_xmote_th = meta_go_sync, |
436 | .go_drop_th = meta_go_sync, | ||
437 | .go_type = LM_TYPE_META, | 367 | .go_type = LM_TYPE_META, |
438 | }; | 368 | }; |
439 | 369 | ||
440 | const struct gfs2_glock_operations gfs2_inode_glops = { | 370 | const struct gfs2_glock_operations gfs2_inode_glops = { |
441 | .go_xmote_th = inode_go_xmote_th, | 371 | .go_xmote_th = inode_go_sync, |
442 | .go_xmote_bh = inode_go_xmote_bh, | 372 | .go_xmote_bh = inode_go_xmote_bh, |
443 | .go_drop_th = inode_go_drop_th, | ||
444 | .go_inval = inode_go_inval, | 373 | .go_inval = inode_go_inval, |
445 | .go_demote_ok = inode_go_demote_ok, | 374 | .go_demote_ok = inode_go_demote_ok, |
446 | .go_lock = inode_go_lock, | 375 | .go_lock = inode_go_lock, |
447 | .go_unlock = inode_go_unlock, | ||
448 | .go_type = LM_TYPE_INODE, | 376 | .go_type = LM_TYPE_INODE, |
449 | .go_min_hold_time = HZ / 10, | 377 | .go_min_hold_time = HZ / 10, |
450 | }; | 378 | }; |
451 | 379 | ||
452 | const struct gfs2_glock_operations gfs2_rgrp_glops = { | 380 | const struct gfs2_glock_operations gfs2_rgrp_glops = { |
453 | .go_xmote_th = meta_go_sync, | 381 | .go_xmote_th = meta_go_sync, |
454 | .go_drop_th = meta_go_sync, | ||
455 | .go_inval = meta_go_inval, | 382 | .go_inval = meta_go_inval, |
456 | .go_demote_ok = rgrp_go_demote_ok, | 383 | .go_demote_ok = rgrp_go_demote_ok, |
457 | .go_lock = rgrp_go_lock, | 384 | .go_lock = rgrp_go_lock, |
@@ -461,9 +388,8 @@ const struct gfs2_glock_operations gfs2_rgrp_glops = { | |||
461 | }; | 388 | }; |
462 | 389 | ||
463 | const struct gfs2_glock_operations gfs2_trans_glops = { | 390 | const struct gfs2_glock_operations gfs2_trans_glops = { |
464 | .go_xmote_th = trans_go_xmote_th, | 391 | .go_xmote_th = trans_go_sync, |
465 | .go_xmote_bh = trans_go_xmote_bh, | 392 | .go_xmote_bh = trans_go_xmote_bh, |
466 | .go_drop_th = trans_go_drop_th, | ||
467 | .go_type = LM_TYPE_NONDISK, | 393 | .go_type = LM_TYPE_NONDISK, |
468 | }; | 394 | }; |
469 | 395 | ||
diff --git a/fs/gfs2/incore.h b/fs/gfs2/incore.h index eaddfb5a8e6..513aaf0dc0a 100644 --- a/fs/gfs2/incore.h +++ b/fs/gfs2/incore.h | |||
@@ -1,6 +1,6 @@ | |||
1 | /* | 1 | /* |
2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. |
3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | 3 | * Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved. |
4 | * | 4 | * |
5 | * This copyrighted material is made available to anyone wishing to use, | 5 | * This copyrighted material is made available to anyone wishing to use, |
6 | * modify, copy, or redistribute it subject to the terms and conditions | 6 | * modify, copy, or redistribute it subject to the terms and conditions |
@@ -131,7 +131,6 @@ struct gfs2_bufdata { | |||
131 | struct gfs2_glock_operations { | 131 | struct gfs2_glock_operations { |
132 | void (*go_xmote_th) (struct gfs2_glock *gl); | 132 | void (*go_xmote_th) (struct gfs2_glock *gl); |
133 | void (*go_xmote_bh) (struct gfs2_glock *gl); | 133 | void (*go_xmote_bh) (struct gfs2_glock *gl); |
134 | void (*go_drop_th) (struct gfs2_glock *gl); | ||
135 | void (*go_inval) (struct gfs2_glock *gl, int flags); | 134 | void (*go_inval) (struct gfs2_glock *gl, int flags); |
136 | int (*go_demote_ok) (struct gfs2_glock *gl); | 135 | int (*go_demote_ok) (struct gfs2_glock *gl); |
137 | int (*go_lock) (struct gfs2_holder *gh); | 136 | int (*go_lock) (struct gfs2_holder *gh); |
@@ -141,10 +140,6 @@ struct gfs2_glock_operations { | |||
141 | }; | 140 | }; |
142 | 141 | ||
143 | enum { | 142 | enum { |
144 | /* Actions */ | ||
145 | HIF_MUTEX = 0, | ||
146 | HIF_PROMOTE = 1, | ||
147 | |||
148 | /* States */ | 143 | /* States */ |
149 | HIF_HOLDER = 6, | 144 | HIF_HOLDER = 6, |
150 | HIF_FIRST = 7, | 145 | HIF_FIRST = 7, |
@@ -171,6 +166,8 @@ enum { | |||
171 | GLF_DEMOTE = 3, | 166 | GLF_DEMOTE = 3, |
172 | GLF_PENDING_DEMOTE = 4, | 167 | GLF_PENDING_DEMOTE = 4, |
173 | GLF_DIRTY = 5, | 168 | GLF_DIRTY = 5, |
169 | GLF_DEMOTE_IN_PROGRESS = 6, | ||
170 | GLF_LFLUSH = 7, | ||
174 | }; | 171 | }; |
175 | 172 | ||
176 | struct gfs2_glock { | 173 | struct gfs2_glock { |
@@ -190,6 +187,7 @@ struct gfs2_glock { | |||
190 | struct list_head gl_holders; | 187 | struct list_head gl_holders; |
191 | struct list_head gl_waiters1; /* HIF_MUTEX */ | 188 | struct list_head gl_waiters1; /* HIF_MUTEX */ |
192 | struct list_head gl_waiters3; /* HIF_PROMOTE */ | 189 | struct list_head gl_waiters3; /* HIF_PROMOTE */ |
190 | int gl_waiters2; /* GIF_DEMOTE */ | ||
193 | 191 | ||
194 | const struct gfs2_glock_operations *gl_ops; | 192 | const struct gfs2_glock_operations *gl_ops; |
195 | 193 | ||
@@ -210,7 +208,6 @@ struct gfs2_glock { | |||
210 | struct gfs2_sbd *gl_sbd; | 208 | struct gfs2_sbd *gl_sbd; |
211 | 209 | ||
212 | struct inode *gl_aspace; | 210 | struct inode *gl_aspace; |
213 | struct gfs2_log_element gl_le; | ||
214 | struct list_head gl_ail_list; | 211 | struct list_head gl_ail_list; |
215 | atomic_t gl_ail_count; | 212 | atomic_t gl_ail_count; |
216 | struct delayed_work gl_work; | 213 | struct delayed_work gl_work; |
@@ -239,7 +236,6 @@ struct gfs2_alloc { | |||
239 | enum { | 236 | enum { |
240 | GIF_INVALID = 0, | 237 | GIF_INVALID = 0, |
241 | GIF_QD_LOCKED = 1, | 238 | GIF_QD_LOCKED = 1, |
242 | GIF_PAGED = 2, | ||
243 | GIF_SW_PAGED = 3, | 239 | GIF_SW_PAGED = 3, |
244 | }; | 240 | }; |
245 | 241 | ||
@@ -268,14 +264,10 @@ struct gfs2_inode { | |||
268 | struct gfs2_glock *i_gl; /* Move into i_gh? */ | 264 | struct gfs2_glock *i_gl; /* Move into i_gh? */ |
269 | struct gfs2_holder i_iopen_gh; | 265 | struct gfs2_holder i_iopen_gh; |
270 | struct gfs2_holder i_gh; /* for prepare/commit_write only */ | 266 | struct gfs2_holder i_gh; /* for prepare/commit_write only */ |
271 | struct gfs2_alloc i_alloc; | 267 | struct gfs2_alloc *i_alloc; |
272 | u64 i_last_rg_alloc; | 268 | u64 i_last_rg_alloc; |
273 | 269 | ||
274 | spinlock_t i_spin; | ||
275 | struct rw_semaphore i_rw_mutex; | 270 | struct rw_semaphore i_rw_mutex; |
276 | unsigned long i_last_pfault; | ||
277 | |||
278 | struct buffer_head *i_cache[GFS2_MAX_META_HEIGHT]; | ||
279 | }; | 271 | }; |
280 | 272 | ||
281 | /* | 273 | /* |
@@ -287,19 +279,12 @@ static inline struct gfs2_inode *GFS2_I(struct inode *inode) | |||
287 | return container_of(inode, struct gfs2_inode, i_inode); | 279 | return container_of(inode, struct gfs2_inode, i_inode); |
288 | } | 280 | } |
289 | 281 | ||
290 | /* To be removed? */ | 282 | static inline struct gfs2_sbd *GFS2_SB(const struct inode *inode) |
291 | static inline struct gfs2_sbd *GFS2_SB(struct inode *inode) | ||
292 | { | 283 | { |
293 | return inode->i_sb->s_fs_info; | 284 | return inode->i_sb->s_fs_info; |
294 | } | 285 | } |
295 | 286 | ||
296 | enum { | ||
297 | GFF_DID_DIRECT_ALLOC = 0, | ||
298 | GFF_EXLOCK = 1, | ||
299 | }; | ||
300 | |||
301 | struct gfs2_file { | 287 | struct gfs2_file { |
302 | unsigned long f_flags; /* GFF_... */ | ||
303 | struct mutex f_fl_mutex; | 288 | struct mutex f_fl_mutex; |
304 | struct gfs2_holder f_fl_gh; | 289 | struct gfs2_holder f_fl_gh; |
305 | }; | 290 | }; |
@@ -373,8 +358,17 @@ struct gfs2_ail { | |||
373 | u64 ai_sync_gen; | 358 | u64 ai_sync_gen; |
374 | }; | 359 | }; |
375 | 360 | ||
361 | struct gfs2_journal_extent { | ||
362 | struct list_head extent_list; | ||
363 | |||
364 | unsigned int lblock; /* First logical block */ | ||
365 | u64 dblock; /* First disk block */ | ||
366 | u64 blocks; | ||
367 | }; | ||
368 | |||
376 | struct gfs2_jdesc { | 369 | struct gfs2_jdesc { |
377 | struct list_head jd_list; | 370 | struct list_head jd_list; |
371 | struct list_head extent_list; | ||
378 | 372 | ||
379 | struct inode *jd_inode; | 373 | struct inode *jd_inode; |
380 | unsigned int jd_jid; | 374 | unsigned int jd_jid; |
@@ -421,13 +415,9 @@ struct gfs2_args { | |||
421 | struct gfs2_tune { | 415 | struct gfs2_tune { |
422 | spinlock_t gt_spin; | 416 | spinlock_t gt_spin; |
423 | 417 | ||
424 | unsigned int gt_ilimit; | ||
425 | unsigned int gt_ilimit_tries; | ||
426 | unsigned int gt_ilimit_min; | ||
427 | unsigned int gt_demote_secs; /* Cache retention for unheld glock */ | 418 | unsigned int gt_demote_secs; /* Cache retention for unheld glock */ |
428 | unsigned int gt_incore_log_blocks; | 419 | unsigned int gt_incore_log_blocks; |
429 | unsigned int gt_log_flush_secs; | 420 | unsigned int gt_log_flush_secs; |
430 | unsigned int gt_jindex_refresh_secs; /* Check for new journal index */ | ||
431 | 421 | ||
432 | unsigned int gt_recoverd_secs; | 422 | unsigned int gt_recoverd_secs; |
433 | unsigned int gt_logd_secs; | 423 | unsigned int gt_logd_secs; |
@@ -443,10 +433,8 @@ struct gfs2_tune { | |||
443 | unsigned int gt_new_files_jdata; | 433 | unsigned int gt_new_files_jdata; |
444 | unsigned int gt_new_files_directio; | 434 | unsigned int gt_new_files_directio; |
445 | unsigned int gt_max_readahead; /* Max bytes to read-ahead from disk */ | 435 | unsigned int gt_max_readahead; /* Max bytes to read-ahead from disk */ |
446 | unsigned int gt_lockdump_size; | ||
447 | unsigned int gt_stall_secs; /* Detects trouble! */ | 436 | unsigned int gt_stall_secs; /* Detects trouble! */ |
448 | unsigned int gt_complain_secs; | 437 | unsigned int gt_complain_secs; |
449 | unsigned int gt_reclaim_limit; /* Max num of glocks in reclaim list */ | ||
450 | unsigned int gt_statfs_quantum; | 438 | unsigned int gt_statfs_quantum; |
451 | unsigned int gt_statfs_slow; | 439 | unsigned int gt_statfs_slow; |
452 | }; | 440 | }; |
@@ -539,7 +527,6 @@ struct gfs2_sbd { | |||
539 | /* StatFS stuff */ | 527 | /* StatFS stuff */ |
540 | 528 | ||
541 | spinlock_t sd_statfs_spin; | 529 | spinlock_t sd_statfs_spin; |
542 | struct mutex sd_statfs_mutex; | ||
543 | struct gfs2_statfs_change_host sd_statfs_master; | 530 | struct gfs2_statfs_change_host sd_statfs_master; |
544 | struct gfs2_statfs_change_host sd_statfs_local; | 531 | struct gfs2_statfs_change_host sd_statfs_local; |
545 | unsigned long sd_statfs_sync_time; | 532 | unsigned long sd_statfs_sync_time; |
@@ -602,20 +589,18 @@ struct gfs2_sbd { | |||
602 | unsigned int sd_log_commited_databuf; | 589 | unsigned int sd_log_commited_databuf; |
603 | unsigned int sd_log_commited_revoke; | 590 | unsigned int sd_log_commited_revoke; |
604 | 591 | ||
605 | unsigned int sd_log_num_gl; | ||
606 | unsigned int sd_log_num_buf; | 592 | unsigned int sd_log_num_buf; |
607 | unsigned int sd_log_num_revoke; | 593 | unsigned int sd_log_num_revoke; |
608 | unsigned int sd_log_num_rg; | 594 | unsigned int sd_log_num_rg; |
609 | unsigned int sd_log_num_databuf; | 595 | unsigned int sd_log_num_databuf; |
610 | 596 | ||
611 | struct list_head sd_log_le_gl; | ||
612 | struct list_head sd_log_le_buf; | 597 | struct list_head sd_log_le_buf; |
613 | struct list_head sd_log_le_revoke; | 598 | struct list_head sd_log_le_revoke; |
614 | struct list_head sd_log_le_rg; | 599 | struct list_head sd_log_le_rg; |
615 | struct list_head sd_log_le_databuf; | 600 | struct list_head sd_log_le_databuf; |
616 | struct list_head sd_log_le_ordered; | 601 | struct list_head sd_log_le_ordered; |
617 | 602 | ||
618 | unsigned int sd_log_blks_free; | 603 | atomic_t sd_log_blks_free; |
619 | struct mutex sd_log_reserve_mutex; | 604 | struct mutex sd_log_reserve_mutex; |
620 | 605 | ||
621 | u64 sd_log_sequence; | 606 | u64 sd_log_sequence; |
diff --git a/fs/gfs2/inode.c b/fs/gfs2/inode.c index 5f6dc32946c..728d3169e7b 100644 --- a/fs/gfs2/inode.c +++ b/fs/gfs2/inode.c | |||
@@ -31,7 +31,6 @@ | |||
31 | #include "log.h" | 31 | #include "log.h" |
32 | #include "meta_io.h" | 32 | #include "meta_io.h" |
33 | #include "ops_address.h" | 33 | #include "ops_address.h" |
34 | #include "ops_file.h" | ||
35 | #include "ops_inode.h" | 34 | #include "ops_inode.h" |
36 | #include "quota.h" | 35 | #include "quota.h" |
37 | #include "rgrp.h" | 36 | #include "rgrp.h" |
@@ -132,15 +131,21 @@ static struct inode *gfs2_iget_skip(struct super_block *sb, | |||
132 | 131 | ||
133 | void gfs2_set_iop(struct inode *inode) | 132 | void gfs2_set_iop(struct inode *inode) |
134 | { | 133 | { |
134 | struct gfs2_sbd *sdp = GFS2_SB(inode); | ||
135 | umode_t mode = inode->i_mode; | 135 | umode_t mode = inode->i_mode; |
136 | 136 | ||
137 | if (S_ISREG(mode)) { | 137 | if (S_ISREG(mode)) { |
138 | inode->i_op = &gfs2_file_iops; | 138 | inode->i_op = &gfs2_file_iops; |
139 | inode->i_fop = &gfs2_file_fops; | 139 | if (sdp->sd_args.ar_localflocks) |
140 | inode->i_mapping->a_ops = &gfs2_file_aops; | 140 | inode->i_fop = &gfs2_file_fops_nolock; |
141 | else | ||
142 | inode->i_fop = &gfs2_file_fops; | ||
141 | } else if (S_ISDIR(mode)) { | 143 | } else if (S_ISDIR(mode)) { |
142 | inode->i_op = &gfs2_dir_iops; | 144 | inode->i_op = &gfs2_dir_iops; |
143 | inode->i_fop = &gfs2_dir_fops; | 145 | if (sdp->sd_args.ar_localflocks) |
146 | inode->i_fop = &gfs2_dir_fops_nolock; | ||
147 | else | ||
148 | inode->i_fop = &gfs2_dir_fops; | ||
144 | } else if (S_ISLNK(mode)) { | 149 | } else if (S_ISLNK(mode)) { |
145 | inode->i_op = &gfs2_symlink_iops; | 150 | inode->i_op = &gfs2_symlink_iops; |
146 | } else { | 151 | } else { |
@@ -291,12 +296,10 @@ static int gfs2_dinode_in(struct gfs2_inode *ip, const void *buf) | |||
291 | di->di_entries = be32_to_cpu(str->di_entries); | 296 | di->di_entries = be32_to_cpu(str->di_entries); |
292 | 297 | ||
293 | di->di_eattr = be64_to_cpu(str->di_eattr); | 298 | di->di_eattr = be64_to_cpu(str->di_eattr); |
294 | return 0; | 299 | if (S_ISREG(ip->i_inode.i_mode)) |
295 | } | 300 | gfs2_set_aops(&ip->i_inode); |
296 | 301 | ||
297 | static void gfs2_inode_bh(struct gfs2_inode *ip, struct buffer_head *bh) | 302 | return 0; |
298 | { | ||
299 | ip->i_cache[0] = bh; | ||
300 | } | 303 | } |
301 | 304 | ||
302 | /** | 305 | /** |
@@ -366,7 +369,8 @@ int gfs2_dinode_dealloc(struct gfs2_inode *ip) | |||
366 | if (error) | 369 | if (error) |
367 | goto out_rg_gunlock; | 370 | goto out_rg_gunlock; |
368 | 371 | ||
369 | gfs2_trans_add_gl(ip->i_gl); | 372 | set_bit(GLF_DIRTY, &ip->i_gl->gl_flags); |
373 | set_bit(GLF_LFLUSH, &ip->i_gl->gl_flags); | ||
370 | 374 | ||
371 | gfs2_free_di(rgd, ip); | 375 | gfs2_free_di(rgd, ip); |
372 | 376 | ||
@@ -707,9 +711,10 @@ static int alloc_dinode(struct gfs2_inode *dip, u64 *no_addr, u64 *generation) | |||
707 | struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); | 711 | struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); |
708 | int error; | 712 | int error; |
709 | 713 | ||
710 | gfs2_alloc_get(dip); | 714 | if (gfs2_alloc_get(dip) == NULL) |
715 | return -ENOMEM; | ||
711 | 716 | ||
712 | dip->i_alloc.al_requested = RES_DINODE; | 717 | dip->i_alloc->al_requested = RES_DINODE; |
713 | error = gfs2_inplace_reserve(dip); | 718 | error = gfs2_inplace_reserve(dip); |
714 | if (error) | 719 | if (error) |
715 | goto out; | 720 | goto out; |
@@ -855,7 +860,7 @@ static int link_dinode(struct gfs2_inode *dip, const struct qstr *name, | |||
855 | 860 | ||
856 | error = alloc_required = gfs2_diradd_alloc_required(&dip->i_inode, name); | 861 | error = alloc_required = gfs2_diradd_alloc_required(&dip->i_inode, name); |
857 | if (alloc_required < 0) | 862 | if (alloc_required < 0) |
858 | goto fail; | 863 | goto fail_quota_locks; |
859 | if (alloc_required) { | 864 | if (alloc_required) { |
860 | error = gfs2_quota_check(dip, dip->i_inode.i_uid, dip->i_inode.i_gid); | 865 | error = gfs2_quota_check(dip, dip->i_inode.i_uid, dip->i_inode.i_gid); |
861 | if (error) | 866 | if (error) |
@@ -896,7 +901,7 @@ fail_end_trans: | |||
896 | gfs2_trans_end(sdp); | 901 | gfs2_trans_end(sdp); |
897 | 902 | ||
898 | fail_ipreserv: | 903 | fail_ipreserv: |
899 | if (dip->i_alloc.al_rgd) | 904 | if (dip->i_alloc->al_rgd) |
900 | gfs2_inplace_release(dip); | 905 | gfs2_inplace_release(dip); |
901 | 906 | ||
902 | fail_quota_locks: | 907 | fail_quota_locks: |
@@ -966,7 +971,7 @@ struct inode *gfs2_createi(struct gfs2_holder *ghs, const struct qstr *name, | |||
966 | struct gfs2_inum_host inum = { .no_addr = 0, .no_formal_ino = 0 }; | 971 | struct gfs2_inum_host inum = { .no_addr = 0, .no_formal_ino = 0 }; |
967 | int error; | 972 | int error; |
968 | u64 generation; | 973 | u64 generation; |
969 | struct buffer_head *bh=NULL; | 974 | struct buffer_head *bh = NULL; |
970 | 975 | ||
971 | if (!name->len || name->len > GFS2_FNAMESIZE) | 976 | if (!name->len || name->len > GFS2_FNAMESIZE) |
972 | return ERR_PTR(-ENAMETOOLONG); | 977 | return ERR_PTR(-ENAMETOOLONG); |
@@ -1003,8 +1008,6 @@ struct inode *gfs2_createi(struct gfs2_holder *ghs, const struct qstr *name, | |||
1003 | if (IS_ERR(inode)) | 1008 | if (IS_ERR(inode)) |
1004 | goto fail_gunlock2; | 1009 | goto fail_gunlock2; |
1005 | 1010 | ||
1006 | gfs2_inode_bh(GFS2_I(inode), bh); | ||
1007 | |||
1008 | error = gfs2_inode_refresh(GFS2_I(inode)); | 1011 | error = gfs2_inode_refresh(GFS2_I(inode)); |
1009 | if (error) | 1012 | if (error) |
1010 | goto fail_gunlock2; | 1013 | goto fail_gunlock2; |
@@ -1021,6 +1024,8 @@ struct inode *gfs2_createi(struct gfs2_holder *ghs, const struct qstr *name, | |||
1021 | if (error) | 1024 | if (error) |
1022 | goto fail_gunlock2; | 1025 | goto fail_gunlock2; |
1023 | 1026 | ||
1027 | if (bh) | ||
1028 | brelse(bh); | ||
1024 | if (!inode) | 1029 | if (!inode) |
1025 | return ERR_PTR(-ENOMEM); | 1030 | return ERR_PTR(-ENOMEM); |
1026 | return inode; | 1031 | return inode; |
@@ -1032,6 +1037,8 @@ fail_gunlock2: | |||
1032 | fail_gunlock: | 1037 | fail_gunlock: |
1033 | gfs2_glock_dq(ghs); | 1038 | gfs2_glock_dq(ghs); |
1034 | fail: | 1039 | fail: |
1040 | if (bh) | ||
1041 | brelse(bh); | ||
1035 | return ERR_PTR(error); | 1042 | return ERR_PTR(error); |
1036 | } | 1043 | } |
1037 | 1044 | ||
diff --git a/fs/gfs2/inode.h b/fs/gfs2/inode.h index 351ac87ab38..d4465066261 100644 --- a/fs/gfs2/inode.h +++ b/fs/gfs2/inode.h | |||
@@ -20,6 +20,18 @@ static inline int gfs2_is_jdata(const struct gfs2_inode *ip) | |||
20 | return ip->i_di.di_flags & GFS2_DIF_JDATA; | 20 | return ip->i_di.di_flags & GFS2_DIF_JDATA; |
21 | } | 21 | } |
22 | 22 | ||
23 | static inline int gfs2_is_writeback(const struct gfs2_inode *ip) | ||
24 | { | ||
25 | const struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | ||
26 | return (sdp->sd_args.ar_data == GFS2_DATA_WRITEBACK) && !gfs2_is_jdata(ip); | ||
27 | } | ||
28 | |||
29 | static inline int gfs2_is_ordered(const struct gfs2_inode *ip) | ||
30 | { | ||
31 | const struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | ||
32 | return (sdp->sd_args.ar_data == GFS2_DATA_ORDERED) && !gfs2_is_jdata(ip); | ||
33 | } | ||
34 | |||
23 | static inline int gfs2_is_dir(const struct gfs2_inode *ip) | 35 | static inline int gfs2_is_dir(const struct gfs2_inode *ip) |
24 | { | 36 | { |
25 | return S_ISDIR(ip->i_inode.i_mode); | 37 | return S_ISDIR(ip->i_inode.i_mode); |
diff --git a/fs/gfs2/locking/dlm/mount.c b/fs/gfs2/locking/dlm/mount.c index 41c5b04caab..f2efff42422 100644 --- a/fs/gfs2/locking/dlm/mount.c +++ b/fs/gfs2/locking/dlm/mount.c | |||
@@ -67,6 +67,11 @@ static int make_args(struct gdlm_ls *ls, char *data_arg, int *nodir) | |||
67 | memset(data, 0, 256); | 67 | memset(data, 0, 256); |
68 | strncpy(data, data_arg, 255); | 68 | strncpy(data, data_arg, 255); |
69 | 69 | ||
70 | if (!strlen(data)) { | ||
71 | log_error("no mount options, (u)mount helpers not installed"); | ||
72 | return -EINVAL; | ||
73 | } | ||
74 | |||
70 | for (options = data; (x = strsep(&options, ":")); ) { | 75 | for (options = data; (x = strsep(&options, ":")); ) { |
71 | if (!*x) | 76 | if (!*x) |
72 | continue; | 77 | continue; |
diff --git a/fs/gfs2/locking/dlm/plock.c b/fs/gfs2/locking/dlm/plock.c index 1f7b038530b..2ebd374b314 100644 --- a/fs/gfs2/locking/dlm/plock.c +++ b/fs/gfs2/locking/dlm/plock.c | |||
@@ -89,15 +89,19 @@ int gdlm_plock(void *lockspace, struct lm_lockname *name, | |||
89 | op->info.number = name->ln_number; | 89 | op->info.number = name->ln_number; |
90 | op->info.start = fl->fl_start; | 90 | op->info.start = fl->fl_start; |
91 | op->info.end = fl->fl_end; | 91 | op->info.end = fl->fl_end; |
92 | op->info.owner = (__u64)(long) fl->fl_owner; | ||
93 | if (fl->fl_lmops && fl->fl_lmops->fl_grant) { | 92 | if (fl->fl_lmops && fl->fl_lmops->fl_grant) { |
93 | /* fl_owner is lockd which doesn't distinguish | ||
94 | processes on the nfs client */ | ||
95 | op->info.owner = (__u64) fl->fl_pid; | ||
94 | xop->callback = fl->fl_lmops->fl_grant; | 96 | xop->callback = fl->fl_lmops->fl_grant; |
95 | locks_init_lock(&xop->flc); | 97 | locks_init_lock(&xop->flc); |
96 | locks_copy_lock(&xop->flc, fl); | 98 | locks_copy_lock(&xop->flc, fl); |
97 | xop->fl = fl; | 99 | xop->fl = fl; |
98 | xop->file = file; | 100 | xop->file = file; |
99 | } else | 101 | } else { |
102 | op->info.owner = (__u64)(long) fl->fl_owner; | ||
100 | xop->callback = NULL; | 103 | xop->callback = NULL; |
104 | } | ||
101 | 105 | ||
102 | send_op(op); | 106 | send_op(op); |
103 | 107 | ||
@@ -203,7 +207,10 @@ int gdlm_punlock(void *lockspace, struct lm_lockname *name, | |||
203 | op->info.number = name->ln_number; | 207 | op->info.number = name->ln_number; |
204 | op->info.start = fl->fl_start; | 208 | op->info.start = fl->fl_start; |
205 | op->info.end = fl->fl_end; | 209 | op->info.end = fl->fl_end; |
206 | op->info.owner = (__u64)(long) fl->fl_owner; | 210 | if (fl->fl_lmops && fl->fl_lmops->fl_grant) |
211 | op->info.owner = (__u64) fl->fl_pid; | ||
212 | else | ||
213 | op->info.owner = (__u64)(long) fl->fl_owner; | ||
207 | 214 | ||
208 | send_op(op); | 215 | send_op(op); |
209 | wait_event(recv_wq, (op->done != 0)); | 216 | wait_event(recv_wq, (op->done != 0)); |
@@ -242,7 +249,10 @@ int gdlm_plock_get(void *lockspace, struct lm_lockname *name, | |||
242 | op->info.number = name->ln_number; | 249 | op->info.number = name->ln_number; |
243 | op->info.start = fl->fl_start; | 250 | op->info.start = fl->fl_start; |
244 | op->info.end = fl->fl_end; | 251 | op->info.end = fl->fl_end; |
245 | op->info.owner = (__u64)(long) fl->fl_owner; | 252 | if (fl->fl_lmops && fl->fl_lmops->fl_grant) |
253 | op->info.owner = (__u64) fl->fl_pid; | ||
254 | else | ||
255 | op->info.owner = (__u64)(long) fl->fl_owner; | ||
246 | 256 | ||
247 | send_op(op); | 257 | send_op(op); |
248 | wait_event(recv_wq, (op->done != 0)); | 258 | wait_event(recv_wq, (op->done != 0)); |
diff --git a/fs/gfs2/locking/dlm/sysfs.c b/fs/gfs2/locking/dlm/sysfs.c index ae9e6a25fe2..a87b0983976 100644 --- a/fs/gfs2/locking/dlm/sysfs.c +++ b/fs/gfs2/locking/dlm/sysfs.c | |||
@@ -189,51 +189,39 @@ static struct kobj_type gdlm_ktype = { | |||
189 | .sysfs_ops = &gdlm_attr_ops, | 189 | .sysfs_ops = &gdlm_attr_ops, |
190 | }; | 190 | }; |
191 | 191 | ||
192 | static struct kset gdlm_kset = { | 192 | static struct kset *gdlm_kset; |
193 | .ktype = &gdlm_ktype, | ||
194 | }; | ||
195 | 193 | ||
196 | int gdlm_kobject_setup(struct gdlm_ls *ls, struct kobject *fskobj) | 194 | int gdlm_kobject_setup(struct gdlm_ls *ls, struct kobject *fskobj) |
197 | { | 195 | { |
198 | int error; | 196 | int error; |
199 | 197 | ||
200 | error = kobject_set_name(&ls->kobj, "%s", "lock_module"); | 198 | ls->kobj.kset = gdlm_kset; |
201 | if (error) { | 199 | error = kobject_init_and_add(&ls->kobj, &gdlm_ktype, fskobj, |
202 | log_error("can't set kobj name %d", error); | 200 | "lock_module"); |
203 | return error; | ||
204 | } | ||
205 | |||
206 | ls->kobj.kset = &gdlm_kset; | ||
207 | ls->kobj.ktype = &gdlm_ktype; | ||
208 | ls->kobj.parent = fskobj; | ||
209 | |||
210 | error = kobject_register(&ls->kobj); | ||
211 | if (error) | 201 | if (error) |
212 | log_error("can't register kobj %d", error); | 202 | log_error("can't register kobj %d", error); |
203 | kobject_uevent(&ls->kobj, KOBJ_ADD); | ||
213 | 204 | ||
214 | return error; | 205 | return error; |
215 | } | 206 | } |
216 | 207 | ||
217 | void gdlm_kobject_release(struct gdlm_ls *ls) | 208 | void gdlm_kobject_release(struct gdlm_ls *ls) |
218 | { | 209 | { |
219 | kobject_unregister(&ls->kobj); | 210 | kobject_put(&ls->kobj); |
220 | } | 211 | } |
221 | 212 | ||
222 | int gdlm_sysfs_init(void) | 213 | int gdlm_sysfs_init(void) |
223 | { | 214 | { |
224 | int error; | 215 | gdlm_kset = kset_create_and_add("lock_dlm", NULL, kernel_kobj); |
225 | 216 | if (!gdlm_kset) { | |
226 | kobject_set_name(&gdlm_kset.kobj, "lock_dlm"); | 217 | printk(KERN_WARNING "%s: can not create kset\n", __FUNCTION__); |
227 | kobj_set_kset_s(&gdlm_kset, kernel_subsys); | 218 | return -ENOMEM; |
228 | error = kset_register(&gdlm_kset); | 219 | } |
229 | if (error) | 220 | return 0; |
230 | printk("lock_dlm: cannot register kset %d\n", error); | ||
231 | |||
232 | return error; | ||
233 | } | 221 | } |
234 | 222 | ||
235 | void gdlm_sysfs_exit(void) | 223 | void gdlm_sysfs_exit(void) |
236 | { | 224 | { |
237 | kset_unregister(&gdlm_kset); | 225 | kset_unregister(gdlm_kset); |
238 | } | 226 | } |
239 | 227 | ||
diff --git a/fs/gfs2/locking/dlm/thread.c b/fs/gfs2/locking/dlm/thread.c index bd938f06481..521694fc19d 100644 --- a/fs/gfs2/locking/dlm/thread.c +++ b/fs/gfs2/locking/dlm/thread.c | |||
@@ -273,18 +273,13 @@ static int gdlm_thread(void *data, int blist) | |||
273 | struct gdlm_ls *ls = (struct gdlm_ls *) data; | 273 | struct gdlm_ls *ls = (struct gdlm_ls *) data; |
274 | struct gdlm_lock *lp = NULL; | 274 | struct gdlm_lock *lp = NULL; |
275 | uint8_t complete, blocking, submit, drop; | 275 | uint8_t complete, blocking, submit, drop; |
276 | DECLARE_WAITQUEUE(wait, current); | ||
277 | 276 | ||
278 | /* Only thread1 is allowed to do blocking callbacks since gfs | 277 | /* Only thread1 is allowed to do blocking callbacks since gfs |
279 | may wait for a completion callback within a blocking cb. */ | 278 | may wait for a completion callback within a blocking cb. */ |
280 | 279 | ||
281 | while (!kthread_should_stop()) { | 280 | while (!kthread_should_stop()) { |
282 | set_current_state(TASK_INTERRUPTIBLE); | 281 | wait_event_interruptible(ls->thread_wait, |
283 | add_wait_queue(&ls->thread_wait, &wait); | 282 | !no_work(ls, blist) || kthread_should_stop()); |
284 | if (no_work(ls, blist)) | ||
285 | schedule(); | ||
286 | remove_wait_queue(&ls->thread_wait, &wait); | ||
287 | set_current_state(TASK_RUNNING); | ||
288 | 283 | ||
289 | complete = blocking = submit = drop = 0; | 284 | complete = blocking = submit = drop = 0; |
290 | 285 | ||
diff --git a/fs/gfs2/log.c b/fs/gfs2/log.c index 7df70247325..161ab6f2058 100644 --- a/fs/gfs2/log.c +++ b/fs/gfs2/log.c | |||
@@ -1,6 +1,6 @@ | |||
1 | /* | 1 | /* |
2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. |
3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | 3 | * Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved. |
4 | * | 4 | * |
5 | * This copyrighted material is made available to anyone wishing to use, | 5 | * This copyrighted material is made available to anyone wishing to use, |
6 | * modify, copy, or redistribute it subject to the terms and conditions | 6 | * modify, copy, or redistribute it subject to the terms and conditions |
@@ -16,6 +16,8 @@ | |||
16 | #include <linux/crc32.h> | 16 | #include <linux/crc32.h> |
17 | #include <linux/lm_interface.h> | 17 | #include <linux/lm_interface.h> |
18 | #include <linux/delay.h> | 18 | #include <linux/delay.h> |
19 | #include <linux/kthread.h> | ||
20 | #include <linux/freezer.h> | ||
19 | 21 | ||
20 | #include "gfs2.h" | 22 | #include "gfs2.h" |
21 | #include "incore.h" | 23 | #include "incore.h" |
@@ -68,14 +70,12 @@ unsigned int gfs2_struct2blk(struct gfs2_sbd *sdp, unsigned int nstruct, | |||
68 | * | 70 | * |
69 | */ | 71 | */ |
70 | 72 | ||
71 | void gfs2_remove_from_ail(struct address_space *mapping, struct gfs2_bufdata *bd) | 73 | void gfs2_remove_from_ail(struct gfs2_bufdata *bd) |
72 | { | 74 | { |
73 | bd->bd_ail = NULL; | 75 | bd->bd_ail = NULL; |
74 | list_del_init(&bd->bd_ail_st_list); | 76 | list_del_init(&bd->bd_ail_st_list); |
75 | list_del_init(&bd->bd_ail_gl_list); | 77 | list_del_init(&bd->bd_ail_gl_list); |
76 | atomic_dec(&bd->bd_gl->gl_ail_count); | 78 | atomic_dec(&bd->bd_gl->gl_ail_count); |
77 | if (mapping) | ||
78 | gfs2_meta_cache_flush(GFS2_I(mapping->host)); | ||
79 | brelse(bd->bd_bh); | 79 | brelse(bd->bd_bh); |
80 | } | 80 | } |
81 | 81 | ||
@@ -92,8 +92,6 @@ static void gfs2_ail1_start_one(struct gfs2_sbd *sdp, struct gfs2_ail *ai) | |||
92 | struct buffer_head *bh; | 92 | struct buffer_head *bh; |
93 | int retry; | 93 | int retry; |
94 | 94 | ||
95 | BUG_ON(!spin_is_locked(&sdp->sd_log_lock)); | ||
96 | |||
97 | do { | 95 | do { |
98 | retry = 0; | 96 | retry = 0; |
99 | 97 | ||
@@ -210,7 +208,7 @@ static void gfs2_ail1_start(struct gfs2_sbd *sdp, int flags) | |||
210 | gfs2_log_unlock(sdp); | 208 | gfs2_log_unlock(sdp); |
211 | } | 209 | } |
212 | 210 | ||
213 | int gfs2_ail1_empty(struct gfs2_sbd *sdp, int flags) | 211 | static int gfs2_ail1_empty(struct gfs2_sbd *sdp, int flags) |
214 | { | 212 | { |
215 | struct gfs2_ail *ai, *s; | 213 | struct gfs2_ail *ai, *s; |
216 | int ret; | 214 | int ret; |
@@ -248,7 +246,7 @@ static void gfs2_ail2_empty_one(struct gfs2_sbd *sdp, struct gfs2_ail *ai) | |||
248 | bd = list_entry(head->prev, struct gfs2_bufdata, | 246 | bd = list_entry(head->prev, struct gfs2_bufdata, |
249 | bd_ail_st_list); | 247 | bd_ail_st_list); |
250 | gfs2_assert(sdp, bd->bd_ail == ai); | 248 | gfs2_assert(sdp, bd->bd_ail == ai); |
251 | gfs2_remove_from_ail(bd->bd_bh->b_page->mapping, bd); | 249 | gfs2_remove_from_ail(bd); |
252 | } | 250 | } |
253 | } | 251 | } |
254 | 252 | ||
@@ -303,7 +301,7 @@ int gfs2_log_reserve(struct gfs2_sbd *sdp, unsigned int blks) | |||
303 | 301 | ||
304 | mutex_lock(&sdp->sd_log_reserve_mutex); | 302 | mutex_lock(&sdp->sd_log_reserve_mutex); |
305 | gfs2_log_lock(sdp); | 303 | gfs2_log_lock(sdp); |
306 | while(sdp->sd_log_blks_free <= (blks + reserved_blks)) { | 304 | while(atomic_read(&sdp->sd_log_blks_free) <= (blks + reserved_blks)) { |
307 | gfs2_log_unlock(sdp); | 305 | gfs2_log_unlock(sdp); |
308 | gfs2_ail1_empty(sdp, 0); | 306 | gfs2_ail1_empty(sdp, 0); |
309 | gfs2_log_flush(sdp, NULL); | 307 | gfs2_log_flush(sdp, NULL); |
@@ -312,7 +310,7 @@ int gfs2_log_reserve(struct gfs2_sbd *sdp, unsigned int blks) | |||
312 | gfs2_ail1_start(sdp, 0); | 310 | gfs2_ail1_start(sdp, 0); |
313 | gfs2_log_lock(sdp); | 311 | gfs2_log_lock(sdp); |
314 | } | 312 | } |
315 | sdp->sd_log_blks_free -= blks; | 313 | atomic_sub(blks, &sdp->sd_log_blks_free); |
316 | gfs2_log_unlock(sdp); | 314 | gfs2_log_unlock(sdp); |
317 | mutex_unlock(&sdp->sd_log_reserve_mutex); | 315 | mutex_unlock(&sdp->sd_log_reserve_mutex); |
318 | 316 | ||
@@ -332,27 +330,23 @@ void gfs2_log_release(struct gfs2_sbd *sdp, unsigned int blks) | |||
332 | { | 330 | { |
333 | 331 | ||
334 | gfs2_log_lock(sdp); | 332 | gfs2_log_lock(sdp); |
335 | sdp->sd_log_blks_free += blks; | 333 | atomic_add(blks, &sdp->sd_log_blks_free); |
336 | gfs2_assert_withdraw(sdp, | 334 | gfs2_assert_withdraw(sdp, |
337 | sdp->sd_log_blks_free <= sdp->sd_jdesc->jd_blocks); | 335 | atomic_read(&sdp->sd_log_blks_free) <= sdp->sd_jdesc->jd_blocks); |
338 | gfs2_log_unlock(sdp); | 336 | gfs2_log_unlock(sdp); |
339 | up_read(&sdp->sd_log_flush_lock); | 337 | up_read(&sdp->sd_log_flush_lock); |
340 | } | 338 | } |
341 | 339 | ||
342 | static u64 log_bmap(struct gfs2_sbd *sdp, unsigned int lbn) | 340 | static u64 log_bmap(struct gfs2_sbd *sdp, unsigned int lbn) |
343 | { | 341 | { |
344 | struct inode *inode = sdp->sd_jdesc->jd_inode; | 342 | struct gfs2_journal_extent *je; |
345 | int error; | 343 | |
346 | struct buffer_head bh_map = { .b_state = 0, .b_blocknr = 0 }; | 344 | list_for_each_entry(je, &sdp->sd_jdesc->extent_list, extent_list) { |
347 | 345 | if (lbn >= je->lblock && lbn < je->lblock + je->blocks) | |
348 | bh_map.b_size = 1 << inode->i_blkbits; | 346 | return je->dblock + lbn - je->lblock; |
349 | error = gfs2_block_map(inode, lbn, 0, &bh_map); | 347 | } |
350 | if (error || !bh_map.b_blocknr) | 348 | |
351 | printk(KERN_INFO "error=%d, dbn=%llu lbn=%u", error, | 349 | return -1; |
352 | (unsigned long long)bh_map.b_blocknr, lbn); | ||
353 | gfs2_assert_withdraw(sdp, !error && bh_map.b_blocknr); | ||
354 | |||
355 | return bh_map.b_blocknr; | ||
356 | } | 350 | } |
357 | 351 | ||
358 | /** | 352 | /** |
@@ -561,8 +555,8 @@ static void log_pull_tail(struct gfs2_sbd *sdp, unsigned int new_tail) | |||
561 | ail2_empty(sdp, new_tail); | 555 | ail2_empty(sdp, new_tail); |
562 | 556 | ||
563 | gfs2_log_lock(sdp); | 557 | gfs2_log_lock(sdp); |
564 | sdp->sd_log_blks_free += dist; | 558 | atomic_add(dist, &sdp->sd_log_blks_free); |
565 | gfs2_assert_withdraw(sdp, sdp->sd_log_blks_free <= sdp->sd_jdesc->jd_blocks); | 559 | gfs2_assert_withdraw(sdp, atomic_read(&sdp->sd_log_blks_free) <= sdp->sd_jdesc->jd_blocks); |
566 | gfs2_log_unlock(sdp); | 560 | gfs2_log_unlock(sdp); |
567 | 561 | ||
568 | sdp->sd_log_tail = new_tail; | 562 | sdp->sd_log_tail = new_tail; |
@@ -652,7 +646,7 @@ static void gfs2_ordered_write(struct gfs2_sbd *sdp) | |||
652 | get_bh(bh); | 646 | get_bh(bh); |
653 | gfs2_log_unlock(sdp); | 647 | gfs2_log_unlock(sdp); |
654 | lock_buffer(bh); | 648 | lock_buffer(bh); |
655 | if (test_clear_buffer_dirty(bh)) { | 649 | if (buffer_mapped(bh) && test_clear_buffer_dirty(bh)) { |
656 | bh->b_end_io = end_buffer_write_sync; | 650 | bh->b_end_io = end_buffer_write_sync; |
657 | submit_bh(WRITE, bh); | 651 | submit_bh(WRITE, bh); |
658 | } else { | 652 | } else { |
@@ -694,20 +688,16 @@ static void gfs2_ordered_wait(struct gfs2_sbd *sdp) | |||
694 | * | 688 | * |
695 | */ | 689 | */ |
696 | 690 | ||
697 | void gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl) | 691 | void __gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl) |
698 | { | 692 | { |
699 | struct gfs2_ail *ai; | 693 | struct gfs2_ail *ai; |
700 | 694 | ||
701 | down_write(&sdp->sd_log_flush_lock); | 695 | down_write(&sdp->sd_log_flush_lock); |
702 | 696 | ||
703 | if (gl) { | 697 | /* Log might have been flushed while we waited for the flush lock */ |
704 | gfs2_log_lock(sdp); | 698 | if (gl && !test_bit(GLF_LFLUSH, &gl->gl_flags)) { |
705 | if (list_empty(&gl->gl_le.le_list)) { | 699 | up_write(&sdp->sd_log_flush_lock); |
706 | gfs2_log_unlock(sdp); | 700 | return; |
707 | up_write(&sdp->sd_log_flush_lock); | ||
708 | return; | ||
709 | } | ||
710 | gfs2_log_unlock(sdp); | ||
711 | } | 701 | } |
712 | 702 | ||
713 | ai = kzalloc(sizeof(struct gfs2_ail), GFP_NOFS | __GFP_NOFAIL); | 703 | ai = kzalloc(sizeof(struct gfs2_ail), GFP_NOFS | __GFP_NOFAIL); |
@@ -739,7 +729,7 @@ void gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl) | |||
739 | log_flush_commit(sdp); | 729 | log_flush_commit(sdp); |
740 | else if (sdp->sd_log_tail != current_tail(sdp) && !sdp->sd_log_idle){ | 730 | else if (sdp->sd_log_tail != current_tail(sdp) && !sdp->sd_log_idle){ |
741 | gfs2_log_lock(sdp); | 731 | gfs2_log_lock(sdp); |
742 | sdp->sd_log_blks_free--; /* Adjust for unreserved buffer */ | 732 | atomic_dec(&sdp->sd_log_blks_free); /* Adjust for unreserved buffer */ |
743 | gfs2_log_unlock(sdp); | 733 | gfs2_log_unlock(sdp); |
744 | log_write_header(sdp, 0, PULL); | 734 | log_write_header(sdp, 0, PULL); |
745 | } | 735 | } |
@@ -767,7 +757,7 @@ void gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl) | |||
767 | static void log_refund(struct gfs2_sbd *sdp, struct gfs2_trans *tr) | 757 | static void log_refund(struct gfs2_sbd *sdp, struct gfs2_trans *tr) |
768 | { | 758 | { |
769 | unsigned int reserved; | 759 | unsigned int reserved; |
770 | unsigned int old; | 760 | unsigned int unused; |
771 | 761 | ||
772 | gfs2_log_lock(sdp); | 762 | gfs2_log_lock(sdp); |
773 | 763 | ||
@@ -779,14 +769,11 @@ static void log_refund(struct gfs2_sbd *sdp, struct gfs2_trans *tr) | |||
779 | sdp->sd_log_commited_revoke += tr->tr_num_revoke - tr->tr_num_revoke_rm; | 769 | sdp->sd_log_commited_revoke += tr->tr_num_revoke - tr->tr_num_revoke_rm; |
780 | gfs2_assert_withdraw(sdp, ((int)sdp->sd_log_commited_revoke) >= 0); | 770 | gfs2_assert_withdraw(sdp, ((int)sdp->sd_log_commited_revoke) >= 0); |
781 | reserved = calc_reserved(sdp); | 771 | reserved = calc_reserved(sdp); |
782 | old = sdp->sd_log_blks_free; | 772 | unused = sdp->sd_log_blks_reserved - reserved + tr->tr_reserved; |
783 | sdp->sd_log_blks_free += tr->tr_reserved - | 773 | gfs2_assert_withdraw(sdp, unused >= 0); |
784 | (reserved - sdp->sd_log_blks_reserved); | 774 | atomic_add(unused, &sdp->sd_log_blks_free); |
785 | 775 | gfs2_assert_withdraw(sdp, atomic_read(&sdp->sd_log_blks_free) <= | |
786 | gfs2_assert_withdraw(sdp, sdp->sd_log_blks_free >= old); | ||
787 | gfs2_assert_withdraw(sdp, sdp->sd_log_blks_free <= | ||
788 | sdp->sd_jdesc->jd_blocks); | 776 | sdp->sd_jdesc->jd_blocks); |
789 | |||
790 | sdp->sd_log_blks_reserved = reserved; | 777 | sdp->sd_log_blks_reserved = reserved; |
791 | 778 | ||
792 | gfs2_log_unlock(sdp); | 779 | gfs2_log_unlock(sdp); |
@@ -825,7 +812,6 @@ void gfs2_log_shutdown(struct gfs2_sbd *sdp) | |||
825 | down_write(&sdp->sd_log_flush_lock); | 812 | down_write(&sdp->sd_log_flush_lock); |
826 | 813 | ||
827 | gfs2_assert_withdraw(sdp, !sdp->sd_log_blks_reserved); | 814 | gfs2_assert_withdraw(sdp, !sdp->sd_log_blks_reserved); |
828 | gfs2_assert_withdraw(sdp, !sdp->sd_log_num_gl); | ||
829 | gfs2_assert_withdraw(sdp, !sdp->sd_log_num_buf); | 815 | gfs2_assert_withdraw(sdp, !sdp->sd_log_num_buf); |
830 | gfs2_assert_withdraw(sdp, !sdp->sd_log_num_revoke); | 816 | gfs2_assert_withdraw(sdp, !sdp->sd_log_num_revoke); |
831 | gfs2_assert_withdraw(sdp, !sdp->sd_log_num_rg); | 817 | gfs2_assert_withdraw(sdp, !sdp->sd_log_num_rg); |
@@ -838,7 +824,7 @@ void gfs2_log_shutdown(struct gfs2_sbd *sdp) | |||
838 | log_write_header(sdp, GFS2_LOG_HEAD_UNMOUNT, | 824 | log_write_header(sdp, GFS2_LOG_HEAD_UNMOUNT, |
839 | (sdp->sd_log_tail == current_tail(sdp)) ? 0 : PULL); | 825 | (sdp->sd_log_tail == current_tail(sdp)) ? 0 : PULL); |
840 | 826 | ||
841 | gfs2_assert_warn(sdp, sdp->sd_log_blks_free == sdp->sd_jdesc->jd_blocks); | 827 | gfs2_assert_warn(sdp, atomic_read(&sdp->sd_log_blks_free) == sdp->sd_jdesc->jd_blocks); |
842 | gfs2_assert_warn(sdp, sdp->sd_log_head == sdp->sd_log_tail); | 828 | gfs2_assert_warn(sdp, sdp->sd_log_head == sdp->sd_log_tail); |
843 | gfs2_assert_warn(sdp, list_empty(&sdp->sd_ail2_list)); | 829 | gfs2_assert_warn(sdp, list_empty(&sdp->sd_ail2_list)); |
844 | 830 | ||
@@ -866,3 +852,42 @@ void gfs2_meta_syncfs(struct gfs2_sbd *sdp) | |||
866 | } | 852 | } |
867 | } | 853 | } |
868 | 854 | ||
855 | |||
856 | /** | ||
857 | * gfs2_logd - Update log tail as Active Items get flushed to in-place blocks | ||
858 | * @sdp: Pointer to GFS2 superblock | ||
859 | * | ||
860 | * Also, periodically check to make sure that we're using the most recent | ||
861 | * journal index. | ||
862 | */ | ||
863 | |||
864 | int gfs2_logd(void *data) | ||
865 | { | ||
866 | struct gfs2_sbd *sdp = data; | ||
867 | unsigned long t; | ||
868 | int need_flush; | ||
869 | |||
870 | while (!kthread_should_stop()) { | ||
871 | /* Advance the log tail */ | ||
872 | |||
873 | t = sdp->sd_log_flush_time + | ||
874 | gfs2_tune_get(sdp, gt_log_flush_secs) * HZ; | ||
875 | |||
876 | gfs2_ail1_empty(sdp, DIO_ALL); | ||
877 | gfs2_log_lock(sdp); | ||
878 | need_flush = sdp->sd_log_num_buf > gfs2_tune_get(sdp, gt_incore_log_blocks); | ||
879 | gfs2_log_unlock(sdp); | ||
880 | if (need_flush || time_after_eq(jiffies, t)) { | ||
881 | gfs2_log_flush(sdp, NULL); | ||
882 | sdp->sd_log_flush_time = jiffies; | ||
883 | } | ||
884 | |||
885 | t = gfs2_tune_get(sdp, gt_logd_secs) * HZ; | ||
886 | if (freezing(current)) | ||
887 | refrigerator(); | ||
888 | schedule_timeout_interruptible(t); | ||
889 | } | ||
890 | |||
891 | return 0; | ||
892 | } | ||
893 | |||
diff --git a/fs/gfs2/log.h b/fs/gfs2/log.h index dae28240062..77115281650 100644 --- a/fs/gfs2/log.h +++ b/fs/gfs2/log.h | |||
@@ -48,8 +48,6 @@ static inline void gfs2_log_pointers_init(struct gfs2_sbd *sdp, | |||
48 | unsigned int gfs2_struct2blk(struct gfs2_sbd *sdp, unsigned int nstruct, | 48 | unsigned int gfs2_struct2blk(struct gfs2_sbd *sdp, unsigned int nstruct, |
49 | unsigned int ssize); | 49 | unsigned int ssize); |
50 | 50 | ||
51 | int gfs2_ail1_empty(struct gfs2_sbd *sdp, int flags); | ||
52 | |||
53 | int gfs2_log_reserve(struct gfs2_sbd *sdp, unsigned int blks); | 51 | int gfs2_log_reserve(struct gfs2_sbd *sdp, unsigned int blks); |
54 | void gfs2_log_release(struct gfs2_sbd *sdp, unsigned int blks); | 52 | void gfs2_log_release(struct gfs2_sbd *sdp, unsigned int blks); |
55 | void gfs2_log_incr_head(struct gfs2_sbd *sdp); | 53 | void gfs2_log_incr_head(struct gfs2_sbd *sdp); |
@@ -57,11 +55,19 @@ void gfs2_log_incr_head(struct gfs2_sbd *sdp); | |||
57 | struct buffer_head *gfs2_log_get_buf(struct gfs2_sbd *sdp); | 55 | struct buffer_head *gfs2_log_get_buf(struct gfs2_sbd *sdp); |
58 | struct buffer_head *gfs2_log_fake_buf(struct gfs2_sbd *sdp, | 56 | struct buffer_head *gfs2_log_fake_buf(struct gfs2_sbd *sdp, |
59 | struct buffer_head *real); | 57 | struct buffer_head *real); |
60 | void gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl); | 58 | void __gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl); |
59 | |||
60 | static inline void gfs2_log_flush(struct gfs2_sbd *sbd, struct gfs2_glock *gl) | ||
61 | { | ||
62 | if (!gl || test_bit(GLF_LFLUSH, &gl->gl_flags)) | ||
63 | __gfs2_log_flush(sbd, gl); | ||
64 | } | ||
65 | |||
61 | void gfs2_log_commit(struct gfs2_sbd *sdp, struct gfs2_trans *trans); | 66 | void gfs2_log_commit(struct gfs2_sbd *sdp, struct gfs2_trans *trans); |
62 | void gfs2_remove_from_ail(struct address_space *mapping, struct gfs2_bufdata *bd); | 67 | void gfs2_remove_from_ail(struct gfs2_bufdata *bd); |
63 | 68 | ||
64 | void gfs2_log_shutdown(struct gfs2_sbd *sdp); | 69 | void gfs2_log_shutdown(struct gfs2_sbd *sdp); |
65 | void gfs2_meta_syncfs(struct gfs2_sbd *sdp); | 70 | void gfs2_meta_syncfs(struct gfs2_sbd *sdp); |
71 | int gfs2_logd(void *data); | ||
66 | 72 | ||
67 | #endif /* __LOG_DOT_H__ */ | 73 | #endif /* __LOG_DOT_H__ */ |
diff --git a/fs/gfs2/lops.c b/fs/gfs2/lops.c index 6c27cea761c..fae59d69d01 100644 --- a/fs/gfs2/lops.c +++ b/fs/gfs2/lops.c | |||
@@ -87,6 +87,7 @@ static void gfs2_unpin(struct gfs2_sbd *sdp, struct buffer_head *bh, | |||
87 | } | 87 | } |
88 | bd->bd_ail = ai; | 88 | bd->bd_ail = ai; |
89 | list_add(&bd->bd_ail_st_list, &ai->ai_ail1_list); | 89 | list_add(&bd->bd_ail_st_list, &ai->ai_ail1_list); |
90 | clear_bit(GLF_LFLUSH, &bd->bd_gl->gl_flags); | ||
90 | gfs2_log_unlock(sdp); | 91 | gfs2_log_unlock(sdp); |
91 | unlock_buffer(bh); | 92 | unlock_buffer(bh); |
92 | } | 93 | } |
@@ -124,49 +125,6 @@ static struct buffer_head *gfs2_get_log_desc(struct gfs2_sbd *sdp, u32 ld_type) | |||
124 | return bh; | 125 | return bh; |
125 | } | 126 | } |
126 | 127 | ||
127 | static void __glock_lo_add(struct gfs2_sbd *sdp, struct gfs2_log_element *le) | ||
128 | { | ||
129 | struct gfs2_glock *gl; | ||
130 | struct gfs2_trans *tr = current->journal_info; | ||
131 | |||
132 | tr->tr_touched = 1; | ||
133 | |||
134 | gl = container_of(le, struct gfs2_glock, gl_le); | ||
135 | if (gfs2_assert_withdraw(sdp, gfs2_glock_is_held_excl(gl))) | ||
136 | return; | ||
137 | |||
138 | if (!list_empty(&le->le_list)) | ||
139 | return; | ||
140 | |||
141 | gfs2_glock_hold(gl); | ||
142 | set_bit(GLF_DIRTY, &gl->gl_flags); | ||
143 | sdp->sd_log_num_gl++; | ||
144 | list_add(&le->le_list, &sdp->sd_log_le_gl); | ||
145 | } | ||
146 | |||
147 | static void glock_lo_add(struct gfs2_sbd *sdp, struct gfs2_log_element *le) | ||
148 | { | ||
149 | gfs2_log_lock(sdp); | ||
150 | __glock_lo_add(sdp, le); | ||
151 | gfs2_log_unlock(sdp); | ||
152 | } | ||
153 | |||
154 | static void glock_lo_after_commit(struct gfs2_sbd *sdp, struct gfs2_ail *ai) | ||
155 | { | ||
156 | struct list_head *head = &sdp->sd_log_le_gl; | ||
157 | struct gfs2_glock *gl; | ||
158 | |||
159 | while (!list_empty(head)) { | ||
160 | gl = list_entry(head->next, struct gfs2_glock, gl_le.le_list); | ||
161 | list_del_init(&gl->gl_le.le_list); | ||
162 | sdp->sd_log_num_gl--; | ||
163 | |||
164 | gfs2_assert_withdraw(sdp, gfs2_glock_is_held_excl(gl)); | ||
165 | gfs2_glock_put(gl); | ||
166 | } | ||
167 | gfs2_assert_warn(sdp, !sdp->sd_log_num_gl); | ||
168 | } | ||
169 | |||
170 | static void buf_lo_add(struct gfs2_sbd *sdp, struct gfs2_log_element *le) | 128 | static void buf_lo_add(struct gfs2_sbd *sdp, struct gfs2_log_element *le) |
171 | { | 129 | { |
172 | struct gfs2_bufdata *bd = container_of(le, struct gfs2_bufdata, bd_le); | 130 | struct gfs2_bufdata *bd = container_of(le, struct gfs2_bufdata, bd_le); |
@@ -182,7 +140,8 @@ static void buf_lo_add(struct gfs2_sbd *sdp, struct gfs2_log_element *le) | |||
182 | list_add(&bd->bd_list_tr, &tr->tr_list_buf); | 140 | list_add(&bd->bd_list_tr, &tr->tr_list_buf); |
183 | if (!list_empty(&le->le_list)) | 141 | if (!list_empty(&le->le_list)) |
184 | goto out; | 142 | goto out; |
185 | __glock_lo_add(sdp, &bd->bd_gl->gl_le); | 143 | set_bit(GLF_LFLUSH, &bd->bd_gl->gl_flags); |
144 | set_bit(GLF_DIRTY, &bd->bd_gl->gl_flags); | ||
186 | gfs2_meta_check(sdp, bd->bd_bh); | 145 | gfs2_meta_check(sdp, bd->bd_bh); |
187 | gfs2_pin(sdp, bd->bd_bh); | 146 | gfs2_pin(sdp, bd->bd_bh); |
188 | sdp->sd_log_num_buf++; | 147 | sdp->sd_log_num_buf++; |
@@ -556,17 +515,20 @@ static void databuf_lo_add(struct gfs2_sbd *sdp, struct gfs2_log_element *le) | |||
556 | 515 | ||
557 | lock_buffer(bd->bd_bh); | 516 | lock_buffer(bd->bd_bh); |
558 | gfs2_log_lock(sdp); | 517 | gfs2_log_lock(sdp); |
559 | if (!list_empty(&bd->bd_list_tr)) | 518 | if (tr) { |
560 | goto out; | 519 | if (!list_empty(&bd->bd_list_tr)) |
561 | tr->tr_touched = 1; | 520 | goto out; |
562 | if (gfs2_is_jdata(ip)) { | 521 | tr->tr_touched = 1; |
563 | tr->tr_num_buf++; | 522 | if (gfs2_is_jdata(ip)) { |
564 | list_add(&bd->bd_list_tr, &tr->tr_list_buf); | 523 | tr->tr_num_buf++; |
524 | list_add(&bd->bd_list_tr, &tr->tr_list_buf); | ||
525 | } | ||
565 | } | 526 | } |
566 | if (!list_empty(&le->le_list)) | 527 | if (!list_empty(&le->le_list)) |
567 | goto out; | 528 | goto out; |
568 | 529 | ||
569 | __glock_lo_add(sdp, &bd->bd_gl->gl_le); | 530 | set_bit(GLF_LFLUSH, &bd->bd_gl->gl_flags); |
531 | set_bit(GLF_DIRTY, &bd->bd_gl->gl_flags); | ||
570 | if (gfs2_is_jdata(ip)) { | 532 | if (gfs2_is_jdata(ip)) { |
571 | gfs2_pin(sdp, bd->bd_bh); | 533 | gfs2_pin(sdp, bd->bd_bh); |
572 | tr->tr_num_databuf_new++; | 534 | tr->tr_num_databuf_new++; |
@@ -773,12 +735,6 @@ static void databuf_lo_after_commit(struct gfs2_sbd *sdp, struct gfs2_ail *ai) | |||
773 | } | 735 | } |
774 | 736 | ||
775 | 737 | ||
776 | const struct gfs2_log_operations gfs2_glock_lops = { | ||
777 | .lo_add = glock_lo_add, | ||
778 | .lo_after_commit = glock_lo_after_commit, | ||
779 | .lo_name = "glock", | ||
780 | }; | ||
781 | |||
782 | const struct gfs2_log_operations gfs2_buf_lops = { | 738 | const struct gfs2_log_operations gfs2_buf_lops = { |
783 | .lo_add = buf_lo_add, | 739 | .lo_add = buf_lo_add, |
784 | .lo_incore_commit = buf_lo_incore_commit, | 740 | .lo_incore_commit = buf_lo_incore_commit, |
@@ -816,7 +772,6 @@ const struct gfs2_log_operations gfs2_databuf_lops = { | |||
816 | }; | 772 | }; |
817 | 773 | ||
818 | const struct gfs2_log_operations *gfs2_log_ops[] = { | 774 | const struct gfs2_log_operations *gfs2_log_ops[] = { |
819 | &gfs2_glock_lops, | ||
820 | &gfs2_databuf_lops, | 775 | &gfs2_databuf_lops, |
821 | &gfs2_buf_lops, | 776 | &gfs2_buf_lops, |
822 | &gfs2_rg_lops, | 777 | &gfs2_rg_lops, |
diff --git a/fs/gfs2/main.c b/fs/gfs2/main.c index 7ecfe0d3a49..9c7765c12d6 100644 --- a/fs/gfs2/main.c +++ b/fs/gfs2/main.c | |||
@@ -29,9 +29,8 @@ static void gfs2_init_inode_once(struct kmem_cache *cachep, void *foo) | |||
29 | struct gfs2_inode *ip = foo; | 29 | struct gfs2_inode *ip = foo; |
30 | 30 | ||
31 | inode_init_once(&ip->i_inode); | 31 | inode_init_once(&ip->i_inode); |
32 | spin_lock_init(&ip->i_spin); | ||
33 | init_rwsem(&ip->i_rw_mutex); | 32 | init_rwsem(&ip->i_rw_mutex); |
34 | memset(ip->i_cache, 0, sizeof(ip->i_cache)); | 33 | ip->i_alloc = NULL; |
35 | } | 34 | } |
36 | 35 | ||
37 | static void gfs2_init_glock_once(struct kmem_cache *cachep, void *foo) | 36 | static void gfs2_init_glock_once(struct kmem_cache *cachep, void *foo) |
diff --git a/fs/gfs2/meta_io.c b/fs/gfs2/meta_io.c index 4da423985e4..85aea27b4a8 100644 --- a/fs/gfs2/meta_io.c +++ b/fs/gfs2/meta_io.c | |||
@@ -50,6 +50,7 @@ static int gfs2_aspace_writepage(struct page *page, | |||
50 | static const struct address_space_operations aspace_aops = { | 50 | static const struct address_space_operations aspace_aops = { |
51 | .writepage = gfs2_aspace_writepage, | 51 | .writepage = gfs2_aspace_writepage, |
52 | .releasepage = gfs2_releasepage, | 52 | .releasepage = gfs2_releasepage, |
53 | .sync_page = block_sync_page, | ||
53 | }; | 54 | }; |
54 | 55 | ||
55 | /** | 56 | /** |
@@ -221,13 +222,14 @@ int gfs2_meta_read(struct gfs2_glock *gl, u64 blkno, int flags, | |||
221 | struct buffer_head **bhp) | 222 | struct buffer_head **bhp) |
222 | { | 223 | { |
223 | *bhp = getbuf(gl, blkno, CREATE); | 224 | *bhp = getbuf(gl, blkno, CREATE); |
224 | if (!buffer_uptodate(*bhp)) | 225 | if (!buffer_uptodate(*bhp)) { |
225 | ll_rw_block(READ_META, 1, bhp); | 226 | ll_rw_block(READ_META, 1, bhp); |
226 | if (flags & DIO_WAIT) { | 227 | if (flags & DIO_WAIT) { |
227 | int error = gfs2_meta_wait(gl->gl_sbd, *bhp); | 228 | int error = gfs2_meta_wait(gl->gl_sbd, *bhp); |
228 | if (error) { | 229 | if (error) { |
229 | brelse(*bhp); | 230 | brelse(*bhp); |
230 | return error; | 231 | return error; |
232 | } | ||
231 | } | 233 | } |
232 | } | 234 | } |
233 | 235 | ||
@@ -282,7 +284,7 @@ void gfs2_attach_bufdata(struct gfs2_glock *gl, struct buffer_head *bh, | |||
282 | return; | 284 | return; |
283 | } | 285 | } |
284 | 286 | ||
285 | bd = kmem_cache_zalloc(gfs2_bufdata_cachep, GFP_NOFS | __GFP_NOFAIL), | 287 | bd = kmem_cache_zalloc(gfs2_bufdata_cachep, GFP_NOFS | __GFP_NOFAIL); |
286 | bd->bd_bh = bh; | 288 | bd->bd_bh = bh; |
287 | bd->bd_gl = gl; | 289 | bd->bd_gl = gl; |
288 | 290 | ||
@@ -317,7 +319,7 @@ void gfs2_remove_from_journal(struct buffer_head *bh, struct gfs2_trans *tr, int | |||
317 | } | 319 | } |
318 | if (bd) { | 320 | if (bd) { |
319 | if (bd->bd_ail) { | 321 | if (bd->bd_ail) { |
320 | gfs2_remove_from_ail(NULL, bd); | 322 | gfs2_remove_from_ail(bd); |
321 | bh->b_private = NULL; | 323 | bh->b_private = NULL; |
322 | bd->bd_bh = NULL; | 324 | bd->bd_bh = NULL; |
323 | bd->bd_blkno = bh->b_blocknr; | 325 | bd->bd_blkno = bh->b_blocknr; |
@@ -358,32 +360,6 @@ void gfs2_meta_wipe(struct gfs2_inode *ip, u64 bstart, u32 blen) | |||
358 | } | 360 | } |
359 | 361 | ||
360 | /** | 362 | /** |
361 | * gfs2_meta_cache_flush - get rid of any references on buffers for this inode | ||
362 | * @ip: The GFS2 inode | ||
363 | * | ||
364 | * This releases buffers that are in the most-recently-used array of | ||
365 | * blocks used for indirect block addressing for this inode. | ||
366 | */ | ||
367 | |||
368 | void gfs2_meta_cache_flush(struct gfs2_inode *ip) | ||
369 | { | ||
370 | struct buffer_head **bh_slot; | ||
371 | unsigned int x; | ||
372 | |||
373 | spin_lock(&ip->i_spin); | ||
374 | |||
375 | for (x = 0; x < GFS2_MAX_META_HEIGHT; x++) { | ||
376 | bh_slot = &ip->i_cache[x]; | ||
377 | if (*bh_slot) { | ||
378 | brelse(*bh_slot); | ||
379 | *bh_slot = NULL; | ||
380 | } | ||
381 | } | ||
382 | |||
383 | spin_unlock(&ip->i_spin); | ||
384 | } | ||
385 | |||
386 | /** | ||
387 | * gfs2_meta_indirect_buffer - Get a metadata buffer | 363 | * gfs2_meta_indirect_buffer - Get a metadata buffer |
388 | * @ip: The GFS2 inode | 364 | * @ip: The GFS2 inode |
389 | * @height: The level of this buf in the metadata (indir addr) tree (if any) | 365 | * @height: The level of this buf in the metadata (indir addr) tree (if any) |
@@ -391,8 +367,6 @@ void gfs2_meta_cache_flush(struct gfs2_inode *ip) | |||
391 | * @new: Non-zero if we may create a new buffer | 367 | * @new: Non-zero if we may create a new buffer |
392 | * @bhp: the buffer is returned here | 368 | * @bhp: the buffer is returned here |
393 | * | 369 | * |
394 | * Try to use the gfs2_inode's MRU metadata tree cache. | ||
395 | * | ||
396 | * Returns: errno | 370 | * Returns: errno |
397 | */ | 371 | */ |
398 | 372 | ||
@@ -401,58 +375,25 @@ int gfs2_meta_indirect_buffer(struct gfs2_inode *ip, int height, u64 num, | |||
401 | { | 375 | { |
402 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 376 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
403 | struct gfs2_glock *gl = ip->i_gl; | 377 | struct gfs2_glock *gl = ip->i_gl; |
404 | struct buffer_head *bh = NULL, **bh_slot = ip->i_cache + height; | 378 | struct buffer_head *bh; |
405 | int in_cache = 0; | 379 | int ret = 0; |
406 | |||
407 | BUG_ON(!gl); | ||
408 | BUG_ON(!sdp); | ||
409 | |||
410 | spin_lock(&ip->i_spin); | ||
411 | if (*bh_slot && (*bh_slot)->b_blocknr == num) { | ||
412 | bh = *bh_slot; | ||
413 | get_bh(bh); | ||
414 | in_cache = 1; | ||
415 | } | ||
416 | spin_unlock(&ip->i_spin); | ||
417 | |||
418 | if (!bh) | ||
419 | bh = getbuf(gl, num, CREATE); | ||
420 | |||
421 | if (!bh) | ||
422 | return -ENOBUFS; | ||
423 | 380 | ||
424 | if (new) { | 381 | if (new) { |
425 | if (gfs2_assert_warn(sdp, height)) | 382 | BUG_ON(height == 0); |
426 | goto err; | 383 | bh = gfs2_meta_new(gl, num); |
427 | meta_prep_new(bh); | ||
428 | gfs2_trans_add_bh(ip->i_gl, bh, 1); | 384 | gfs2_trans_add_bh(ip->i_gl, bh, 1); |
429 | gfs2_metatype_set(bh, GFS2_METATYPE_IN, GFS2_FORMAT_IN); | 385 | gfs2_metatype_set(bh, GFS2_METATYPE_IN, GFS2_FORMAT_IN); |
430 | gfs2_buffer_clear_tail(bh, sizeof(struct gfs2_meta_header)); | 386 | gfs2_buffer_clear_tail(bh, sizeof(struct gfs2_meta_header)); |
431 | } else { | 387 | } else { |
432 | u32 mtype = height ? GFS2_METATYPE_IN : GFS2_METATYPE_DI; | 388 | u32 mtype = height ? GFS2_METATYPE_IN : GFS2_METATYPE_DI; |
433 | if (!buffer_uptodate(bh)) { | 389 | ret = gfs2_meta_read(gl, num, DIO_WAIT, &bh); |
434 | ll_rw_block(READ_META, 1, &bh); | 390 | if (ret == 0 && gfs2_metatype_check(sdp, bh, mtype)) { |
435 | if (gfs2_meta_wait(sdp, bh)) | 391 | brelse(bh); |
436 | goto err; | 392 | ret = -EIO; |
437 | } | 393 | } |
438 | if (gfs2_metatype_check(sdp, bh, mtype)) | ||
439 | goto err; | ||
440 | } | ||
441 | |||
442 | if (!in_cache) { | ||
443 | spin_lock(&ip->i_spin); | ||
444 | if (*bh_slot) | ||
445 | brelse(*bh_slot); | ||
446 | *bh_slot = bh; | ||
447 | get_bh(bh); | ||
448 | spin_unlock(&ip->i_spin); | ||
449 | } | 394 | } |
450 | |||
451 | *bhp = bh; | 395 | *bhp = bh; |
452 | return 0; | 396 | return ret; |
453 | err: | ||
454 | brelse(bh); | ||
455 | return -EIO; | ||
456 | } | 397 | } |
457 | 398 | ||
458 | /** | 399 | /** |
diff --git a/fs/gfs2/meta_io.h b/fs/gfs2/meta_io.h index b7048222ebb..73e3b1c76fe 100644 --- a/fs/gfs2/meta_io.h +++ b/fs/gfs2/meta_io.h | |||
@@ -56,7 +56,6 @@ void gfs2_remove_from_journal(struct buffer_head *bh, struct gfs2_trans *tr, | |||
56 | 56 | ||
57 | void gfs2_meta_wipe(struct gfs2_inode *ip, u64 bstart, u32 blen); | 57 | void gfs2_meta_wipe(struct gfs2_inode *ip, u64 bstart, u32 blen); |
58 | 58 | ||
59 | void gfs2_meta_cache_flush(struct gfs2_inode *ip); | ||
60 | int gfs2_meta_indirect_buffer(struct gfs2_inode *ip, int height, u64 num, | 59 | int gfs2_meta_indirect_buffer(struct gfs2_inode *ip, int height, u64 num, |
61 | int new, struct buffer_head **bhp); | 60 | int new, struct buffer_head **bhp); |
62 | 61 | ||
diff --git a/fs/gfs2/ops_address.c b/fs/gfs2/ops_address.c index 9679f8b9870..38dbe99a30e 100644 --- a/fs/gfs2/ops_address.c +++ b/fs/gfs2/ops_address.c | |||
@@ -20,6 +20,8 @@ | |||
20 | #include <linux/swap.h> | 20 | #include <linux/swap.h> |
21 | #include <linux/gfs2_ondisk.h> | 21 | #include <linux/gfs2_ondisk.h> |
22 | #include <linux/lm_interface.h> | 22 | #include <linux/lm_interface.h> |
23 | #include <linux/backing-dev.h> | ||
24 | #include <linux/pagevec.h> | ||
23 | 25 | ||
24 | #include "gfs2.h" | 26 | #include "gfs2.h" |
25 | #include "incore.h" | 27 | #include "incore.h" |
@@ -32,7 +34,6 @@ | |||
32 | #include "quota.h" | 34 | #include "quota.h" |
33 | #include "trans.h" | 35 | #include "trans.h" |
34 | #include "rgrp.h" | 36 | #include "rgrp.h" |
35 | #include "ops_file.h" | ||
36 | #include "super.h" | 37 | #include "super.h" |
37 | #include "util.h" | 38 | #include "util.h" |
38 | #include "glops.h" | 39 | #include "glops.h" |
@@ -58,22 +59,6 @@ static void gfs2_page_add_databufs(struct gfs2_inode *ip, struct page *page, | |||
58 | } | 59 | } |
59 | 60 | ||
60 | /** | 61 | /** |
61 | * gfs2_get_block - Fills in a buffer head with details about a block | ||
62 | * @inode: The inode | ||
63 | * @lblock: The block number to look up | ||
64 | * @bh_result: The buffer head to return the result in | ||
65 | * @create: Non-zero if we may add block to the file | ||
66 | * | ||
67 | * Returns: errno | ||
68 | */ | ||
69 | |||
70 | int gfs2_get_block(struct inode *inode, sector_t lblock, | ||
71 | struct buffer_head *bh_result, int create) | ||
72 | { | ||
73 | return gfs2_block_map(inode, lblock, create, bh_result); | ||
74 | } | ||
75 | |||
76 | /** | ||
77 | * gfs2_get_block_noalloc - Fills in a buffer head with details about a block | 62 | * gfs2_get_block_noalloc - Fills in a buffer head with details about a block |
78 | * @inode: The inode | 63 | * @inode: The inode |
79 | * @lblock: The block number to look up | 64 | * @lblock: The block number to look up |
@@ -88,7 +73,7 @@ static int gfs2_get_block_noalloc(struct inode *inode, sector_t lblock, | |||
88 | { | 73 | { |
89 | int error; | 74 | int error; |
90 | 75 | ||
91 | error = gfs2_block_map(inode, lblock, 0, bh_result); | 76 | error = gfs2_block_map(inode, lblock, bh_result, 0); |
92 | if (error) | 77 | if (error) |
93 | return error; | 78 | return error; |
94 | if (!buffer_mapped(bh_result)) | 79 | if (!buffer_mapped(bh_result)) |
@@ -99,20 +84,19 @@ static int gfs2_get_block_noalloc(struct inode *inode, sector_t lblock, | |||
99 | static int gfs2_get_block_direct(struct inode *inode, sector_t lblock, | 84 | static int gfs2_get_block_direct(struct inode *inode, sector_t lblock, |
100 | struct buffer_head *bh_result, int create) | 85 | struct buffer_head *bh_result, int create) |
101 | { | 86 | { |
102 | return gfs2_block_map(inode, lblock, 0, bh_result); | 87 | return gfs2_block_map(inode, lblock, bh_result, 0); |
103 | } | 88 | } |
104 | 89 | ||
105 | /** | 90 | /** |
106 | * gfs2_writepage - Write complete page | 91 | * gfs2_writepage_common - Common bits of writepage |
107 | * @page: Page to write | 92 | * @page: The page to be written |
93 | * @wbc: The writeback control | ||
108 | * | 94 | * |
109 | * Returns: errno | 95 | * Returns: 1 if writepage is ok, otherwise an error code or zero if no error. |
110 | * | ||
111 | * Some of this is copied from block_write_full_page() although we still | ||
112 | * call it to do most of the work. | ||
113 | */ | 96 | */ |
114 | 97 | ||
115 | static int gfs2_writepage(struct page *page, struct writeback_control *wbc) | 98 | static int gfs2_writepage_common(struct page *page, |
99 | struct writeback_control *wbc) | ||
116 | { | 100 | { |
117 | struct inode *inode = page->mapping->host; | 101 | struct inode *inode = page->mapping->host; |
118 | struct gfs2_inode *ip = GFS2_I(inode); | 102 | struct gfs2_inode *ip = GFS2_I(inode); |
@@ -120,41 +104,133 @@ static int gfs2_writepage(struct page *page, struct writeback_control *wbc) | |||
120 | loff_t i_size = i_size_read(inode); | 104 | loff_t i_size = i_size_read(inode); |
121 | pgoff_t end_index = i_size >> PAGE_CACHE_SHIFT; | 105 | pgoff_t end_index = i_size >> PAGE_CACHE_SHIFT; |
122 | unsigned offset; | 106 | unsigned offset; |
123 | int error; | 107 | int ret = -EIO; |
124 | int done_trans = 0; | ||
125 | 108 | ||
126 | if (gfs2_assert_withdraw(sdp, gfs2_glock_is_held_excl(ip->i_gl))) { | 109 | if (gfs2_assert_withdraw(sdp, gfs2_glock_is_held_excl(ip->i_gl))) |
127 | unlock_page(page); | 110 | goto out; |
128 | return -EIO; | 111 | ret = 0; |
129 | } | ||
130 | if (current->journal_info) | 112 | if (current->journal_info) |
131 | goto out_ignore; | 113 | goto redirty; |
132 | |||
133 | /* Is the page fully outside i_size? (truncate in progress) */ | 114 | /* Is the page fully outside i_size? (truncate in progress) */ |
134 | offset = i_size & (PAGE_CACHE_SIZE-1); | 115 | offset = i_size & (PAGE_CACHE_SIZE-1); |
135 | if (page->index > end_index || (page->index == end_index && !offset)) { | 116 | if (page->index > end_index || (page->index == end_index && !offset)) { |
136 | page->mapping->a_ops->invalidatepage(page, 0); | 117 | page->mapping->a_ops->invalidatepage(page, 0); |
137 | unlock_page(page); | 118 | goto out; |
138 | return 0; /* don't care */ | 119 | } |
120 | return 1; | ||
121 | redirty: | ||
122 | redirty_page_for_writepage(wbc, page); | ||
123 | out: | ||
124 | unlock_page(page); | ||
125 | return 0; | ||
126 | } | ||
127 | |||
128 | /** | ||
129 | * gfs2_writeback_writepage - Write page for writeback mappings | ||
130 | * @page: The page | ||
131 | * @wbc: The writeback control | ||
132 | * | ||
133 | */ | ||
134 | |||
135 | static int gfs2_writeback_writepage(struct page *page, | ||
136 | struct writeback_control *wbc) | ||
137 | { | ||
138 | int ret; | ||
139 | |||
140 | ret = gfs2_writepage_common(page, wbc); | ||
141 | if (ret <= 0) | ||
142 | return ret; | ||
143 | |||
144 | ret = mpage_writepage(page, gfs2_get_block_noalloc, wbc); | ||
145 | if (ret == -EAGAIN) | ||
146 | ret = block_write_full_page(page, gfs2_get_block_noalloc, wbc); | ||
147 | return ret; | ||
148 | } | ||
149 | |||
150 | /** | ||
151 | * gfs2_ordered_writepage - Write page for ordered data files | ||
152 | * @page: The page to write | ||
153 | * @wbc: The writeback control | ||
154 | * | ||
155 | */ | ||
156 | |||
157 | static int gfs2_ordered_writepage(struct page *page, | ||
158 | struct writeback_control *wbc) | ||
159 | { | ||
160 | struct inode *inode = page->mapping->host; | ||
161 | struct gfs2_inode *ip = GFS2_I(inode); | ||
162 | int ret; | ||
163 | |||
164 | ret = gfs2_writepage_common(page, wbc); | ||
165 | if (ret <= 0) | ||
166 | return ret; | ||
167 | |||
168 | if (!page_has_buffers(page)) { | ||
169 | create_empty_buffers(page, inode->i_sb->s_blocksize, | ||
170 | (1 << BH_Dirty)|(1 << BH_Uptodate)); | ||
139 | } | 171 | } |
172 | gfs2_page_add_databufs(ip, page, 0, inode->i_sb->s_blocksize-1); | ||
173 | return block_write_full_page(page, gfs2_get_block_noalloc, wbc); | ||
174 | } | ||
140 | 175 | ||
141 | if ((sdp->sd_args.ar_data == GFS2_DATA_ORDERED || gfs2_is_jdata(ip)) && | 176 | /** |
142 | PageChecked(page)) { | 177 | * __gfs2_jdata_writepage - The core of jdata writepage |
178 | * @page: The page to write | ||
179 | * @wbc: The writeback control | ||
180 | * | ||
181 | * This is shared between writepage and writepages and implements the | ||
182 | * core of the writepage operation. If a transaction is required then | ||
183 | * PageChecked will have been set and the transaction will have | ||
184 | * already been started before this is called. | ||
185 | */ | ||
186 | |||
187 | static int __gfs2_jdata_writepage(struct page *page, struct writeback_control *wbc) | ||
188 | { | ||
189 | struct inode *inode = page->mapping->host; | ||
190 | struct gfs2_inode *ip = GFS2_I(inode); | ||
191 | struct gfs2_sbd *sdp = GFS2_SB(inode); | ||
192 | |||
193 | if (PageChecked(page)) { | ||
143 | ClearPageChecked(page); | 194 | ClearPageChecked(page); |
144 | error = gfs2_trans_begin(sdp, RES_DINODE + 1, 0); | ||
145 | if (error) | ||
146 | goto out_ignore; | ||
147 | if (!page_has_buffers(page)) { | 195 | if (!page_has_buffers(page)) { |
148 | create_empty_buffers(page, inode->i_sb->s_blocksize, | 196 | create_empty_buffers(page, inode->i_sb->s_blocksize, |
149 | (1 << BH_Dirty)|(1 << BH_Uptodate)); | 197 | (1 << BH_Dirty)|(1 << BH_Uptodate)); |
150 | } | 198 | } |
151 | gfs2_page_add_databufs(ip, page, 0, sdp->sd_vfs->s_blocksize-1); | 199 | gfs2_page_add_databufs(ip, page, 0, sdp->sd_vfs->s_blocksize-1); |
200 | } | ||
201 | return block_write_full_page(page, gfs2_get_block_noalloc, wbc); | ||
202 | } | ||
203 | |||
204 | /** | ||
205 | * gfs2_jdata_writepage - Write complete page | ||
206 | * @page: Page to write | ||
207 | * | ||
208 | * Returns: errno | ||
209 | * | ||
210 | */ | ||
211 | |||
212 | static int gfs2_jdata_writepage(struct page *page, struct writeback_control *wbc) | ||
213 | { | ||
214 | struct inode *inode = page->mapping->host; | ||
215 | struct gfs2_sbd *sdp = GFS2_SB(inode); | ||
216 | int error; | ||
217 | int done_trans = 0; | ||
218 | |||
219 | error = gfs2_writepage_common(page, wbc); | ||
220 | if (error <= 0) | ||
221 | return error; | ||
222 | |||
223 | if (PageChecked(page)) { | ||
224 | if (wbc->sync_mode != WB_SYNC_ALL) | ||
225 | goto out_ignore; | ||
226 | error = gfs2_trans_begin(sdp, RES_DINODE + 1, 0); | ||
227 | if (error) | ||
228 | goto out_ignore; | ||
152 | done_trans = 1; | 229 | done_trans = 1; |
153 | } | 230 | } |
154 | error = block_write_full_page(page, gfs2_get_block_noalloc, wbc); | 231 | error = __gfs2_jdata_writepage(page, wbc); |
155 | if (done_trans) | 232 | if (done_trans) |
156 | gfs2_trans_end(sdp); | 233 | gfs2_trans_end(sdp); |
157 | gfs2_meta_cache_flush(ip); | ||
158 | return error; | 234 | return error; |
159 | 235 | ||
160 | out_ignore: | 236 | out_ignore: |
@@ -164,29 +240,190 @@ out_ignore: | |||
164 | } | 240 | } |
165 | 241 | ||
166 | /** | 242 | /** |
167 | * gfs2_writepages - Write a bunch of dirty pages back to disk | 243 | * gfs2_writeback_writepages - Write a bunch of dirty pages back to disk |
168 | * @mapping: The mapping to write | 244 | * @mapping: The mapping to write |
169 | * @wbc: Write-back control | 245 | * @wbc: Write-back control |
170 | * | 246 | * |
171 | * For journaled files and/or ordered writes this just falls back to the | 247 | * For the data=writeback case we can already ignore buffer heads |
172 | * kernel's default writepages path for now. We will probably want to change | ||
173 | * that eventually (i.e. when we look at allocate on flush). | ||
174 | * | ||
175 | * For the data=writeback case though we can already ignore buffer heads | ||
176 | * and write whole extents at once. This is a big reduction in the | 248 | * and write whole extents at once. This is a big reduction in the |
177 | * number of I/O requests we send and the bmap calls we make in this case. | 249 | * number of I/O requests we send and the bmap calls we make in this case. |
178 | */ | 250 | */ |
179 | static int gfs2_writepages(struct address_space *mapping, | 251 | static int gfs2_writeback_writepages(struct address_space *mapping, |
180 | struct writeback_control *wbc) | 252 | struct writeback_control *wbc) |
253 | { | ||
254 | return mpage_writepages(mapping, wbc, gfs2_get_block_noalloc); | ||
255 | } | ||
256 | |||
257 | /** | ||
258 | * gfs2_write_jdata_pagevec - Write back a pagevec's worth of pages | ||
259 | * @mapping: The mapping | ||
260 | * @wbc: The writeback control | ||
261 | * @writepage: The writepage function to call for each page | ||
262 | * @pvec: The vector of pages | ||
263 | * @nr_pages: The number of pages to write | ||
264 | * | ||
265 | * Returns: non-zero if loop should terminate, zero otherwise | ||
266 | */ | ||
267 | |||
268 | static int gfs2_write_jdata_pagevec(struct address_space *mapping, | ||
269 | struct writeback_control *wbc, | ||
270 | struct pagevec *pvec, | ||
271 | int nr_pages, pgoff_t end) | ||
181 | { | 272 | { |
182 | struct inode *inode = mapping->host; | 273 | struct inode *inode = mapping->host; |
183 | struct gfs2_inode *ip = GFS2_I(inode); | ||
184 | struct gfs2_sbd *sdp = GFS2_SB(inode); | 274 | struct gfs2_sbd *sdp = GFS2_SB(inode); |
275 | loff_t i_size = i_size_read(inode); | ||
276 | pgoff_t end_index = i_size >> PAGE_CACHE_SHIFT; | ||
277 | unsigned offset = i_size & (PAGE_CACHE_SIZE-1); | ||
278 | unsigned nrblocks = nr_pages * (PAGE_CACHE_SIZE/inode->i_sb->s_blocksize); | ||
279 | struct backing_dev_info *bdi = mapping->backing_dev_info; | ||
280 | int i; | ||
281 | int ret; | ||
282 | |||
283 | ret = gfs2_trans_begin(sdp, nrblocks, 0); | ||
284 | if (ret < 0) | ||
285 | return ret; | ||
286 | |||
287 | for(i = 0; i < nr_pages; i++) { | ||
288 | struct page *page = pvec->pages[i]; | ||
289 | |||
290 | lock_page(page); | ||
291 | |||
292 | if (unlikely(page->mapping != mapping)) { | ||
293 | unlock_page(page); | ||
294 | continue; | ||
295 | } | ||
296 | |||
297 | if (!wbc->range_cyclic && page->index > end) { | ||
298 | ret = 1; | ||
299 | unlock_page(page); | ||
300 | continue; | ||
301 | } | ||
302 | |||
303 | if (wbc->sync_mode != WB_SYNC_NONE) | ||
304 | wait_on_page_writeback(page); | ||
305 | |||
306 | if (PageWriteback(page) || | ||
307 | !clear_page_dirty_for_io(page)) { | ||
308 | unlock_page(page); | ||
309 | continue; | ||
310 | } | ||
311 | |||
312 | /* Is the page fully outside i_size? (truncate in progress) */ | ||
313 | if (page->index > end_index || (page->index == end_index && !offset)) { | ||
314 | page->mapping->a_ops->invalidatepage(page, 0); | ||
315 | unlock_page(page); | ||
316 | continue; | ||
317 | } | ||
318 | |||
319 | ret = __gfs2_jdata_writepage(page, wbc); | ||
320 | |||
321 | if (ret || (--(wbc->nr_to_write) <= 0)) | ||
322 | ret = 1; | ||
323 | if (wbc->nonblocking && bdi_write_congested(bdi)) { | ||
324 | wbc->encountered_congestion = 1; | ||
325 | ret = 1; | ||
326 | } | ||
327 | |||
328 | } | ||
329 | gfs2_trans_end(sdp); | ||
330 | return ret; | ||
331 | } | ||
332 | |||
333 | /** | ||
334 | * gfs2_write_cache_jdata - Like write_cache_pages but different | ||
335 | * @mapping: The mapping to write | ||
336 | * @wbc: The writeback control | ||
337 | * @writepage: The writepage function to call | ||
338 | * @data: The data to pass to writepage | ||
339 | * | ||
340 | * The reason that we use our own function here is that we need to | ||
341 | * start transactions before we grab page locks. This allows us | ||
342 | * to get the ordering right. | ||
343 | */ | ||
344 | |||
345 | static int gfs2_write_cache_jdata(struct address_space *mapping, | ||
346 | struct writeback_control *wbc) | ||
347 | { | ||
348 | struct backing_dev_info *bdi = mapping->backing_dev_info; | ||
349 | int ret = 0; | ||
350 | int done = 0; | ||
351 | struct pagevec pvec; | ||
352 | int nr_pages; | ||
353 | pgoff_t index; | ||
354 | pgoff_t end; | ||
355 | int scanned = 0; | ||
356 | int range_whole = 0; | ||
357 | |||
358 | if (wbc->nonblocking && bdi_write_congested(bdi)) { | ||
359 | wbc->encountered_congestion = 1; | ||
360 | return 0; | ||
361 | } | ||
362 | |||
363 | pagevec_init(&pvec, 0); | ||
364 | if (wbc->range_cyclic) { | ||
365 | index = mapping->writeback_index; /* Start from prev offset */ | ||
366 | end = -1; | ||
367 | } else { | ||
368 | index = wbc->range_start >> PAGE_CACHE_SHIFT; | ||
369 | end = wbc->range_end >> PAGE_CACHE_SHIFT; | ||
370 | if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) | ||
371 | range_whole = 1; | ||
372 | scanned = 1; | ||
373 | } | ||
185 | 374 | ||
186 | if (sdp->sd_args.ar_data == GFS2_DATA_WRITEBACK && !gfs2_is_jdata(ip)) | 375 | retry: |
187 | return mpage_writepages(mapping, wbc, gfs2_get_block_noalloc); | 376 | while (!done && (index <= end) && |
377 | (nr_pages = pagevec_lookup_tag(&pvec, mapping, &index, | ||
378 | PAGECACHE_TAG_DIRTY, | ||
379 | min(end - index, (pgoff_t)PAGEVEC_SIZE-1) + 1))) { | ||
380 | scanned = 1; | ||
381 | ret = gfs2_write_jdata_pagevec(mapping, wbc, &pvec, nr_pages, end); | ||
382 | if (ret) | ||
383 | done = 1; | ||
384 | if (ret > 0) | ||
385 | ret = 0; | ||
386 | |||
387 | pagevec_release(&pvec); | ||
388 | cond_resched(); | ||
389 | } | ||
390 | |||
391 | if (!scanned && !done) { | ||
392 | /* | ||
393 | * We hit the last page and there is more work to be done: wrap | ||
394 | * back to the start of the file | ||
395 | */ | ||
396 | scanned = 1; | ||
397 | index = 0; | ||
398 | goto retry; | ||
399 | } | ||
400 | |||
401 | if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) | ||
402 | mapping->writeback_index = index; | ||
403 | return ret; | ||
404 | } | ||
405 | |||
406 | |||
407 | /** | ||
408 | * gfs2_jdata_writepages - Write a bunch of dirty pages back to disk | ||
409 | * @mapping: The mapping to write | ||
410 | * @wbc: The writeback control | ||
411 | * | ||
412 | */ | ||
188 | 413 | ||
189 | return generic_writepages(mapping, wbc); | 414 | static int gfs2_jdata_writepages(struct address_space *mapping, |
415 | struct writeback_control *wbc) | ||
416 | { | ||
417 | struct gfs2_inode *ip = GFS2_I(mapping->host); | ||
418 | struct gfs2_sbd *sdp = GFS2_SB(mapping->host); | ||
419 | int ret; | ||
420 | |||
421 | ret = gfs2_write_cache_jdata(mapping, wbc); | ||
422 | if (ret == 0 && wbc->sync_mode == WB_SYNC_ALL) { | ||
423 | gfs2_log_flush(sdp, ip->i_gl); | ||
424 | ret = gfs2_write_cache_jdata(mapping, wbc); | ||
425 | } | ||
426 | return ret; | ||
190 | } | 427 | } |
191 | 428 | ||
192 | /** | 429 | /** |
@@ -231,62 +468,107 @@ static int stuffed_readpage(struct gfs2_inode *ip, struct page *page) | |||
231 | 468 | ||
232 | 469 | ||
233 | /** | 470 | /** |
234 | * gfs2_readpage - readpage with locking | 471 | * __gfs2_readpage - readpage |
235 | * @file: The file to read a page for. N.B. This may be NULL if we are | 472 | * @file: The file to read a page for |
236 | * reading an internal file. | ||
237 | * @page: The page to read | 473 | * @page: The page to read |
238 | * | 474 | * |
239 | * Returns: errno | 475 | * This is the core of gfs2's readpage. Its used by the internal file |
476 | * reading code as in that case we already hold the glock. Also its | ||
477 | * called by gfs2_readpage() once the required lock has been granted. | ||
478 | * | ||
240 | */ | 479 | */ |
241 | 480 | ||
242 | static int gfs2_readpage(struct file *file, struct page *page) | 481 | static int __gfs2_readpage(void *file, struct page *page) |
243 | { | 482 | { |
244 | struct gfs2_inode *ip = GFS2_I(page->mapping->host); | 483 | struct gfs2_inode *ip = GFS2_I(page->mapping->host); |
245 | struct gfs2_sbd *sdp = GFS2_SB(page->mapping->host); | 484 | struct gfs2_sbd *sdp = GFS2_SB(page->mapping->host); |
246 | struct gfs2_file *gf = NULL; | ||
247 | struct gfs2_holder gh; | ||
248 | int error; | 485 | int error; |
249 | int do_unlock = 0; | ||
250 | |||
251 | if (likely(file != &gfs2_internal_file_sentinel)) { | ||
252 | if (file) { | ||
253 | gf = file->private_data; | ||
254 | if (test_bit(GFF_EXLOCK, &gf->f_flags)) | ||
255 | /* gfs2_sharewrite_fault has grabbed the ip->i_gl already */ | ||
256 | goto skip_lock; | ||
257 | } | ||
258 | gfs2_holder_init(ip->i_gl, LM_ST_SHARED, GL_ATIME|LM_FLAG_TRY_1CB, &gh); | ||
259 | do_unlock = 1; | ||
260 | error = gfs2_glock_nq_atime(&gh); | ||
261 | if (unlikely(error)) | ||
262 | goto out_unlock; | ||
263 | } | ||
264 | 486 | ||
265 | skip_lock: | ||
266 | if (gfs2_is_stuffed(ip)) { | 487 | if (gfs2_is_stuffed(ip)) { |
267 | error = stuffed_readpage(ip, page); | 488 | error = stuffed_readpage(ip, page); |
268 | unlock_page(page); | 489 | unlock_page(page); |
269 | } else | 490 | } else { |
270 | error = mpage_readpage(page, gfs2_get_block); | 491 | error = mpage_readpage(page, gfs2_block_map); |
492 | } | ||
271 | 493 | ||
272 | if (unlikely(test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) | 494 | if (unlikely(test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) |
273 | error = -EIO; | 495 | return -EIO; |
496 | |||
497 | return error; | ||
498 | } | ||
499 | |||
500 | /** | ||
501 | * gfs2_readpage - read a page of a file | ||
502 | * @file: The file to read | ||
503 | * @page: The page of the file | ||
504 | * | ||
505 | * This deals with the locking required. We use a trylock in order to | ||
506 | * avoid the page lock / glock ordering problems returning AOP_TRUNCATED_PAGE | ||
507 | * in the event that we are unable to get the lock. | ||
508 | */ | ||
509 | |||
510 | static int gfs2_readpage(struct file *file, struct page *page) | ||
511 | { | ||
512 | struct gfs2_inode *ip = GFS2_I(page->mapping->host); | ||
513 | struct gfs2_holder gh; | ||
514 | int error; | ||
274 | 515 | ||
275 | if (do_unlock) { | 516 | gfs2_holder_init(ip->i_gl, LM_ST_SHARED, GL_ATIME|LM_FLAG_TRY_1CB, &gh); |
276 | gfs2_glock_dq_m(1, &gh); | 517 | error = gfs2_glock_nq_atime(&gh); |
277 | gfs2_holder_uninit(&gh); | 518 | if (unlikely(error)) { |
519 | unlock_page(page); | ||
520 | goto out; | ||
278 | } | 521 | } |
522 | error = __gfs2_readpage(file, page); | ||
523 | gfs2_glock_dq(&gh); | ||
279 | out: | 524 | out: |
280 | return error; | 525 | gfs2_holder_uninit(&gh); |
281 | out_unlock: | ||
282 | unlock_page(page); | ||
283 | if (error == GLR_TRYFAILED) { | 526 | if (error == GLR_TRYFAILED) { |
284 | error = AOP_TRUNCATED_PAGE; | ||
285 | yield(); | 527 | yield(); |
528 | return AOP_TRUNCATED_PAGE; | ||
286 | } | 529 | } |
287 | if (do_unlock) | 530 | return error; |
288 | gfs2_holder_uninit(&gh); | 531 | } |
289 | goto out; | 532 | |
533 | /** | ||
534 | * gfs2_internal_read - read an internal file | ||
535 | * @ip: The gfs2 inode | ||
536 | * @ra_state: The readahead state (or NULL for no readahead) | ||
537 | * @buf: The buffer to fill | ||
538 | * @pos: The file position | ||
539 | * @size: The amount to read | ||
540 | * | ||
541 | */ | ||
542 | |||
543 | int gfs2_internal_read(struct gfs2_inode *ip, struct file_ra_state *ra_state, | ||
544 | char *buf, loff_t *pos, unsigned size) | ||
545 | { | ||
546 | struct address_space *mapping = ip->i_inode.i_mapping; | ||
547 | unsigned long index = *pos / PAGE_CACHE_SIZE; | ||
548 | unsigned offset = *pos & (PAGE_CACHE_SIZE - 1); | ||
549 | unsigned copied = 0; | ||
550 | unsigned amt; | ||
551 | struct page *page; | ||
552 | void *p; | ||
553 | |||
554 | do { | ||
555 | amt = size - copied; | ||
556 | if (offset + size > PAGE_CACHE_SIZE) | ||
557 | amt = PAGE_CACHE_SIZE - offset; | ||
558 | page = read_cache_page(mapping, index, __gfs2_readpage, NULL); | ||
559 | if (IS_ERR(page)) | ||
560 | return PTR_ERR(page); | ||
561 | p = kmap_atomic(page, KM_USER0); | ||
562 | memcpy(buf + copied, p + offset, amt); | ||
563 | kunmap_atomic(p, KM_USER0); | ||
564 | mark_page_accessed(page); | ||
565 | page_cache_release(page); | ||
566 | copied += amt; | ||
567 | index++; | ||
568 | offset = 0; | ||
569 | } while(copied < size); | ||
570 | (*pos) += size; | ||
571 | return size; | ||
290 | } | 572 | } |
291 | 573 | ||
292 | /** | 574 | /** |
@@ -300,10 +582,9 @@ out_unlock: | |||
300 | * Any I/O we ignore at this time will be done via readpage later. | 582 | * Any I/O we ignore at this time will be done via readpage later. |
301 | * 2. We don't handle stuffed files here we let readpage do the honours. | 583 | * 2. We don't handle stuffed files here we let readpage do the honours. |
302 | * 3. mpage_readpages() does most of the heavy lifting in the common case. | 584 | * 3. mpage_readpages() does most of the heavy lifting in the common case. |
303 | * 4. gfs2_get_block() is relied upon to set BH_Boundary in the right places. | 585 | * 4. gfs2_block_map() is relied upon to set BH_Boundary in the right places. |
304 | * 5. We use LM_FLAG_TRY_1CB here, effectively we then have lock-ahead as | ||
305 | * well as read-ahead. | ||
306 | */ | 586 | */ |
587 | |||
307 | static int gfs2_readpages(struct file *file, struct address_space *mapping, | 588 | static int gfs2_readpages(struct file *file, struct address_space *mapping, |
308 | struct list_head *pages, unsigned nr_pages) | 589 | struct list_head *pages, unsigned nr_pages) |
309 | { | 590 | { |
@@ -311,42 +592,20 @@ static int gfs2_readpages(struct file *file, struct address_space *mapping, | |||
311 | struct gfs2_inode *ip = GFS2_I(inode); | 592 | struct gfs2_inode *ip = GFS2_I(inode); |
312 | struct gfs2_sbd *sdp = GFS2_SB(inode); | 593 | struct gfs2_sbd *sdp = GFS2_SB(inode); |
313 | struct gfs2_holder gh; | 594 | struct gfs2_holder gh; |
314 | int ret = 0; | 595 | int ret; |
315 | int do_unlock = 0; | ||
316 | 596 | ||
317 | if (likely(file != &gfs2_internal_file_sentinel)) { | 597 | gfs2_holder_init(ip->i_gl, LM_ST_SHARED, GL_ATIME, &gh); |
318 | if (file) { | 598 | ret = gfs2_glock_nq_atime(&gh); |
319 | struct gfs2_file *gf = file->private_data; | 599 | if (unlikely(ret)) |
320 | if (test_bit(GFF_EXLOCK, &gf->f_flags)) | 600 | goto out_uninit; |
321 | goto skip_lock; | ||
322 | } | ||
323 | gfs2_holder_init(ip->i_gl, LM_ST_SHARED, | ||
324 | LM_FLAG_TRY_1CB|GL_ATIME, &gh); | ||
325 | do_unlock = 1; | ||
326 | ret = gfs2_glock_nq_atime(&gh); | ||
327 | if (ret == GLR_TRYFAILED) | ||
328 | goto out_noerror; | ||
329 | if (unlikely(ret)) | ||
330 | goto out_unlock; | ||
331 | } | ||
332 | skip_lock: | ||
333 | if (!gfs2_is_stuffed(ip)) | 601 | if (!gfs2_is_stuffed(ip)) |
334 | ret = mpage_readpages(mapping, pages, nr_pages, gfs2_get_block); | 602 | ret = mpage_readpages(mapping, pages, nr_pages, gfs2_block_map); |
335 | 603 | gfs2_glock_dq(&gh); | |
336 | if (do_unlock) { | 604 | out_uninit: |
337 | gfs2_glock_dq_m(1, &gh); | 605 | gfs2_holder_uninit(&gh); |
338 | gfs2_holder_uninit(&gh); | ||
339 | } | ||
340 | out: | ||
341 | if (unlikely(test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) | 606 | if (unlikely(test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) |
342 | ret = -EIO; | 607 | ret = -EIO; |
343 | return ret; | 608 | return ret; |
344 | out_noerror: | ||
345 | ret = 0; | ||
346 | out_unlock: | ||
347 | if (do_unlock) | ||
348 | gfs2_holder_uninit(&gh); | ||
349 | goto out; | ||
350 | } | 609 | } |
351 | 610 | ||
352 | /** | 611 | /** |
@@ -382,20 +641,11 @@ static int gfs2_write_begin(struct file *file, struct address_space *mapping, | |||
382 | if (unlikely(error)) | 641 | if (unlikely(error)) |
383 | goto out_uninit; | 642 | goto out_uninit; |
384 | 643 | ||
385 | error = -ENOMEM; | ||
386 | page = __grab_cache_page(mapping, index); | ||
387 | *pagep = page; | ||
388 | if (!page) | ||
389 | goto out_unlock; | ||
390 | |||
391 | gfs2_write_calc_reserv(ip, len, &data_blocks, &ind_blocks); | 644 | gfs2_write_calc_reserv(ip, len, &data_blocks, &ind_blocks); |
392 | |||
393 | error = gfs2_write_alloc_required(ip, pos, len, &alloc_required); | 645 | error = gfs2_write_alloc_required(ip, pos, len, &alloc_required); |
394 | if (error) | 646 | if (error) |
395 | goto out_putpage; | 647 | goto out_unlock; |
396 | |||
397 | 648 | ||
398 | ip->i_alloc.al_requested = 0; | ||
399 | if (alloc_required) { | 649 | if (alloc_required) { |
400 | al = gfs2_alloc_get(ip); | 650 | al = gfs2_alloc_get(ip); |
401 | 651 | ||
@@ -424,40 +674,47 @@ static int gfs2_write_begin(struct file *file, struct address_space *mapping, | |||
424 | if (error) | 674 | if (error) |
425 | goto out_trans_fail; | 675 | goto out_trans_fail; |
426 | 676 | ||
677 | error = -ENOMEM; | ||
678 | page = __grab_cache_page(mapping, index); | ||
679 | *pagep = page; | ||
680 | if (unlikely(!page)) | ||
681 | goto out_endtrans; | ||
682 | |||
427 | if (gfs2_is_stuffed(ip)) { | 683 | if (gfs2_is_stuffed(ip)) { |
684 | error = 0; | ||
428 | if (pos + len > sdp->sd_sb.sb_bsize - sizeof(struct gfs2_dinode)) { | 685 | if (pos + len > sdp->sd_sb.sb_bsize - sizeof(struct gfs2_dinode)) { |
429 | error = gfs2_unstuff_dinode(ip, page); | 686 | error = gfs2_unstuff_dinode(ip, page); |
430 | if (error == 0) | 687 | if (error == 0) |
431 | goto prepare_write; | 688 | goto prepare_write; |
432 | } else if (!PageUptodate(page)) | 689 | } else if (!PageUptodate(page)) { |
433 | error = stuffed_readpage(ip, page); | 690 | error = stuffed_readpage(ip, page); |
691 | } | ||
434 | goto out; | 692 | goto out; |
435 | } | 693 | } |
436 | 694 | ||
437 | prepare_write: | 695 | prepare_write: |
438 | error = block_prepare_write(page, from, to, gfs2_get_block); | 696 | error = block_prepare_write(page, from, to, gfs2_block_map); |
439 | |||
440 | out: | 697 | out: |
441 | if (error) { | 698 | if (error == 0) |
442 | gfs2_trans_end(sdp); | 699 | return 0; |
700 | |||
701 | page_cache_release(page); | ||
702 | if (pos + len > ip->i_inode.i_size) | ||
703 | vmtruncate(&ip->i_inode, ip->i_inode.i_size); | ||
704 | out_endtrans: | ||
705 | gfs2_trans_end(sdp); | ||
443 | out_trans_fail: | 706 | out_trans_fail: |
444 | if (alloc_required) { | 707 | if (alloc_required) { |
445 | gfs2_inplace_release(ip); | 708 | gfs2_inplace_release(ip); |
446 | out_qunlock: | 709 | out_qunlock: |
447 | gfs2_quota_unlock(ip); | 710 | gfs2_quota_unlock(ip); |
448 | out_alloc_put: | 711 | out_alloc_put: |
449 | gfs2_alloc_put(ip); | 712 | gfs2_alloc_put(ip); |
450 | } | 713 | } |
451 | out_putpage: | ||
452 | page_cache_release(page); | ||
453 | if (pos + len > ip->i_inode.i_size) | ||
454 | vmtruncate(&ip->i_inode, ip->i_inode.i_size); | ||
455 | out_unlock: | 714 | out_unlock: |
456 | gfs2_glock_dq_m(1, &ip->i_gh); | 715 | gfs2_glock_dq(&ip->i_gh); |
457 | out_uninit: | 716 | out_uninit: |
458 | gfs2_holder_uninit(&ip->i_gh); | 717 | gfs2_holder_uninit(&ip->i_gh); |
459 | } | ||
460 | |||
461 | return error; | 718 | return error; |
462 | } | 719 | } |
463 | 720 | ||
@@ -565,7 +822,7 @@ static int gfs2_write_end(struct file *file, struct address_space *mapping, | |||
565 | struct gfs2_inode *ip = GFS2_I(inode); | 822 | struct gfs2_inode *ip = GFS2_I(inode); |
566 | struct gfs2_sbd *sdp = GFS2_SB(inode); | 823 | struct gfs2_sbd *sdp = GFS2_SB(inode); |
567 | struct buffer_head *dibh; | 824 | struct buffer_head *dibh; |
568 | struct gfs2_alloc *al = &ip->i_alloc; | 825 | struct gfs2_alloc *al = ip->i_alloc; |
569 | struct gfs2_dinode *di; | 826 | struct gfs2_dinode *di; |
570 | unsigned int from = pos & (PAGE_CACHE_SIZE - 1); | 827 | unsigned int from = pos & (PAGE_CACHE_SIZE - 1); |
571 | unsigned int to = from + len; | 828 | unsigned int to = from + len; |
@@ -585,19 +842,16 @@ static int gfs2_write_end(struct file *file, struct address_space *mapping, | |||
585 | if (gfs2_is_stuffed(ip)) | 842 | if (gfs2_is_stuffed(ip)) |
586 | return gfs2_stuffed_write_end(inode, dibh, pos, len, copied, page); | 843 | return gfs2_stuffed_write_end(inode, dibh, pos, len, copied, page); |
587 | 844 | ||
588 | if (sdp->sd_args.ar_data == GFS2_DATA_ORDERED || gfs2_is_jdata(ip)) | 845 | if (!gfs2_is_writeback(ip)) |
589 | gfs2_page_add_databufs(ip, page, from, to); | 846 | gfs2_page_add_databufs(ip, page, from, to); |
590 | 847 | ||
591 | ret = generic_write_end(file, mapping, pos, len, copied, page, fsdata); | 848 | ret = generic_write_end(file, mapping, pos, len, copied, page, fsdata); |
592 | 849 | ||
593 | if (likely(ret >= 0)) { | 850 | if (likely(ret >= 0) && (inode->i_size > ip->i_di.di_size)) { |
594 | copied = ret; | 851 | di = (struct gfs2_dinode *)dibh->b_data; |
595 | if ((pos + copied) > inode->i_size) { | 852 | ip->i_di.di_size = inode->i_size; |
596 | di = (struct gfs2_dinode *)dibh->b_data; | 853 | di->di_size = cpu_to_be64(inode->i_size); |
597 | ip->i_di.di_size = inode->i_size; | 854 | mark_inode_dirty(inode); |
598 | di->di_size = cpu_to_be64(inode->i_size); | ||
599 | mark_inode_dirty(inode); | ||
600 | } | ||
601 | } | 855 | } |
602 | 856 | ||
603 | if (inode == sdp->sd_rindex) | 857 | if (inode == sdp->sd_rindex) |
@@ -606,7 +860,7 @@ static int gfs2_write_end(struct file *file, struct address_space *mapping, | |||
606 | brelse(dibh); | 860 | brelse(dibh); |
607 | gfs2_trans_end(sdp); | 861 | gfs2_trans_end(sdp); |
608 | failed: | 862 | failed: |
609 | if (al->al_requested) { | 863 | if (al) { |
610 | gfs2_inplace_release(ip); | 864 | gfs2_inplace_release(ip); |
611 | gfs2_quota_unlock(ip); | 865 | gfs2_quota_unlock(ip); |
612 | gfs2_alloc_put(ip); | 866 | gfs2_alloc_put(ip); |
@@ -625,11 +879,7 @@ failed: | |||
625 | 879 | ||
626 | static int gfs2_set_page_dirty(struct page *page) | 880 | static int gfs2_set_page_dirty(struct page *page) |
627 | { | 881 | { |
628 | struct gfs2_inode *ip = GFS2_I(page->mapping->host); | 882 | SetPageChecked(page); |
629 | struct gfs2_sbd *sdp = GFS2_SB(page->mapping->host); | ||
630 | |||
631 | if (sdp->sd_args.ar_data == GFS2_DATA_ORDERED || gfs2_is_jdata(ip)) | ||
632 | SetPageChecked(page); | ||
633 | return __set_page_dirty_buffers(page); | 883 | return __set_page_dirty_buffers(page); |
634 | } | 884 | } |
635 | 885 | ||
@@ -653,7 +903,7 @@ static sector_t gfs2_bmap(struct address_space *mapping, sector_t lblock) | |||
653 | return 0; | 903 | return 0; |
654 | 904 | ||
655 | if (!gfs2_is_stuffed(ip)) | 905 | if (!gfs2_is_stuffed(ip)) |
656 | dblock = generic_block_bmap(mapping, lblock, gfs2_get_block); | 906 | dblock = generic_block_bmap(mapping, lblock, gfs2_block_map); |
657 | 907 | ||
658 | gfs2_glock_dq_uninit(&i_gh); | 908 | gfs2_glock_dq_uninit(&i_gh); |
659 | 909 | ||
@@ -719,13 +969,9 @@ static int gfs2_ok_for_dio(struct gfs2_inode *ip, int rw, loff_t offset) | |||
719 | { | 969 | { |
720 | /* | 970 | /* |
721 | * Should we return an error here? I can't see that O_DIRECT for | 971 | * Should we return an error here? I can't see that O_DIRECT for |
722 | * a journaled file makes any sense. For now we'll silently fall | 972 | * a stuffed file makes any sense. For now we'll silently fall |
723 | * back to buffered I/O, likewise we do the same for stuffed | 973 | * back to buffered I/O |
724 | * files since they are (a) small and (b) unaligned. | ||
725 | */ | 974 | */ |
726 | if (gfs2_is_jdata(ip)) | ||
727 | return 0; | ||
728 | |||
729 | if (gfs2_is_stuffed(ip)) | 975 | if (gfs2_is_stuffed(ip)) |
730 | return 0; | 976 | return 0; |
731 | 977 | ||
@@ -836,9 +1082,23 @@ cannot_release: | |||
836 | return 0; | 1082 | return 0; |
837 | } | 1083 | } |
838 | 1084 | ||
839 | const struct address_space_operations gfs2_file_aops = { | 1085 | static const struct address_space_operations gfs2_writeback_aops = { |
840 | .writepage = gfs2_writepage, | 1086 | .writepage = gfs2_writeback_writepage, |
841 | .writepages = gfs2_writepages, | 1087 | .writepages = gfs2_writeback_writepages, |
1088 | .readpage = gfs2_readpage, | ||
1089 | .readpages = gfs2_readpages, | ||
1090 | .sync_page = block_sync_page, | ||
1091 | .write_begin = gfs2_write_begin, | ||
1092 | .write_end = gfs2_write_end, | ||
1093 | .bmap = gfs2_bmap, | ||
1094 | .invalidatepage = gfs2_invalidatepage, | ||
1095 | .releasepage = gfs2_releasepage, | ||
1096 | .direct_IO = gfs2_direct_IO, | ||
1097 | .migratepage = buffer_migrate_page, | ||
1098 | }; | ||
1099 | |||
1100 | static const struct address_space_operations gfs2_ordered_aops = { | ||
1101 | .writepage = gfs2_ordered_writepage, | ||
842 | .readpage = gfs2_readpage, | 1102 | .readpage = gfs2_readpage, |
843 | .readpages = gfs2_readpages, | 1103 | .readpages = gfs2_readpages, |
844 | .sync_page = block_sync_page, | 1104 | .sync_page = block_sync_page, |
@@ -849,5 +1109,34 @@ const struct address_space_operations gfs2_file_aops = { | |||
849 | .invalidatepage = gfs2_invalidatepage, | 1109 | .invalidatepage = gfs2_invalidatepage, |
850 | .releasepage = gfs2_releasepage, | 1110 | .releasepage = gfs2_releasepage, |
851 | .direct_IO = gfs2_direct_IO, | 1111 | .direct_IO = gfs2_direct_IO, |
1112 | .migratepage = buffer_migrate_page, | ||
852 | }; | 1113 | }; |
853 | 1114 | ||
1115 | static const struct address_space_operations gfs2_jdata_aops = { | ||
1116 | .writepage = gfs2_jdata_writepage, | ||
1117 | .writepages = gfs2_jdata_writepages, | ||
1118 | .readpage = gfs2_readpage, | ||
1119 | .readpages = gfs2_readpages, | ||
1120 | .sync_page = block_sync_page, | ||
1121 | .write_begin = gfs2_write_begin, | ||
1122 | .write_end = gfs2_write_end, | ||
1123 | .set_page_dirty = gfs2_set_page_dirty, | ||
1124 | .bmap = gfs2_bmap, | ||
1125 | .invalidatepage = gfs2_invalidatepage, | ||
1126 | .releasepage = gfs2_releasepage, | ||
1127 | }; | ||
1128 | |||
1129 | void gfs2_set_aops(struct inode *inode) | ||
1130 | { | ||
1131 | struct gfs2_inode *ip = GFS2_I(inode); | ||
1132 | |||
1133 | if (gfs2_is_writeback(ip)) | ||
1134 | inode->i_mapping->a_ops = &gfs2_writeback_aops; | ||
1135 | else if (gfs2_is_ordered(ip)) | ||
1136 | inode->i_mapping->a_ops = &gfs2_ordered_aops; | ||
1137 | else if (gfs2_is_jdata(ip)) | ||
1138 | inode->i_mapping->a_ops = &gfs2_jdata_aops; | ||
1139 | else | ||
1140 | BUG(); | ||
1141 | } | ||
1142 | |||
diff --git a/fs/gfs2/ops_address.h b/fs/gfs2/ops_address.h index fa1b5b3d28b..5da21285bba 100644 --- a/fs/gfs2/ops_address.h +++ b/fs/gfs2/ops_address.h | |||
@@ -14,9 +14,10 @@ | |||
14 | #include <linux/buffer_head.h> | 14 | #include <linux/buffer_head.h> |
15 | #include <linux/mm.h> | 15 | #include <linux/mm.h> |
16 | 16 | ||
17 | extern const struct address_space_operations gfs2_file_aops; | ||
18 | extern int gfs2_get_block(struct inode *inode, sector_t lblock, | ||
19 | struct buffer_head *bh_result, int create); | ||
20 | extern int gfs2_releasepage(struct page *page, gfp_t gfp_mask); | 17 | extern int gfs2_releasepage(struct page *page, gfp_t gfp_mask); |
18 | extern int gfs2_internal_read(struct gfs2_inode *ip, | ||
19 | struct file_ra_state *ra_state, | ||
20 | char *buf, loff_t *pos, unsigned size); | ||
21 | extern void gfs2_set_aops(struct inode *inode); | ||
21 | 22 | ||
22 | #endif /* __OPS_ADDRESS_DOT_H__ */ | 23 | #endif /* __OPS_ADDRESS_DOT_H__ */ |
diff --git a/fs/gfs2/ops_file.c b/fs/gfs2/ops_file.c index bb11fd6752d..f4842f2548c 100644 --- a/fs/gfs2/ops_file.c +++ b/fs/gfs2/ops_file.c | |||
@@ -33,57 +33,12 @@ | |||
33 | #include "lm.h" | 33 | #include "lm.h" |
34 | #include "log.h" | 34 | #include "log.h" |
35 | #include "meta_io.h" | 35 | #include "meta_io.h" |
36 | #include "ops_file.h" | ||
37 | #include "ops_vm.h" | ||
38 | #include "quota.h" | 36 | #include "quota.h" |
39 | #include "rgrp.h" | 37 | #include "rgrp.h" |
40 | #include "trans.h" | 38 | #include "trans.h" |
41 | #include "util.h" | 39 | #include "util.h" |
42 | #include "eaops.h" | 40 | #include "eaops.h" |
43 | 41 | #include "ops_address.h" | |
44 | /* | ||
45 | * Most fields left uninitialised to catch anybody who tries to | ||
46 | * use them. f_flags set to prevent file_accessed() from touching | ||
47 | * any other part of this. Its use is purely as a flag so that we | ||
48 | * know (in readpage()) whether or not do to locking. | ||
49 | */ | ||
50 | struct file gfs2_internal_file_sentinel = { | ||
51 | .f_flags = O_NOATIME|O_RDONLY, | ||
52 | }; | ||
53 | |||
54 | static int gfs2_read_actor(read_descriptor_t *desc, struct page *page, | ||
55 | unsigned long offset, unsigned long size) | ||
56 | { | ||
57 | char *kaddr; | ||
58 | unsigned long count = desc->count; | ||
59 | |||
60 | if (size > count) | ||
61 | size = count; | ||
62 | |||
63 | kaddr = kmap(page); | ||
64 | memcpy(desc->arg.data, kaddr + offset, size); | ||
65 | kunmap(page); | ||
66 | |||
67 | desc->count = count - size; | ||
68 | desc->written += size; | ||
69 | desc->arg.buf += size; | ||
70 | return size; | ||
71 | } | ||
72 | |||
73 | int gfs2_internal_read(struct gfs2_inode *ip, struct file_ra_state *ra_state, | ||
74 | char *buf, loff_t *pos, unsigned size) | ||
75 | { | ||
76 | struct inode *inode = &ip->i_inode; | ||
77 | read_descriptor_t desc; | ||
78 | desc.written = 0; | ||
79 | desc.arg.data = buf; | ||
80 | desc.count = size; | ||
81 | desc.error = 0; | ||
82 | do_generic_mapping_read(inode->i_mapping, ra_state, | ||
83 | &gfs2_internal_file_sentinel, pos, &desc, | ||
84 | gfs2_read_actor); | ||
85 | return desc.written ? desc.written : desc.error; | ||
86 | } | ||
87 | 42 | ||
88 | /** | 43 | /** |
89 | * gfs2_llseek - seek to a location in a file | 44 | * gfs2_llseek - seek to a location in a file |
@@ -214,7 +169,7 @@ static int gfs2_get_flags(struct file *filp, u32 __user *ptr) | |||
214 | if (put_user(fsflags, ptr)) | 169 | if (put_user(fsflags, ptr)) |
215 | error = -EFAULT; | 170 | error = -EFAULT; |
216 | 171 | ||
217 | gfs2_glock_dq_m(1, &gh); | 172 | gfs2_glock_dq(&gh); |
218 | gfs2_holder_uninit(&gh); | 173 | gfs2_holder_uninit(&gh); |
219 | return error; | 174 | return error; |
220 | } | 175 | } |
@@ -291,7 +246,16 @@ static int do_gfs2_set_flags(struct file *filp, u32 reqflags, u32 mask) | |||
291 | if (error) | 246 | if (error) |
292 | goto out; | 247 | goto out; |
293 | } | 248 | } |
294 | 249 | if ((flags ^ new_flags) & GFS2_DIF_JDATA) { | |
250 | if (flags & GFS2_DIF_JDATA) | ||
251 | gfs2_log_flush(sdp, ip->i_gl); | ||
252 | error = filemap_fdatawrite(inode->i_mapping); | ||
253 | if (error) | ||
254 | goto out; | ||
255 | error = filemap_fdatawait(inode->i_mapping); | ||
256 | if (error) | ||
257 | goto out; | ||
258 | } | ||
295 | error = gfs2_trans_begin(sdp, RES_DINODE, 0); | 259 | error = gfs2_trans_begin(sdp, RES_DINODE, 0); |
296 | if (error) | 260 | if (error) |
297 | goto out; | 261 | goto out; |
@@ -303,6 +267,7 @@ static int do_gfs2_set_flags(struct file *filp, u32 reqflags, u32 mask) | |||
303 | gfs2_dinode_out(ip, bh->b_data); | 267 | gfs2_dinode_out(ip, bh->b_data); |
304 | brelse(bh); | 268 | brelse(bh); |
305 | gfs2_set_inode_flags(inode); | 269 | gfs2_set_inode_flags(inode); |
270 | gfs2_set_aops(inode); | ||
306 | out_trans_end: | 271 | out_trans_end: |
307 | gfs2_trans_end(sdp); | 272 | gfs2_trans_end(sdp); |
308 | out: | 273 | out: |
@@ -338,6 +303,128 @@ static long gfs2_ioctl(struct file *filp, unsigned int cmd, unsigned long arg) | |||
338 | return -ENOTTY; | 303 | return -ENOTTY; |
339 | } | 304 | } |
340 | 305 | ||
306 | /** | ||
307 | * gfs2_allocate_page_backing - Use bmap to allocate blocks | ||
308 | * @page: The (locked) page to allocate backing for | ||
309 | * | ||
310 | * We try to allocate all the blocks required for the page in | ||
311 | * one go. This might fail for various reasons, so we keep | ||
312 | * trying until all the blocks to back this page are allocated. | ||
313 | * If some of the blocks are already allocated, thats ok too. | ||
314 | */ | ||
315 | |||
316 | static int gfs2_allocate_page_backing(struct page *page) | ||
317 | { | ||
318 | struct inode *inode = page->mapping->host; | ||
319 | struct buffer_head bh; | ||
320 | unsigned long size = PAGE_CACHE_SIZE; | ||
321 | u64 lblock = page->index << (PAGE_CACHE_SHIFT - inode->i_blkbits); | ||
322 | |||
323 | do { | ||
324 | bh.b_state = 0; | ||
325 | bh.b_size = size; | ||
326 | gfs2_block_map(inode, lblock, &bh, 1); | ||
327 | if (!buffer_mapped(&bh)) | ||
328 | return -EIO; | ||
329 | size -= bh.b_size; | ||
330 | lblock += (bh.b_size >> inode->i_blkbits); | ||
331 | } while(size > 0); | ||
332 | return 0; | ||
333 | } | ||
334 | |||
335 | /** | ||
336 | * gfs2_page_mkwrite - Make a shared, mmap()ed, page writable | ||
337 | * @vma: The virtual memory area | ||
338 | * @page: The page which is about to become writable | ||
339 | * | ||
340 | * When the page becomes writable, we need to ensure that we have | ||
341 | * blocks allocated on disk to back that page. | ||
342 | */ | ||
343 | |||
344 | static int gfs2_page_mkwrite(struct vm_area_struct *vma, struct page *page) | ||
345 | { | ||
346 | struct inode *inode = vma->vm_file->f_path.dentry->d_inode; | ||
347 | struct gfs2_inode *ip = GFS2_I(inode); | ||
348 | struct gfs2_sbd *sdp = GFS2_SB(inode); | ||
349 | unsigned long last_index; | ||
350 | u64 pos = page->index << (PAGE_CACHE_SIZE - inode->i_blkbits); | ||
351 | unsigned int data_blocks, ind_blocks, rblocks; | ||
352 | int alloc_required = 0; | ||
353 | struct gfs2_holder gh; | ||
354 | struct gfs2_alloc *al; | ||
355 | int ret; | ||
356 | |||
357 | gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, GL_ATIME, &gh); | ||
358 | ret = gfs2_glock_nq_atime(&gh); | ||
359 | if (ret) | ||
360 | goto out; | ||
361 | |||
362 | set_bit(GIF_SW_PAGED, &ip->i_flags); | ||
363 | gfs2_write_calc_reserv(ip, PAGE_CACHE_SIZE, &data_blocks, &ind_blocks); | ||
364 | ret = gfs2_write_alloc_required(ip, pos, PAGE_CACHE_SIZE, &alloc_required); | ||
365 | if (ret || !alloc_required) | ||
366 | goto out_unlock; | ||
367 | ret = -ENOMEM; | ||
368 | al = gfs2_alloc_get(ip); | ||
369 | if (al == NULL) | ||
370 | goto out_unlock; | ||
371 | |||
372 | ret = gfs2_quota_lock(ip, NO_QUOTA_CHANGE, NO_QUOTA_CHANGE); | ||
373 | if (ret) | ||
374 | goto out_alloc_put; | ||
375 | ret = gfs2_quota_check(ip, ip->i_inode.i_uid, ip->i_inode.i_gid); | ||
376 | if (ret) | ||
377 | goto out_quota_unlock; | ||
378 | al->al_requested = data_blocks + ind_blocks; | ||
379 | ret = gfs2_inplace_reserve(ip); | ||
380 | if (ret) | ||
381 | goto out_quota_unlock; | ||
382 | |||
383 | rblocks = RES_DINODE + ind_blocks; | ||
384 | if (gfs2_is_jdata(ip)) | ||
385 | rblocks += data_blocks ? data_blocks : 1; | ||
386 | if (ind_blocks || data_blocks) | ||
387 | rblocks += RES_STATFS + RES_QUOTA; | ||
388 | ret = gfs2_trans_begin(sdp, rblocks, 0); | ||
389 | if (ret) | ||
390 | goto out_trans_fail; | ||
391 | |||
392 | lock_page(page); | ||
393 | ret = -EINVAL; | ||
394 | last_index = ip->i_inode.i_size >> PAGE_CACHE_SHIFT; | ||
395 | if (page->index > last_index) | ||
396 | goto out_unlock_page; | ||
397 | ret = 0; | ||
398 | if (!PageUptodate(page) || page->mapping != ip->i_inode.i_mapping) | ||
399 | goto out_unlock_page; | ||
400 | if (gfs2_is_stuffed(ip)) { | ||
401 | ret = gfs2_unstuff_dinode(ip, page); | ||
402 | if (ret) | ||
403 | goto out_unlock_page; | ||
404 | } | ||
405 | ret = gfs2_allocate_page_backing(page); | ||
406 | |||
407 | out_unlock_page: | ||
408 | unlock_page(page); | ||
409 | gfs2_trans_end(sdp); | ||
410 | out_trans_fail: | ||
411 | gfs2_inplace_release(ip); | ||
412 | out_quota_unlock: | ||
413 | gfs2_quota_unlock(ip); | ||
414 | out_alloc_put: | ||
415 | gfs2_alloc_put(ip); | ||
416 | out_unlock: | ||
417 | gfs2_glock_dq(&gh); | ||
418 | out: | ||
419 | gfs2_holder_uninit(&gh); | ||
420 | return ret; | ||
421 | } | ||
422 | |||
423 | static struct vm_operations_struct gfs2_vm_ops = { | ||
424 | .fault = filemap_fault, | ||
425 | .page_mkwrite = gfs2_page_mkwrite, | ||
426 | }; | ||
427 | |||
341 | 428 | ||
342 | /** | 429 | /** |
343 | * gfs2_mmap - | 430 | * gfs2_mmap - |
@@ -360,14 +447,7 @@ static int gfs2_mmap(struct file *file, struct vm_area_struct *vma) | |||
360 | return error; | 447 | return error; |
361 | } | 448 | } |
362 | 449 | ||
363 | /* This is VM_MAYWRITE instead of VM_WRITE because a call | 450 | vma->vm_ops = &gfs2_vm_ops; |
364 | to mprotect() can turn on VM_WRITE later. */ | ||
365 | |||
366 | if ((vma->vm_flags & (VM_MAYSHARE | VM_MAYWRITE)) == | ||
367 | (VM_MAYSHARE | VM_MAYWRITE)) | ||
368 | vma->vm_ops = &gfs2_vm_ops_sharewrite; | ||
369 | else | ||
370 | vma->vm_ops = &gfs2_vm_ops_private; | ||
371 | 451 | ||
372 | gfs2_glock_dq_uninit(&i_gh); | 452 | gfs2_glock_dq_uninit(&i_gh); |
373 | 453 | ||
@@ -538,15 +618,6 @@ static int gfs2_lock(struct file *file, int cmd, struct file_lock *fl) | |||
538 | if (__mandatory_lock(&ip->i_inode)) | 618 | if (__mandatory_lock(&ip->i_inode)) |
539 | return -ENOLCK; | 619 | return -ENOLCK; |
540 | 620 | ||
541 | if (sdp->sd_args.ar_localflocks) { | ||
542 | if (IS_GETLK(cmd)) { | ||
543 | posix_test_lock(file, fl); | ||
544 | return 0; | ||
545 | } else { | ||
546 | return posix_lock_file_wait(file, fl); | ||
547 | } | ||
548 | } | ||
549 | |||
550 | if (cmd == F_CANCELLK) { | 621 | if (cmd == F_CANCELLK) { |
551 | /* Hack: */ | 622 | /* Hack: */ |
552 | cmd = F_SETLK; | 623 | cmd = F_SETLK; |
@@ -632,16 +703,12 @@ static void do_unflock(struct file *file, struct file_lock *fl) | |||
632 | static int gfs2_flock(struct file *file, int cmd, struct file_lock *fl) | 703 | static int gfs2_flock(struct file *file, int cmd, struct file_lock *fl) |
633 | { | 704 | { |
634 | struct gfs2_inode *ip = GFS2_I(file->f_mapping->host); | 705 | struct gfs2_inode *ip = GFS2_I(file->f_mapping->host); |
635 | struct gfs2_sbd *sdp = GFS2_SB(file->f_mapping->host); | ||
636 | 706 | ||
637 | if (!(fl->fl_flags & FL_FLOCK)) | 707 | if (!(fl->fl_flags & FL_FLOCK)) |
638 | return -ENOLCK; | 708 | return -ENOLCK; |
639 | if (__mandatory_lock(&ip->i_inode)) | 709 | if (__mandatory_lock(&ip->i_inode)) |
640 | return -ENOLCK; | 710 | return -ENOLCK; |
641 | 711 | ||
642 | if (sdp->sd_args.ar_localflocks) | ||
643 | return flock_lock_file_wait(file, fl); | ||
644 | |||
645 | if (fl->fl_type == F_UNLCK) { | 712 | if (fl->fl_type == F_UNLCK) { |
646 | do_unflock(file, fl); | 713 | do_unflock(file, fl); |
647 | return 0; | 714 | return 0; |
@@ -678,3 +745,27 @@ const struct file_operations gfs2_dir_fops = { | |||
678 | .flock = gfs2_flock, | 745 | .flock = gfs2_flock, |
679 | }; | 746 | }; |
680 | 747 | ||
748 | const struct file_operations gfs2_file_fops_nolock = { | ||
749 | .llseek = gfs2_llseek, | ||
750 | .read = do_sync_read, | ||
751 | .aio_read = generic_file_aio_read, | ||
752 | .write = do_sync_write, | ||
753 | .aio_write = generic_file_aio_write, | ||
754 | .unlocked_ioctl = gfs2_ioctl, | ||
755 | .mmap = gfs2_mmap, | ||
756 | .open = gfs2_open, | ||
757 | .release = gfs2_close, | ||
758 | .fsync = gfs2_fsync, | ||
759 | .splice_read = generic_file_splice_read, | ||
760 | .splice_write = generic_file_splice_write, | ||
761 | .setlease = gfs2_setlease, | ||
762 | }; | ||
763 | |||
764 | const struct file_operations gfs2_dir_fops_nolock = { | ||
765 | .readdir = gfs2_readdir, | ||
766 | .unlocked_ioctl = gfs2_ioctl, | ||
767 | .open = gfs2_open, | ||
768 | .release = gfs2_close, | ||
769 | .fsync = gfs2_fsync, | ||
770 | }; | ||
771 | |||
diff --git a/fs/gfs2/ops_file.h b/fs/gfs2/ops_file.h deleted file mode 100644 index 7e5d8ec9c84..00000000000 --- a/fs/gfs2/ops_file.h +++ /dev/null | |||
@@ -1,24 +0,0 @@ | |||
1 | /* | ||
2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | ||
3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | ||
4 | * | ||
5 | * This copyrighted material is made available to anyone wishing to use, | ||
6 | * modify, copy, or redistribute it subject to the terms and conditions | ||
7 | * of the GNU General Public License version 2. | ||
8 | */ | ||
9 | |||
10 | #ifndef __OPS_FILE_DOT_H__ | ||
11 | #define __OPS_FILE_DOT_H__ | ||
12 | |||
13 | #include <linux/fs.h> | ||
14 | struct gfs2_inode; | ||
15 | |||
16 | extern struct file gfs2_internal_file_sentinel; | ||
17 | extern int gfs2_internal_read(struct gfs2_inode *ip, | ||
18 | struct file_ra_state *ra_state, | ||
19 | char *buf, loff_t *pos, unsigned size); | ||
20 | extern void gfs2_set_inode_flags(struct inode *inode); | ||
21 | extern const struct file_operations gfs2_file_fops; | ||
22 | extern const struct file_operations gfs2_dir_fops; | ||
23 | |||
24 | #endif /* __OPS_FILE_DOT_H__ */ | ||
diff --git a/fs/gfs2/ops_fstype.c b/fs/gfs2/ops_fstype.c index 17de58e83d9..43d511bba52 100644 --- a/fs/gfs2/ops_fstype.c +++ b/fs/gfs2/ops_fstype.c | |||
@@ -1,6 +1,6 @@ | |||
1 | /* | 1 | /* |
2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. |
3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | 3 | * Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved. |
4 | * | 4 | * |
5 | * This copyrighted material is made available to anyone wishing to use, | 5 | * This copyrighted material is made available to anyone wishing to use, |
6 | * modify, copy, or redistribute it subject to the terms and conditions | 6 | * modify, copy, or redistribute it subject to the terms and conditions |
@@ -21,6 +21,7 @@ | |||
21 | 21 | ||
22 | #include "gfs2.h" | 22 | #include "gfs2.h" |
23 | #include "incore.h" | 23 | #include "incore.h" |
24 | #include "bmap.h" | ||
24 | #include "daemon.h" | 25 | #include "daemon.h" |
25 | #include "glock.h" | 26 | #include "glock.h" |
26 | #include "glops.h" | 27 | #include "glops.h" |
@@ -59,7 +60,6 @@ static struct gfs2_sbd *init_sbd(struct super_block *sb) | |||
59 | 60 | ||
60 | mutex_init(&sdp->sd_inum_mutex); | 61 | mutex_init(&sdp->sd_inum_mutex); |
61 | spin_lock_init(&sdp->sd_statfs_spin); | 62 | spin_lock_init(&sdp->sd_statfs_spin); |
62 | mutex_init(&sdp->sd_statfs_mutex); | ||
63 | 63 | ||
64 | spin_lock_init(&sdp->sd_rindex_spin); | 64 | spin_lock_init(&sdp->sd_rindex_spin); |
65 | mutex_init(&sdp->sd_rindex_mutex); | 65 | mutex_init(&sdp->sd_rindex_mutex); |
@@ -77,7 +77,6 @@ static struct gfs2_sbd *init_sbd(struct super_block *sb) | |||
77 | 77 | ||
78 | spin_lock_init(&sdp->sd_log_lock); | 78 | spin_lock_init(&sdp->sd_log_lock); |
79 | 79 | ||
80 | INIT_LIST_HEAD(&sdp->sd_log_le_gl); | ||
81 | INIT_LIST_HEAD(&sdp->sd_log_le_buf); | 80 | INIT_LIST_HEAD(&sdp->sd_log_le_buf); |
82 | INIT_LIST_HEAD(&sdp->sd_log_le_revoke); | 81 | INIT_LIST_HEAD(&sdp->sd_log_le_revoke); |
83 | INIT_LIST_HEAD(&sdp->sd_log_le_rg); | 82 | INIT_LIST_HEAD(&sdp->sd_log_le_rg); |
@@ -303,6 +302,67 @@ out: | |||
303 | return error; | 302 | return error; |
304 | } | 303 | } |
305 | 304 | ||
305 | /** | ||
306 | * map_journal_extents - create a reusable "extent" mapping from all logical | ||
307 | * blocks to all physical blocks for the given journal. This will save | ||
308 | * us time when writing journal blocks. Most journals will have only one | ||
309 | * extent that maps all their logical blocks. That's because gfs2.mkfs | ||
310 | * arranges the journal blocks sequentially to maximize performance. | ||
311 | * So the extent would map the first block for the entire file length. | ||
312 | * However, gfs2_jadd can happen while file activity is happening, so | ||
313 | * those journals may not be sequential. Less likely is the case where | ||
314 | * the users created their own journals by mounting the metafs and | ||
315 | * laying it out. But it's still possible. These journals might have | ||
316 | * several extents. | ||
317 | * | ||
318 | * TODO: This should be done in bigger chunks rather than one block at a time, | ||
319 | * but since it's only done at mount time, I'm not worried about the | ||
320 | * time it takes. | ||
321 | */ | ||
322 | static int map_journal_extents(struct gfs2_sbd *sdp) | ||
323 | { | ||
324 | struct gfs2_jdesc *jd = sdp->sd_jdesc; | ||
325 | unsigned int lb; | ||
326 | u64 db, prev_db; /* logical block, disk block, prev disk block */ | ||
327 | struct gfs2_inode *ip = GFS2_I(jd->jd_inode); | ||
328 | struct gfs2_journal_extent *jext = NULL; | ||
329 | struct buffer_head bh; | ||
330 | int rc = 0; | ||
331 | |||
332 | prev_db = 0; | ||
333 | |||
334 | for (lb = 0; lb < ip->i_di.di_size >> sdp->sd_sb.sb_bsize_shift; lb++) { | ||
335 | bh.b_state = 0; | ||
336 | bh.b_blocknr = 0; | ||
337 | bh.b_size = 1 << ip->i_inode.i_blkbits; | ||
338 | rc = gfs2_block_map(jd->jd_inode, lb, &bh, 0); | ||
339 | db = bh.b_blocknr; | ||
340 | if (rc || !db) { | ||
341 | printk(KERN_INFO "GFS2 journal mapping error %d: lb=" | ||
342 | "%u db=%llu\n", rc, lb, (unsigned long long)db); | ||
343 | break; | ||
344 | } | ||
345 | if (!prev_db || db != prev_db + 1) { | ||
346 | jext = kzalloc(sizeof(struct gfs2_journal_extent), | ||
347 | GFP_KERNEL); | ||
348 | if (!jext) { | ||
349 | printk(KERN_INFO "GFS2 error: out of memory " | ||
350 | "mapping journal extents.\n"); | ||
351 | rc = -ENOMEM; | ||
352 | break; | ||
353 | } | ||
354 | jext->dblock = db; | ||
355 | jext->lblock = lb; | ||
356 | jext->blocks = 1; | ||
357 | list_add_tail(&jext->extent_list, &jd->extent_list); | ||
358 | } else { | ||
359 | jext->blocks++; | ||
360 | } | ||
361 | prev_db = db; | ||
362 | } | ||
363 | return rc; | ||
364 | } | ||
365 | |||
306 | static int init_journal(struct gfs2_sbd *sdp, int undo) | 366 | static int init_journal(struct gfs2_sbd *sdp, int undo) |
307 | { | 367 | { |
308 | struct gfs2_holder ji_gh; | 368 | struct gfs2_holder ji_gh; |
@@ -340,7 +400,7 @@ static int init_journal(struct gfs2_sbd *sdp, int undo) | |||
340 | 400 | ||
341 | if (sdp->sd_args.ar_spectator) { | 401 | if (sdp->sd_args.ar_spectator) { |
342 | sdp->sd_jdesc = gfs2_jdesc_find(sdp, 0); | 402 | sdp->sd_jdesc = gfs2_jdesc_find(sdp, 0); |
343 | sdp->sd_log_blks_free = sdp->sd_jdesc->jd_blocks; | 403 | atomic_set(&sdp->sd_log_blks_free, sdp->sd_jdesc->jd_blocks); |
344 | } else { | 404 | } else { |
345 | if (sdp->sd_lockstruct.ls_jid >= gfs2_jindex_size(sdp)) { | 405 | if (sdp->sd_lockstruct.ls_jid >= gfs2_jindex_size(sdp)) { |
346 | fs_err(sdp, "can't mount journal #%u\n", | 406 | fs_err(sdp, "can't mount journal #%u\n", |
@@ -377,7 +437,10 @@ static int init_journal(struct gfs2_sbd *sdp, int undo) | |||
377 | sdp->sd_jdesc->jd_jid, error); | 437 | sdp->sd_jdesc->jd_jid, error); |
378 | goto fail_jinode_gh; | 438 | goto fail_jinode_gh; |
379 | } | 439 | } |
380 | sdp->sd_log_blks_free = sdp->sd_jdesc->jd_blocks; | 440 | atomic_set(&sdp->sd_log_blks_free, sdp->sd_jdesc->jd_blocks); |
441 | |||
442 | /* Map the extents for this journal's blocks */ | ||
443 | map_journal_extents(sdp); | ||
381 | } | 444 | } |
382 | 445 | ||
383 | if (sdp->sd_lockstruct.ls_first) { | 446 | if (sdp->sd_lockstruct.ls_first) { |
diff --git a/fs/gfs2/ops_inode.c b/fs/gfs2/ops_inode.c index 291f0c7eaa3..9f71372c175 100644 --- a/fs/gfs2/ops_inode.c +++ b/fs/gfs2/ops_inode.c | |||
@@ -61,7 +61,7 @@ static int gfs2_create(struct inode *dir, struct dentry *dentry, | |||
61 | inode = gfs2_createi(ghs, &dentry->d_name, S_IFREG | mode, 0); | 61 | inode = gfs2_createi(ghs, &dentry->d_name, S_IFREG | mode, 0); |
62 | if (!IS_ERR(inode)) { | 62 | if (!IS_ERR(inode)) { |
63 | gfs2_trans_end(sdp); | 63 | gfs2_trans_end(sdp); |
64 | if (dip->i_alloc.al_rgd) | 64 | if (dip->i_alloc->al_rgd) |
65 | gfs2_inplace_release(dip); | 65 | gfs2_inplace_release(dip); |
66 | gfs2_quota_unlock(dip); | 66 | gfs2_quota_unlock(dip); |
67 | gfs2_alloc_put(dip); | 67 | gfs2_alloc_put(dip); |
@@ -113,8 +113,18 @@ static struct dentry *gfs2_lookup(struct inode *dir, struct dentry *dentry, | |||
113 | if (inode && IS_ERR(inode)) | 113 | if (inode && IS_ERR(inode)) |
114 | return ERR_PTR(PTR_ERR(inode)); | 114 | return ERR_PTR(PTR_ERR(inode)); |
115 | 115 | ||
116 | if (inode) | 116 | if (inode) { |
117 | struct gfs2_glock *gl = GFS2_I(inode)->i_gl; | ||
118 | struct gfs2_holder gh; | ||
119 | int error; | ||
120 | error = gfs2_glock_nq_init(gl, LM_ST_SHARED, LM_FLAG_ANY, &gh); | ||
121 | if (error) { | ||
122 | iput(inode); | ||
123 | return ERR_PTR(error); | ||
124 | } | ||
125 | gfs2_glock_dq_uninit(&gh); | ||
117 | return d_splice_alias(inode, dentry); | 126 | return d_splice_alias(inode, dentry); |
127 | } | ||
118 | d_add(dentry, inode); | 128 | d_add(dentry, inode); |
119 | 129 | ||
120 | return NULL; | 130 | return NULL; |
@@ -366,7 +376,7 @@ static int gfs2_symlink(struct inode *dir, struct dentry *dentry, | |||
366 | } | 376 | } |
367 | 377 | ||
368 | gfs2_trans_end(sdp); | 378 | gfs2_trans_end(sdp); |
369 | if (dip->i_alloc.al_rgd) | 379 | if (dip->i_alloc->al_rgd) |
370 | gfs2_inplace_release(dip); | 380 | gfs2_inplace_release(dip); |
371 | gfs2_quota_unlock(dip); | 381 | gfs2_quota_unlock(dip); |
372 | gfs2_alloc_put(dip); | 382 | gfs2_alloc_put(dip); |
@@ -442,7 +452,7 @@ static int gfs2_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
442 | gfs2_assert_withdraw(sdp, !error); /* dip already pinned */ | 452 | gfs2_assert_withdraw(sdp, !error); /* dip already pinned */ |
443 | 453 | ||
444 | gfs2_trans_end(sdp); | 454 | gfs2_trans_end(sdp); |
445 | if (dip->i_alloc.al_rgd) | 455 | if (dip->i_alloc->al_rgd) |
446 | gfs2_inplace_release(dip); | 456 | gfs2_inplace_release(dip); |
447 | gfs2_quota_unlock(dip); | 457 | gfs2_quota_unlock(dip); |
448 | gfs2_alloc_put(dip); | 458 | gfs2_alloc_put(dip); |
@@ -548,7 +558,7 @@ static int gfs2_mknod(struct inode *dir, struct dentry *dentry, int mode, | |||
548 | } | 558 | } |
549 | 559 | ||
550 | gfs2_trans_end(sdp); | 560 | gfs2_trans_end(sdp); |
551 | if (dip->i_alloc.al_rgd) | 561 | if (dip->i_alloc->al_rgd) |
552 | gfs2_inplace_release(dip); | 562 | gfs2_inplace_release(dip); |
553 | gfs2_quota_unlock(dip); | 563 | gfs2_quota_unlock(dip); |
554 | gfs2_alloc_put(dip); | 564 | gfs2_alloc_put(dip); |
diff --git a/fs/gfs2/ops_inode.h b/fs/gfs2/ops_inode.h index 34f0caac1a0..fd8cee231e1 100644 --- a/fs/gfs2/ops_inode.h +++ b/fs/gfs2/ops_inode.h | |||
@@ -16,5 +16,11 @@ extern const struct inode_operations gfs2_file_iops; | |||
16 | extern const struct inode_operations gfs2_dir_iops; | 16 | extern const struct inode_operations gfs2_dir_iops; |
17 | extern const struct inode_operations gfs2_symlink_iops; | 17 | extern const struct inode_operations gfs2_symlink_iops; |
18 | extern const struct inode_operations gfs2_dev_iops; | 18 | extern const struct inode_operations gfs2_dev_iops; |
19 | extern const struct file_operations gfs2_file_fops; | ||
20 | extern const struct file_operations gfs2_dir_fops; | ||
21 | extern const struct file_operations gfs2_file_fops_nolock; | ||
22 | extern const struct file_operations gfs2_dir_fops_nolock; | ||
23 | |||
24 | extern void gfs2_set_inode_flags(struct inode *inode); | ||
19 | 25 | ||
20 | #endif /* __OPS_INODE_DOT_H__ */ | 26 | #endif /* __OPS_INODE_DOT_H__ */ |
diff --git a/fs/gfs2/ops_super.c b/fs/gfs2/ops_super.c index 950f31460e8..5e524217944 100644 --- a/fs/gfs2/ops_super.c +++ b/fs/gfs2/ops_super.c | |||
@@ -487,7 +487,6 @@ static struct inode *gfs2_alloc_inode(struct super_block *sb) | |||
487 | if (ip) { | 487 | if (ip) { |
488 | ip->i_flags = 0; | 488 | ip->i_flags = 0; |
489 | ip->i_gl = NULL; | 489 | ip->i_gl = NULL; |
490 | ip->i_last_pfault = jiffies; | ||
491 | } | 490 | } |
492 | return &ip->i_inode; | 491 | return &ip->i_inode; |
493 | } | 492 | } |
diff --git a/fs/gfs2/ops_vm.c b/fs/gfs2/ops_vm.c deleted file mode 100644 index 927d739d468..00000000000 --- a/fs/gfs2/ops_vm.c +++ /dev/null | |||
@@ -1,169 +0,0 @@ | |||
1 | /* | ||
2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | ||
3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | ||
4 | * | ||
5 | * This copyrighted material is made available to anyone wishing to use, | ||
6 | * modify, copy, or redistribute it subject to the terms and conditions | ||
7 | * of the GNU General Public License version 2. | ||
8 | */ | ||
9 | |||
10 | #include <linux/slab.h> | ||
11 | #include <linux/spinlock.h> | ||
12 | #include <linux/completion.h> | ||
13 | #include <linux/buffer_head.h> | ||
14 | #include <linux/mm.h> | ||
15 | #include <linux/pagemap.h> | ||
16 | #include <linux/gfs2_ondisk.h> | ||
17 | #include <linux/lm_interface.h> | ||
18 | |||
19 | #include "gfs2.h" | ||
20 | #include "incore.h" | ||
21 | #include "bmap.h" | ||
22 | #include "glock.h" | ||
23 | #include "inode.h" | ||
24 | #include "ops_vm.h" | ||
25 | #include "quota.h" | ||
26 | #include "rgrp.h" | ||
27 | #include "trans.h" | ||
28 | #include "util.h" | ||
29 | |||
30 | static int gfs2_private_fault(struct vm_area_struct *vma, struct vm_fault *vmf) | ||
31 | { | ||
32 | struct gfs2_inode *ip = GFS2_I(vma->vm_file->f_mapping->host); | ||
33 | |||
34 | set_bit(GIF_PAGED, &ip->i_flags); | ||
35 | return filemap_fault(vma, vmf); | ||
36 | } | ||
37 | |||
38 | static int alloc_page_backing(struct gfs2_inode *ip, struct page *page) | ||
39 | { | ||
40 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | ||
41 | unsigned long index = page->index; | ||
42 | u64 lblock = index << (PAGE_CACHE_SHIFT - | ||
43 | sdp->sd_sb.sb_bsize_shift); | ||
44 | unsigned int blocks = PAGE_CACHE_SIZE >> sdp->sd_sb.sb_bsize_shift; | ||
45 | struct gfs2_alloc *al; | ||
46 | unsigned int data_blocks, ind_blocks; | ||
47 | unsigned int x; | ||
48 | int error; | ||
49 | |||
50 | al = gfs2_alloc_get(ip); | ||
51 | |||
52 | error = gfs2_quota_lock(ip, NO_QUOTA_CHANGE, NO_QUOTA_CHANGE); | ||
53 | if (error) | ||
54 | goto out; | ||
55 | |||
56 | error = gfs2_quota_check(ip, ip->i_inode.i_uid, ip->i_inode.i_gid); | ||
57 | if (error) | ||
58 | goto out_gunlock_q; | ||
59 | |||
60 | gfs2_write_calc_reserv(ip, PAGE_CACHE_SIZE, &data_blocks, &ind_blocks); | ||
61 | |||
62 | al->al_requested = data_blocks + ind_blocks; | ||
63 | |||
64 | error = gfs2_inplace_reserve(ip); | ||
65 | if (error) | ||
66 | goto out_gunlock_q; | ||
67 | |||
68 | error = gfs2_trans_begin(sdp, al->al_rgd->rd_length + | ||
69 | ind_blocks + RES_DINODE + | ||
70 | RES_STATFS + RES_QUOTA, 0); | ||
71 | if (error) | ||
72 | goto out_ipres; | ||
73 | |||
74 | if (gfs2_is_stuffed(ip)) { | ||
75 | error = gfs2_unstuff_dinode(ip, NULL); | ||
76 | if (error) | ||
77 | goto out_trans; | ||
78 | } | ||
79 | |||
80 | for (x = 0; x < blocks; ) { | ||
81 | u64 dblock; | ||
82 | unsigned int extlen; | ||
83 | int new = 1; | ||
84 | |||
85 | error = gfs2_extent_map(&ip->i_inode, lblock, &new, &dblock, &extlen); | ||
86 | if (error) | ||
87 | goto out_trans; | ||
88 | |||
89 | lblock += extlen; | ||
90 | x += extlen; | ||
91 | } | ||
92 | |||
93 | gfs2_assert_warn(sdp, al->al_alloced); | ||
94 | |||
95 | out_trans: | ||
96 | gfs2_trans_end(sdp); | ||
97 | out_ipres: | ||
98 | gfs2_inplace_release(ip); | ||
99 | out_gunlock_q: | ||
100 | gfs2_quota_unlock(ip); | ||
101 | out: | ||
102 | gfs2_alloc_put(ip); | ||
103 | return error; | ||
104 | } | ||
105 | |||
106 | static int gfs2_sharewrite_fault(struct vm_area_struct *vma, | ||
107 | struct vm_fault *vmf) | ||
108 | { | ||
109 | struct file *file = vma->vm_file; | ||
110 | struct gfs2_file *gf = file->private_data; | ||
111 | struct gfs2_inode *ip = GFS2_I(file->f_mapping->host); | ||
112 | struct gfs2_holder i_gh; | ||
113 | int alloc_required; | ||
114 | int error; | ||
115 | int ret = 0; | ||
116 | |||
117 | error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh); | ||
118 | if (error) | ||
119 | goto out; | ||
120 | |||
121 | set_bit(GIF_PAGED, &ip->i_flags); | ||
122 | set_bit(GIF_SW_PAGED, &ip->i_flags); | ||
123 | |||
124 | error = gfs2_write_alloc_required(ip, | ||
125 | (u64)vmf->pgoff << PAGE_CACHE_SHIFT, | ||
126 | PAGE_CACHE_SIZE, &alloc_required); | ||
127 | if (error) { | ||
128 | ret = VM_FAULT_OOM; /* XXX: are these right? */ | ||
129 | goto out_unlock; | ||
130 | } | ||
131 | |||
132 | set_bit(GFF_EXLOCK, &gf->f_flags); | ||
133 | ret = filemap_fault(vma, vmf); | ||
134 | clear_bit(GFF_EXLOCK, &gf->f_flags); | ||
135 | if (ret & VM_FAULT_ERROR) | ||
136 | goto out_unlock; | ||
137 | |||
138 | if (alloc_required) { | ||
139 | /* XXX: do we need to drop page lock around alloc_page_backing?*/ | ||
140 | error = alloc_page_backing(ip, vmf->page); | ||
141 | if (error) { | ||
142 | /* | ||
143 | * VM_FAULT_LOCKED should always be the case for | ||
144 | * filemap_fault, but it may not be in a future | ||
145 | * implementation. | ||
146 | */ | ||
147 | if (ret & VM_FAULT_LOCKED) | ||
148 | unlock_page(vmf->page); | ||
149 | page_cache_release(vmf->page); | ||
150 | ret = VM_FAULT_OOM; | ||
151 | goto out_unlock; | ||
152 | } | ||
153 | set_page_dirty(vmf->page); | ||
154 | } | ||
155 | |||
156 | out_unlock: | ||
157 | gfs2_glock_dq_uninit(&i_gh); | ||
158 | out: | ||
159 | return ret; | ||
160 | } | ||
161 | |||
162 | struct vm_operations_struct gfs2_vm_ops_private = { | ||
163 | .fault = gfs2_private_fault, | ||
164 | }; | ||
165 | |||
166 | struct vm_operations_struct gfs2_vm_ops_sharewrite = { | ||
167 | .fault = gfs2_sharewrite_fault, | ||
168 | }; | ||
169 | |||
diff --git a/fs/gfs2/ops_vm.h b/fs/gfs2/ops_vm.h deleted file mode 100644 index 4ae8f43ed5e..00000000000 --- a/fs/gfs2/ops_vm.h +++ /dev/null | |||
@@ -1,18 +0,0 @@ | |||
1 | /* | ||
2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | ||
3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | ||
4 | * | ||
5 | * This copyrighted material is made available to anyone wishing to use, | ||
6 | * modify, copy, or redistribute it subject to the terms and conditions | ||
7 | * of the GNU General Public License version 2. | ||
8 | */ | ||
9 | |||
10 | #ifndef __OPS_VM_DOT_H__ | ||
11 | #define __OPS_VM_DOT_H__ | ||
12 | |||
13 | #include <linux/mm.h> | ||
14 | |||
15 | extern struct vm_operations_struct gfs2_vm_ops_private; | ||
16 | extern struct vm_operations_struct gfs2_vm_ops_sharewrite; | ||
17 | |||
18 | #endif /* __OPS_VM_DOT_H__ */ | ||
diff --git a/fs/gfs2/quota.c b/fs/gfs2/quota.c index addb51e0f13..a08dabd6ce9 100644 --- a/fs/gfs2/quota.c +++ b/fs/gfs2/quota.c | |||
@@ -1,6 +1,6 @@ | |||
1 | /* | 1 | /* |
2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. |
3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | 3 | * Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved. |
4 | * | 4 | * |
5 | * This copyrighted material is made available to anyone wishing to use, | 5 | * This copyrighted material is made available to anyone wishing to use, |
6 | * modify, copy, or redistribute it subject to the terms and conditions | 6 | * modify, copy, or redistribute it subject to the terms and conditions |
@@ -59,7 +59,6 @@ | |||
59 | #include "super.h" | 59 | #include "super.h" |
60 | #include "trans.h" | 60 | #include "trans.h" |
61 | #include "inode.h" | 61 | #include "inode.h" |
62 | #include "ops_file.h" | ||
63 | #include "ops_address.h" | 62 | #include "ops_address.h" |
64 | #include "util.h" | 63 | #include "util.h" |
65 | 64 | ||
@@ -274,10 +273,10 @@ static int bh_get(struct gfs2_quota_data *qd) | |||
274 | } | 273 | } |
275 | 274 | ||
276 | block = qd->qd_slot / sdp->sd_qc_per_block; | 275 | block = qd->qd_slot / sdp->sd_qc_per_block; |
277 | offset = qd->qd_slot % sdp->sd_qc_per_block;; | 276 | offset = qd->qd_slot % sdp->sd_qc_per_block; |
278 | 277 | ||
279 | bh_map.b_size = 1 << ip->i_inode.i_blkbits; | 278 | bh_map.b_size = 1 << ip->i_inode.i_blkbits; |
280 | error = gfs2_block_map(&ip->i_inode, block, 0, &bh_map); | 279 | error = gfs2_block_map(&ip->i_inode, block, &bh_map, 0); |
281 | if (error) | 280 | if (error) |
282 | goto fail; | 281 | goto fail; |
283 | error = gfs2_meta_read(ip->i_gl, bh_map.b_blocknr, DIO_WAIT, &bh); | 282 | error = gfs2_meta_read(ip->i_gl, bh_map.b_blocknr, DIO_WAIT, &bh); |
@@ -454,7 +453,7 @@ static void qdsb_put(struct gfs2_quota_data *qd) | |||
454 | int gfs2_quota_hold(struct gfs2_inode *ip, u32 uid, u32 gid) | 453 | int gfs2_quota_hold(struct gfs2_inode *ip, u32 uid, u32 gid) |
455 | { | 454 | { |
456 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 455 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
457 | struct gfs2_alloc *al = &ip->i_alloc; | 456 | struct gfs2_alloc *al = ip->i_alloc; |
458 | struct gfs2_quota_data **qd = al->al_qd; | 457 | struct gfs2_quota_data **qd = al->al_qd; |
459 | int error; | 458 | int error; |
460 | 459 | ||
@@ -502,7 +501,7 @@ out: | |||
502 | void gfs2_quota_unhold(struct gfs2_inode *ip) | 501 | void gfs2_quota_unhold(struct gfs2_inode *ip) |
503 | { | 502 | { |
504 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 503 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
505 | struct gfs2_alloc *al = &ip->i_alloc; | 504 | struct gfs2_alloc *al = ip->i_alloc; |
506 | unsigned int x; | 505 | unsigned int x; |
507 | 506 | ||
508 | gfs2_assert_warn(sdp, !test_bit(GIF_QD_LOCKED, &ip->i_flags)); | 507 | gfs2_assert_warn(sdp, !test_bit(GIF_QD_LOCKED, &ip->i_flags)); |
@@ -646,7 +645,7 @@ static int gfs2_adjust_quota(struct gfs2_inode *ip, loff_t loc, | |||
646 | } | 645 | } |
647 | 646 | ||
648 | if (!buffer_mapped(bh)) { | 647 | if (!buffer_mapped(bh)) { |
649 | gfs2_get_block(inode, iblock, bh, 1); | 648 | gfs2_block_map(inode, iblock, bh, 1); |
650 | if (!buffer_mapped(bh)) | 649 | if (!buffer_mapped(bh)) |
651 | goto unlock; | 650 | goto unlock; |
652 | } | 651 | } |
@@ -793,11 +792,9 @@ static int do_glock(struct gfs2_quota_data *qd, int force_refresh, | |||
793 | struct gfs2_holder i_gh; | 792 | struct gfs2_holder i_gh; |
794 | struct gfs2_quota_host q; | 793 | struct gfs2_quota_host q; |
795 | char buf[sizeof(struct gfs2_quota)]; | 794 | char buf[sizeof(struct gfs2_quota)]; |
796 | struct file_ra_state ra_state; | ||
797 | int error; | 795 | int error; |
798 | struct gfs2_quota_lvb *qlvb; | 796 | struct gfs2_quota_lvb *qlvb; |
799 | 797 | ||
800 | file_ra_state_init(&ra_state, sdp->sd_quota_inode->i_mapping); | ||
801 | restart: | 798 | restart: |
802 | error = gfs2_glock_nq_init(qd->qd_gl, LM_ST_SHARED, 0, q_gh); | 799 | error = gfs2_glock_nq_init(qd->qd_gl, LM_ST_SHARED, 0, q_gh); |
803 | if (error) | 800 | if (error) |
@@ -820,8 +817,8 @@ restart: | |||
820 | 817 | ||
821 | memset(buf, 0, sizeof(struct gfs2_quota)); | 818 | memset(buf, 0, sizeof(struct gfs2_quota)); |
822 | pos = qd2offset(qd); | 819 | pos = qd2offset(qd); |
823 | error = gfs2_internal_read(ip, &ra_state, buf, | 820 | error = gfs2_internal_read(ip, NULL, buf, &pos, |
824 | &pos, sizeof(struct gfs2_quota)); | 821 | sizeof(struct gfs2_quota)); |
825 | if (error < 0) | 822 | if (error < 0) |
826 | goto fail_gunlock; | 823 | goto fail_gunlock; |
827 | 824 | ||
@@ -856,7 +853,7 @@ fail: | |||
856 | int gfs2_quota_lock(struct gfs2_inode *ip, u32 uid, u32 gid) | 853 | int gfs2_quota_lock(struct gfs2_inode *ip, u32 uid, u32 gid) |
857 | { | 854 | { |
858 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 855 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
859 | struct gfs2_alloc *al = &ip->i_alloc; | 856 | struct gfs2_alloc *al = ip->i_alloc; |
860 | unsigned int x; | 857 | unsigned int x; |
861 | int error = 0; | 858 | int error = 0; |
862 | 859 | ||
@@ -924,7 +921,7 @@ static int need_sync(struct gfs2_quota_data *qd) | |||
924 | 921 | ||
925 | void gfs2_quota_unlock(struct gfs2_inode *ip) | 922 | void gfs2_quota_unlock(struct gfs2_inode *ip) |
926 | { | 923 | { |
927 | struct gfs2_alloc *al = &ip->i_alloc; | 924 | struct gfs2_alloc *al = ip->i_alloc; |
928 | struct gfs2_quota_data *qda[4]; | 925 | struct gfs2_quota_data *qda[4]; |
929 | unsigned int count = 0; | 926 | unsigned int count = 0; |
930 | unsigned int x; | 927 | unsigned int x; |
@@ -972,7 +969,7 @@ static int print_message(struct gfs2_quota_data *qd, char *type) | |||
972 | int gfs2_quota_check(struct gfs2_inode *ip, u32 uid, u32 gid) | 969 | int gfs2_quota_check(struct gfs2_inode *ip, u32 uid, u32 gid) |
973 | { | 970 | { |
974 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 971 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
975 | struct gfs2_alloc *al = &ip->i_alloc; | 972 | struct gfs2_alloc *al = ip->i_alloc; |
976 | struct gfs2_quota_data *qd; | 973 | struct gfs2_quota_data *qd; |
977 | s64 value; | 974 | s64 value; |
978 | unsigned int x; | 975 | unsigned int x; |
@@ -1016,10 +1013,9 @@ int gfs2_quota_check(struct gfs2_inode *ip, u32 uid, u32 gid) | |||
1016 | void gfs2_quota_change(struct gfs2_inode *ip, s64 change, | 1013 | void gfs2_quota_change(struct gfs2_inode *ip, s64 change, |
1017 | u32 uid, u32 gid) | 1014 | u32 uid, u32 gid) |
1018 | { | 1015 | { |
1019 | struct gfs2_alloc *al = &ip->i_alloc; | 1016 | struct gfs2_alloc *al = ip->i_alloc; |
1020 | struct gfs2_quota_data *qd; | 1017 | struct gfs2_quota_data *qd; |
1021 | unsigned int x; | 1018 | unsigned int x; |
1022 | unsigned int found = 0; | ||
1023 | 1019 | ||
1024 | if (gfs2_assert_warn(GFS2_SB(&ip->i_inode), change)) | 1020 | if (gfs2_assert_warn(GFS2_SB(&ip->i_inode), change)) |
1025 | return; | 1021 | return; |
@@ -1032,7 +1028,6 @@ void gfs2_quota_change(struct gfs2_inode *ip, s64 change, | |||
1032 | if ((qd->qd_id == uid && test_bit(QDF_USER, &qd->qd_flags)) || | 1028 | if ((qd->qd_id == uid && test_bit(QDF_USER, &qd->qd_flags)) || |
1033 | (qd->qd_id == gid && !test_bit(QDF_USER, &qd->qd_flags))) { | 1029 | (qd->qd_id == gid && !test_bit(QDF_USER, &qd->qd_flags))) { |
1034 | do_qc(qd, change); | 1030 | do_qc(qd, change); |
1035 | found++; | ||
1036 | } | 1031 | } |
1037 | } | 1032 | } |
1038 | } | 1033 | } |
diff --git a/fs/gfs2/recovery.c b/fs/gfs2/recovery.c index beb6c7ac008..b249e294a95 100644 --- a/fs/gfs2/recovery.c +++ b/fs/gfs2/recovery.c | |||
@@ -391,7 +391,7 @@ static int clean_journal(struct gfs2_jdesc *jd, struct gfs2_log_header_host *hea | |||
391 | lblock = head->lh_blkno; | 391 | lblock = head->lh_blkno; |
392 | gfs2_replay_incr_blk(sdp, &lblock); | 392 | gfs2_replay_incr_blk(sdp, &lblock); |
393 | bh_map.b_size = 1 << ip->i_inode.i_blkbits; | 393 | bh_map.b_size = 1 << ip->i_inode.i_blkbits; |
394 | error = gfs2_block_map(&ip->i_inode, lblock, 0, &bh_map); | 394 | error = gfs2_block_map(&ip->i_inode, lblock, &bh_map, 0); |
395 | if (error) | 395 | if (error) |
396 | return error; | 396 | return error; |
397 | if (!bh_map.b_blocknr) { | 397 | if (!bh_map.b_blocknr) { |
@@ -504,13 +504,21 @@ int gfs2_recover_journal(struct gfs2_jdesc *jd) | |||
504 | if (!test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags)) | 504 | if (!test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags)) |
505 | ro = 1; | 505 | ro = 1; |
506 | } else { | 506 | } else { |
507 | if (sdp->sd_vfs->s_flags & MS_RDONLY) | 507 | if (sdp->sd_vfs->s_flags & MS_RDONLY) { |
508 | ro = 1; | 508 | /* check if device itself is read-only */ |
509 | ro = bdev_read_only(sdp->sd_vfs->s_bdev); | ||
510 | if (!ro) { | ||
511 | fs_info(sdp, "recovery required on " | ||
512 | "read-only filesystem.\n"); | ||
513 | fs_info(sdp, "write access will be " | ||
514 | "enabled during recovery.\n"); | ||
515 | } | ||
516 | } | ||
509 | } | 517 | } |
510 | 518 | ||
511 | if (ro) { | 519 | if (ro) { |
512 | fs_warn(sdp, "jid=%u: Can't replay: read-only FS\n", | 520 | fs_warn(sdp, "jid=%u: Can't replay: read-only block " |
513 | jd->jd_jid); | 521 | "device\n", jd->jd_jid); |
514 | error = -EROFS; | 522 | error = -EROFS; |
515 | goto fail_gunlock_tr; | 523 | goto fail_gunlock_tr; |
516 | } | 524 | } |
diff --git a/fs/gfs2/rgrp.c b/fs/gfs2/rgrp.c index 708c287e1d0..3552110b2e5 100644 --- a/fs/gfs2/rgrp.c +++ b/fs/gfs2/rgrp.c | |||
@@ -25,10 +25,10 @@ | |||
25 | #include "rgrp.h" | 25 | #include "rgrp.h" |
26 | #include "super.h" | 26 | #include "super.h" |
27 | #include "trans.h" | 27 | #include "trans.h" |
28 | #include "ops_file.h" | ||
29 | #include "util.h" | 28 | #include "util.h" |
30 | #include "log.h" | 29 | #include "log.h" |
31 | #include "inode.h" | 30 | #include "inode.h" |
31 | #include "ops_address.h" | ||
32 | 32 | ||
33 | #define BFITNOENT ((u32)~0) | 33 | #define BFITNOENT ((u32)~0) |
34 | #define NO_BLOCK ((u64)~0) | 34 | #define NO_BLOCK ((u64)~0) |
@@ -126,41 +126,43 @@ static unsigned char gfs2_testbit(struct gfs2_rgrpd *rgd, unsigned char *buffer, | |||
126 | * Return: the block number (bitmap buffer scope) that was found | 126 | * Return: the block number (bitmap buffer scope) that was found |
127 | */ | 127 | */ |
128 | 128 | ||
129 | static u32 gfs2_bitfit(struct gfs2_rgrpd *rgd, unsigned char *buffer, | 129 | static u32 gfs2_bitfit(unsigned char *buffer, unsigned int buflen, u32 goal, |
130 | unsigned int buflen, u32 goal, | 130 | unsigned char old_state) |
131 | unsigned char old_state) | ||
132 | { | 131 | { |
133 | unsigned char *byte, *end, alloc; | 132 | unsigned char *byte; |
134 | u32 blk = goal; | 133 | u32 blk = goal; |
135 | unsigned int bit; | 134 | unsigned int bit, bitlong; |
135 | unsigned long *plong, plong55; | ||
136 | 136 | ||
137 | byte = buffer + (goal / GFS2_NBBY); | 137 | byte = buffer + (goal / GFS2_NBBY); |
138 | plong = (unsigned long *)(buffer + (goal / GFS2_NBBY)); | ||
138 | bit = (goal % GFS2_NBBY) * GFS2_BIT_SIZE; | 139 | bit = (goal % GFS2_NBBY) * GFS2_BIT_SIZE; |
139 | end = buffer + buflen; | 140 | bitlong = bit; |
140 | alloc = (old_state == GFS2_BLKST_FREE) ? 0x55 : 0; | 141 | #if BITS_PER_LONG == 32 |
141 | 142 | plong55 = 0x55555555; | |
142 | while (byte < end) { | 143 | #else |
143 | /* If we're looking for a free block we can eliminate all | 144 | plong55 = 0x5555555555555555; |
144 | bitmap settings with 0x55, which represents four data | 145 | #endif |
145 | blocks in a row. If we're looking for a data block, we can | 146 | while (byte < buffer + buflen) { |
146 | eliminate 0x00 which corresponds to four free blocks. */ | 147 | |
147 | if ((*byte & 0x55) == alloc) { | 148 | if (bitlong == 0 && old_state == 0 && *plong == plong55) { |
148 | blk += (8 - bit) >> 1; | 149 | plong++; |
149 | 150 | byte += sizeof(unsigned long); | |
150 | bit = 0; | 151 | blk += sizeof(unsigned long) * GFS2_NBBY; |
151 | byte++; | ||
152 | |||
153 | continue; | 152 | continue; |
154 | } | 153 | } |
155 | |||
156 | if (((*byte >> bit) & GFS2_BIT_MASK) == old_state) | 154 | if (((*byte >> bit) & GFS2_BIT_MASK) == old_state) |
157 | return blk; | 155 | return blk; |
158 | |||
159 | bit += GFS2_BIT_SIZE; | 156 | bit += GFS2_BIT_SIZE; |
160 | if (bit >= 8) { | 157 | if (bit >= 8) { |
161 | bit = 0; | 158 | bit = 0; |
162 | byte++; | 159 | byte++; |
163 | } | 160 | } |
161 | bitlong += GFS2_BIT_SIZE; | ||
162 | if (bitlong >= sizeof(unsigned long) * 8) { | ||
163 | bitlong = 0; | ||
164 | plong++; | ||
165 | } | ||
164 | 166 | ||
165 | blk++; | 167 | blk++; |
166 | } | 168 | } |
@@ -817,11 +819,9 @@ void gfs2_rgrp_repolish_clones(struct gfs2_rgrpd *rgd) | |||
817 | 819 | ||
818 | struct gfs2_alloc *gfs2_alloc_get(struct gfs2_inode *ip) | 820 | struct gfs2_alloc *gfs2_alloc_get(struct gfs2_inode *ip) |
819 | { | 821 | { |
820 | struct gfs2_alloc *al = &ip->i_alloc; | 822 | BUG_ON(ip->i_alloc != NULL); |
821 | 823 | ip->i_alloc = kzalloc(sizeof(struct gfs2_alloc), GFP_KERNEL); | |
822 | /* FIXME: Should assert that the correct locks are held here... */ | 824 | return ip->i_alloc; |
823 | memset(al, 0, sizeof(*al)); | ||
824 | return al; | ||
825 | } | 825 | } |
826 | 826 | ||
827 | /** | 827 | /** |
@@ -1059,26 +1059,34 @@ static struct inode *get_local_rgrp(struct gfs2_inode *ip, u64 *last_unlinked) | |||
1059 | struct inode *inode = NULL; | 1059 | struct inode *inode = NULL; |
1060 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 1060 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
1061 | struct gfs2_rgrpd *rgd, *begin = NULL; | 1061 | struct gfs2_rgrpd *rgd, *begin = NULL; |
1062 | struct gfs2_alloc *al = &ip->i_alloc; | 1062 | struct gfs2_alloc *al = ip->i_alloc; |
1063 | int flags = LM_FLAG_TRY; | 1063 | int flags = LM_FLAG_TRY; |
1064 | int skipped = 0; | 1064 | int skipped = 0; |
1065 | int loops = 0; | 1065 | int loops = 0; |
1066 | int error; | 1066 | int error, rg_locked; |
1067 | 1067 | ||
1068 | /* Try recently successful rgrps */ | 1068 | /* Try recently successful rgrps */ |
1069 | 1069 | ||
1070 | rgd = recent_rgrp_first(sdp, ip->i_last_rg_alloc); | 1070 | rgd = recent_rgrp_first(sdp, ip->i_last_rg_alloc); |
1071 | 1071 | ||
1072 | while (rgd) { | 1072 | while (rgd) { |
1073 | error = gfs2_glock_nq_init(rgd->rd_gl, LM_ST_EXCLUSIVE, | 1073 | rg_locked = 0; |
1074 | LM_FLAG_TRY, &al->al_rgd_gh); | 1074 | |
1075 | if (gfs2_glock_is_locked_by_me(rgd->rd_gl)) { | ||
1076 | rg_locked = 1; | ||
1077 | error = 0; | ||
1078 | } else { | ||
1079 | error = gfs2_glock_nq_init(rgd->rd_gl, LM_ST_EXCLUSIVE, | ||
1080 | LM_FLAG_TRY, &al->al_rgd_gh); | ||
1081 | } | ||
1075 | switch (error) { | 1082 | switch (error) { |
1076 | case 0: | 1083 | case 0: |
1077 | if (try_rgrp_fit(rgd, al)) | 1084 | if (try_rgrp_fit(rgd, al)) |
1078 | goto out; | 1085 | goto out; |
1079 | if (rgd->rd_flags & GFS2_RDF_CHECK) | 1086 | if (rgd->rd_flags & GFS2_RDF_CHECK) |
1080 | inode = try_rgrp_unlink(rgd, last_unlinked); | 1087 | inode = try_rgrp_unlink(rgd, last_unlinked); |
1081 | gfs2_glock_dq_uninit(&al->al_rgd_gh); | 1088 | if (!rg_locked) |
1089 | gfs2_glock_dq_uninit(&al->al_rgd_gh); | ||
1082 | if (inode) | 1090 | if (inode) |
1083 | return inode; | 1091 | return inode; |
1084 | rgd = recent_rgrp_next(rgd, 1); | 1092 | rgd = recent_rgrp_next(rgd, 1); |
@@ -1098,15 +1106,23 @@ static struct inode *get_local_rgrp(struct gfs2_inode *ip, u64 *last_unlinked) | |||
1098 | begin = rgd = forward_rgrp_get(sdp); | 1106 | begin = rgd = forward_rgrp_get(sdp); |
1099 | 1107 | ||
1100 | for (;;) { | 1108 | for (;;) { |
1101 | error = gfs2_glock_nq_init(rgd->rd_gl, LM_ST_EXCLUSIVE, flags, | 1109 | rg_locked = 0; |
1102 | &al->al_rgd_gh); | 1110 | |
1111 | if (gfs2_glock_is_locked_by_me(rgd->rd_gl)) { | ||
1112 | rg_locked = 1; | ||
1113 | error = 0; | ||
1114 | } else { | ||
1115 | error = gfs2_glock_nq_init(rgd->rd_gl, LM_ST_EXCLUSIVE, flags, | ||
1116 | &al->al_rgd_gh); | ||
1117 | } | ||
1103 | switch (error) { | 1118 | switch (error) { |
1104 | case 0: | 1119 | case 0: |
1105 | if (try_rgrp_fit(rgd, al)) | 1120 | if (try_rgrp_fit(rgd, al)) |
1106 | goto out; | 1121 | goto out; |
1107 | if (rgd->rd_flags & GFS2_RDF_CHECK) | 1122 | if (rgd->rd_flags & GFS2_RDF_CHECK) |
1108 | inode = try_rgrp_unlink(rgd, last_unlinked); | 1123 | inode = try_rgrp_unlink(rgd, last_unlinked); |
1109 | gfs2_glock_dq_uninit(&al->al_rgd_gh); | 1124 | if (!rg_locked) |
1125 | gfs2_glock_dq_uninit(&al->al_rgd_gh); | ||
1110 | if (inode) | 1126 | if (inode) |
1111 | return inode; | 1127 | return inode; |
1112 | break; | 1128 | break; |
@@ -1158,7 +1174,7 @@ out: | |||
1158 | int gfs2_inplace_reserve_i(struct gfs2_inode *ip, char *file, unsigned int line) | 1174 | int gfs2_inplace_reserve_i(struct gfs2_inode *ip, char *file, unsigned int line) |
1159 | { | 1175 | { |
1160 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 1176 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
1161 | struct gfs2_alloc *al = &ip->i_alloc; | 1177 | struct gfs2_alloc *al = ip->i_alloc; |
1162 | struct inode *inode; | 1178 | struct inode *inode; |
1163 | int error = 0; | 1179 | int error = 0; |
1164 | u64 last_unlinked = NO_BLOCK; | 1180 | u64 last_unlinked = NO_BLOCK; |
@@ -1204,7 +1220,7 @@ try_again: | |||
1204 | void gfs2_inplace_release(struct gfs2_inode *ip) | 1220 | void gfs2_inplace_release(struct gfs2_inode *ip) |
1205 | { | 1221 | { |
1206 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 1222 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
1207 | struct gfs2_alloc *al = &ip->i_alloc; | 1223 | struct gfs2_alloc *al = ip->i_alloc; |
1208 | 1224 | ||
1209 | if (gfs2_assert_warn(sdp, al->al_alloced <= al->al_requested) == -1) | 1225 | if (gfs2_assert_warn(sdp, al->al_alloced <= al->al_requested) == -1) |
1210 | fs_warn(sdp, "al_alloced = %u, al_requested = %u " | 1226 | fs_warn(sdp, "al_alloced = %u, al_requested = %u " |
@@ -1213,7 +1229,8 @@ void gfs2_inplace_release(struct gfs2_inode *ip) | |||
1213 | al->al_line); | 1229 | al->al_line); |
1214 | 1230 | ||
1215 | al->al_rgd = NULL; | 1231 | al->al_rgd = NULL; |
1216 | gfs2_glock_dq_uninit(&al->al_rgd_gh); | 1232 | if (al->al_rgd_gh.gh_gl) |
1233 | gfs2_glock_dq_uninit(&al->al_rgd_gh); | ||
1217 | if (ip != GFS2_I(sdp->sd_rindex)) | 1234 | if (ip != GFS2_I(sdp->sd_rindex)) |
1218 | gfs2_glock_dq_uninit(&al->al_ri_gh); | 1235 | gfs2_glock_dq_uninit(&al->al_ri_gh); |
1219 | } | 1236 | } |
@@ -1301,11 +1318,10 @@ static u32 rgblk_search(struct gfs2_rgrpd *rgd, u32 goal, | |||
1301 | /* The GFS2_BLKST_UNLINKED state doesn't apply to the clone | 1318 | /* The GFS2_BLKST_UNLINKED state doesn't apply to the clone |
1302 | bitmaps, so we must search the originals for that. */ | 1319 | bitmaps, so we must search the originals for that. */ |
1303 | if (old_state != GFS2_BLKST_UNLINKED && bi->bi_clone) | 1320 | if (old_state != GFS2_BLKST_UNLINKED && bi->bi_clone) |
1304 | blk = gfs2_bitfit(rgd, bi->bi_clone + bi->bi_offset, | 1321 | blk = gfs2_bitfit(bi->bi_clone + bi->bi_offset, |
1305 | bi->bi_len, goal, old_state); | 1322 | bi->bi_len, goal, old_state); |
1306 | else | 1323 | else |
1307 | blk = gfs2_bitfit(rgd, | 1324 | blk = gfs2_bitfit(bi->bi_bh->b_data + bi->bi_offset, |
1308 | bi->bi_bh->b_data + bi->bi_offset, | ||
1309 | bi->bi_len, goal, old_state); | 1325 | bi->bi_len, goal, old_state); |
1310 | if (blk != BFITNOENT) | 1326 | if (blk != BFITNOENT) |
1311 | break; | 1327 | break; |
@@ -1394,7 +1410,7 @@ static struct gfs2_rgrpd *rgblk_free(struct gfs2_sbd *sdp, u64 bstart, | |||
1394 | u64 gfs2_alloc_data(struct gfs2_inode *ip) | 1410 | u64 gfs2_alloc_data(struct gfs2_inode *ip) |
1395 | { | 1411 | { |
1396 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 1412 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
1397 | struct gfs2_alloc *al = &ip->i_alloc; | 1413 | struct gfs2_alloc *al = ip->i_alloc; |
1398 | struct gfs2_rgrpd *rgd = al->al_rgd; | 1414 | struct gfs2_rgrpd *rgd = al->al_rgd; |
1399 | u32 goal, blk; | 1415 | u32 goal, blk; |
1400 | u64 block; | 1416 | u64 block; |
@@ -1439,7 +1455,7 @@ u64 gfs2_alloc_data(struct gfs2_inode *ip) | |||
1439 | u64 gfs2_alloc_meta(struct gfs2_inode *ip) | 1455 | u64 gfs2_alloc_meta(struct gfs2_inode *ip) |
1440 | { | 1456 | { |
1441 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); | 1457 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
1442 | struct gfs2_alloc *al = &ip->i_alloc; | 1458 | struct gfs2_alloc *al = ip->i_alloc; |
1443 | struct gfs2_rgrpd *rgd = al->al_rgd; | 1459 | struct gfs2_rgrpd *rgd = al->al_rgd; |
1444 | u32 goal, blk; | 1460 | u32 goal, blk; |
1445 | u64 block; | 1461 | u64 block; |
@@ -1485,7 +1501,7 @@ u64 gfs2_alloc_meta(struct gfs2_inode *ip) | |||
1485 | u64 gfs2_alloc_di(struct gfs2_inode *dip, u64 *generation) | 1501 | u64 gfs2_alloc_di(struct gfs2_inode *dip, u64 *generation) |
1486 | { | 1502 | { |
1487 | struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); | 1503 | struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); |
1488 | struct gfs2_alloc *al = &dip->i_alloc; | 1504 | struct gfs2_alloc *al = dip->i_alloc; |
1489 | struct gfs2_rgrpd *rgd = al->al_rgd; | 1505 | struct gfs2_rgrpd *rgd = al->al_rgd; |
1490 | u32 blk; | 1506 | u32 blk; |
1491 | u64 block; | 1507 | u64 block; |
diff --git a/fs/gfs2/rgrp.h b/fs/gfs2/rgrp.h index b4c6adfc6f2..149bb161f4b 100644 --- a/fs/gfs2/rgrp.h +++ b/fs/gfs2/rgrp.h | |||
@@ -32,7 +32,9 @@ void gfs2_rgrp_repolish_clones(struct gfs2_rgrpd *rgd); | |||
32 | struct gfs2_alloc *gfs2_alloc_get(struct gfs2_inode *ip); | 32 | struct gfs2_alloc *gfs2_alloc_get(struct gfs2_inode *ip); |
33 | static inline void gfs2_alloc_put(struct gfs2_inode *ip) | 33 | static inline void gfs2_alloc_put(struct gfs2_inode *ip) |
34 | { | 34 | { |
35 | return; /* So we can see where ip->i_alloc is used */ | 35 | BUG_ON(ip->i_alloc == NULL); |
36 | kfree(ip->i_alloc); | ||
37 | ip->i_alloc = NULL; | ||
36 | } | 38 | } |
37 | 39 | ||
38 | int gfs2_inplace_reserve_i(struct gfs2_inode *ip, | 40 | int gfs2_inplace_reserve_i(struct gfs2_inode *ip, |
diff --git a/fs/gfs2/super.c b/fs/gfs2/super.c index dd3e737f528..ef0562c3bc7 100644 --- a/fs/gfs2/super.c +++ b/fs/gfs2/super.c | |||
@@ -1,6 +1,6 @@ | |||
1 | /* | 1 | /* |
2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | 2 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. |
3 | * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. | 3 | * Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved. |
4 | * | 4 | * |
5 | * This copyrighted material is made available to anyone wishing to use, | 5 | * This copyrighted material is made available to anyone wishing to use, |
6 | * modify, copy, or redistribute it subject to the terms and conditions | 6 | * modify, copy, or redistribute it subject to the terms and conditions |
@@ -51,13 +51,9 @@ void gfs2_tune_init(struct gfs2_tune *gt) | |||
51 | { | 51 | { |
52 | spin_lock_init(>->gt_spin); | 52 | spin_lock_init(>->gt_spin); |
53 | 53 | ||
54 | gt->gt_ilimit = 100; | ||
55 | gt->gt_ilimit_tries = 3; | ||
56 | gt->gt_ilimit_min = 1; | ||
57 | gt->gt_demote_secs = 300; | 54 | gt->gt_demote_secs = 300; |
58 | gt->gt_incore_log_blocks = 1024; | 55 | gt->gt_incore_log_blocks = 1024; |
59 | gt->gt_log_flush_secs = 60; | 56 | gt->gt_log_flush_secs = 60; |
60 | gt->gt_jindex_refresh_secs = 60; | ||
61 | gt->gt_recoverd_secs = 60; | 57 | gt->gt_recoverd_secs = 60; |
62 | gt->gt_logd_secs = 1; | 58 | gt->gt_logd_secs = 1; |
63 | gt->gt_quotad_secs = 5; | 59 | gt->gt_quotad_secs = 5; |
@@ -71,10 +67,8 @@ void gfs2_tune_init(struct gfs2_tune *gt) | |||
71 | gt->gt_new_files_jdata = 0; | 67 | gt->gt_new_files_jdata = 0; |
72 | gt->gt_new_files_directio = 0; | 68 | gt->gt_new_files_directio = 0; |
73 | gt->gt_max_readahead = 1 << 18; | 69 | gt->gt_max_readahead = 1 << 18; |
74 | gt->gt_lockdump_size = 131072; | ||
75 | gt->gt_stall_secs = 600; | 70 | gt->gt_stall_secs = 600; |
76 | gt->gt_complain_secs = 10; | 71 | gt->gt_complain_secs = 10; |
77 | gt->gt_reclaim_limit = 5000; | ||
78 | gt->gt_statfs_quantum = 30; | 72 | gt->gt_statfs_quantum = 30; |
79 | gt->gt_statfs_slow = 0; | 73 | gt->gt_statfs_slow = 0; |
80 | } | 74 | } |
@@ -393,6 +387,7 @@ int gfs2_jindex_hold(struct gfs2_sbd *sdp, struct gfs2_holder *ji_gh) | |||
393 | if (!jd) | 387 | if (!jd) |
394 | break; | 388 | break; |
395 | 389 | ||
390 | INIT_LIST_HEAD(&jd->extent_list); | ||
396 | jd->jd_inode = gfs2_lookupi(sdp->sd_jindex, &name, 1, NULL); | 391 | jd->jd_inode = gfs2_lookupi(sdp->sd_jindex, &name, 1, NULL); |
397 | if (!jd->jd_inode || IS_ERR(jd->jd_inode)) { | 392 | if (!jd->jd_inode || IS_ERR(jd->jd_inode)) { |
398 | if (!jd->jd_inode) | 393 | if (!jd->jd_inode) |
@@ -422,8 +417,9 @@ int gfs2_jindex_hold(struct gfs2_sbd *sdp, struct gfs2_holder *ji_gh) | |||
422 | 417 | ||
423 | void gfs2_jindex_free(struct gfs2_sbd *sdp) | 418 | void gfs2_jindex_free(struct gfs2_sbd *sdp) |
424 | { | 419 | { |
425 | struct list_head list; | 420 | struct list_head list, *head; |
426 | struct gfs2_jdesc *jd; | 421 | struct gfs2_jdesc *jd; |
422 | struct gfs2_journal_extent *jext; | ||
427 | 423 | ||
428 | spin_lock(&sdp->sd_jindex_spin); | 424 | spin_lock(&sdp->sd_jindex_spin); |
429 | list_add(&list, &sdp->sd_jindex_list); | 425 | list_add(&list, &sdp->sd_jindex_list); |
@@ -433,6 +429,14 @@ void gfs2_jindex_free(struct gfs2_sbd *sdp) | |||
433 | 429 | ||
434 | while (!list_empty(&list)) { | 430 | while (!list_empty(&list)) { |
435 | jd = list_entry(list.next, struct gfs2_jdesc, jd_list); | 431 | jd = list_entry(list.next, struct gfs2_jdesc, jd_list); |
432 | head = &jd->extent_list; | ||
433 | while (!list_empty(head)) { | ||
434 | jext = list_entry(head->next, | ||
435 | struct gfs2_journal_extent, | ||
436 | extent_list); | ||
437 | list_del(&jext->extent_list); | ||
438 | kfree(jext); | ||
439 | } | ||
436 | list_del(&jd->jd_list); | 440 | list_del(&jd->jd_list); |
437 | iput(jd->jd_inode); | 441 | iput(jd->jd_inode); |
438 | kfree(jd); | 442 | kfree(jd); |
@@ -543,7 +547,6 @@ int gfs2_make_fs_rw(struct gfs2_sbd *sdp) | |||
543 | if (error) | 547 | if (error) |
544 | return error; | 548 | return error; |
545 | 549 | ||
546 | gfs2_meta_cache_flush(ip); | ||
547 | j_gl->gl_ops->go_inval(j_gl, DIO_METADATA); | 550 | j_gl->gl_ops->go_inval(j_gl, DIO_METADATA); |
548 | 551 | ||
549 | error = gfs2_find_jhead(sdp->sd_jdesc, &head); | 552 | error = gfs2_find_jhead(sdp->sd_jdesc, &head); |
@@ -686,9 +689,7 @@ void gfs2_statfs_change(struct gfs2_sbd *sdp, s64 total, s64 free, | |||
686 | if (error) | 689 | if (error) |
687 | return; | 690 | return; |
688 | 691 | ||
689 | mutex_lock(&sdp->sd_statfs_mutex); | ||
690 | gfs2_trans_add_bh(l_ip->i_gl, l_bh, 1); | 692 | gfs2_trans_add_bh(l_ip->i_gl, l_bh, 1); |
691 | mutex_unlock(&sdp->sd_statfs_mutex); | ||
692 | 693 | ||
693 | spin_lock(&sdp->sd_statfs_spin); | 694 | spin_lock(&sdp->sd_statfs_spin); |
694 | l_sc->sc_total += total; | 695 | l_sc->sc_total += total; |
@@ -736,9 +737,7 @@ int gfs2_statfs_sync(struct gfs2_sbd *sdp) | |||
736 | if (error) | 737 | if (error) |
737 | goto out_bh2; | 738 | goto out_bh2; |
738 | 739 | ||
739 | mutex_lock(&sdp->sd_statfs_mutex); | ||
740 | gfs2_trans_add_bh(l_ip->i_gl, l_bh, 1); | 740 | gfs2_trans_add_bh(l_ip->i_gl, l_bh, 1); |
741 | mutex_unlock(&sdp->sd_statfs_mutex); | ||
742 | 741 | ||
743 | spin_lock(&sdp->sd_statfs_spin); | 742 | spin_lock(&sdp->sd_statfs_spin); |
744 | m_sc->sc_total += l_sc->sc_total; | 743 | m_sc->sc_total += l_sc->sc_total; |
diff --git a/fs/gfs2/sys.c b/fs/gfs2/sys.c index 06e0b7768d9..eaa3b7b2f99 100644 --- a/fs/gfs2/sys.c +++ b/fs/gfs2/sys.c | |||
@@ -32,7 +32,8 @@ spinlock_t gfs2_sys_margs_lock; | |||
32 | 32 | ||
33 | static ssize_t id_show(struct gfs2_sbd *sdp, char *buf) | 33 | static ssize_t id_show(struct gfs2_sbd *sdp, char *buf) |
34 | { | 34 | { |
35 | return snprintf(buf, PAGE_SIZE, "%s\n", sdp->sd_vfs->s_id); | 35 | return snprintf(buf, PAGE_SIZE, "%u:%u\n", |
36 | MAJOR(sdp->sd_vfs->s_dev), MINOR(sdp->sd_vfs->s_dev)); | ||
36 | } | 37 | } |
37 | 38 | ||
38 | static ssize_t fsname_show(struct gfs2_sbd *sdp, char *buf) | 39 | static ssize_t fsname_show(struct gfs2_sbd *sdp, char *buf) |
@@ -221,9 +222,7 @@ static struct kobj_type gfs2_ktype = { | |||
221 | .sysfs_ops = &gfs2_attr_ops, | 222 | .sysfs_ops = &gfs2_attr_ops, |
222 | }; | 223 | }; |
223 | 224 | ||
224 | static struct kset gfs2_kset = { | 225 | static struct kset *gfs2_kset; |
225 | .ktype = &gfs2_ktype, | ||
226 | }; | ||
227 | 226 | ||
228 | /* | 227 | /* |
229 | * display struct lm_lockstruct fields | 228 | * display struct lm_lockstruct fields |
@@ -427,13 +426,11 @@ TUNE_ATTR_2(name, name##_store) | |||
427 | TUNE_ATTR(demote_secs, 0); | 426 | TUNE_ATTR(demote_secs, 0); |
428 | TUNE_ATTR(incore_log_blocks, 0); | 427 | TUNE_ATTR(incore_log_blocks, 0); |
429 | TUNE_ATTR(log_flush_secs, 0); | 428 | TUNE_ATTR(log_flush_secs, 0); |
430 | TUNE_ATTR(jindex_refresh_secs, 0); | ||
431 | TUNE_ATTR(quota_warn_period, 0); | 429 | TUNE_ATTR(quota_warn_period, 0); |
432 | TUNE_ATTR(quota_quantum, 0); | 430 | TUNE_ATTR(quota_quantum, 0); |
433 | TUNE_ATTR(atime_quantum, 0); | 431 | TUNE_ATTR(atime_quantum, 0); |
434 | TUNE_ATTR(max_readahead, 0); | 432 | TUNE_ATTR(max_readahead, 0); |
435 | TUNE_ATTR(complain_secs, 0); | 433 | TUNE_ATTR(complain_secs, 0); |
436 | TUNE_ATTR(reclaim_limit, 0); | ||
437 | TUNE_ATTR(statfs_slow, 0); | 434 | TUNE_ATTR(statfs_slow, 0); |
438 | TUNE_ATTR(new_files_jdata, 0); | 435 | TUNE_ATTR(new_files_jdata, 0); |
439 | TUNE_ATTR(new_files_directio, 0); | 436 | TUNE_ATTR(new_files_directio, 0); |
@@ -450,13 +447,11 @@ static struct attribute *tune_attrs[] = { | |||
450 | &tune_attr_demote_secs.attr, | 447 | &tune_attr_demote_secs.attr, |
451 | &tune_attr_incore_log_blocks.attr, | 448 | &tune_attr_incore_log_blocks.attr, |
452 | &tune_attr_log_flush_secs.attr, | 449 | &tune_attr_log_flush_secs.attr, |
453 | &tune_attr_jindex_refresh_secs.attr, | ||
454 | &tune_attr_quota_warn_period.attr, | 450 | &tune_attr_quota_warn_period.attr, |
455 | &tune_attr_quota_quantum.attr, | 451 | &tune_attr_quota_quantum.attr, |
456 | &tune_attr_atime_quantum.attr, | 452 | &tune_attr_atime_quantum.attr, |
457 | &tune_attr_max_readahead.attr, | 453 | &tune_attr_max_readahead.attr, |
458 | &tune_attr_complain_secs.attr, | 454 | &tune_attr_complain_secs.attr, |
459 | &tune_attr_reclaim_limit.attr, | ||
460 | &tune_attr_statfs_slow.attr, | 455 | &tune_attr_statfs_slow.attr, |
461 | &tune_attr_quota_simul_sync.attr, | 456 | &tune_attr_quota_simul_sync.attr, |
462 | &tune_attr_quota_cache_secs.attr, | 457 | &tune_attr_quota_cache_secs.attr, |
@@ -495,14 +490,9 @@ int gfs2_sys_fs_add(struct gfs2_sbd *sdp) | |||
495 | { | 490 | { |
496 | int error; | 491 | int error; |
497 | 492 | ||
498 | sdp->sd_kobj.kset = &gfs2_kset; | 493 | sdp->sd_kobj.kset = gfs2_kset; |
499 | sdp->sd_kobj.ktype = &gfs2_ktype; | 494 | error = kobject_init_and_add(&sdp->sd_kobj, &gfs2_ktype, NULL, |
500 | 495 | "%s", sdp->sd_table_name); | |
501 | error = kobject_set_name(&sdp->sd_kobj, "%s", sdp->sd_table_name); | ||
502 | if (error) | ||
503 | goto fail; | ||
504 | |||
505 | error = kobject_register(&sdp->sd_kobj); | ||
506 | if (error) | 496 | if (error) |
507 | goto fail; | 497 | goto fail; |
508 | 498 | ||
@@ -522,6 +512,7 @@ int gfs2_sys_fs_add(struct gfs2_sbd *sdp) | |||
522 | if (error) | 512 | if (error) |
523 | goto fail_args; | 513 | goto fail_args; |
524 | 514 | ||
515 | kobject_uevent(&sdp->sd_kobj, KOBJ_ADD); | ||
525 | return 0; | 516 | return 0; |
526 | 517 | ||
527 | fail_args: | 518 | fail_args: |
@@ -531,7 +522,7 @@ fail_counters: | |||
531 | fail_lockstruct: | 522 | fail_lockstruct: |
532 | sysfs_remove_group(&sdp->sd_kobj, &lockstruct_group); | 523 | sysfs_remove_group(&sdp->sd_kobj, &lockstruct_group); |
533 | fail_reg: | 524 | fail_reg: |
534 | kobject_unregister(&sdp->sd_kobj); | 525 | kobject_put(&sdp->sd_kobj); |
535 | fail: | 526 | fail: |
536 | fs_err(sdp, "error %d adding sysfs files", error); | 527 | fs_err(sdp, "error %d adding sysfs files", error); |
537 | return error; | 528 | return error; |
@@ -543,21 +534,22 @@ void gfs2_sys_fs_del(struct gfs2_sbd *sdp) | |||
543 | sysfs_remove_group(&sdp->sd_kobj, &args_group); | 534 | sysfs_remove_group(&sdp->sd_kobj, &args_group); |
544 | sysfs_remove_group(&sdp->sd_kobj, &counters_group); | 535 | sysfs_remove_group(&sdp->sd_kobj, &counters_group); |
545 | sysfs_remove_group(&sdp->sd_kobj, &lockstruct_group); | 536 | sysfs_remove_group(&sdp->sd_kobj, &lockstruct_group); |
546 | kobject_unregister(&sdp->sd_kobj); | 537 | kobject_put(&sdp->sd_kobj); |
547 | } | 538 | } |
548 | 539 | ||
549 | int gfs2_sys_init(void) | 540 | int gfs2_sys_init(void) |
550 | { | 541 | { |
551 | gfs2_sys_margs = NULL; | 542 | gfs2_sys_margs = NULL; |
552 | spin_lock_init(&gfs2_sys_margs_lock); | 543 | spin_lock_init(&gfs2_sys_margs_lock); |
553 | kobject_set_name(&gfs2_kset.kobj, "gfs2"); | 544 | gfs2_kset = kset_create_and_add("gfs2", NULL, fs_kobj); |
554 | kobj_set_kset_s(&gfs2_kset, fs_subsys); | 545 | if (!gfs2_kset) |
555 | return kset_register(&gfs2_kset); | 546 | return -ENOMEM; |
547 | return 0; | ||
556 | } | 548 | } |
557 | 549 | ||
558 | void gfs2_sys_uninit(void) | 550 | void gfs2_sys_uninit(void) |
559 | { | 551 | { |
560 | kfree(gfs2_sys_margs); | 552 | kfree(gfs2_sys_margs); |
561 | kset_unregister(&gfs2_kset); | 553 | kset_unregister(gfs2_kset); |
562 | } | 554 | } |
563 | 555 | ||
diff --git a/fs/gfs2/trans.c b/fs/gfs2/trans.c index 717983e2c2a..73e5d92a657 100644 --- a/fs/gfs2/trans.c +++ b/fs/gfs2/trans.c | |||
@@ -114,11 +114,6 @@ void gfs2_trans_end(struct gfs2_sbd *sdp) | |||
114 | gfs2_log_flush(sdp, NULL); | 114 | gfs2_log_flush(sdp, NULL); |
115 | } | 115 | } |
116 | 116 | ||
117 | void gfs2_trans_add_gl(struct gfs2_glock *gl) | ||
118 | { | ||
119 | lops_add(gl->gl_sbd, &gl->gl_le); | ||
120 | } | ||
121 | |||
122 | /** | 117 | /** |
123 | * gfs2_trans_add_bh - Add a to-be-modified buffer to the current transaction | 118 | * gfs2_trans_add_bh - Add a to-be-modified buffer to the current transaction |
124 | * @gl: the glock the buffer belongs to | 119 | * @gl: the glock the buffer belongs to |
diff --git a/fs/gfs2/trans.h b/fs/gfs2/trans.h index 043d5f4b9c4..e826f0dab80 100644 --- a/fs/gfs2/trans.h +++ b/fs/gfs2/trans.h | |||
@@ -30,7 +30,6 @@ int gfs2_trans_begin(struct gfs2_sbd *sdp, unsigned int blocks, | |||
30 | 30 | ||
31 | void gfs2_trans_end(struct gfs2_sbd *sdp); | 31 | void gfs2_trans_end(struct gfs2_sbd *sdp); |
32 | 32 | ||
33 | void gfs2_trans_add_gl(struct gfs2_glock *gl); | ||
34 | void gfs2_trans_add_bh(struct gfs2_glock *gl, struct buffer_head *bh, int meta); | 33 | void gfs2_trans_add_bh(struct gfs2_glock *gl, struct buffer_head *bh, int meta); |
35 | void gfs2_trans_add_revoke(struct gfs2_sbd *sdp, struct gfs2_bufdata *bd); | 34 | void gfs2_trans_add_revoke(struct gfs2_sbd *sdp, struct gfs2_bufdata *bd); |
36 | void gfs2_trans_add_unrevoke(struct gfs2_sbd *sdp, u64 blkno); | 35 | void gfs2_trans_add_unrevoke(struct gfs2_sbd *sdp, u64 blkno); |