diff options
Diffstat (limited to 'net')
-rw-r--r-- | net/ceph/messenger.c | 6 | ||||
-rw-r--r-- | net/ceph/pagelist.c | 4 | ||||
-rw-r--r-- | net/ceph/pagevec.c | 30 | ||||
-rw-r--r-- | net/sunrpc/auth_gss/auth_gss.c | 8 | ||||
-rw-r--r-- | net/sunrpc/auth_gss/gss_krb5_crypto.c | 2 | ||||
-rw-r--r-- | net/sunrpc/auth_gss/gss_krb5_wrap.c | 4 | ||||
-rw-r--r-- | net/sunrpc/cache.c | 4 | ||||
-rw-r--r-- | net/sunrpc/rpc_pipe.c | 4 | ||||
-rw-r--r-- | net/sunrpc/socklib.c | 6 | ||||
-rw-r--r-- | net/sunrpc/xdr.c | 48 |
10 files changed, 58 insertions, 58 deletions
diff --git a/net/ceph/messenger.c b/net/ceph/messenger.c index 1831f6353622..a5502898ea33 100644 --- a/net/ceph/messenger.c +++ b/net/ceph/messenger.c | |||
@@ -269,7 +269,7 @@ static void _ceph_msgr_exit(void) | |||
269 | } | 269 | } |
270 | 270 | ||
271 | BUG_ON(zero_page == NULL); | 271 | BUG_ON(zero_page == NULL); |
272 | page_cache_release(zero_page); | 272 | put_page(zero_page); |
273 | zero_page = NULL; | 273 | zero_page = NULL; |
274 | 274 | ||
275 | ceph_msgr_slab_exit(); | 275 | ceph_msgr_slab_exit(); |
@@ -282,7 +282,7 @@ int ceph_msgr_init(void) | |||
282 | 282 | ||
283 | BUG_ON(zero_page != NULL); | 283 | BUG_ON(zero_page != NULL); |
284 | zero_page = ZERO_PAGE(0); | 284 | zero_page = ZERO_PAGE(0); |
285 | page_cache_get(zero_page); | 285 | get_page(zero_page); |
286 | 286 | ||
287 | /* | 287 | /* |
288 | * The number of active work items is limited by the number of | 288 | * The number of active work items is limited by the number of |
@@ -1602,7 +1602,7 @@ static int write_partial_skip(struct ceph_connection *con) | |||
1602 | 1602 | ||
1603 | dout("%s %p %d left\n", __func__, con, con->out_skip); | 1603 | dout("%s %p %d left\n", __func__, con, con->out_skip); |
1604 | while (con->out_skip > 0) { | 1604 | while (con->out_skip > 0) { |
1605 | size_t size = min(con->out_skip, (int) PAGE_CACHE_SIZE); | 1605 | size_t size = min(con->out_skip, (int) PAGE_SIZE); |
1606 | 1606 | ||
1607 | ret = ceph_tcp_sendpage(con->sock, zero_page, 0, size, true); | 1607 | ret = ceph_tcp_sendpage(con->sock, zero_page, 0, size, true); |
1608 | if (ret <= 0) | 1608 | if (ret <= 0) |
diff --git a/net/ceph/pagelist.c b/net/ceph/pagelist.c index c7c220a736e5..6864007e64fc 100644 --- a/net/ceph/pagelist.c +++ b/net/ceph/pagelist.c | |||
@@ -56,7 +56,7 @@ int ceph_pagelist_append(struct ceph_pagelist *pl, const void *buf, size_t len) | |||
56 | size_t bit = pl->room; | 56 | size_t bit = pl->room; |
57 | int ret; | 57 | int ret; |
58 | 58 | ||
59 | memcpy(pl->mapped_tail + (pl->length & ~PAGE_CACHE_MASK), | 59 | memcpy(pl->mapped_tail + (pl->length & ~PAGE_MASK), |
60 | buf, bit); | 60 | buf, bit); |
61 | pl->length += bit; | 61 | pl->length += bit; |
62 | pl->room -= bit; | 62 | pl->room -= bit; |
@@ -67,7 +67,7 @@ int ceph_pagelist_append(struct ceph_pagelist *pl, const void *buf, size_t len) | |||
67 | return ret; | 67 | return ret; |
68 | } | 68 | } |
69 | 69 | ||
70 | memcpy(pl->mapped_tail + (pl->length & ~PAGE_CACHE_MASK), buf, len); | 70 | memcpy(pl->mapped_tail + (pl->length & ~PAGE_MASK), buf, len); |
71 | pl->length += len; | 71 | pl->length += len; |
72 | pl->room -= len; | 72 | pl->room -= len; |
73 | return 0; | 73 | return 0; |
diff --git a/net/ceph/pagevec.c b/net/ceph/pagevec.c index 10297f7a89ba..00d2601407c5 100644 --- a/net/ceph/pagevec.c +++ b/net/ceph/pagevec.c | |||
@@ -95,19 +95,19 @@ int ceph_copy_user_to_page_vector(struct page **pages, | |||
95 | loff_t off, size_t len) | 95 | loff_t off, size_t len) |
96 | { | 96 | { |
97 | int i = 0; | 97 | int i = 0; |
98 | int po = off & ~PAGE_CACHE_MASK; | 98 | int po = off & ~PAGE_MASK; |
99 | int left = len; | 99 | int left = len; |
100 | int l, bad; | 100 | int l, bad; |
101 | 101 | ||
102 | while (left > 0) { | 102 | while (left > 0) { |
103 | l = min_t(int, PAGE_CACHE_SIZE-po, left); | 103 | l = min_t(int, PAGE_SIZE-po, left); |
104 | bad = copy_from_user(page_address(pages[i]) + po, data, l); | 104 | bad = copy_from_user(page_address(pages[i]) + po, data, l); |
105 | if (bad == l) | 105 | if (bad == l) |
106 | return -EFAULT; | 106 | return -EFAULT; |
107 | data += l - bad; | 107 | data += l - bad; |
108 | left -= l - bad; | 108 | left -= l - bad; |
109 | po += l - bad; | 109 | po += l - bad; |
110 | if (po == PAGE_CACHE_SIZE) { | 110 | if (po == PAGE_SIZE) { |
111 | po = 0; | 111 | po = 0; |
112 | i++; | 112 | i++; |
113 | } | 113 | } |
@@ -121,17 +121,17 @@ void ceph_copy_to_page_vector(struct page **pages, | |||
121 | loff_t off, size_t len) | 121 | loff_t off, size_t len) |
122 | { | 122 | { |
123 | int i = 0; | 123 | int i = 0; |
124 | size_t po = off & ~PAGE_CACHE_MASK; | 124 | size_t po = off & ~PAGE_MASK; |
125 | size_t left = len; | 125 | size_t left = len; |
126 | 126 | ||
127 | while (left > 0) { | 127 | while (left > 0) { |
128 | size_t l = min_t(size_t, PAGE_CACHE_SIZE-po, left); | 128 | size_t l = min_t(size_t, PAGE_SIZE-po, left); |
129 | 129 | ||
130 | memcpy(page_address(pages[i]) + po, data, l); | 130 | memcpy(page_address(pages[i]) + po, data, l); |
131 | data += l; | 131 | data += l; |
132 | left -= l; | 132 | left -= l; |
133 | po += l; | 133 | po += l; |
134 | if (po == PAGE_CACHE_SIZE) { | 134 | if (po == PAGE_SIZE) { |
135 | po = 0; | 135 | po = 0; |
136 | i++; | 136 | i++; |
137 | } | 137 | } |
@@ -144,17 +144,17 @@ void ceph_copy_from_page_vector(struct page **pages, | |||
144 | loff_t off, size_t len) | 144 | loff_t off, size_t len) |
145 | { | 145 | { |
146 | int i = 0; | 146 | int i = 0; |
147 | size_t po = off & ~PAGE_CACHE_MASK; | 147 | size_t po = off & ~PAGE_MASK; |
148 | size_t left = len; | 148 | size_t left = len; |
149 | 149 | ||
150 | while (left > 0) { | 150 | while (left > 0) { |
151 | size_t l = min_t(size_t, PAGE_CACHE_SIZE-po, left); | 151 | size_t l = min_t(size_t, PAGE_SIZE-po, left); |
152 | 152 | ||
153 | memcpy(data, page_address(pages[i]) + po, l); | 153 | memcpy(data, page_address(pages[i]) + po, l); |
154 | data += l; | 154 | data += l; |
155 | left -= l; | 155 | left -= l; |
156 | po += l; | 156 | po += l; |
157 | if (po == PAGE_CACHE_SIZE) { | 157 | if (po == PAGE_SIZE) { |
158 | po = 0; | 158 | po = 0; |
159 | i++; | 159 | i++; |
160 | } | 160 | } |
@@ -168,25 +168,25 @@ EXPORT_SYMBOL(ceph_copy_from_page_vector); | |||
168 | */ | 168 | */ |
169 | void ceph_zero_page_vector_range(int off, int len, struct page **pages) | 169 | void ceph_zero_page_vector_range(int off, int len, struct page **pages) |
170 | { | 170 | { |
171 | int i = off >> PAGE_CACHE_SHIFT; | 171 | int i = off >> PAGE_SHIFT; |
172 | 172 | ||
173 | off &= ~PAGE_CACHE_MASK; | 173 | off &= ~PAGE_MASK; |
174 | 174 | ||
175 | dout("zero_page_vector_page %u~%u\n", off, len); | 175 | dout("zero_page_vector_page %u~%u\n", off, len); |
176 | 176 | ||
177 | /* leading partial page? */ | 177 | /* leading partial page? */ |
178 | if (off) { | 178 | if (off) { |
179 | int end = min((int)PAGE_CACHE_SIZE, off + len); | 179 | int end = min((int)PAGE_SIZE, off + len); |
180 | dout("zeroing %d %p head from %d\n", i, pages[i], | 180 | dout("zeroing %d %p head from %d\n", i, pages[i], |
181 | (int)off); | 181 | (int)off); |
182 | zero_user_segment(pages[i], off, end); | 182 | zero_user_segment(pages[i], off, end); |
183 | len -= (end - off); | 183 | len -= (end - off); |
184 | i++; | 184 | i++; |
185 | } | 185 | } |
186 | while (len >= PAGE_CACHE_SIZE) { | 186 | while (len >= PAGE_SIZE) { |
187 | dout("zeroing %d %p len=%d\n", i, pages[i], len); | 187 | dout("zeroing %d %p len=%d\n", i, pages[i], len); |
188 | zero_user_segment(pages[i], 0, PAGE_CACHE_SIZE); | 188 | zero_user_segment(pages[i], 0, PAGE_SIZE); |
189 | len -= PAGE_CACHE_SIZE; | 189 | len -= PAGE_SIZE; |
190 | i++; | 190 | i++; |
191 | } | 191 | } |
192 | /* trailing partial page? */ | 192 | /* trailing partial page? */ |
diff --git a/net/sunrpc/auth_gss/auth_gss.c b/net/sunrpc/auth_gss/auth_gss.c index 8c6bc795f060..15612ffa8d57 100644 --- a/net/sunrpc/auth_gss/auth_gss.c +++ b/net/sunrpc/auth_gss/auth_gss.c | |||
@@ -1728,8 +1728,8 @@ alloc_enc_pages(struct rpc_rqst *rqstp) | |||
1728 | return 0; | 1728 | return 0; |
1729 | } | 1729 | } |
1730 | 1730 | ||
1731 | first = snd_buf->page_base >> PAGE_CACHE_SHIFT; | 1731 | first = snd_buf->page_base >> PAGE_SHIFT; |
1732 | last = (snd_buf->page_base + snd_buf->page_len - 1) >> PAGE_CACHE_SHIFT; | 1732 | last = (snd_buf->page_base + snd_buf->page_len - 1) >> PAGE_SHIFT; |
1733 | rqstp->rq_enc_pages_num = last - first + 1 + 1; | 1733 | rqstp->rq_enc_pages_num = last - first + 1 + 1; |
1734 | rqstp->rq_enc_pages | 1734 | rqstp->rq_enc_pages |
1735 | = kmalloc(rqstp->rq_enc_pages_num * sizeof(struct page *), | 1735 | = kmalloc(rqstp->rq_enc_pages_num * sizeof(struct page *), |
@@ -1775,10 +1775,10 @@ gss_wrap_req_priv(struct rpc_cred *cred, struct gss_cl_ctx *ctx, | |||
1775 | status = alloc_enc_pages(rqstp); | 1775 | status = alloc_enc_pages(rqstp); |
1776 | if (status) | 1776 | if (status) |
1777 | return status; | 1777 | return status; |
1778 | first = snd_buf->page_base >> PAGE_CACHE_SHIFT; | 1778 | first = snd_buf->page_base >> PAGE_SHIFT; |
1779 | inpages = snd_buf->pages + first; | 1779 | inpages = snd_buf->pages + first; |
1780 | snd_buf->pages = rqstp->rq_enc_pages; | 1780 | snd_buf->pages = rqstp->rq_enc_pages; |
1781 | snd_buf->page_base -= first << PAGE_CACHE_SHIFT; | 1781 | snd_buf->page_base -= first << PAGE_SHIFT; |
1782 | /* | 1782 | /* |
1783 | * Give the tail its own page, in case we need extra space in the | 1783 | * Give the tail its own page, in case we need extra space in the |
1784 | * head when wrapping: | 1784 | * head when wrapping: |
diff --git a/net/sunrpc/auth_gss/gss_krb5_crypto.c b/net/sunrpc/auth_gss/gss_krb5_crypto.c index d94a8e1e9f05..045e11ecd332 100644 --- a/net/sunrpc/auth_gss/gss_krb5_crypto.c +++ b/net/sunrpc/auth_gss/gss_krb5_crypto.c | |||
@@ -465,7 +465,7 @@ encryptor(struct scatterlist *sg, void *data) | |||
465 | page_pos = desc->pos - outbuf->head[0].iov_len; | 465 | page_pos = desc->pos - outbuf->head[0].iov_len; |
466 | if (page_pos >= 0 && page_pos < outbuf->page_len) { | 466 | if (page_pos >= 0 && page_pos < outbuf->page_len) { |
467 | /* pages are not in place: */ | 467 | /* pages are not in place: */ |
468 | int i = (page_pos + outbuf->page_base) >> PAGE_CACHE_SHIFT; | 468 | int i = (page_pos + outbuf->page_base) >> PAGE_SHIFT; |
469 | in_page = desc->pages[i]; | 469 | in_page = desc->pages[i]; |
470 | } else { | 470 | } else { |
471 | in_page = sg_page(sg); | 471 | in_page = sg_page(sg); |
diff --git a/net/sunrpc/auth_gss/gss_krb5_wrap.c b/net/sunrpc/auth_gss/gss_krb5_wrap.c index 765088e4ad84..a737c2da0837 100644 --- a/net/sunrpc/auth_gss/gss_krb5_wrap.c +++ b/net/sunrpc/auth_gss/gss_krb5_wrap.c | |||
@@ -79,9 +79,9 @@ gss_krb5_remove_padding(struct xdr_buf *buf, int blocksize) | |||
79 | len -= buf->head[0].iov_len; | 79 | len -= buf->head[0].iov_len; |
80 | if (len <= buf->page_len) { | 80 | if (len <= buf->page_len) { |
81 | unsigned int last = (buf->page_base + len - 1) | 81 | unsigned int last = (buf->page_base + len - 1) |
82 | >>PAGE_CACHE_SHIFT; | 82 | >>PAGE_SHIFT; |
83 | unsigned int offset = (buf->page_base + len - 1) | 83 | unsigned int offset = (buf->page_base + len - 1) |
84 | & (PAGE_CACHE_SIZE - 1); | 84 | & (PAGE_SIZE - 1); |
85 | ptr = kmap_atomic(buf->pages[last]); | 85 | ptr = kmap_atomic(buf->pages[last]); |
86 | pad = *(ptr + offset); | 86 | pad = *(ptr + offset); |
87 | kunmap_atomic(ptr); | 87 | kunmap_atomic(ptr); |
diff --git a/net/sunrpc/cache.c b/net/sunrpc/cache.c index 008c25d1b9f9..553bf95f7003 100644 --- a/net/sunrpc/cache.c +++ b/net/sunrpc/cache.c | |||
@@ -881,7 +881,7 @@ static ssize_t cache_downcall(struct address_space *mapping, | |||
881 | char *kaddr; | 881 | char *kaddr; |
882 | ssize_t ret = -ENOMEM; | 882 | ssize_t ret = -ENOMEM; |
883 | 883 | ||
884 | if (count >= PAGE_CACHE_SIZE) | 884 | if (count >= PAGE_SIZE) |
885 | goto out_slow; | 885 | goto out_slow; |
886 | 886 | ||
887 | page = find_or_create_page(mapping, 0, GFP_KERNEL); | 887 | page = find_or_create_page(mapping, 0, GFP_KERNEL); |
@@ -892,7 +892,7 @@ static ssize_t cache_downcall(struct address_space *mapping, | |||
892 | ret = cache_do_downcall(kaddr, buf, count, cd); | 892 | ret = cache_do_downcall(kaddr, buf, count, cd); |
893 | kunmap(page); | 893 | kunmap(page); |
894 | unlock_page(page); | 894 | unlock_page(page); |
895 | page_cache_release(page); | 895 | put_page(page); |
896 | return ret; | 896 | return ret; |
897 | out_slow: | 897 | out_slow: |
898 | return cache_slow_downcall(buf, count, cd); | 898 | return cache_slow_downcall(buf, count, cd); |
diff --git a/net/sunrpc/rpc_pipe.c b/net/sunrpc/rpc_pipe.c index 31789ef3e614..fc48eca21fd2 100644 --- a/net/sunrpc/rpc_pipe.c +++ b/net/sunrpc/rpc_pipe.c | |||
@@ -1390,8 +1390,8 @@ rpc_fill_super(struct super_block *sb, void *data, int silent) | |||
1390 | struct sunrpc_net *sn = net_generic(net, sunrpc_net_id); | 1390 | struct sunrpc_net *sn = net_generic(net, sunrpc_net_id); |
1391 | int err; | 1391 | int err; |
1392 | 1392 | ||
1393 | sb->s_blocksize = PAGE_CACHE_SIZE; | 1393 | sb->s_blocksize = PAGE_SIZE; |
1394 | sb->s_blocksize_bits = PAGE_CACHE_SHIFT; | 1394 | sb->s_blocksize_bits = PAGE_SHIFT; |
1395 | sb->s_magic = RPCAUTH_GSSMAGIC; | 1395 | sb->s_magic = RPCAUTH_GSSMAGIC; |
1396 | sb->s_op = &s_ops; | 1396 | sb->s_op = &s_ops; |
1397 | sb->s_d_op = &simple_dentry_operations; | 1397 | sb->s_d_op = &simple_dentry_operations; |
diff --git a/net/sunrpc/socklib.c b/net/sunrpc/socklib.c index 2df87f78e518..de70c78025d7 100644 --- a/net/sunrpc/socklib.c +++ b/net/sunrpc/socklib.c | |||
@@ -96,8 +96,8 @@ ssize_t xdr_partial_copy_from_skb(struct xdr_buf *xdr, unsigned int base, struct | |||
96 | if (base || xdr->page_base) { | 96 | if (base || xdr->page_base) { |
97 | pglen -= base; | 97 | pglen -= base; |
98 | base += xdr->page_base; | 98 | base += xdr->page_base; |
99 | ppage += base >> PAGE_CACHE_SHIFT; | 99 | ppage += base >> PAGE_SHIFT; |
100 | base &= ~PAGE_CACHE_MASK; | 100 | base &= ~PAGE_MASK; |
101 | } | 101 | } |
102 | do { | 102 | do { |
103 | char *kaddr; | 103 | char *kaddr; |
@@ -113,7 +113,7 @@ ssize_t xdr_partial_copy_from_skb(struct xdr_buf *xdr, unsigned int base, struct | |||
113 | } | 113 | } |
114 | } | 114 | } |
115 | 115 | ||
116 | len = PAGE_CACHE_SIZE; | 116 | len = PAGE_SIZE; |
117 | kaddr = kmap_atomic(*ppage); | 117 | kaddr = kmap_atomic(*ppage); |
118 | if (base) { | 118 | if (base) { |
119 | len -= base; | 119 | len -= base; |
diff --git a/net/sunrpc/xdr.c b/net/sunrpc/xdr.c index 4439ac4c1b53..1e6aba0501a1 100644 --- a/net/sunrpc/xdr.c +++ b/net/sunrpc/xdr.c | |||
@@ -181,20 +181,20 @@ _shift_data_right_pages(struct page **pages, size_t pgto_base, | |||
181 | pgto_base += len; | 181 | pgto_base += len; |
182 | pgfrom_base += len; | 182 | pgfrom_base += len; |
183 | 183 | ||
184 | pgto = pages + (pgto_base >> PAGE_CACHE_SHIFT); | 184 | pgto = pages + (pgto_base >> PAGE_SHIFT); |
185 | pgfrom = pages + (pgfrom_base >> PAGE_CACHE_SHIFT); | 185 | pgfrom = pages + (pgfrom_base >> PAGE_SHIFT); |
186 | 186 | ||
187 | pgto_base &= ~PAGE_CACHE_MASK; | 187 | pgto_base &= ~PAGE_MASK; |
188 | pgfrom_base &= ~PAGE_CACHE_MASK; | 188 | pgfrom_base &= ~PAGE_MASK; |
189 | 189 | ||
190 | do { | 190 | do { |
191 | /* Are any pointers crossing a page boundary? */ | 191 | /* Are any pointers crossing a page boundary? */ |
192 | if (pgto_base == 0) { | 192 | if (pgto_base == 0) { |
193 | pgto_base = PAGE_CACHE_SIZE; | 193 | pgto_base = PAGE_SIZE; |
194 | pgto--; | 194 | pgto--; |
195 | } | 195 | } |
196 | if (pgfrom_base == 0) { | 196 | if (pgfrom_base == 0) { |
197 | pgfrom_base = PAGE_CACHE_SIZE; | 197 | pgfrom_base = PAGE_SIZE; |
198 | pgfrom--; | 198 | pgfrom--; |
199 | } | 199 | } |
200 | 200 | ||
@@ -236,11 +236,11 @@ _copy_to_pages(struct page **pages, size_t pgbase, const char *p, size_t len) | |||
236 | char *vto; | 236 | char *vto; |
237 | size_t copy; | 237 | size_t copy; |
238 | 238 | ||
239 | pgto = pages + (pgbase >> PAGE_CACHE_SHIFT); | 239 | pgto = pages + (pgbase >> PAGE_SHIFT); |
240 | pgbase &= ~PAGE_CACHE_MASK; | 240 | pgbase &= ~PAGE_MASK; |
241 | 241 | ||
242 | for (;;) { | 242 | for (;;) { |
243 | copy = PAGE_CACHE_SIZE - pgbase; | 243 | copy = PAGE_SIZE - pgbase; |
244 | if (copy > len) | 244 | if (copy > len) |
245 | copy = len; | 245 | copy = len; |
246 | 246 | ||
@@ -253,7 +253,7 @@ _copy_to_pages(struct page **pages, size_t pgbase, const char *p, size_t len) | |||
253 | break; | 253 | break; |
254 | 254 | ||
255 | pgbase += copy; | 255 | pgbase += copy; |
256 | if (pgbase == PAGE_CACHE_SIZE) { | 256 | if (pgbase == PAGE_SIZE) { |
257 | flush_dcache_page(*pgto); | 257 | flush_dcache_page(*pgto); |
258 | pgbase = 0; | 258 | pgbase = 0; |
259 | pgto++; | 259 | pgto++; |
@@ -280,11 +280,11 @@ _copy_from_pages(char *p, struct page **pages, size_t pgbase, size_t len) | |||
280 | char *vfrom; | 280 | char *vfrom; |
281 | size_t copy; | 281 | size_t copy; |
282 | 282 | ||
283 | pgfrom = pages + (pgbase >> PAGE_CACHE_SHIFT); | 283 | pgfrom = pages + (pgbase >> PAGE_SHIFT); |
284 | pgbase &= ~PAGE_CACHE_MASK; | 284 | pgbase &= ~PAGE_MASK; |
285 | 285 | ||
286 | do { | 286 | do { |
287 | copy = PAGE_CACHE_SIZE - pgbase; | 287 | copy = PAGE_SIZE - pgbase; |
288 | if (copy > len) | 288 | if (copy > len) |
289 | copy = len; | 289 | copy = len; |
290 | 290 | ||
@@ -293,7 +293,7 @@ _copy_from_pages(char *p, struct page **pages, size_t pgbase, size_t len) | |||
293 | kunmap_atomic(vfrom); | 293 | kunmap_atomic(vfrom); |
294 | 294 | ||
295 | pgbase += copy; | 295 | pgbase += copy; |
296 | if (pgbase == PAGE_CACHE_SIZE) { | 296 | if (pgbase == PAGE_SIZE) { |
297 | pgbase = 0; | 297 | pgbase = 0; |
298 | pgfrom++; | 298 | pgfrom++; |
299 | } | 299 | } |
@@ -1038,8 +1038,8 @@ xdr_buf_subsegment(struct xdr_buf *buf, struct xdr_buf *subbuf, | |||
1038 | if (base < buf->page_len) { | 1038 | if (base < buf->page_len) { |
1039 | subbuf->page_len = min(buf->page_len - base, len); | 1039 | subbuf->page_len = min(buf->page_len - base, len); |
1040 | base += buf->page_base; | 1040 | base += buf->page_base; |
1041 | subbuf->page_base = base & ~PAGE_CACHE_MASK; | 1041 | subbuf->page_base = base & ~PAGE_MASK; |
1042 | subbuf->pages = &buf->pages[base >> PAGE_CACHE_SHIFT]; | 1042 | subbuf->pages = &buf->pages[base >> PAGE_SHIFT]; |
1043 | len -= subbuf->page_len; | 1043 | len -= subbuf->page_len; |
1044 | base = 0; | 1044 | base = 0; |
1045 | } else { | 1045 | } else { |
@@ -1297,9 +1297,9 @@ xdr_xcode_array2(struct xdr_buf *buf, unsigned int base, | |||
1297 | todo -= avail_here; | 1297 | todo -= avail_here; |
1298 | 1298 | ||
1299 | base += buf->page_base; | 1299 | base += buf->page_base; |
1300 | ppages = buf->pages + (base >> PAGE_CACHE_SHIFT); | 1300 | ppages = buf->pages + (base >> PAGE_SHIFT); |
1301 | base &= ~PAGE_CACHE_MASK; | 1301 | base &= ~PAGE_MASK; |
1302 | avail_page = min_t(unsigned int, PAGE_CACHE_SIZE - base, | 1302 | avail_page = min_t(unsigned int, PAGE_SIZE - base, |
1303 | avail_here); | 1303 | avail_here); |
1304 | c = kmap(*ppages) + base; | 1304 | c = kmap(*ppages) + base; |
1305 | 1305 | ||
@@ -1383,7 +1383,7 @@ xdr_xcode_array2(struct xdr_buf *buf, unsigned int base, | |||
1383 | } | 1383 | } |
1384 | 1384 | ||
1385 | avail_page = min(avail_here, | 1385 | avail_page = min(avail_here, |
1386 | (unsigned int) PAGE_CACHE_SIZE); | 1386 | (unsigned int) PAGE_SIZE); |
1387 | } | 1387 | } |
1388 | base = buf->page_len; /* align to start of tail */ | 1388 | base = buf->page_len; /* align to start of tail */ |
1389 | } | 1389 | } |
@@ -1479,9 +1479,9 @@ xdr_process_buf(struct xdr_buf *buf, unsigned int offset, unsigned int len, | |||
1479 | if (page_len > len) | 1479 | if (page_len > len) |
1480 | page_len = len; | 1480 | page_len = len; |
1481 | len -= page_len; | 1481 | len -= page_len; |
1482 | page_offset = (offset + buf->page_base) & (PAGE_CACHE_SIZE - 1); | 1482 | page_offset = (offset + buf->page_base) & (PAGE_SIZE - 1); |
1483 | i = (offset + buf->page_base) >> PAGE_CACHE_SHIFT; | 1483 | i = (offset + buf->page_base) >> PAGE_SHIFT; |
1484 | thislen = PAGE_CACHE_SIZE - page_offset; | 1484 | thislen = PAGE_SIZE - page_offset; |
1485 | do { | 1485 | do { |
1486 | if (thislen > page_len) | 1486 | if (thislen > page_len) |
1487 | thislen = page_len; | 1487 | thislen = page_len; |
@@ -1492,7 +1492,7 @@ xdr_process_buf(struct xdr_buf *buf, unsigned int offset, unsigned int len, | |||
1492 | page_len -= thislen; | 1492 | page_len -= thislen; |
1493 | i++; | 1493 | i++; |
1494 | page_offset = 0; | 1494 | page_offset = 0; |
1495 | thislen = PAGE_CACHE_SIZE; | 1495 | thislen = PAGE_SIZE; |
1496 | } while (page_len != 0); | 1496 | } while (page_len != 0); |
1497 | offset = 0; | 1497 | offset = 0; |
1498 | } | 1498 | } |