diff options
Diffstat (limited to 'fs')
227 files changed, 6106 insertions, 4420 deletions
diff --git a/fs/9p/Kconfig b/fs/9p/Kconfig new file mode 100644 index 000000000000..74e0723e90bc --- /dev/null +++ b/fs/9p/Kconfig | |||
@@ -0,0 +1,10 @@ | |||
1 | config 9P_FS | ||
2 | tristate "Plan 9 Resource Sharing Support (9P2000) (Experimental)" | ||
3 | depends on INET && NET_9P && EXPERIMENTAL | ||
4 | help | ||
5 | If you say Y here, you will get experimental support for | ||
6 | Plan 9 resource sharing via the 9P2000 protocol. | ||
7 | |||
8 | See <http://v9fs.sf.net> for more information. | ||
9 | |||
10 | If unsure, say N. | ||
diff --git a/fs/Kconfig b/fs/Kconfig index 51307b0fdf0f..93945dd0b1ae 100644 --- a/fs/Kconfig +++ b/fs/Kconfig | |||
@@ -27,141 +27,8 @@ config FS_MBCACHE | |||
27 | default y if EXT4_FS=y && EXT4_FS_XATTR | 27 | default y if EXT4_FS=y && EXT4_FS_XATTR |
28 | default m if EXT2_FS_XATTR || EXT3_FS_XATTR || EXT4_FS_XATTR | 28 | default m if EXT2_FS_XATTR || EXT3_FS_XATTR || EXT4_FS_XATTR |
29 | 29 | ||
30 | config REISERFS_FS | 30 | source "fs/reiserfs/Kconfig" |
31 | tristate "Reiserfs support" | 31 | source "fs/jfs/Kconfig" |
32 | help | ||
33 | Stores not just filenames but the files themselves in a balanced | ||
34 | tree. Uses journalling. | ||
35 | |||
36 | Balanced trees are more efficient than traditional file system | ||
37 | architectural foundations. | ||
38 | |||
39 | In general, ReiserFS is as fast as ext2, but is very efficient with | ||
40 | large directories and small files. Additional patches are needed | ||
41 | for NFS and quotas, please see <http://www.namesys.com/> for links. | ||
42 | |||
43 | It is more easily extended to have features currently found in | ||
44 | database and keyword search systems than block allocation based file | ||
45 | systems are. The next version will be so extended, and will support | ||
46 | plugins consistent with our motto ``It takes more than a license to | ||
47 | make source code open.'' | ||
48 | |||
49 | Read <http://www.namesys.com/> to learn more about reiserfs. | ||
50 | |||
51 | Sponsored by Threshold Networks, Emusic.com, and Bigstorage.com. | ||
52 | |||
53 | If you like it, you can pay us to add new features to it that you | ||
54 | need, buy a support contract, or pay us to port it to another OS. | ||
55 | |||
56 | config REISERFS_CHECK | ||
57 | bool "Enable reiserfs debug mode" | ||
58 | depends on REISERFS_FS | ||
59 | help | ||
60 | If you set this to Y, then ReiserFS will perform every check it can | ||
61 | possibly imagine of its internal consistency throughout its | ||
62 | operation. It will also go substantially slower. More than once we | ||
63 | have forgotten that this was on, and then gone despondent over the | ||
64 | latest benchmarks.:-) Use of this option allows our team to go all | ||
65 | out in checking for consistency when debugging without fear of its | ||
66 | effect on end users. If you are on the verge of sending in a bug | ||
67 | report, say Y and you might get a useful error message. Almost | ||
68 | everyone should say N. | ||
69 | |||
70 | config REISERFS_PROC_INFO | ||
71 | bool "Stats in /proc/fs/reiserfs" | ||
72 | depends on REISERFS_FS && PROC_FS | ||
73 | help | ||
74 | Create under /proc/fs/reiserfs a hierarchy of files, displaying | ||
75 | various ReiserFS statistics and internal data at the expense of | ||
76 | making your kernel or module slightly larger (+8 KB). This also | ||
77 | increases the amount of kernel memory required for each mount. | ||
78 | Almost everyone but ReiserFS developers and people fine-tuning | ||
79 | reiserfs or tracing problems should say N. | ||
80 | |||
81 | config REISERFS_FS_XATTR | ||
82 | bool "ReiserFS extended attributes" | ||
83 | depends on REISERFS_FS | ||
84 | help | ||
85 | Extended attributes are name:value pairs associated with inodes by | ||
86 | the kernel or by users (see the attr(5) manual page, or visit | ||
87 | <http://acl.bestbits.at/> for details). | ||
88 | |||
89 | If unsure, say N. | ||
90 | |||
91 | config REISERFS_FS_POSIX_ACL | ||
92 | bool "ReiserFS POSIX Access Control Lists" | ||
93 | depends on REISERFS_FS_XATTR | ||
94 | select FS_POSIX_ACL | ||
95 | help | ||
96 | Posix Access Control Lists (ACLs) support permissions for users and | ||
97 | groups beyond the owner/group/world scheme. | ||
98 | |||
99 | To learn more about Access Control Lists, visit the Posix ACLs for | ||
100 | Linux website <http://acl.bestbits.at/>. | ||
101 | |||
102 | If you don't know what Access Control Lists are, say N | ||
103 | |||
104 | config REISERFS_FS_SECURITY | ||
105 | bool "ReiserFS Security Labels" | ||
106 | depends on REISERFS_FS_XATTR | ||
107 | help | ||
108 | Security labels support alternative access control models | ||
109 | implemented by security modules like SELinux. This option | ||
110 | enables an extended attribute handler for file security | ||
111 | labels in the ReiserFS filesystem. | ||
112 | |||
113 | If you are not using a security module that requires using | ||
114 | extended attributes for file security labels, say N. | ||
115 | |||
116 | config JFS_FS | ||
117 | tristate "JFS filesystem support" | ||
118 | select NLS | ||
119 | help | ||
120 | This is a port of IBM's Journaled Filesystem . More information is | ||
121 | available in the file <file:Documentation/filesystems/jfs.txt>. | ||
122 | |||
123 | If you do not intend to use the JFS filesystem, say N. | ||
124 | |||
125 | config JFS_POSIX_ACL | ||
126 | bool "JFS POSIX Access Control Lists" | ||
127 | depends on JFS_FS | ||
128 | select FS_POSIX_ACL | ||
129 | help | ||
130 | Posix Access Control Lists (ACLs) support permissions for users and | ||
131 | groups beyond the owner/group/world scheme. | ||
132 | |||
133 | To learn more about Access Control Lists, visit the Posix ACLs for | ||
134 | Linux website <http://acl.bestbits.at/>. | ||
135 | |||
136 | If you don't know what Access Control Lists are, say N | ||
137 | |||
138 | config JFS_SECURITY | ||
139 | bool "JFS Security Labels" | ||
140 | depends on JFS_FS | ||
141 | help | ||
142 | Security labels support alternative access control models | ||
143 | implemented by security modules like SELinux. This option | ||
144 | enables an extended attribute handler for file security | ||
145 | labels in the jfs filesystem. | ||
146 | |||
147 | If you are not using a security module that requires using | ||
148 | extended attributes for file security labels, say N. | ||
149 | |||
150 | config JFS_DEBUG | ||
151 | bool "JFS debugging" | ||
152 | depends on JFS_FS | ||
153 | help | ||
154 | If you are experiencing any problems with the JFS filesystem, say | ||
155 | Y here. This will result in additional debugging messages to be | ||
156 | written to the system log. Under normal circumstances, this | ||
157 | results in very little overhead. | ||
158 | |||
159 | config JFS_STATISTICS | ||
160 | bool "JFS statistics" | ||
161 | depends on JFS_FS | ||
162 | help | ||
163 | Enabling this option will cause statistics from the JFS file system | ||
164 | to be made available to the user in the /proc/fs/jfs/ directory. | ||
165 | 32 | ||
166 | config FS_POSIX_ACL | 33 | config FS_POSIX_ACL |
167 | # Posix ACL utility routines (for now, only ext2/ext3/jfs/reiserfs/nfs4) | 34 | # Posix ACL utility routines (for now, only ext2/ext3/jfs/reiserfs/nfs4) |
@@ -182,111 +49,8 @@ config FILE_LOCKING | |||
182 | 49 | ||
183 | source "fs/xfs/Kconfig" | 50 | source "fs/xfs/Kconfig" |
184 | source "fs/gfs2/Kconfig" | 51 | source "fs/gfs2/Kconfig" |
185 | 52 | source "fs/ocfs2/Kconfig" | |
186 | config OCFS2_FS | 53 | source "fs/btrfs/Kconfig" |
187 | tristate "OCFS2 file system support" | ||
188 | depends on NET && SYSFS | ||
189 | select CONFIGFS_FS | ||
190 | select JBD2 | ||
191 | select CRC32 | ||
192 | select QUOTA | ||
193 | select QUOTA_TREE | ||
194 | help | ||
195 | OCFS2 is a general purpose extent based shared disk cluster file | ||
196 | system with many similarities to ext3. It supports 64 bit inode | ||
197 | numbers, and has automatically extending metadata groups which may | ||
198 | also make it attractive for non-clustered use. | ||
199 | |||
200 | You'll want to install the ocfs2-tools package in order to at least | ||
201 | get "mount.ocfs2". | ||
202 | |||
203 | Project web page: http://oss.oracle.com/projects/ocfs2 | ||
204 | Tools web page: http://oss.oracle.com/projects/ocfs2-tools | ||
205 | OCFS2 mailing lists: http://oss.oracle.com/projects/ocfs2/mailman/ | ||
206 | |||
207 | For more information on OCFS2, see the file | ||
208 | <file:Documentation/filesystems/ocfs2.txt>. | ||
209 | |||
210 | config OCFS2_FS_O2CB | ||
211 | tristate "O2CB Kernelspace Clustering" | ||
212 | depends on OCFS2_FS | ||
213 | default y | ||
214 | help | ||
215 | OCFS2 includes a simple kernelspace clustering package, the OCFS2 | ||
216 | Cluster Base. It only requires a very small userspace component | ||
217 | to configure it. This comes with the standard ocfs2-tools package. | ||
218 | O2CB is limited to maintaining a cluster for OCFS2 file systems. | ||
219 | It cannot manage any other cluster applications. | ||
220 | |||
221 | It is always safe to say Y here, as the clustering method is | ||
222 | run-time selectable. | ||
223 | |||
224 | config OCFS2_FS_USERSPACE_CLUSTER | ||
225 | tristate "OCFS2 Userspace Clustering" | ||
226 | depends on OCFS2_FS && DLM | ||
227 | default y | ||
228 | help | ||
229 | This option will allow OCFS2 to use userspace clustering services | ||
230 | in conjunction with the DLM in fs/dlm. If you are using a | ||
231 | userspace cluster manager, say Y here. | ||
232 | |||
233 | It is safe to say Y, as the clustering method is run-time | ||
234 | selectable. | ||
235 | |||
236 | config OCFS2_FS_STATS | ||
237 | bool "OCFS2 statistics" | ||
238 | depends on OCFS2_FS | ||
239 | default y | ||
240 | help | ||
241 | This option allows some fs statistics to be captured. Enabling | ||
242 | this option may increase the memory consumption. | ||
243 | |||
244 | config OCFS2_DEBUG_MASKLOG | ||
245 | bool "OCFS2 logging support" | ||
246 | depends on OCFS2_FS | ||
247 | default y | ||
248 | help | ||
249 | The ocfs2 filesystem has an extensive logging system. The system | ||
250 | allows selection of events to log via files in /sys/o2cb/logmask/. | ||
251 | This option will enlarge your kernel, but it allows debugging of | ||
252 | ocfs2 filesystem issues. | ||
253 | |||
254 | config OCFS2_DEBUG_FS | ||
255 | bool "OCFS2 expensive checks" | ||
256 | depends on OCFS2_FS | ||
257 | default n | ||
258 | help | ||
259 | This option will enable expensive consistency checks. Enable | ||
260 | this option for debugging only as it is likely to decrease | ||
261 | performance of the filesystem. | ||
262 | |||
263 | config OCFS2_FS_POSIX_ACL | ||
264 | bool "OCFS2 POSIX Access Control Lists" | ||
265 | depends on OCFS2_FS | ||
266 | select FS_POSIX_ACL | ||
267 | default n | ||
268 | help | ||
269 | Posix Access Control Lists (ACLs) support permissions for users and | ||
270 | groups beyond the owner/group/world scheme. | ||
271 | |||
272 | config BTRFS_FS | ||
273 | tristate "Btrfs filesystem (EXPERIMENTAL) Unstable disk format" | ||
274 | depends on EXPERIMENTAL | ||
275 | select LIBCRC32C | ||
276 | select ZLIB_INFLATE | ||
277 | select ZLIB_DEFLATE | ||
278 | help | ||
279 | Btrfs is a new filesystem with extents, writable snapshotting, | ||
280 | support for multiple devices and many more features. | ||
281 | |||
282 | Btrfs is highly experimental, and THE DISK FORMAT IS NOT YET | ||
283 | FINALIZED. You should say N here unless you are interested in | ||
284 | testing Btrfs with non-critical data. | ||
285 | |||
286 | To compile this file system support as a module, choose M here. The | ||
287 | module will be called btrfs. | ||
288 | |||
289 | If unsure, say N. | ||
290 | 54 | ||
291 | endif # BLOCK | 55 | endif # BLOCK |
292 | 56 | ||
@@ -348,64 +112,9 @@ config QUOTACTL | |||
348 | depends on XFS_QUOTA || QUOTA | 112 | depends on XFS_QUOTA || QUOTA |
349 | default y | 113 | default y |
350 | 114 | ||
351 | config AUTOFS_FS | 115 | source "fs/autofs/Kconfig" |
352 | tristate "Kernel automounter support" | 116 | source "fs/autofs4/Kconfig" |
353 | help | 117 | source "fs/fuse/Kconfig" |
354 | The automounter is a tool to automatically mount remote file systems | ||
355 | on demand. This implementation is partially kernel-based to reduce | ||
356 | overhead in the already-mounted case; this is unlike the BSD | ||
357 | automounter (amd), which is a pure user space daemon. | ||
358 | |||
359 | To use the automounter you need the user-space tools from the autofs | ||
360 | package; you can find the location in <file:Documentation/Changes>. | ||
361 | You also want to answer Y to "NFS file system support", below. | ||
362 | |||
363 | If you want to use the newer version of the automounter with more | ||
364 | features, say N here and say Y to "Kernel automounter v4 support", | ||
365 | below. | ||
366 | |||
367 | To compile this support as a module, choose M here: the module will be | ||
368 | called autofs. | ||
369 | |||
370 | If you are not a part of a fairly large, distributed network, you | ||
371 | probably do not need an automounter, and can say N here. | ||
372 | |||
373 | config AUTOFS4_FS | ||
374 | tristate "Kernel automounter version 4 support (also supports v3)" | ||
375 | help | ||
376 | The automounter is a tool to automatically mount remote file systems | ||
377 | on demand. This implementation is partially kernel-based to reduce | ||
378 | overhead in the already-mounted case; this is unlike the BSD | ||
379 | automounter (amd), which is a pure user space daemon. | ||
380 | |||
381 | To use the automounter you need the user-space tools from | ||
382 | <ftp://ftp.kernel.org/pub/linux/daemons/autofs/v4/>; you also | ||
383 | want to answer Y to "NFS file system support", below. | ||
384 | |||
385 | To compile this support as a module, choose M here: the module will be | ||
386 | called autofs4. You will need to add "alias autofs autofs4" to your | ||
387 | modules configuration file. | ||
388 | |||
389 | If you are not a part of a fairly large, distributed network or | ||
390 | don't have a laptop which needs to dynamically reconfigure to the | ||
391 | local network, you probably do not need an automounter, and can say | ||
392 | N here. | ||
393 | |||
394 | config FUSE_FS | ||
395 | tristate "FUSE (Filesystem in Userspace) support" | ||
396 | help | ||
397 | With FUSE it is possible to implement a fully functional filesystem | ||
398 | in a userspace program. | ||
399 | |||
400 | There's also companion library: libfuse. This library along with | ||
401 | utilities is available from the FUSE homepage: | ||
402 | <http://fuse.sourceforge.net/> | ||
403 | |||
404 | See <file:Documentation/filesystems/fuse.txt> for more information. | ||
405 | See <file:Documentation/Changes> for needed library/utility version. | ||
406 | |||
407 | If you want to develop a userspace FS, or if you want to use | ||
408 | a filesystem based on FUSE, answer Y or M. | ||
409 | 118 | ||
410 | config GENERIC_ACL | 119 | config GENERIC_ACL |
411 | bool | 120 | bool |
@@ -414,64 +123,8 @@ config GENERIC_ACL | |||
414 | if BLOCK | 123 | if BLOCK |
415 | menu "CD-ROM/DVD Filesystems" | 124 | menu "CD-ROM/DVD Filesystems" |
416 | 125 | ||
417 | config ISO9660_FS | 126 | source "fs/isofs/Kconfig" |
418 | tristate "ISO 9660 CDROM file system support" | 127 | source "fs/udf/Kconfig" |
419 | help | ||
420 | This is the standard file system used on CD-ROMs. It was previously | ||
421 | known as "High Sierra File System" and is called "hsfs" on other | ||
422 | Unix systems. The so-called Rock-Ridge extensions which allow for | ||
423 | long Unix filenames and symbolic links are also supported by this | ||
424 | driver. If you have a CD-ROM drive and want to do more with it than | ||
425 | just listen to audio CDs and watch its LEDs, say Y (and read | ||
426 | <file:Documentation/filesystems/isofs.txt> and the CD-ROM-HOWTO, | ||
427 | available from <http://www.tldp.org/docs.html#howto>), thereby | ||
428 | enlarging your kernel by about 27 KB; otherwise say N. | ||
429 | |||
430 | To compile this file system support as a module, choose M here: the | ||
431 | module will be called isofs. | ||
432 | |||
433 | config JOLIET | ||
434 | bool "Microsoft Joliet CDROM extensions" | ||
435 | depends on ISO9660_FS | ||
436 | select NLS | ||
437 | help | ||
438 | Joliet is a Microsoft extension for the ISO 9660 CD-ROM file system | ||
439 | which allows for long filenames in unicode format (unicode is the | ||
440 | new 16 bit character code, successor to ASCII, which encodes the | ||
441 | characters of almost all languages of the world; see | ||
442 | <http://www.unicode.org/> for more information). Say Y here if you | ||
443 | want to be able to read Joliet CD-ROMs under Linux. | ||
444 | |||
445 | config ZISOFS | ||
446 | bool "Transparent decompression extension" | ||
447 | depends on ISO9660_FS | ||
448 | select ZLIB_INFLATE | ||
449 | help | ||
450 | This is a Linux-specific extension to RockRidge which lets you store | ||
451 | data in compressed form on a CD-ROM and have it transparently | ||
452 | decompressed when the CD-ROM is accessed. See | ||
453 | <http://www.kernel.org/pub/linux/utils/fs/zisofs/> for the tools | ||
454 | necessary to create such a filesystem. Say Y here if you want to be | ||
455 | able to read such compressed CD-ROMs. | ||
456 | |||
457 | config UDF_FS | ||
458 | tristate "UDF file system support" | ||
459 | select CRC_ITU_T | ||
460 | help | ||
461 | This is the new file system used on some CD-ROMs and DVDs. Say Y if | ||
462 | you intend to mount DVD discs or CDRW's written in packet mode, or | ||
463 | if written to by other UDF utilities, such as DirectCD. | ||
464 | Please read <file:Documentation/filesystems/udf.txt>. | ||
465 | |||
466 | To compile this file system support as a module, choose M here: the | ||
467 | module will be called udf. | ||
468 | |||
469 | If unsure, say N. | ||
470 | |||
471 | config UDF_NLS | ||
472 | bool | ||
473 | default y | ||
474 | depends on (UDF_FS=m && NLS) || (UDF_FS=y && NLS=y) | ||
475 | 128 | ||
476 | endmenu | 129 | endmenu |
477 | endif # BLOCK | 130 | endif # BLOCK |
@@ -479,182 +132,8 @@ endif # BLOCK | |||
479 | if BLOCK | 132 | if BLOCK |
480 | menu "DOS/FAT/NT Filesystems" | 133 | menu "DOS/FAT/NT Filesystems" |
481 | 134 | ||
482 | config FAT_FS | 135 | source "fs/fat/Kconfig" |
483 | tristate | 136 | source "fs/ntfs/Kconfig" |
484 | select NLS | ||
485 | help | ||
486 | If you want to use one of the FAT-based file systems (the MS-DOS and | ||
487 | VFAT (Windows 95) file systems), then you must say Y or M here | ||
488 | to include FAT support. You will then be able to mount partitions or | ||
489 | diskettes with FAT-based file systems and transparently access the | ||
490 | files on them, i.e. MSDOS files will look and behave just like all | ||
491 | other Unix files. | ||
492 | |||
493 | This FAT support is not a file system in itself, it only provides | ||
494 | the foundation for the other file systems. You will have to say Y or | ||
495 | M to at least one of "MSDOS fs support" or "VFAT fs support" in | ||
496 | order to make use of it. | ||
497 | |||
498 | Another way to read and write MSDOS floppies and hard drive | ||
499 | partitions from within Linux (but not transparently) is with the | ||
500 | mtools ("man mtools") program suite. You don't need to say Y here in | ||
501 | order to do that. | ||
502 | |||
503 | If you need to move large files on floppies between a DOS and a | ||
504 | Linux box, say Y here, mount the floppy under Linux with an MSDOS | ||
505 | file system and use GNU tar's M option. GNU tar is a program | ||
506 | available for Unix and DOS ("man tar" or "info tar"). | ||
507 | |||
508 | The FAT support will enlarge your kernel by about 37 KB. If unsure, | ||
509 | say Y. | ||
510 | |||
511 | To compile this as a module, choose M here: the module will be called | ||
512 | fat. Note that if you compile the FAT support as a module, you | ||
513 | cannot compile any of the FAT-based file systems into the kernel | ||
514 | -- they will have to be modules as well. | ||
515 | |||
516 | config MSDOS_FS | ||
517 | tristate "MSDOS fs support" | ||
518 | select FAT_FS | ||
519 | help | ||
520 | This allows you to mount MSDOS partitions of your hard drive (unless | ||
521 | they are compressed; to access compressed MSDOS partitions under | ||
522 | Linux, you can either use the DOS emulator DOSEMU, described in the | ||
523 | DOSEMU-HOWTO, available from | ||
524 | <http://www.tldp.org/docs.html#howto>, or try dmsdosfs in | ||
525 | <ftp://ibiblio.org/pub/Linux/system/filesystems/dosfs/>. If you | ||
526 | intend to use dosemu with a non-compressed MSDOS partition, say Y | ||
527 | here) and MSDOS floppies. This means that file access becomes | ||
528 | transparent, i.e. the MSDOS files look and behave just like all | ||
529 | other Unix files. | ||
530 | |||
531 | If you have Windows 95 or Windows NT installed on your MSDOS | ||
532 | partitions, you should use the VFAT file system (say Y to "VFAT fs | ||
533 | support" below), or you will not be able to see the long filenames | ||
534 | generated by Windows 95 / Windows NT. | ||
535 | |||
536 | This option will enlarge your kernel by about 7 KB. If unsure, | ||
537 | answer Y. This will only work if you said Y to "DOS FAT fs support" | ||
538 | as well. To compile this as a module, choose M here: the module will | ||
539 | be called msdos. | ||
540 | |||
541 | config VFAT_FS | ||
542 | tristate "VFAT (Windows-95) fs support" | ||
543 | select FAT_FS | ||
544 | help | ||
545 | This option provides support for normal Windows file systems with | ||
546 | long filenames. That includes non-compressed FAT-based file systems | ||
547 | used by Windows 95, Windows 98, Windows NT 4.0, and the Unix | ||
548 | programs from the mtools package. | ||
549 | |||
550 | The VFAT support enlarges your kernel by about 10 KB and it only | ||
551 | works if you said Y to the "DOS FAT fs support" above. Please read | ||
552 | the file <file:Documentation/filesystems/vfat.txt> for details. If | ||
553 | unsure, say Y. | ||
554 | |||
555 | To compile this as a module, choose M here: the module will be called | ||
556 | vfat. | ||
557 | |||
558 | config FAT_DEFAULT_CODEPAGE | ||
559 | int "Default codepage for FAT" | ||
560 | depends on MSDOS_FS || VFAT_FS | ||
561 | default 437 | ||
562 | help | ||
563 | This option should be set to the codepage of your FAT filesystems. | ||
564 | It can be overridden with the "codepage" mount option. | ||
565 | See <file:Documentation/filesystems/vfat.txt> for more information. | ||
566 | |||
567 | config FAT_DEFAULT_IOCHARSET | ||
568 | string "Default iocharset for FAT" | ||
569 | depends on VFAT_FS | ||
570 | default "iso8859-1" | ||
571 | help | ||
572 | Set this to the default input/output character set you'd | ||
573 | like FAT to use. It should probably match the character set | ||
574 | that most of your FAT filesystems use, and can be overridden | ||
575 | with the "iocharset" mount option for FAT filesystems. | ||
576 | Note that "utf8" is not recommended for FAT filesystems. | ||
577 | If unsure, you shouldn't set "utf8" here. | ||
578 | See <file:Documentation/filesystems/vfat.txt> for more information. | ||
579 | |||
580 | config NTFS_FS | ||
581 | tristate "NTFS file system support" | ||
582 | select NLS | ||
583 | help | ||
584 | NTFS is the file system of Microsoft Windows NT, 2000, XP and 2003. | ||
585 | |||
586 | Saying Y or M here enables read support. There is partial, but | ||
587 | safe, write support available. For write support you must also | ||
588 | say Y to "NTFS write support" below. | ||
589 | |||
590 | There are also a number of user-space tools available, called | ||
591 | ntfsprogs. These include ntfsundelete and ntfsresize, that work | ||
592 | without NTFS support enabled in the kernel. | ||
593 | |||
594 | This is a rewrite from scratch of Linux NTFS support and replaced | ||
595 | the old NTFS code starting with Linux 2.5.11. A backport to | ||
596 | the Linux 2.4 kernel series is separately available as a patch | ||
597 | from the project web site. | ||
598 | |||
599 | For more information see <file:Documentation/filesystems/ntfs.txt> | ||
600 | and <http://www.linux-ntfs.org/>. | ||
601 | |||
602 | To compile this file system support as a module, choose M here: the | ||
603 | module will be called ntfs. | ||
604 | |||
605 | If you are not using Windows NT, 2000, XP or 2003 in addition to | ||
606 | Linux on your computer it is safe to say N. | ||
607 | |||
608 | config NTFS_DEBUG | ||
609 | bool "NTFS debugging support" | ||
610 | depends on NTFS_FS | ||
611 | help | ||
612 | If you are experiencing any problems with the NTFS file system, say | ||
613 | Y here. This will result in additional consistency checks to be | ||
614 | performed by the driver as well as additional debugging messages to | ||
615 | be written to the system log. Note that debugging messages are | ||
616 | disabled by default. To enable them, supply the option debug_msgs=1 | ||
617 | at the kernel command line when booting the kernel or as an option | ||
618 | to insmod when loading the ntfs module. Once the driver is active, | ||
619 | you can enable debugging messages by doing (as root): | ||
620 | echo 1 > /proc/sys/fs/ntfs-debug | ||
621 | Replacing the "1" with "0" would disable debug messages. | ||
622 | |||
623 | If you leave debugging messages disabled, this results in little | ||
624 | overhead, but enabling debug messages results in very significant | ||
625 | slowdown of the system. | ||
626 | |||
627 | When reporting bugs, please try to have available a full dump of | ||
628 | debugging messages while the misbehaviour was occurring. | ||
629 | |||
630 | config NTFS_RW | ||
631 | bool "NTFS write support" | ||
632 | depends on NTFS_FS | ||
633 | help | ||
634 | This enables the partial, but safe, write support in the NTFS driver. | ||
635 | |||
636 | The only supported operation is overwriting existing files, without | ||
637 | changing the file length. No file or directory creation, deletion or | ||
638 | renaming is possible. Note only non-resident files can be written to | ||
639 | so you may find that some very small files (<500 bytes or so) cannot | ||
640 | be written to. | ||
641 | |||
642 | While we cannot guarantee that it will not damage any data, we have | ||
643 | so far not received a single report where the driver would have | ||
644 | damaged someones data so we assume it is perfectly safe to use. | ||
645 | |||
646 | Note: While write support is safe in this version (a rewrite from | ||
647 | scratch of the NTFS support), it should be noted that the old NTFS | ||
648 | write support, included in Linux 2.5.10 and before (since 1997), | ||
649 | is not safe. | ||
650 | |||
651 | This is currently useful with TopologiLinux. TopologiLinux is run | ||
652 | on top of any DOS/Microsoft Windows system without partitioning your | ||
653 | hard disk. Unlike other Linux distributions TopologiLinux does not | ||
654 | need its own partition. For more information see | ||
655 | <http://topologi-linux.sourceforge.net/> | ||
656 | |||
657 | It is perfectly safe to say N here. | ||
658 | 137 | ||
659 | endmenu | 138 | endmenu |
660 | endif # BLOCK | 139 | endif # BLOCK |
@@ -662,30 +141,7 @@ endif # BLOCK | |||
662 | menu "Pseudo filesystems" | 141 | menu "Pseudo filesystems" |
663 | 142 | ||
664 | source "fs/proc/Kconfig" | 143 | source "fs/proc/Kconfig" |
665 | 144 | source "fs/sysfs/Kconfig" | |
666 | config SYSFS | ||
667 | bool "sysfs file system support" if EMBEDDED | ||
668 | default y | ||
669 | help | ||
670 | The sysfs filesystem is a virtual filesystem that the kernel uses to | ||
671 | export internal kernel objects, their attributes, and their | ||
672 | relationships to one another. | ||
673 | |||
674 | Users can use sysfs to ascertain useful information about the running | ||
675 | kernel, such as the devices the kernel has discovered on each bus and | ||
676 | which driver each is bound to. sysfs can also be used to tune devices | ||
677 | and other kernel subsystems. | ||
678 | |||
679 | Some system agents rely on the information in sysfs to operate. | ||
680 | /sbin/hotplug uses device and object attributes in sysfs to assist in | ||
681 | delegating policy decisions, like persistently naming devices. | ||
682 | |||
683 | sysfs is currently used by the block subsystem to mount the root | ||
684 | partition. If sysfs is disabled you must specify the boot device on | ||
685 | the kernel boot command line via its major and minor numbers. For | ||
686 | example, "root=03:01" for /dev/hda1. | ||
687 | |||
688 | Designers of embedded systems may wish to say N here to conserve space. | ||
689 | 145 | ||
690 | config TMPFS | 146 | config TMPFS |
691 | bool "Virtual memory file system support (former shm fs)" | 147 | bool "Virtual memory file system support (former shm fs)" |
@@ -726,17 +182,7 @@ config HUGETLBFS | |||
726 | config HUGETLB_PAGE | 182 | config HUGETLB_PAGE |
727 | def_bool HUGETLBFS | 183 | def_bool HUGETLBFS |
728 | 184 | ||
729 | config CONFIGFS_FS | 185 | source "fs/configfs/Kconfig" |
730 | tristate "Userspace-driven configuration filesystem" | ||
731 | depends on SYSFS | ||
732 | help | ||
733 | configfs is a ram-based filesystem that provides the converse | ||
734 | of sysfs's functionality. Where sysfs is a filesystem-based | ||
735 | view of kernel objects, configfs is a filesystem-based manager | ||
736 | of kernel objects, or config_items. | ||
737 | |||
738 | Both sysfs and configfs can and should exist together on the | ||
739 | same system. One is not a replacement for the other. | ||
740 | 186 | ||
741 | endmenu | 187 | endmenu |
742 | 188 | ||
@@ -755,425 +201,27 @@ menuconfig MISC_FILESYSTEMS | |||
755 | 201 | ||
756 | if MISC_FILESYSTEMS | 202 | if MISC_FILESYSTEMS |
757 | 203 | ||
758 | config ADFS_FS | 204 | source "fs/adfs/Kconfig" |
759 | tristate "ADFS file system support (EXPERIMENTAL)" | 205 | source "fs/affs/Kconfig" |
760 | depends on BLOCK && EXPERIMENTAL | 206 | source "fs/ecryptfs/Kconfig" |
761 | help | 207 | source "fs/hfs/Kconfig" |
762 | The Acorn Disc Filing System is the standard file system of the | 208 | source "fs/hfsplus/Kconfig" |
763 | RiscOS operating system which runs on Acorn's ARM-based Risc PC | 209 | source "fs/befs/Kconfig" |
764 | systems and the Acorn Archimedes range of machines. If you say Y | 210 | source "fs/bfs/Kconfig" |
765 | here, Linux will be able to read from ADFS partitions on hard drives | 211 | source "fs/efs/Kconfig" |
766 | and from ADFS-formatted floppy discs. If you also want to be able to | ||
767 | write to those devices, say Y to "ADFS write support" below. | ||
768 | |||
769 | The ADFS partition should be the first partition (i.e., | ||
770 | /dev/[hs]d?1) on each of your drives. Please read the file | ||
771 | <file:Documentation/filesystems/adfs.txt> for further details. | ||
772 | |||
773 | To compile this code as a module, choose M here: the module will be | ||
774 | called adfs. | ||
775 | |||
776 | If unsure, say N. | ||
777 | |||
778 | config ADFS_FS_RW | ||
779 | bool "ADFS write support (DANGEROUS)" | ||
780 | depends on ADFS_FS | ||
781 | help | ||
782 | If you say Y here, you will be able to write to ADFS partitions on | ||
783 | hard drives and ADFS-formatted floppy disks. This is experimental | ||
784 | codes, so if you're unsure, say N. | ||
785 | |||
786 | config AFFS_FS | ||
787 | tristate "Amiga FFS file system support (EXPERIMENTAL)" | ||
788 | depends on BLOCK && EXPERIMENTAL | ||
789 | help | ||
790 | The Fast File System (FFS) is the common file system used on hard | ||
791 | disks by Amiga(tm) systems since AmigaOS Version 1.3 (34.20). Say Y | ||
792 | if you want to be able to read and write files from and to an Amiga | ||
793 | FFS partition on your hard drive. Amiga floppies however cannot be | ||
794 | read with this driver due to an incompatibility of the floppy | ||
795 | controller used in an Amiga and the standard floppy controller in | ||
796 | PCs and workstations. Read <file:Documentation/filesystems/affs.txt> | ||
797 | and <file:fs/affs/Changes>. | ||
798 | |||
799 | With this driver you can also mount disk files used by Bernd | ||
800 | Schmidt's Un*X Amiga Emulator | ||
801 | (<http://www.freiburg.linux.de/~uae/>). | ||
802 | If you want to do this, you will also need to say Y or M to "Loop | ||
803 | device support", above. | ||
804 | |||
805 | To compile this file system support as a module, choose M here: the | ||
806 | module will be called affs. If unsure, say N. | ||
807 | |||
808 | config ECRYPT_FS | ||
809 | tristate "eCrypt filesystem layer support (EXPERIMENTAL)" | ||
810 | depends on EXPERIMENTAL && KEYS && CRYPTO && NET | ||
811 | help | ||
812 | Encrypted filesystem that operates on the VFS layer. See | ||
813 | <file:Documentation/filesystems/ecryptfs.txt> to learn more about | ||
814 | eCryptfs. Userspace components are required and can be | ||
815 | obtained from <http://ecryptfs.sf.net>. | ||
816 | |||
817 | To compile this file system support as a module, choose M here: the | ||
818 | module will be called ecryptfs. | ||
819 | |||
820 | config HFS_FS | ||
821 | tristate "Apple Macintosh file system support (EXPERIMENTAL)" | ||
822 | depends on BLOCK && EXPERIMENTAL | ||
823 | select NLS | ||
824 | help | ||
825 | If you say Y here, you will be able to mount Macintosh-formatted | ||
826 | floppy disks and hard drive partitions with full read-write access. | ||
827 | Please read <file:Documentation/filesystems/hfs.txt> to learn about | ||
828 | the available mount options. | ||
829 | |||
830 | To compile this file system support as a module, choose M here: the | ||
831 | module will be called hfs. | ||
832 | |||
833 | config HFSPLUS_FS | ||
834 | tristate "Apple Extended HFS file system support" | ||
835 | depends on BLOCK | ||
836 | select NLS | ||
837 | select NLS_UTF8 | ||
838 | help | ||
839 | If you say Y here, you will be able to mount extended format | ||
840 | Macintosh-formatted hard drive partitions with full read-write access. | ||
841 | |||
842 | This file system is often called HFS+ and was introduced with | ||
843 | MacOS 8. It includes all Mac specific filesystem data such as | ||
844 | data forks and creator codes, but it also has several UNIX | ||
845 | style features such as file ownership and permissions. | ||
846 | |||
847 | config BEFS_FS | ||
848 | tristate "BeOS file system (BeFS) support (read only) (EXPERIMENTAL)" | ||
849 | depends on BLOCK && EXPERIMENTAL | ||
850 | select NLS | ||
851 | help | ||
852 | The BeOS File System (BeFS) is the native file system of Be, Inc's | ||
853 | BeOS. Notable features include support for arbitrary attributes | ||
854 | on files and directories, and database-like indices on selected | ||
855 | attributes. (Also note that this driver doesn't make those features | ||
856 | available at this time). It is a 64 bit filesystem, so it supports | ||
857 | extremely large volumes and files. | ||
858 | |||
859 | If you use this filesystem, you should also say Y to at least one | ||
860 | of the NLS (native language support) options below. | ||
861 | |||
862 | If you don't know what this is about, say N. | ||
863 | |||
864 | To compile this as a module, choose M here: the module will be | ||
865 | called befs. | ||
866 | |||
867 | config BEFS_DEBUG | ||
868 | bool "Debug BeFS" | ||
869 | depends on BEFS_FS | ||
870 | help | ||
871 | If you say Y here, you can use the 'debug' mount option to enable | ||
872 | debugging output from the driver. | ||
873 | |||
874 | config BFS_FS | ||
875 | tristate "BFS file system support (EXPERIMENTAL)" | ||
876 | depends on BLOCK && EXPERIMENTAL | ||
877 | help | ||
878 | Boot File System (BFS) is a file system used under SCO UnixWare to | ||
879 | allow the bootloader access to the kernel image and other important | ||
880 | files during the boot process. It is usually mounted under /stand | ||
881 | and corresponds to the slice marked as "STAND" in the UnixWare | ||
882 | partition. You should say Y if you want to read or write the files | ||
883 | on your /stand slice from within Linux. You then also need to say Y | ||
884 | to "UnixWare slices support", below. More information about the BFS | ||
885 | file system is contained in the file | ||
886 | <file:Documentation/filesystems/bfs.txt>. | ||
887 | |||
888 | If you don't know what this is about, say N. | ||
889 | |||
890 | To compile this as a module, choose M here: the module will be called | ||
891 | bfs. Note that the file system of your root partition (the one | ||
892 | containing the directory /) cannot be compiled as a module. | ||
893 | |||
894 | |||
895 | |||
896 | config EFS_FS | ||
897 | tristate "EFS file system support (read only) (EXPERIMENTAL)" | ||
898 | depends on BLOCK && EXPERIMENTAL | ||
899 | help | ||
900 | EFS is an older file system used for non-ISO9660 CD-ROMs and hard | ||
901 | disk partitions by SGI's IRIX operating system (IRIX 6.0 and newer | ||
902 | uses the XFS file system for hard disk partitions however). | ||
903 | |||
904 | This implementation only offers read-only access. If you don't know | ||
905 | what all this is about, it's safe to say N. For more information | ||
906 | about EFS see its home page at <http://aeschi.ch.eu.org/efs/>. | ||
907 | |||
908 | To compile the EFS file system support as a module, choose M here: the | ||
909 | module will be called efs. | ||
910 | |||
911 | source "fs/jffs2/Kconfig" | 212 | source "fs/jffs2/Kconfig" |
912 | # UBIFS File system configuration | 213 | # UBIFS File system configuration |
913 | source "fs/ubifs/Kconfig" | 214 | source "fs/ubifs/Kconfig" |
914 | 215 | source "fs/cramfs/Kconfig" | |
915 | config CRAMFS | 216 | source "fs/squashfs/Kconfig" |
916 | tristate "Compressed ROM file system support (cramfs)" | 217 | source "fs/freevxfs/Kconfig" |
917 | depends on BLOCK | 218 | source "fs/minix/Kconfig" |
918 | select ZLIB_INFLATE | 219 | source "fs/omfs/Kconfig" |
919 | help | 220 | source "fs/hpfs/Kconfig" |
920 | Saying Y here includes support for CramFs (Compressed ROM File | 221 | source "fs/qnx4/Kconfig" |
921 | System). CramFs is designed to be a simple, small, and compressed | 222 | source "fs/romfs/Kconfig" |
922 | file system for ROM based embedded systems. CramFs is read-only, | 223 | source "fs/sysv/Kconfig" |
923 | limited to 256MB file systems (with 16MB files), and doesn't support | 224 | source "fs/ufs/Kconfig" |
924 | 16/32 bits uid/gid, hard links and timestamps. | ||
925 | |||
926 | See <file:Documentation/filesystems/cramfs.txt> and | ||
927 | <file:fs/cramfs/README> for further information. | ||
928 | |||
929 | To compile this as a module, choose M here: the module will be called | ||
930 | cramfs. Note that the root file system (the one containing the | ||
931 | directory /) cannot be compiled as a module. | ||
932 | |||
933 | If unsure, say N. | ||
934 | |||
935 | config SQUASHFS | ||
936 | tristate "SquashFS 4.0 - Squashed file system support" | ||
937 | depends on BLOCK | ||
938 | select ZLIB_INFLATE | ||
939 | help | ||
940 | Saying Y here includes support for SquashFS 4.0 (a Compressed | ||
941 | Read-Only File System). Squashfs is a highly compressed read-only | ||
942 | filesystem for Linux. It uses zlib compression to compress both | ||
943 | files, inodes and directories. Inodes in the system are very small | ||
944 | and all blocks are packed to minimise data overhead. Block sizes | ||
945 | greater than 4K are supported up to a maximum of 1 Mbytes (default | ||
946 | block size 128K). SquashFS 4.0 supports 64 bit filesystems and files | ||
947 | (larger than 4GB), full uid/gid information, hard links and | ||
948 | timestamps. | ||
949 | |||
950 | Squashfs is intended for general read-only filesystem use, for | ||
951 | archival use (i.e. in cases where a .tar.gz file may be used), and in | ||
952 | embedded systems where low overhead is needed. Further information | ||
953 | and tools are available from http://squashfs.sourceforge.net. | ||
954 | |||
955 | If you want to compile this as a module ( = code which can be | ||
956 | inserted in and removed from the running kernel whenever you want), | ||
957 | say M here and read <file:Documentation/modules.txt>. The module | ||
958 | will be called squashfs. Note that the root file system (the one | ||
959 | containing the directory /) cannot be compiled as a module. | ||
960 | |||
961 | If unsure, say N. | ||
962 | |||
963 | config SQUASHFS_EMBEDDED | ||
964 | |||
965 | bool "Additional option for memory-constrained systems" | ||
966 | depends on SQUASHFS | ||
967 | default n | ||
968 | help | ||
969 | Saying Y here allows you to specify cache size. | ||
970 | |||
971 | If unsure, say N. | ||
972 | |||
973 | config SQUASHFS_FRAGMENT_CACHE_SIZE | ||
974 | int "Number of fragments cached" if SQUASHFS_EMBEDDED | ||
975 | depends on SQUASHFS | ||
976 | default "3" | ||
977 | help | ||
978 | By default SquashFS caches the last 3 fragments read from | ||
979 | the filesystem. Increasing this amount may mean SquashFS | ||
980 | has to re-read fragments less often from disk, at the expense | ||
981 | of extra system memory. Decreasing this amount will mean | ||
982 | SquashFS uses less memory at the expense of extra reads from disk. | ||
983 | |||
984 | Note there must be at least one cached fragment. Anything | ||
985 | much more than three will probably not make much difference. | ||
986 | |||
987 | config VXFS_FS | ||
988 | tristate "FreeVxFS file system support (VERITAS VxFS(TM) compatible)" | ||
989 | depends on BLOCK | ||
990 | help | ||
991 | FreeVxFS is a file system driver that support the VERITAS VxFS(TM) | ||
992 | file system format. VERITAS VxFS(TM) is the standard file system | ||
993 | of SCO UnixWare (and possibly others) and optionally available | ||
994 | for Sunsoft Solaris, HP-UX and many other operating systems. | ||
995 | Currently only readonly access is supported. | ||
996 | |||
997 | NOTE: the file system type as used by mount(1), mount(2) and | ||
998 | fstab(5) is 'vxfs' as it describes the file system format, not | ||
999 | the actual driver. | ||
1000 | |||
1001 | To compile this as a module, choose M here: the module will be | ||
1002 | called freevxfs. If unsure, say N. | ||
1003 | |||
1004 | config MINIX_FS | ||
1005 | tristate "Minix file system support" | ||
1006 | depends on BLOCK | ||
1007 | help | ||
1008 | Minix is a simple operating system used in many classes about OS's. | ||
1009 | The minix file system (method to organize files on a hard disk | ||
1010 | partition or a floppy disk) was the original file system for Linux, | ||
1011 | but has been superseded by the second extended file system ext2fs. | ||
1012 | You don't want to use the minix file system on your hard disk | ||
1013 | because of certain built-in restrictions, but it is sometimes found | ||
1014 | on older Linux floppy disks. This option will enlarge your kernel | ||
1015 | by about 28 KB. If unsure, say N. | ||
1016 | |||
1017 | To compile this file system support as a module, choose M here: the | ||
1018 | module will be called minix. Note that the file system of your root | ||
1019 | partition (the one containing the directory /) cannot be compiled as | ||
1020 | a module. | ||
1021 | |||
1022 | config OMFS_FS | ||
1023 | tristate "SonicBlue Optimized MPEG File System support" | ||
1024 | depends on BLOCK | ||
1025 | select CRC_ITU_T | ||
1026 | help | ||
1027 | This is the proprietary file system used by the Rio Karma music | ||
1028 | player and ReplayTV DVR. Despite the name, this filesystem is not | ||
1029 | more efficient than a standard FS for MPEG files, in fact likely | ||
1030 | the opposite is true. Say Y if you have either of these devices | ||
1031 | and wish to mount its disk. | ||
1032 | |||
1033 | To compile this file system support as a module, choose M here: the | ||
1034 | module will be called omfs. If unsure, say N. | ||
1035 | |||
1036 | config HPFS_FS | ||
1037 | tristate "OS/2 HPFS file system support" | ||
1038 | depends on BLOCK | ||
1039 | help | ||
1040 | OS/2 is IBM's operating system for PC's, the same as Warp, and HPFS | ||
1041 | is the file system used for organizing files on OS/2 hard disk | ||
1042 | partitions. Say Y if you want to be able to read files from and | ||
1043 | write files to an OS/2 HPFS partition on your hard drive. OS/2 | ||
1044 | floppies however are in regular MSDOS format, so you don't need this | ||
1045 | option in order to be able to read them. Read | ||
1046 | <file:Documentation/filesystems/hpfs.txt>. | ||
1047 | |||
1048 | To compile this file system support as a module, choose M here: the | ||
1049 | module will be called hpfs. If unsure, say N. | ||
1050 | |||
1051 | |||
1052 | config QNX4FS_FS | ||
1053 | tristate "QNX4 file system support (read only)" | ||
1054 | depends on BLOCK | ||
1055 | help | ||
1056 | This is the file system used by the real-time operating systems | ||
1057 | QNX 4 and QNX 6 (the latter is also called QNX RTP). | ||
1058 | Further information is available at <http://www.qnx.com/>. | ||
1059 | Say Y if you intend to mount QNX hard disks or floppies. | ||
1060 | Unless you say Y to "QNX4FS read-write support" below, you will | ||
1061 | only be able to read these file systems. | ||
1062 | |||
1063 | To compile this file system support as a module, choose M here: the | ||
1064 | module will be called qnx4. | ||
1065 | |||
1066 | If you don't know whether you need it, then you don't need it: | ||
1067 | answer N. | ||
1068 | |||
1069 | config QNX4FS_RW | ||
1070 | bool "QNX4FS write support (DANGEROUS)" | ||
1071 | depends on QNX4FS_FS && EXPERIMENTAL && BROKEN | ||
1072 | help | ||
1073 | Say Y if you want to test write support for QNX4 file systems. | ||
1074 | |||
1075 | It's currently broken, so for now: | ||
1076 | answer N. | ||
1077 | |||
1078 | config ROMFS_FS | ||
1079 | tristate "ROM file system support" | ||
1080 | depends on BLOCK | ||
1081 | ---help--- | ||
1082 | This is a very small read-only file system mainly intended for | ||
1083 | initial ram disks of installation disks, but it could be used for | ||
1084 | other read-only media as well. Read | ||
1085 | <file:Documentation/filesystems/romfs.txt> for details. | ||
1086 | |||
1087 | To compile this file system support as a module, choose M here: the | ||
1088 | module will be called romfs. Note that the file system of your | ||
1089 | root partition (the one containing the directory /) cannot be a | ||
1090 | module. | ||
1091 | |||
1092 | If you don't know whether you need it, then you don't need it: | ||
1093 | answer N. | ||
1094 | |||
1095 | |||
1096 | config SYSV_FS | ||
1097 | tristate "System V/Xenix/V7/Coherent file system support" | ||
1098 | depends on BLOCK | ||
1099 | help | ||
1100 | SCO, Xenix and Coherent are commercial Unix systems for Intel | ||
1101 | machines, and Version 7 was used on the DEC PDP-11. Saying Y | ||
1102 | here would allow you to read from their floppies and hard disk | ||
1103 | partitions. | ||
1104 | |||
1105 | If you have floppies or hard disk partitions like that, it is likely | ||
1106 | that they contain binaries from those other Unix systems; in order | ||
1107 | to run these binaries, you will want to install linux-abi which is | ||
1108 | a set of kernel modules that lets you run SCO, Xenix, Wyse, | ||
1109 | UnixWare, Dell Unix and System V programs under Linux. It is | ||
1110 | available via FTP (user: ftp) from | ||
1111 | <ftp://ftp.openlinux.org/pub/people/hch/linux-abi/>). | ||
1112 | NOTE: that will work only for binaries from Intel-based systems; | ||
1113 | PDP ones will have to wait until somebody ports Linux to -11 ;-) | ||
1114 | |||
1115 | If you only intend to mount files from some other Unix over the | ||
1116 | network using NFS, you don't need the System V file system support | ||
1117 | (but you need NFS file system support obviously). | ||
1118 | |||
1119 | Note that this option is generally not needed for floppies, since a | ||
1120 | good portable way to transport files and directories between unixes | ||
1121 | (and even other operating systems) is given by the tar program ("man | ||
1122 | tar" or preferably "info tar"). Note also that this option has | ||
1123 | nothing whatsoever to do with the option "System V IPC". Read about | ||
1124 | the System V file system in | ||
1125 | <file:Documentation/filesystems/sysv-fs.txt>. | ||
1126 | Saying Y here will enlarge your kernel by about 27 KB. | ||
1127 | |||
1128 | To compile this as a module, choose M here: the module will be called | ||
1129 | sysv. | ||
1130 | |||
1131 | If you haven't heard about all of this before, it's safe to say N. | ||
1132 | |||
1133 | |||
1134 | config UFS_FS | ||
1135 | tristate "UFS file system support (read only)" | ||
1136 | depends on BLOCK | ||
1137 | help | ||
1138 | BSD and derivate versions of Unix (such as SunOS, FreeBSD, NetBSD, | ||
1139 | OpenBSD and NeXTstep) use a file system called UFS. Some System V | ||
1140 | Unixes can create and mount hard disk partitions and diskettes using | ||
1141 | this file system as well. Saying Y here will allow you to read from | ||
1142 | these partitions; if you also want to write to them, say Y to the | ||
1143 | experimental "UFS file system write support", below. Please read the | ||
1144 | file <file:Documentation/filesystems/ufs.txt> for more information. | ||
1145 | |||
1146 | The recently released UFS2 variant (used in FreeBSD 5.x) is | ||
1147 | READ-ONLY supported. | ||
1148 | |||
1149 | Note that this option is generally not needed for floppies, since a | ||
1150 | good portable way to transport files and directories between unixes | ||
1151 | (and even other operating systems) is given by the tar program ("man | ||
1152 | tar" or preferably "info tar"). | ||
1153 | |||
1154 | When accessing NeXTstep files, you may need to convert them from the | ||
1155 | NeXT character set to the Latin1 character set; use the program | ||
1156 | recode ("info recode") for this purpose. | ||
1157 | |||
1158 | To compile the UFS file system support as a module, choose M here: the | ||
1159 | module will be called ufs. | ||
1160 | |||
1161 | If you haven't heard about all of this before, it's safe to say N. | ||
1162 | |||
1163 | config UFS_FS_WRITE | ||
1164 | bool "UFS file system write support (DANGEROUS)" | ||
1165 | depends on UFS_FS && EXPERIMENTAL | ||
1166 | help | ||
1167 | Say Y here if you want to try writing to UFS partitions. This is | ||
1168 | experimental, so you should back up your UFS partitions beforehand. | ||
1169 | |||
1170 | config UFS_DEBUG | ||
1171 | bool "UFS debugging" | ||
1172 | depends on UFS_FS | ||
1173 | help | ||
1174 | If you are experiencing any problems with the UFS filesystem, say | ||
1175 | Y here. This will result in _many_ additional debugging messages to be | ||
1176 | written to the system log. | ||
1177 | 225 | ||
1178 | endif # MISC_FILESYSTEMS | 226 | endif # MISC_FILESYSTEMS |
1179 | 227 | ||
@@ -1193,173 +241,8 @@ menuconfig NETWORK_FILESYSTEMS | |||
1193 | 241 | ||
1194 | if NETWORK_FILESYSTEMS | 242 | if NETWORK_FILESYSTEMS |
1195 | 243 | ||
1196 | config NFS_FS | 244 | source "fs/nfs/Kconfig" |
1197 | tristate "NFS client support" | 245 | source "fs/nfsd/Kconfig" |
1198 | depends on INET | ||
1199 | select LOCKD | ||
1200 | select SUNRPC | ||
1201 | select NFS_ACL_SUPPORT if NFS_V3_ACL | ||
1202 | help | ||
1203 | Choose Y here if you want to access files residing on other | ||
1204 | computers using Sun's Network File System protocol. To compile | ||
1205 | this file system support as a module, choose M here: the module | ||
1206 | will be called nfs. | ||
1207 | |||
1208 | To mount file systems exported by NFS servers, you also need to | ||
1209 | install the user space mount.nfs command which can be found in | ||
1210 | the Linux nfs-utils package, available from http://linux-nfs.org/. | ||
1211 | Information about using the mount command is available in the | ||
1212 | mount(8) man page. More detail about the Linux NFS client | ||
1213 | implementation is available via the nfs(5) man page. | ||
1214 | |||
1215 | Below you can choose which versions of the NFS protocol are | ||
1216 | available in the kernel to mount NFS servers. Support for NFS | ||
1217 | version 2 (RFC 1094) is always available when NFS_FS is selected. | ||
1218 | |||
1219 | To configure a system which mounts its root file system via NFS | ||
1220 | at boot time, say Y here, select "Kernel level IP | ||
1221 | autoconfiguration" in the NETWORK menu, and select "Root file | ||
1222 | system on NFS" below. You cannot compile this file system as a | ||
1223 | module in this case. | ||
1224 | |||
1225 | If unsure, say N. | ||
1226 | |||
1227 | config NFS_V3 | ||
1228 | bool "NFS client support for NFS version 3" | ||
1229 | depends on NFS_FS | ||
1230 | help | ||
1231 | This option enables support for version 3 of the NFS protocol | ||
1232 | (RFC 1813) in the kernel's NFS client. | ||
1233 | |||
1234 | If unsure, say Y. | ||
1235 | |||
1236 | config NFS_V3_ACL | ||
1237 | bool "NFS client support for the NFSv3 ACL protocol extension" | ||
1238 | depends on NFS_V3 | ||
1239 | help | ||
1240 | Some NFS servers support an auxiliary NFSv3 ACL protocol that | ||
1241 | Sun added to Solaris but never became an official part of the | ||
1242 | NFS version 3 protocol. This protocol extension allows | ||
1243 | applications on NFS clients to manipulate POSIX Access Control | ||
1244 | Lists on files residing on NFS servers. NFS servers enforce | ||
1245 | ACLs on local files whether this protocol is available or not. | ||
1246 | |||
1247 | Choose Y here if your NFS server supports the Solaris NFSv3 ACL | ||
1248 | protocol extension and you want your NFS client to allow | ||
1249 | applications to access and modify ACLs on files on the server. | ||
1250 | |||
1251 | Most NFS servers don't support the Solaris NFSv3 ACL protocol | ||
1252 | extension. You can choose N here or specify the "noacl" mount | ||
1253 | option to prevent your NFS client from trying to use the NFSv3 | ||
1254 | ACL protocol. | ||
1255 | |||
1256 | If unsure, say N. | ||
1257 | |||
1258 | config NFS_V4 | ||
1259 | bool "NFS client support for NFS version 4 (EXPERIMENTAL)" | ||
1260 | depends on NFS_FS && EXPERIMENTAL | ||
1261 | select RPCSEC_GSS_KRB5 | ||
1262 | help | ||
1263 | This option enables support for version 4 of the NFS protocol | ||
1264 | (RFC 3530) in the kernel's NFS client. | ||
1265 | |||
1266 | To mount NFS servers using NFSv4, you also need to install user | ||
1267 | space programs which can be found in the Linux nfs-utils package, | ||
1268 | available from http://linux-nfs.org/. | ||
1269 | |||
1270 | If unsure, say N. | ||
1271 | |||
1272 | config ROOT_NFS | ||
1273 | bool "Root file system on NFS" | ||
1274 | depends on NFS_FS=y && IP_PNP | ||
1275 | help | ||
1276 | If you want your system to mount its root file system via NFS, | ||
1277 | choose Y here. This is common practice for managing systems | ||
1278 | without local permanent storage. For details, read | ||
1279 | <file:Documentation/filesystems/nfsroot.txt>. | ||
1280 | |||
1281 | Most people say N here. | ||
1282 | |||
1283 | config NFSD | ||
1284 | tristate "NFS server support" | ||
1285 | depends on INET | ||
1286 | select LOCKD | ||
1287 | select SUNRPC | ||
1288 | select EXPORTFS | ||
1289 | select NFS_ACL_SUPPORT if NFSD_V2_ACL | ||
1290 | help | ||
1291 | Choose Y here if you want to allow other computers to access | ||
1292 | files residing on this system using Sun's Network File System | ||
1293 | protocol. To compile the NFS server support as a module, | ||
1294 | choose M here: the module will be called nfsd. | ||
1295 | |||
1296 | You may choose to use a user-space NFS server instead, in which | ||
1297 | case you can choose N here. | ||
1298 | |||
1299 | To export local file systems using NFS, you also need to install | ||
1300 | user space programs which can be found in the Linux nfs-utils | ||
1301 | package, available from http://linux-nfs.org/. More detail about | ||
1302 | the Linux NFS server implementation is available via the | ||
1303 | exports(5) man page. | ||
1304 | |||
1305 | Below you can choose which versions of the NFS protocol are | ||
1306 | available to clients mounting the NFS server on this system. | ||
1307 | Support for NFS version 2 (RFC 1094) is always available when | ||
1308 | CONFIG_NFSD is selected. | ||
1309 | |||
1310 | If unsure, say N. | ||
1311 | |||
1312 | config NFSD_V2_ACL | ||
1313 | bool | ||
1314 | depends on NFSD | ||
1315 | |||
1316 | config NFSD_V3 | ||
1317 | bool "NFS server support for NFS version 3" | ||
1318 | depends on NFSD | ||
1319 | help | ||
1320 | This option enables support in your system's NFS server for | ||
1321 | version 3 of the NFS protocol (RFC 1813). | ||
1322 | |||
1323 | If unsure, say Y. | ||
1324 | |||
1325 | config NFSD_V3_ACL | ||
1326 | bool "NFS server support for the NFSv3 ACL protocol extension" | ||
1327 | depends on NFSD_V3 | ||
1328 | select NFSD_V2_ACL | ||
1329 | help | ||
1330 | Solaris NFS servers support an auxiliary NFSv3 ACL protocol that | ||
1331 | never became an official part of the NFS version 3 protocol. | ||
1332 | This protocol extension allows applications on NFS clients to | ||
1333 | manipulate POSIX Access Control Lists on files residing on NFS | ||
1334 | servers. NFS servers enforce POSIX ACLs on local files whether | ||
1335 | this protocol is available or not. | ||
1336 | |||
1337 | This option enables support in your system's NFS server for the | ||
1338 | NFSv3 ACL protocol extension allowing NFS clients to manipulate | ||
1339 | POSIX ACLs on files exported by your system's NFS server. NFS | ||
1340 | clients which support the Solaris NFSv3 ACL protocol can then | ||
1341 | access and modify ACLs on your NFS server. | ||
1342 | |||
1343 | To store ACLs on your NFS server, you also need to enable ACL- | ||
1344 | related CONFIG options for your local file systems of choice. | ||
1345 | |||
1346 | If unsure, say N. | ||
1347 | |||
1348 | config NFSD_V4 | ||
1349 | bool "NFS server support for NFS version 4 (EXPERIMENTAL)" | ||
1350 | depends on NFSD && PROC_FS && EXPERIMENTAL | ||
1351 | select NFSD_V3 | ||
1352 | select FS_POSIX_ACL | ||
1353 | select RPCSEC_GSS_KRB5 | ||
1354 | help | ||
1355 | This option enables support in your system's NFS server for | ||
1356 | version 4 of the NFS protocol (RFC 3530). | ||
1357 | |||
1358 | To export files using NFSv4, you need to install additional user | ||
1359 | space programs which can be found in the Linux nfs-utils package, | ||
1360 | available from http://linux-nfs.org/. | ||
1361 | |||
1362 | If unsure, say N. | ||
1363 | 246 | ||
1364 | config LOCKD | 247 | config LOCKD |
1365 | tristate | 248 | tristate |
@@ -1381,221 +264,13 @@ config NFS_COMMON | |||
1381 | depends on NFSD || NFS_FS | 264 | depends on NFSD || NFS_FS |
1382 | default y | 265 | default y |
1383 | 266 | ||
1384 | config SUNRPC | 267 | source "net/sunrpc/Kconfig" |
1385 | tristate | 268 | source "fs/smbfs/Kconfig" |
1386 | |||
1387 | config SUNRPC_GSS | ||
1388 | tristate | ||
1389 | |||
1390 | config SUNRPC_XPRT_RDMA | ||
1391 | tristate | ||
1392 | depends on SUNRPC && INFINIBAND && EXPERIMENTAL | ||
1393 | default SUNRPC && INFINIBAND | ||
1394 | help | ||
1395 | This option enables an RPC client transport capability that | ||
1396 | allows the NFS client to mount servers via an RDMA-enabled | ||
1397 | transport. | ||
1398 | |||
1399 | To compile RPC client RDMA transport support as a module, | ||
1400 | choose M here: the module will be called xprtrdma. | ||
1401 | |||
1402 | If unsure, say N. | ||
1403 | |||
1404 | config SUNRPC_REGISTER_V4 | ||
1405 | bool "Register local RPC services via rpcbind v4 (EXPERIMENTAL)" | ||
1406 | depends on SUNRPC && EXPERIMENTAL | ||
1407 | default n | ||
1408 | help | ||
1409 | Sun added support for registering RPC services at an IPv6 | ||
1410 | address by creating two new versions of the rpcbind protocol | ||
1411 | (RFC 1833). | ||
1412 | |||
1413 | This option enables support in the kernel RPC server for | ||
1414 | registering kernel RPC services via version 4 of the rpcbind | ||
1415 | protocol. If you enable this option, you must run a portmapper | ||
1416 | daemon that supports rpcbind protocol version 4. | ||
1417 | |||
1418 | Serving NFS over IPv6 from knfsd (the kernel's NFS server) | ||
1419 | requires that you enable this option and use a portmapper that | ||
1420 | supports rpcbind version 4. | ||
1421 | |||
1422 | If unsure, say N to get traditional behavior (register kernel | ||
1423 | RPC services using only rpcbind version 2). Distributions | ||
1424 | using the legacy Linux portmapper daemon must say N here. | ||
1425 | |||
1426 | config RPCSEC_GSS_KRB5 | ||
1427 | tristate "Secure RPC: Kerberos V mechanism (EXPERIMENTAL)" | ||
1428 | depends on SUNRPC && EXPERIMENTAL | ||
1429 | select SUNRPC_GSS | ||
1430 | select CRYPTO | ||
1431 | select CRYPTO_MD5 | ||
1432 | select CRYPTO_DES | ||
1433 | select CRYPTO_CBC | ||
1434 | help | ||
1435 | Choose Y here to enable Secure RPC using the Kerberos version 5 | ||
1436 | GSS-API mechanism (RFC 1964). | ||
1437 | |||
1438 | Secure RPC calls with Kerberos require an auxiliary user-space | ||
1439 | daemon which may be found in the Linux nfs-utils package | ||
1440 | available from http://linux-nfs.org/. In addition, user-space | ||
1441 | Kerberos support should be installed. | ||
1442 | |||
1443 | If unsure, say N. | ||
1444 | |||
1445 | config RPCSEC_GSS_SPKM3 | ||
1446 | tristate "Secure RPC: SPKM3 mechanism (EXPERIMENTAL)" | ||
1447 | depends on SUNRPC && EXPERIMENTAL | ||
1448 | select SUNRPC_GSS | ||
1449 | select CRYPTO | ||
1450 | select CRYPTO_MD5 | ||
1451 | select CRYPTO_DES | ||
1452 | select CRYPTO_CAST5 | ||
1453 | select CRYPTO_CBC | ||
1454 | help | ||
1455 | Choose Y here to enable Secure RPC using the SPKM3 public key | ||
1456 | GSS-API mechansim (RFC 2025). | ||
1457 | |||
1458 | Secure RPC calls with SPKM3 require an auxiliary userspace | ||
1459 | daemon which may be found in the Linux nfs-utils package | ||
1460 | available from http://linux-nfs.org/. | ||
1461 | |||
1462 | If unsure, say N. | ||
1463 | |||
1464 | config SMB_FS | ||
1465 | tristate "SMB file system support (OBSOLETE, please use CIFS)" | ||
1466 | depends on INET | ||
1467 | select NLS | ||
1468 | help | ||
1469 | SMB (Server Message Block) is the protocol Windows for Workgroups | ||
1470 | (WfW), Windows 95/98, Windows NT and OS/2 Lan Manager use to share | ||
1471 | files and printers over local networks. Saying Y here allows you to | ||
1472 | mount their file systems (often called "shares" in this context) and | ||
1473 | access them just like any other Unix directory. Currently, this | ||
1474 | works only if the Windows machines use TCP/IP as the underlying | ||
1475 | transport protocol, and not NetBEUI. For details, read | ||
1476 | <file:Documentation/filesystems/smbfs.txt> and the SMB-HOWTO, | ||
1477 | available from <http://www.tldp.org/docs.html#howto>. | ||
1478 | |||
1479 | Note: if you just want your box to act as an SMB *server* and make | ||
1480 | files and printing services available to Windows clients (which need | ||
1481 | to have a TCP/IP stack), you don't need to say Y here; you can use | ||
1482 | the program SAMBA (available from <ftp://ftp.samba.org/pub/samba/>) | ||
1483 | for that. | ||
1484 | |||
1485 | General information about how to connect Linux, Windows machines and | ||
1486 | Macs is on the WWW at <http://www.eats.com/linux_mac_win.html>. | ||
1487 | |||
1488 | To compile the SMB support as a module, choose M here: | ||
1489 | the module will be called smbfs. Most people say N, however. | ||
1490 | |||
1491 | config SMB_NLS_DEFAULT | ||
1492 | bool "Use a default NLS" | ||
1493 | depends on SMB_FS | ||
1494 | help | ||
1495 | Enabling this will make smbfs use nls translations by default. You | ||
1496 | need to specify the local charset (CONFIG_NLS_DEFAULT) in the nls | ||
1497 | settings and you need to give the default nls for the SMB server as | ||
1498 | CONFIG_SMB_NLS_REMOTE. | ||
1499 | |||
1500 | The nls settings can be changed at mount time, if your smbmount | ||
1501 | supports that, using the codepage and iocharset parameters. | ||
1502 | |||
1503 | smbmount from samba 2.2.0 or later supports this. | ||
1504 | |||
1505 | config SMB_NLS_REMOTE | ||
1506 | string "Default Remote NLS Option" | ||
1507 | depends on SMB_NLS_DEFAULT | ||
1508 | default "cp437" | ||
1509 | help | ||
1510 | This setting allows you to specify a default value for which | ||
1511 | codepage the server uses. If this field is left blank no | ||
1512 | translations will be done by default. The local codepage/charset | ||
1513 | default to CONFIG_NLS_DEFAULT. | ||
1514 | |||
1515 | The nls settings can be changed at mount time, if your smbmount | ||
1516 | supports that, using the codepage and iocharset parameters. | ||
1517 | |||
1518 | smbmount from samba 2.2.0 or later supports this. | ||
1519 | |||
1520 | source "fs/cifs/Kconfig" | 269 | source "fs/cifs/Kconfig" |
1521 | |||
1522 | config NCP_FS | ||
1523 | tristate "NCP file system support (to mount NetWare volumes)" | ||
1524 | depends on IPX!=n || INET | ||
1525 | help | ||
1526 | NCP (NetWare Core Protocol) is a protocol that runs over IPX and is | ||
1527 | used by Novell NetWare clients to talk to file servers. It is to | ||
1528 | IPX what NFS is to TCP/IP, if that helps. Saying Y here allows you | ||
1529 | to mount NetWare file server volumes and to access them just like | ||
1530 | any other Unix directory. For details, please read the file | ||
1531 | <file:Documentation/filesystems/ncpfs.txt> in the kernel source and | ||
1532 | the IPX-HOWTO from <http://www.tldp.org/docs.html#howto>. | ||
1533 | |||
1534 | You do not have to say Y here if you want your Linux box to act as a | ||
1535 | file *server* for Novell NetWare clients. | ||
1536 | |||
1537 | General information about how to connect Linux, Windows machines and | ||
1538 | Macs is on the WWW at <http://www.eats.com/linux_mac_win.html>. | ||
1539 | |||
1540 | To compile this as a module, choose M here: the module will be called | ||
1541 | ncpfs. Say N unless you are connected to a Novell network. | ||
1542 | |||
1543 | source "fs/ncpfs/Kconfig" | 270 | source "fs/ncpfs/Kconfig" |
1544 | 271 | source "fs/coda/Kconfig" | |
1545 | config CODA_FS | 272 | source "fs/afs/Kconfig" |
1546 | tristate "Coda file system support (advanced network fs)" | 273 | source "fs/9p/Kconfig" |
1547 | depends on INET | ||
1548 | help | ||
1549 | Coda is an advanced network file system, similar to NFS in that it | ||
1550 | enables you to mount file systems of a remote server and access them | ||
1551 | with regular Unix commands as if they were sitting on your hard | ||
1552 | disk. Coda has several advantages over NFS: support for | ||
1553 | disconnected operation (e.g. for laptops), read/write server | ||
1554 | replication, security model for authentication and encryption, | ||
1555 | persistent client caches and write back caching. | ||
1556 | |||
1557 | If you say Y here, your Linux box will be able to act as a Coda | ||
1558 | *client*. You will need user level code as well, both for the | ||
1559 | client and server. Servers are currently user level, i.e. they need | ||
1560 | no kernel support. Please read | ||
1561 | <file:Documentation/filesystems/coda.txt> and check out the Coda | ||
1562 | home page <http://www.coda.cs.cmu.edu/>. | ||
1563 | |||
1564 | To compile the coda client support as a module, choose M here: the | ||
1565 | module will be called coda. | ||
1566 | |||
1567 | config AFS_FS | ||
1568 | tristate "Andrew File System support (AFS) (EXPERIMENTAL)" | ||
1569 | depends on INET && EXPERIMENTAL | ||
1570 | select AF_RXRPC | ||
1571 | help | ||
1572 | If you say Y here, you will get an experimental Andrew File System | ||
1573 | driver. It currently only supports unsecured read-only AFS access. | ||
1574 | |||
1575 | See <file:Documentation/filesystems/afs.txt> for more information. | ||
1576 | |||
1577 | If unsure, say N. | ||
1578 | |||
1579 | config AFS_DEBUG | ||
1580 | bool "AFS dynamic debugging" | ||
1581 | depends on AFS_FS | ||
1582 | help | ||
1583 | Say Y here to make runtime controllable debugging messages appear. | ||
1584 | |||
1585 | See <file:Documentation/filesystems/afs.txt> for more information. | ||
1586 | |||
1587 | If unsure, say N. | ||
1588 | |||
1589 | config 9P_FS | ||
1590 | tristate "Plan 9 Resource Sharing Support (9P2000) (Experimental)" | ||
1591 | depends on INET && NET_9P && EXPERIMENTAL | ||
1592 | help | ||
1593 | If you say Y here, you will get experimental support for | ||
1594 | Plan 9 resource sharing via the 9P2000 protocol. | ||
1595 | |||
1596 | See <http://v9fs.sf.net> for more information. | ||
1597 | |||
1598 | If unsure, say N. | ||
1599 | 274 | ||
1600 | endif # NETWORK_FILESYSTEMS | 275 | endif # NETWORK_FILESYSTEMS |
1601 | 276 | ||
diff --git a/fs/adfs/Kconfig b/fs/adfs/Kconfig new file mode 100644 index 000000000000..e55182a74605 --- /dev/null +++ b/fs/adfs/Kconfig | |||
@@ -0,0 +1,27 @@ | |||
1 | config ADFS_FS | ||
2 | tristate "ADFS file system support (EXPERIMENTAL)" | ||
3 | depends on BLOCK && EXPERIMENTAL | ||
4 | help | ||
5 | The Acorn Disc Filing System is the standard file system of the | ||
6 | RiscOS operating system which runs on Acorn's ARM-based Risc PC | ||
7 | systems and the Acorn Archimedes range of machines. If you say Y | ||
8 | here, Linux will be able to read from ADFS partitions on hard drives | ||
9 | and from ADFS-formatted floppy discs. If you also want to be able to | ||
10 | write to those devices, say Y to "ADFS write support" below. | ||
11 | |||
12 | The ADFS partition should be the first partition (i.e., | ||
13 | /dev/[hs]d?1) on each of your drives. Please read the file | ||
14 | <file:Documentation/filesystems/adfs.txt> for further details. | ||
15 | |||
16 | To compile this code as a module, choose M here: the module will be | ||
17 | called adfs. | ||
18 | |||
19 | If unsure, say N. | ||
20 | |||
21 | config ADFS_FS_RW | ||
22 | bool "ADFS write support (DANGEROUS)" | ||
23 | depends on ADFS_FS | ||
24 | help | ||
25 | If you say Y here, you will be able to write to ADFS partitions on | ||
26 | hard drives and ADFS-formatted floppy disks. This is experimental | ||
27 | codes, so if you're unsure, say N. | ||
diff --git a/fs/affs/Kconfig b/fs/affs/Kconfig new file mode 100644 index 000000000000..cfad9afb4762 --- /dev/null +++ b/fs/affs/Kconfig | |||
@@ -0,0 +1,21 @@ | |||
1 | config AFFS_FS | ||
2 | tristate "Amiga FFS file system support (EXPERIMENTAL)" | ||
3 | depends on BLOCK && EXPERIMENTAL | ||
4 | help | ||
5 | The Fast File System (FFS) is the common file system used on hard | ||
6 | disks by Amiga(tm) systems since AmigaOS Version 1.3 (34.20). Say Y | ||
7 | if you want to be able to read and write files from and to an Amiga | ||
8 | FFS partition on your hard drive. Amiga floppies however cannot be | ||
9 | read with this driver due to an incompatibility of the floppy | ||
10 | controller used in an Amiga and the standard floppy controller in | ||
11 | PCs and workstations. Read <file:Documentation/filesystems/affs.txt> | ||
12 | and <file:fs/affs/Changes>. | ||
13 | |||
14 | With this driver you can also mount disk files used by Bernd | ||
15 | Schmidt's Un*X Amiga Emulator | ||
16 | (<http://www.freiburg.linux.de/~uae/>). | ||
17 | If you want to do this, you will also need to say Y or M to "Loop | ||
18 | device support", above. | ||
19 | |||
20 | To compile this file system support as a module, choose M here: the | ||
21 | module will be called affs. If unsure, say N. | ||
diff --git a/fs/afs/Kconfig b/fs/afs/Kconfig new file mode 100644 index 000000000000..e7b522fe15e1 --- /dev/null +++ b/fs/afs/Kconfig | |||
@@ -0,0 +1,21 @@ | |||
1 | config AFS_FS | ||
2 | tristate "Andrew File System support (AFS) (EXPERIMENTAL)" | ||
3 | depends on INET && EXPERIMENTAL | ||
4 | select AF_RXRPC | ||
5 | help | ||
6 | If you say Y here, you will get an experimental Andrew File System | ||
7 | driver. It currently only supports unsecured read-only AFS access. | ||
8 | |||
9 | See <file:Documentation/filesystems/afs.txt> for more information. | ||
10 | |||
11 | If unsure, say N. | ||
12 | |||
13 | config AFS_DEBUG | ||
14 | bool "AFS dynamic debugging" | ||
15 | depends on AFS_FS | ||
16 | help | ||
17 | Say Y here to make runtime controllable debugging messages appear. | ||
18 | |||
19 | See <file:Documentation/filesystems/afs.txt> for more information. | ||
20 | |||
21 | If unsure, say N. | ||
@@ -1270,7 +1270,7 @@ static void io_destroy(struct kioctx *ioctx) | |||
1270 | * pointer is passed for ctxp. Will fail with -ENOSYS if not | 1270 | * pointer is passed for ctxp. Will fail with -ENOSYS if not |
1271 | * implemented. | 1271 | * implemented. |
1272 | */ | 1272 | */ |
1273 | asmlinkage long sys_io_setup(unsigned nr_events, aio_context_t __user *ctxp) | 1273 | SYSCALL_DEFINE2(io_setup, unsigned, nr_events, aio_context_t __user *, ctxp) |
1274 | { | 1274 | { |
1275 | struct kioctx *ioctx = NULL; | 1275 | struct kioctx *ioctx = NULL; |
1276 | unsigned long ctx; | 1276 | unsigned long ctx; |
@@ -1308,7 +1308,7 @@ out: | |||
1308 | * implemented. May fail with -EFAULT if the context pointed to | 1308 | * implemented. May fail with -EFAULT if the context pointed to |
1309 | * is invalid. | 1309 | * is invalid. |
1310 | */ | 1310 | */ |
1311 | asmlinkage long sys_io_destroy(aio_context_t ctx) | 1311 | SYSCALL_DEFINE1(io_destroy, aio_context_t, ctx) |
1312 | { | 1312 | { |
1313 | struct kioctx *ioctx = lookup_ioctx(ctx); | 1313 | struct kioctx *ioctx = lookup_ioctx(ctx); |
1314 | if (likely(NULL != ioctx)) { | 1314 | if (likely(NULL != ioctx)) { |
@@ -1662,8 +1662,8 @@ out_put_req: | |||
1662 | * are available to queue any iocbs. Will return 0 if nr is 0. Will | 1662 | * are available to queue any iocbs. Will return 0 if nr is 0. Will |
1663 | * fail with -ENOSYS if not implemented. | 1663 | * fail with -ENOSYS if not implemented. |
1664 | */ | 1664 | */ |
1665 | asmlinkage long sys_io_submit(aio_context_t ctx_id, long nr, | 1665 | SYSCALL_DEFINE3(io_submit, aio_context_t, ctx_id, long, nr, |
1666 | struct iocb __user * __user *iocbpp) | 1666 | struct iocb __user * __user *, iocbpp) |
1667 | { | 1667 | { |
1668 | struct kioctx *ctx; | 1668 | struct kioctx *ctx; |
1669 | long ret = 0; | 1669 | long ret = 0; |
@@ -1737,8 +1737,8 @@ static struct kiocb *lookup_kiocb(struct kioctx *ctx, struct iocb __user *iocb, | |||
1737 | * invalid. May fail with -EAGAIN if the iocb specified was not | 1737 | * invalid. May fail with -EAGAIN if the iocb specified was not |
1738 | * cancelled. Will fail with -ENOSYS if not implemented. | 1738 | * cancelled. Will fail with -ENOSYS if not implemented. |
1739 | */ | 1739 | */ |
1740 | asmlinkage long sys_io_cancel(aio_context_t ctx_id, struct iocb __user *iocb, | 1740 | SYSCALL_DEFINE3(io_cancel, aio_context_t, ctx_id, struct iocb __user *, iocb, |
1741 | struct io_event __user *result) | 1741 | struct io_event __user *, result) |
1742 | { | 1742 | { |
1743 | int (*cancel)(struct kiocb *iocb, struct io_event *res); | 1743 | int (*cancel)(struct kiocb *iocb, struct io_event *res); |
1744 | struct kioctx *ctx; | 1744 | struct kioctx *ctx; |
@@ -1799,11 +1799,11 @@ asmlinkage long sys_io_cancel(aio_context_t ctx_id, struct iocb __user *iocb, | |||
1799 | * will be updated if not NULL and the operation blocks. Will fail | 1799 | * will be updated if not NULL and the operation blocks. Will fail |
1800 | * with -ENOSYS if not implemented. | 1800 | * with -ENOSYS if not implemented. |
1801 | */ | 1801 | */ |
1802 | asmlinkage long sys_io_getevents(aio_context_t ctx_id, | 1802 | SYSCALL_DEFINE5(io_getevents, aio_context_t, ctx_id, |
1803 | long min_nr, | 1803 | long, min_nr, |
1804 | long nr, | 1804 | long, nr, |
1805 | struct io_event __user *events, | 1805 | struct io_event __user *, events, |
1806 | struct timespec __user *timeout) | 1806 | struct timespec __user *, timeout) |
1807 | { | 1807 | { |
1808 | struct kioctx *ioctx = lookup_ioctx(ctx_id); | 1808 | struct kioctx *ioctx = lookup_ioctx(ctx_id); |
1809 | long ret = -EINVAL; | 1809 | long ret = -EINVAL; |
diff --git a/fs/autofs/Kconfig b/fs/autofs/Kconfig new file mode 100644 index 000000000000..5f3bea90911e --- /dev/null +++ b/fs/autofs/Kconfig | |||
@@ -0,0 +1,21 @@ | |||
1 | config AUTOFS_FS | ||
2 | tristate "Kernel automounter support" | ||
3 | help | ||
4 | The automounter is a tool to automatically mount remote file systems | ||
5 | on demand. This implementation is partially kernel-based to reduce | ||
6 | overhead in the already-mounted case; this is unlike the BSD | ||
7 | automounter (amd), which is a pure user space daemon. | ||
8 | |||
9 | To use the automounter you need the user-space tools from the autofs | ||
10 | package; you can find the location in <file:Documentation/Changes>. | ||
11 | You also want to answer Y to "NFS file system support", below. | ||
12 | |||
13 | If you want to use the newer version of the automounter with more | ||
14 | features, say N here and say Y to "Kernel automounter v4 support", | ||
15 | below. | ||
16 | |||
17 | To compile this support as a module, choose M here: the module will be | ||
18 | called autofs. | ||
19 | |||
20 | If you are not a part of a fairly large, distributed network, you | ||
21 | probably do not need an automounter, and can say N here. | ||
diff --git a/fs/autofs4/Kconfig b/fs/autofs4/Kconfig new file mode 100644 index 000000000000..1204d6384d39 --- /dev/null +++ b/fs/autofs4/Kconfig | |||
@@ -0,0 +1,20 @@ | |||
1 | config AUTOFS4_FS | ||
2 | tristate "Kernel automounter version 4 support (also supports v3)" | ||
3 | help | ||
4 | The automounter is a tool to automatically mount remote file systems | ||
5 | on demand. This implementation is partially kernel-based to reduce | ||
6 | overhead in the already-mounted case; this is unlike the BSD | ||
7 | automounter (amd), which is a pure user space daemon. | ||
8 | |||
9 | To use the automounter you need the user-space tools from | ||
10 | <ftp://ftp.kernel.org/pub/linux/daemons/autofs/v4/>; you also | ||
11 | want to answer Y to "NFS file system support", below. | ||
12 | |||
13 | To compile this support as a module, choose M here: the module will be | ||
14 | called autofs4. You will need to add "alias autofs autofs4" to your | ||
15 | modules configuration file. | ||
16 | |||
17 | If you are not a part of a fairly large, distributed network or | ||
18 | don't have a laptop which needs to dynamically reconfigure to the | ||
19 | local network, you probably do not need an automounter, and can say | ||
20 | N here. | ||
diff --git a/fs/befs/Kconfig b/fs/befs/Kconfig new file mode 100644 index 000000000000..7835d30f211f --- /dev/null +++ b/fs/befs/Kconfig | |||
@@ -0,0 +1,26 @@ | |||
1 | config BEFS_FS | ||
2 | tristate "BeOS file system (BeFS) support (read only) (EXPERIMENTAL)" | ||
3 | depends on BLOCK && EXPERIMENTAL | ||
4 | select NLS | ||
5 | help | ||
6 | The BeOS File System (BeFS) is the native file system of Be, Inc's | ||
7 | BeOS. Notable features include support for arbitrary attributes | ||
8 | on files and directories, and database-like indices on selected | ||
9 | attributes. (Also note that this driver doesn't make those features | ||
10 | available at this time). It is a 64 bit filesystem, so it supports | ||
11 | extremely large volumes and files. | ||
12 | |||
13 | If you use this filesystem, you should also say Y to at least one | ||
14 | of the NLS (native language support) options below. | ||
15 | |||
16 | If you don't know what this is about, say N. | ||
17 | |||
18 | To compile this as a module, choose M here: the module will be | ||
19 | called befs. | ||
20 | |||
21 | config BEFS_DEBUG | ||
22 | bool "Debug BeFS" | ||
23 | depends on BEFS_FS | ||
24 | help | ||
25 | If you say Y here, you can use the 'debug' mount option to enable | ||
26 | debugging output from the driver. | ||
diff --git a/fs/bfs/Kconfig b/fs/bfs/Kconfig new file mode 100644 index 000000000000..c2336c62024f --- /dev/null +++ b/fs/bfs/Kconfig | |||
@@ -0,0 +1,19 @@ | |||
1 | config BFS_FS | ||
2 | tristate "BFS file system support (EXPERIMENTAL)" | ||
3 | depends on BLOCK && EXPERIMENTAL | ||
4 | help | ||
5 | Boot File System (BFS) is a file system used under SCO UnixWare to | ||
6 | allow the bootloader access to the kernel image and other important | ||
7 | files during the boot process. It is usually mounted under /stand | ||
8 | and corresponds to the slice marked as "STAND" in the UnixWare | ||
9 | partition. You should say Y if you want to read or write the files | ||
10 | on your /stand slice from within Linux. You then also need to say Y | ||
11 | to "UnixWare slices support", below. More information about the BFS | ||
12 | file system is contained in the file | ||
13 | <file:Documentation/filesystems/bfs.txt>. | ||
14 | |||
15 | If you don't know what this is about, say N. | ||
16 | |||
17 | To compile this as a module, choose M here: the module will be called | ||
18 | bfs. Note that the file system of your root partition (the one | ||
19 | containing the directory /) cannot be compiled as a module. | ||
diff --git a/fs/binfmt_elf.c b/fs/binfmt_elf.c index e3ff2b9e602f..33b7235f853b 100644 --- a/fs/binfmt_elf.c +++ b/fs/binfmt_elf.c | |||
@@ -1208,9 +1208,11 @@ static unsigned long vma_dump_size(struct vm_area_struct *vma, | |||
1208 | * check for an ELF header. If we find one, dump the first page to | 1208 | * check for an ELF header. If we find one, dump the first page to |
1209 | * aid in determining what was mapped here. | 1209 | * aid in determining what was mapped here. |
1210 | */ | 1210 | */ |
1211 | if (FILTER(ELF_HEADERS) && vma->vm_file != NULL && vma->vm_pgoff == 0) { | 1211 | if (FILTER(ELF_HEADERS) && |
1212 | vma->vm_pgoff == 0 && (vma->vm_flags & VM_READ)) { | ||
1212 | u32 __user *header = (u32 __user *) vma->vm_start; | 1213 | u32 __user *header = (u32 __user *) vma->vm_start; |
1213 | u32 word; | 1214 | u32 word; |
1215 | mm_segment_t fs = get_fs(); | ||
1214 | /* | 1216 | /* |
1215 | * Doing it this way gets the constant folded by GCC. | 1217 | * Doing it this way gets the constant folded by GCC. |
1216 | */ | 1218 | */ |
@@ -1223,7 +1225,15 @@ static unsigned long vma_dump_size(struct vm_area_struct *vma, | |||
1223 | magic.elfmag[EI_MAG1] = ELFMAG1; | 1225 | magic.elfmag[EI_MAG1] = ELFMAG1; |
1224 | magic.elfmag[EI_MAG2] = ELFMAG2; | 1226 | magic.elfmag[EI_MAG2] = ELFMAG2; |
1225 | magic.elfmag[EI_MAG3] = ELFMAG3; | 1227 | magic.elfmag[EI_MAG3] = ELFMAG3; |
1226 | if (get_user(word, header) == 0 && word == magic.cmp) | 1228 | /* |
1229 | * Switch to the user "segment" for get_user(), | ||
1230 | * then put back what elf_core_dump() had in place. | ||
1231 | */ | ||
1232 | set_fs(USER_DS); | ||
1233 | if (unlikely(get_user(word, header))) | ||
1234 | word = 0; | ||
1235 | set_fs(fs); | ||
1236 | if (word == magic.cmp) | ||
1227 | return PAGE_SIZE; | 1237 | return PAGE_SIZE; |
1228 | } | 1238 | } |
1229 | 1239 | ||
diff --git a/fs/bio-integrity.c b/fs/bio-integrity.c index 77ebc3c263d6..549b0144da11 100644 --- a/fs/bio-integrity.c +++ b/fs/bio-integrity.c | |||
@@ -140,7 +140,6 @@ int bio_integrity_add_page(struct bio *bio, struct page *page, | |||
140 | 140 | ||
141 | iv = bip_vec_idx(bip, bip->bip_vcnt); | 141 | iv = bip_vec_idx(bip, bip->bip_vcnt); |
142 | BUG_ON(iv == NULL); | 142 | BUG_ON(iv == NULL); |
143 | BUG_ON(iv->bv_page != NULL); | ||
144 | 143 | ||
145 | iv->bv_page = page; | 144 | iv->bv_page = page; |
146 | iv->bv_len = len; | 145 | iv->bv_len = len; |
@@ -465,7 +464,7 @@ static int bio_integrity_verify(struct bio *bio) | |||
465 | 464 | ||
466 | if (ret) { | 465 | if (ret) { |
467 | kunmap_atomic(kaddr, KM_USER0); | 466 | kunmap_atomic(kaddr, KM_USER0); |
468 | break; | 467 | return ret; |
469 | } | 468 | } |
470 | 469 | ||
471 | sectors = bv->bv_len / bi->sector_size; | 470 | sectors = bv->bv_len / bi->sector_size; |
@@ -493,18 +492,13 @@ static void bio_integrity_verify_fn(struct work_struct *work) | |||
493 | struct bio_integrity_payload *bip = | 492 | struct bio_integrity_payload *bip = |
494 | container_of(work, struct bio_integrity_payload, bip_work); | 493 | container_of(work, struct bio_integrity_payload, bip_work); |
495 | struct bio *bio = bip->bip_bio; | 494 | struct bio *bio = bip->bip_bio; |
496 | int error = bip->bip_error; | 495 | int error; |
497 | 496 | ||
498 | if (bio_integrity_verify(bio)) { | 497 | error = bio_integrity_verify(bio); |
499 | clear_bit(BIO_UPTODATE, &bio->bi_flags); | ||
500 | error = -EIO; | ||
501 | } | ||
502 | 498 | ||
503 | /* Restore original bio completion handler */ | 499 | /* Restore original bio completion handler */ |
504 | bio->bi_end_io = bip->bip_end_io; | 500 | bio->bi_end_io = bip->bip_end_io; |
505 | 501 | bio_endio(bio, error); | |
506 | if (bio->bi_end_io) | ||
507 | bio->bi_end_io(bio, error); | ||
508 | } | 502 | } |
509 | 503 | ||
510 | /** | 504 | /** |
@@ -525,7 +519,17 @@ void bio_integrity_endio(struct bio *bio, int error) | |||
525 | 519 | ||
526 | BUG_ON(bip->bip_bio != bio); | 520 | BUG_ON(bip->bip_bio != bio); |
527 | 521 | ||
528 | bip->bip_error = error; | 522 | /* In case of an I/O error there is no point in verifying the |
523 | * integrity metadata. Restore original bio end_io handler | ||
524 | * and run it. | ||
525 | */ | ||
526 | if (error) { | ||
527 | bio->bi_end_io = bip->bip_end_io; | ||
528 | bio_endio(bio, error); | ||
529 | |||
530 | return; | ||
531 | } | ||
532 | |||
529 | INIT_WORK(&bip->bip_work, bio_integrity_verify_fn); | 533 | INIT_WORK(&bip->bip_work, bio_integrity_verify_fn); |
530 | queue_work(kintegrityd_wq, &bip->bip_work); | 534 | queue_work(kintegrityd_wq, &bip->bip_work); |
531 | } | 535 | } |
@@ -302,9 +302,10 @@ void bio_init(struct bio *bio) | |||
302 | struct bio *bio_alloc_bioset(gfp_t gfp_mask, int nr_iovecs, struct bio_set *bs) | 302 | struct bio *bio_alloc_bioset(gfp_t gfp_mask, int nr_iovecs, struct bio_set *bs) |
303 | { | 303 | { |
304 | struct bio *bio = NULL; | 304 | struct bio *bio = NULL; |
305 | void *uninitialized_var(p); | ||
305 | 306 | ||
306 | if (bs) { | 307 | if (bs) { |
307 | void *p = mempool_alloc(bs->bio_pool, gfp_mask); | 308 | p = mempool_alloc(bs->bio_pool, gfp_mask); |
308 | 309 | ||
309 | if (p) | 310 | if (p) |
310 | bio = p + bs->front_pad; | 311 | bio = p + bs->front_pad; |
@@ -329,7 +330,7 @@ struct bio *bio_alloc_bioset(gfp_t gfp_mask, int nr_iovecs, struct bio_set *bs) | |||
329 | } | 330 | } |
330 | if (unlikely(!bvl)) { | 331 | if (unlikely(!bvl)) { |
331 | if (bs) | 332 | if (bs) |
332 | mempool_free(bio, bs->bio_pool); | 333 | mempool_free(p, bs->bio_pool); |
333 | else | 334 | else |
334 | kfree(bio); | 335 | kfree(bio); |
335 | bio = NULL; | 336 | bio = NULL; |
diff --git a/fs/btrfs/Kconfig b/fs/btrfs/Kconfig new file mode 100644 index 000000000000..7bb3c020e570 --- /dev/null +++ b/fs/btrfs/Kconfig | |||
@@ -0,0 +1,31 @@ | |||
1 | config BTRFS_FS | ||
2 | tristate "Btrfs filesystem (EXPERIMENTAL) Unstable disk format" | ||
3 | depends on EXPERIMENTAL | ||
4 | select LIBCRC32C | ||
5 | select ZLIB_INFLATE | ||
6 | select ZLIB_DEFLATE | ||
7 | help | ||
8 | Btrfs is a new filesystem with extents, writable snapshotting, | ||
9 | support for multiple devices and many more features. | ||
10 | |||
11 | Btrfs is highly experimental, and THE DISK FORMAT IS NOT YET | ||
12 | FINALIZED. You should say N here unless you are interested in | ||
13 | testing Btrfs with non-critical data. | ||
14 | |||
15 | To compile this file system support as a module, choose M here. The | ||
16 | module will be called btrfs. | ||
17 | |||
18 | If unsure, say N. | ||
19 | |||
20 | config BTRFS_FS_POSIX_ACL | ||
21 | bool "Btrfs POSIX Access Control Lists" | ||
22 | depends on BTRFS_FS | ||
23 | select FS_POSIX_ACL | ||
24 | help | ||
25 | POSIX Access Control Lists (ACLs) support permissions for users and | ||
26 | groups beyond the owner/group/world scheme. | ||
27 | |||
28 | To learn more about Access Control Lists, visit the POSIX ACLs for | ||
29 | Linux website <http://acl.bestbits.at/>. | ||
30 | |||
31 | If you don't know what Access Control Lists are, say N | ||
diff --git a/fs/btrfs/async-thread.c b/fs/btrfs/async-thread.c index 8e2fec05dbe0..c84ca1f5259a 100644 --- a/fs/btrfs/async-thread.c +++ b/fs/btrfs/async-thread.c | |||
@@ -16,11 +16,11 @@ | |||
16 | * Boston, MA 021110-1307, USA. | 16 | * Boston, MA 021110-1307, USA. |
17 | */ | 17 | */ |
18 | 18 | ||
19 | #include <linux/version.h> | ||
20 | #include <linux/kthread.h> | 19 | #include <linux/kthread.h> |
21 | #include <linux/list.h> | 20 | #include <linux/list.h> |
22 | #include <linux/spinlock.h> | 21 | #include <linux/spinlock.h> |
23 | # include <linux/freezer.h> | 22 | #include <linux/freezer.h> |
23 | #include <linux/ftrace.h> | ||
24 | #include "async-thread.h" | 24 | #include "async-thread.h" |
25 | 25 | ||
26 | #define WORK_QUEUED_BIT 0 | 26 | #define WORK_QUEUED_BIT 0 |
@@ -143,6 +143,7 @@ static int worker_loop(void *arg) | |||
143 | struct btrfs_work *work; | 143 | struct btrfs_work *work; |
144 | do { | 144 | do { |
145 | spin_lock_irq(&worker->lock); | 145 | spin_lock_irq(&worker->lock); |
146 | again_locked: | ||
146 | while (!list_empty(&worker->pending)) { | 147 | while (!list_empty(&worker->pending)) { |
147 | cur = worker->pending.next; | 148 | cur = worker->pending.next; |
148 | work = list_entry(cur, struct btrfs_work, list); | 149 | work = list_entry(cur, struct btrfs_work, list); |
@@ -165,14 +166,50 @@ static int worker_loop(void *arg) | |||
165 | check_idle_worker(worker); | 166 | check_idle_worker(worker); |
166 | 167 | ||
167 | } | 168 | } |
168 | worker->working = 0; | ||
169 | if (freezing(current)) { | 169 | if (freezing(current)) { |
170 | worker->working = 0; | ||
171 | spin_unlock_irq(&worker->lock); | ||
170 | refrigerator(); | 172 | refrigerator(); |
171 | } else { | 173 | } else { |
172 | set_current_state(TASK_INTERRUPTIBLE); | ||
173 | spin_unlock_irq(&worker->lock); | 174 | spin_unlock_irq(&worker->lock); |
174 | if (!kthread_should_stop()) | 175 | if (!kthread_should_stop()) { |
176 | cpu_relax(); | ||
177 | /* | ||
178 | * we've dropped the lock, did someone else | ||
179 | * jump_in? | ||
180 | */ | ||
181 | smp_mb(); | ||
182 | if (!list_empty(&worker->pending)) | ||
183 | continue; | ||
184 | |||
185 | /* | ||
186 | * this short schedule allows more work to | ||
187 | * come in without the queue functions | ||
188 | * needing to go through wake_up_process() | ||
189 | * | ||
190 | * worker->working is still 1, so nobody | ||
191 | * is going to try and wake us up | ||
192 | */ | ||
193 | schedule_timeout(1); | ||
194 | smp_mb(); | ||
195 | if (!list_empty(&worker->pending)) | ||
196 | continue; | ||
197 | |||
198 | /* still no more work?, sleep for real */ | ||
199 | spin_lock_irq(&worker->lock); | ||
200 | set_current_state(TASK_INTERRUPTIBLE); | ||
201 | if (!list_empty(&worker->pending)) | ||
202 | goto again_locked; | ||
203 | |||
204 | /* | ||
205 | * this makes sure we get a wakeup when someone | ||
206 | * adds something new to the queue | ||
207 | */ | ||
208 | worker->working = 0; | ||
209 | spin_unlock_irq(&worker->lock); | ||
210 | |||
175 | schedule(); | 211 | schedule(); |
212 | } | ||
176 | __set_current_state(TASK_RUNNING); | 213 | __set_current_state(TASK_RUNNING); |
177 | } | 214 | } |
178 | } while (!kthread_should_stop()); | 215 | } while (!kthread_should_stop()); |
@@ -350,13 +387,14 @@ int btrfs_requeue_work(struct btrfs_work *work) | |||
350 | { | 387 | { |
351 | struct btrfs_worker_thread *worker = work->worker; | 388 | struct btrfs_worker_thread *worker = work->worker; |
352 | unsigned long flags; | 389 | unsigned long flags; |
390 | int wake = 0; | ||
353 | 391 | ||
354 | if (test_and_set_bit(WORK_QUEUED_BIT, &work->flags)) | 392 | if (test_and_set_bit(WORK_QUEUED_BIT, &work->flags)) |
355 | goto out; | 393 | goto out; |
356 | 394 | ||
357 | spin_lock_irqsave(&worker->lock, flags); | 395 | spin_lock_irqsave(&worker->lock, flags); |
358 | atomic_inc(&worker->num_pending); | ||
359 | list_add_tail(&work->list, &worker->pending); | 396 | list_add_tail(&work->list, &worker->pending); |
397 | atomic_inc(&worker->num_pending); | ||
360 | 398 | ||
361 | /* by definition we're busy, take ourselves off the idle | 399 | /* by definition we're busy, take ourselves off the idle |
362 | * list | 400 | * list |
@@ -368,10 +406,16 @@ int btrfs_requeue_work(struct btrfs_work *work) | |||
368 | &worker->workers->worker_list); | 406 | &worker->workers->worker_list); |
369 | spin_unlock_irqrestore(&worker->workers->lock, flags); | 407 | spin_unlock_irqrestore(&worker->workers->lock, flags); |
370 | } | 408 | } |
409 | if (!worker->working) { | ||
410 | wake = 1; | ||
411 | worker->working = 1; | ||
412 | } | ||
371 | 413 | ||
372 | spin_unlock_irqrestore(&worker->lock, flags); | 414 | spin_unlock_irqrestore(&worker->lock, flags); |
373 | 415 | if (wake) | |
416 | wake_up_process(worker->task); | ||
374 | out: | 417 | out: |
418 | |||
375 | return 0; | 419 | return 0; |
376 | } | 420 | } |
377 | 421 | ||
@@ -398,9 +442,10 @@ int btrfs_queue_worker(struct btrfs_workers *workers, struct btrfs_work *work) | |||
398 | } | 442 | } |
399 | 443 | ||
400 | spin_lock_irqsave(&worker->lock, flags); | 444 | spin_lock_irqsave(&worker->lock, flags); |
445 | |||
446 | list_add_tail(&work->list, &worker->pending); | ||
401 | atomic_inc(&worker->num_pending); | 447 | atomic_inc(&worker->num_pending); |
402 | check_busy_worker(worker); | 448 | check_busy_worker(worker); |
403 | list_add_tail(&work->list, &worker->pending); | ||
404 | 449 | ||
405 | /* | 450 | /* |
406 | * avoid calling into wake_up_process if this thread has already | 451 | * avoid calling into wake_up_process if this thread has already |
diff --git a/fs/btrfs/btrfs_inode.h b/fs/btrfs/btrfs_inode.h index a8c9693b75ac..72677ce2b74f 100644 --- a/fs/btrfs/btrfs_inode.h +++ b/fs/btrfs/btrfs_inode.h | |||
@@ -66,6 +66,9 @@ struct btrfs_inode { | |||
66 | */ | 66 | */ |
67 | struct list_head delalloc_inodes; | 67 | struct list_head delalloc_inodes; |
68 | 68 | ||
69 | /* the space_info for where this inode's data allocations are done */ | ||
70 | struct btrfs_space_info *space_info; | ||
71 | |||
69 | /* full 64 bit generation number, struct vfs_inode doesn't have a big | 72 | /* full 64 bit generation number, struct vfs_inode doesn't have a big |
70 | * enough field for this. | 73 | * enough field for this. |
71 | */ | 74 | */ |
@@ -94,6 +97,11 @@ struct btrfs_inode { | |||
94 | */ | 97 | */ |
95 | u64 delalloc_bytes; | 98 | u64 delalloc_bytes; |
96 | 99 | ||
100 | /* total number of bytes that may be used for this inode for | ||
101 | * delalloc | ||
102 | */ | ||
103 | u64 reserved_bytes; | ||
104 | |||
97 | /* | 105 | /* |
98 | * the size of the file stored in the metadata on disk. data=ordered | 106 | * the size of the file stored in the metadata on disk. data=ordered |
99 | * means the in-memory i_size might be larger than the size on disk | 107 | * means the in-memory i_size might be larger than the size on disk |
diff --git a/fs/btrfs/compression.c b/fs/btrfs/compression.c index ee848d8585d9..ab07627084f1 100644 --- a/fs/btrfs/compression.c +++ b/fs/btrfs/compression.c | |||
@@ -32,7 +32,6 @@ | |||
32 | #include <linux/swap.h> | 32 | #include <linux/swap.h> |
33 | #include <linux/writeback.h> | 33 | #include <linux/writeback.h> |
34 | #include <linux/bit_spinlock.h> | 34 | #include <linux/bit_spinlock.h> |
35 | #include <linux/version.h> | ||
36 | #include <linux/pagevec.h> | 35 | #include <linux/pagevec.h> |
37 | #include "compat.h" | 36 | #include "compat.h" |
38 | #include "ctree.h" | 37 | #include "ctree.h" |
diff --git a/fs/btrfs/ctree.c b/fs/btrfs/ctree.c index 9e46c0776816..42491d728e99 100644 --- a/fs/btrfs/ctree.c +++ b/fs/btrfs/ctree.c | |||
@@ -38,22 +38,64 @@ static int balance_node_right(struct btrfs_trans_handle *trans, | |||
38 | static int del_ptr(struct btrfs_trans_handle *trans, struct btrfs_root *root, | 38 | static int del_ptr(struct btrfs_trans_handle *trans, struct btrfs_root *root, |
39 | struct btrfs_path *path, int level, int slot); | 39 | struct btrfs_path *path, int level, int slot); |
40 | 40 | ||
41 | inline void btrfs_init_path(struct btrfs_path *p) | ||
42 | { | ||
43 | memset(p, 0, sizeof(*p)); | ||
44 | } | ||
45 | |||
46 | struct btrfs_path *btrfs_alloc_path(void) | 41 | struct btrfs_path *btrfs_alloc_path(void) |
47 | { | 42 | { |
48 | struct btrfs_path *path; | 43 | struct btrfs_path *path; |
49 | path = kmem_cache_alloc(btrfs_path_cachep, GFP_NOFS); | 44 | path = kmem_cache_zalloc(btrfs_path_cachep, GFP_NOFS); |
50 | if (path) { | 45 | if (path) |
51 | btrfs_init_path(path); | ||
52 | path->reada = 1; | 46 | path->reada = 1; |
53 | } | ||
54 | return path; | 47 | return path; |
55 | } | 48 | } |
56 | 49 | ||
50 | /* | ||
51 | * set all locked nodes in the path to blocking locks. This should | ||
52 | * be done before scheduling | ||
53 | */ | ||
54 | noinline void btrfs_set_path_blocking(struct btrfs_path *p) | ||
55 | { | ||
56 | int i; | ||
57 | for (i = 0; i < BTRFS_MAX_LEVEL; i++) { | ||
58 | if (p->nodes[i] && p->locks[i]) | ||
59 | btrfs_set_lock_blocking(p->nodes[i]); | ||
60 | } | ||
61 | } | ||
62 | |||
63 | /* | ||
64 | * reset all the locked nodes in the patch to spinning locks. | ||
65 | * | ||
66 | * held is used to keep lockdep happy, when lockdep is enabled | ||
67 | * we set held to a blocking lock before we go around and | ||
68 | * retake all the spinlocks in the path. You can safely use NULL | ||
69 | * for held | ||
70 | */ | ||
71 | noinline void btrfs_clear_path_blocking(struct btrfs_path *p, | ||
72 | struct extent_buffer *held) | ||
73 | { | ||
74 | int i; | ||
75 | |||
76 | #ifdef CONFIG_DEBUG_LOCK_ALLOC | ||
77 | /* lockdep really cares that we take all of these spinlocks | ||
78 | * in the right order. If any of the locks in the path are not | ||
79 | * currently blocking, it is going to complain. So, make really | ||
80 | * really sure by forcing the path to blocking before we clear | ||
81 | * the path blocking. | ||
82 | */ | ||
83 | if (held) | ||
84 | btrfs_set_lock_blocking(held); | ||
85 | btrfs_set_path_blocking(p); | ||
86 | #endif | ||
87 | |||
88 | for (i = BTRFS_MAX_LEVEL - 1; i >= 0; i--) { | ||
89 | if (p->nodes[i] && p->locks[i]) | ||
90 | btrfs_clear_lock_blocking(p->nodes[i]); | ||
91 | } | ||
92 | |||
93 | #ifdef CONFIG_DEBUG_LOCK_ALLOC | ||
94 | if (held) | ||
95 | btrfs_clear_lock_blocking(held); | ||
96 | #endif | ||
97 | } | ||
98 | |||
57 | /* this also releases the path */ | 99 | /* this also releases the path */ |
58 | void btrfs_free_path(struct btrfs_path *p) | 100 | void btrfs_free_path(struct btrfs_path *p) |
59 | { | 101 | { |
@@ -261,7 +303,7 @@ static noinline int __btrfs_cow_block(struct btrfs_trans_handle *trans, | |||
261 | trans->transid, level, &ins); | 303 | trans->transid, level, &ins); |
262 | BUG_ON(ret); | 304 | BUG_ON(ret); |
263 | cow = btrfs_init_new_buffer(trans, root, prealloc_dest, | 305 | cow = btrfs_init_new_buffer(trans, root, prealloc_dest, |
264 | buf->len); | 306 | buf->len, level); |
265 | } else { | 307 | } else { |
266 | cow = btrfs_alloc_free_block(trans, root, buf->len, | 308 | cow = btrfs_alloc_free_block(trans, root, buf->len, |
267 | parent_start, | 309 | parent_start, |
@@ -272,6 +314,8 @@ static noinline int __btrfs_cow_block(struct btrfs_trans_handle *trans, | |||
272 | if (IS_ERR(cow)) | 314 | if (IS_ERR(cow)) |
273 | return PTR_ERR(cow); | 315 | return PTR_ERR(cow); |
274 | 316 | ||
317 | /* cow is set to blocking by btrfs_init_new_buffer */ | ||
318 | |||
275 | copy_extent_buffer(cow, buf, 0, 0, cow->len); | 319 | copy_extent_buffer(cow, buf, 0, 0, cow->len); |
276 | btrfs_set_header_bytenr(cow, cow->start); | 320 | btrfs_set_header_bytenr(cow, cow->start); |
277 | btrfs_set_header_generation(cow, trans->transid); | 321 | btrfs_set_header_generation(cow, trans->transid); |
@@ -388,17 +432,20 @@ noinline int btrfs_cow_block(struct btrfs_trans_handle *trans, | |||
388 | WARN_ON(1); | 432 | WARN_ON(1); |
389 | } | 433 | } |
390 | 434 | ||
391 | spin_lock(&root->fs_info->hash_lock); | ||
392 | if (btrfs_header_generation(buf) == trans->transid && | 435 | if (btrfs_header_generation(buf) == trans->transid && |
393 | btrfs_header_owner(buf) == root->root_key.objectid && | 436 | btrfs_header_owner(buf) == root->root_key.objectid && |
394 | !btrfs_header_flag(buf, BTRFS_HEADER_FLAG_WRITTEN)) { | 437 | !btrfs_header_flag(buf, BTRFS_HEADER_FLAG_WRITTEN)) { |
395 | *cow_ret = buf; | 438 | *cow_ret = buf; |
396 | spin_unlock(&root->fs_info->hash_lock); | ||
397 | WARN_ON(prealloc_dest); | 439 | WARN_ON(prealloc_dest); |
398 | return 0; | 440 | return 0; |
399 | } | 441 | } |
400 | spin_unlock(&root->fs_info->hash_lock); | 442 | |
401 | search_start = buf->start & ~((u64)(1024 * 1024 * 1024) - 1); | 443 | search_start = buf->start & ~((u64)(1024 * 1024 * 1024) - 1); |
444 | |||
445 | if (parent) | ||
446 | btrfs_set_lock_blocking(parent); | ||
447 | btrfs_set_lock_blocking(buf); | ||
448 | |||
402 | ret = __btrfs_cow_block(trans, root, buf, parent, | 449 | ret = __btrfs_cow_block(trans, root, buf, parent, |
403 | parent_slot, cow_ret, search_start, 0, | 450 | parent_slot, cow_ret, search_start, 0, |
404 | prealloc_dest); | 451 | prealloc_dest); |
@@ -504,6 +551,8 @@ int btrfs_realloc_node(struct btrfs_trans_handle *trans, | |||
504 | if (parent_nritems == 1) | 551 | if (parent_nritems == 1) |
505 | return 0; | 552 | return 0; |
506 | 553 | ||
554 | btrfs_set_lock_blocking(parent); | ||
555 | |||
507 | for (i = start_slot; i < end_slot; i++) { | 556 | for (i = start_slot; i < end_slot; i++) { |
508 | int close = 1; | 557 | int close = 1; |
509 | 558 | ||
@@ -564,6 +613,7 @@ int btrfs_realloc_node(struct btrfs_trans_handle *trans, | |||
564 | search_start = last_block; | 613 | search_start = last_block; |
565 | 614 | ||
566 | btrfs_tree_lock(cur); | 615 | btrfs_tree_lock(cur); |
616 | btrfs_set_lock_blocking(cur); | ||
567 | err = __btrfs_cow_block(trans, root, cur, parent, i, | 617 | err = __btrfs_cow_block(trans, root, cur, parent, i, |
568 | &cur, search_start, | 618 | &cur, search_start, |
569 | min(16 * blocksize, | 619 | min(16 * blocksize, |
@@ -862,6 +912,7 @@ static noinline int balance_level(struct btrfs_trans_handle *trans, | |||
862 | return 0; | 912 | return 0; |
863 | 913 | ||
864 | mid = path->nodes[level]; | 914 | mid = path->nodes[level]; |
915 | |||
865 | WARN_ON(!path->locks[level]); | 916 | WARN_ON(!path->locks[level]); |
866 | WARN_ON(btrfs_header_generation(mid) != trans->transid); | 917 | WARN_ON(btrfs_header_generation(mid) != trans->transid); |
867 | 918 | ||
@@ -883,8 +934,9 @@ static noinline int balance_level(struct btrfs_trans_handle *trans, | |||
883 | 934 | ||
884 | /* promote the child to a root */ | 935 | /* promote the child to a root */ |
885 | child = read_node_slot(root, mid, 0); | 936 | child = read_node_slot(root, mid, 0); |
886 | btrfs_tree_lock(child); | ||
887 | BUG_ON(!child); | 937 | BUG_ON(!child); |
938 | btrfs_tree_lock(child); | ||
939 | btrfs_set_lock_blocking(child); | ||
888 | ret = btrfs_cow_block(trans, root, child, mid, 0, &child, 0); | 940 | ret = btrfs_cow_block(trans, root, child, mid, 0, &child, 0); |
889 | BUG_ON(ret); | 941 | BUG_ON(ret); |
890 | 942 | ||
@@ -900,6 +952,7 @@ static noinline int balance_level(struct btrfs_trans_handle *trans, | |||
900 | 952 | ||
901 | add_root_to_dirty_list(root); | 953 | add_root_to_dirty_list(root); |
902 | btrfs_tree_unlock(child); | 954 | btrfs_tree_unlock(child); |
955 | |||
903 | path->locks[level] = 0; | 956 | path->locks[level] = 0; |
904 | path->nodes[level] = NULL; | 957 | path->nodes[level] = NULL; |
905 | clean_tree_block(trans, root, mid); | 958 | clean_tree_block(trans, root, mid); |
@@ -924,6 +977,7 @@ static noinline int balance_level(struct btrfs_trans_handle *trans, | |||
924 | left = read_node_slot(root, parent, pslot - 1); | 977 | left = read_node_slot(root, parent, pslot - 1); |
925 | if (left) { | 978 | if (left) { |
926 | btrfs_tree_lock(left); | 979 | btrfs_tree_lock(left); |
980 | btrfs_set_lock_blocking(left); | ||
927 | wret = btrfs_cow_block(trans, root, left, | 981 | wret = btrfs_cow_block(trans, root, left, |
928 | parent, pslot - 1, &left, 0); | 982 | parent, pslot - 1, &left, 0); |
929 | if (wret) { | 983 | if (wret) { |
@@ -934,6 +988,7 @@ static noinline int balance_level(struct btrfs_trans_handle *trans, | |||
934 | right = read_node_slot(root, parent, pslot + 1); | 988 | right = read_node_slot(root, parent, pslot + 1); |
935 | if (right) { | 989 | if (right) { |
936 | btrfs_tree_lock(right); | 990 | btrfs_tree_lock(right); |
991 | btrfs_set_lock_blocking(right); | ||
937 | wret = btrfs_cow_block(trans, root, right, | 992 | wret = btrfs_cow_block(trans, root, right, |
938 | parent, pslot + 1, &right, 0); | 993 | parent, pslot + 1, &right, 0); |
939 | if (wret) { | 994 | if (wret) { |
@@ -1109,6 +1164,8 @@ static noinline int push_nodes_for_insert(struct btrfs_trans_handle *trans, | |||
1109 | u32 left_nr; | 1164 | u32 left_nr; |
1110 | 1165 | ||
1111 | btrfs_tree_lock(left); | 1166 | btrfs_tree_lock(left); |
1167 | btrfs_set_lock_blocking(left); | ||
1168 | |||
1112 | left_nr = btrfs_header_nritems(left); | 1169 | left_nr = btrfs_header_nritems(left); |
1113 | if (left_nr >= BTRFS_NODEPTRS_PER_BLOCK(root) - 1) { | 1170 | if (left_nr >= BTRFS_NODEPTRS_PER_BLOCK(root) - 1) { |
1114 | wret = 1; | 1171 | wret = 1; |
@@ -1155,7 +1212,10 @@ static noinline int push_nodes_for_insert(struct btrfs_trans_handle *trans, | |||
1155 | */ | 1212 | */ |
1156 | if (right) { | 1213 | if (right) { |
1157 | u32 right_nr; | 1214 | u32 right_nr; |
1215 | |||
1158 | btrfs_tree_lock(right); | 1216 | btrfs_tree_lock(right); |
1217 | btrfs_set_lock_blocking(right); | ||
1218 | |||
1159 | right_nr = btrfs_header_nritems(right); | 1219 | right_nr = btrfs_header_nritems(right); |
1160 | if (right_nr >= BTRFS_NODEPTRS_PER_BLOCK(root) - 1) { | 1220 | if (right_nr >= BTRFS_NODEPTRS_PER_BLOCK(root) - 1) { |
1161 | wret = 1; | 1221 | wret = 1; |
@@ -1210,8 +1270,7 @@ static noinline void reada_for_search(struct btrfs_root *root, | |||
1210 | struct btrfs_disk_key disk_key; | 1270 | struct btrfs_disk_key disk_key; |
1211 | u32 nritems; | 1271 | u32 nritems; |
1212 | u64 search; | 1272 | u64 search; |
1213 | u64 lowest_read; | 1273 | u64 target; |
1214 | u64 highest_read; | ||
1215 | u64 nread = 0; | 1274 | u64 nread = 0; |
1216 | int direction = path->reada; | 1275 | int direction = path->reada; |
1217 | struct extent_buffer *eb; | 1276 | struct extent_buffer *eb; |
@@ -1235,8 +1294,7 @@ static noinline void reada_for_search(struct btrfs_root *root, | |||
1235 | return; | 1294 | return; |
1236 | } | 1295 | } |
1237 | 1296 | ||
1238 | highest_read = search; | 1297 | target = search; |
1239 | lowest_read = search; | ||
1240 | 1298 | ||
1241 | nritems = btrfs_header_nritems(node); | 1299 | nritems = btrfs_header_nritems(node); |
1242 | nr = slot; | 1300 | nr = slot; |
@@ -1256,27 +1314,80 @@ static noinline void reada_for_search(struct btrfs_root *root, | |||
1256 | break; | 1314 | break; |
1257 | } | 1315 | } |
1258 | search = btrfs_node_blockptr(node, nr); | 1316 | search = btrfs_node_blockptr(node, nr); |
1259 | if ((search >= lowest_read && search <= highest_read) || | 1317 | if ((search <= target && target - search <= 65536) || |
1260 | (search < lowest_read && lowest_read - search <= 16384) || | 1318 | (search > target && search - target <= 65536)) { |
1261 | (search > highest_read && search - highest_read <= 16384)) { | ||
1262 | readahead_tree_block(root, search, blocksize, | 1319 | readahead_tree_block(root, search, blocksize, |
1263 | btrfs_node_ptr_generation(node, nr)); | 1320 | btrfs_node_ptr_generation(node, nr)); |
1264 | nread += blocksize; | 1321 | nread += blocksize; |
1265 | } | 1322 | } |
1266 | nscan++; | 1323 | nscan++; |
1267 | if (path->reada < 2 && (nread > (64 * 1024) || nscan > 32)) | 1324 | if ((nread > 65536 || nscan > 32)) |
1268 | break; | 1325 | break; |
1326 | } | ||
1327 | } | ||
1269 | 1328 | ||
1270 | if (nread > (256 * 1024) || nscan > 128) | 1329 | /* |
1271 | break; | 1330 | * returns -EAGAIN if it had to drop the path, or zero if everything was in |
1331 | * cache | ||
1332 | */ | ||
1333 | static noinline int reada_for_balance(struct btrfs_root *root, | ||
1334 | struct btrfs_path *path, int level) | ||
1335 | { | ||
1336 | int slot; | ||
1337 | int nritems; | ||
1338 | struct extent_buffer *parent; | ||
1339 | struct extent_buffer *eb; | ||
1340 | u64 gen; | ||
1341 | u64 block1 = 0; | ||
1342 | u64 block2 = 0; | ||
1343 | int ret = 0; | ||
1344 | int blocksize; | ||
1345 | |||
1346 | parent = path->nodes[level - 1]; | ||
1347 | if (!parent) | ||
1348 | return 0; | ||
1272 | 1349 | ||
1273 | if (search < lowest_read) | 1350 | nritems = btrfs_header_nritems(parent); |
1274 | lowest_read = search; | 1351 | slot = path->slots[level]; |
1275 | if (search > highest_read) | 1352 | blocksize = btrfs_level_size(root, level); |
1276 | highest_read = search; | 1353 | |
1354 | if (slot > 0) { | ||
1355 | block1 = btrfs_node_blockptr(parent, slot - 1); | ||
1356 | gen = btrfs_node_ptr_generation(parent, slot - 1); | ||
1357 | eb = btrfs_find_tree_block(root, block1, blocksize); | ||
1358 | if (eb && btrfs_buffer_uptodate(eb, gen)) | ||
1359 | block1 = 0; | ||
1360 | free_extent_buffer(eb); | ||
1361 | } | ||
1362 | if (slot < nritems) { | ||
1363 | block2 = btrfs_node_blockptr(parent, slot + 1); | ||
1364 | gen = btrfs_node_ptr_generation(parent, slot + 1); | ||
1365 | eb = btrfs_find_tree_block(root, block2, blocksize); | ||
1366 | if (eb && btrfs_buffer_uptodate(eb, gen)) | ||
1367 | block2 = 0; | ||
1368 | free_extent_buffer(eb); | ||
1277 | } | 1369 | } |
1370 | if (block1 || block2) { | ||
1371 | ret = -EAGAIN; | ||
1372 | btrfs_release_path(root, path); | ||
1373 | if (block1) | ||
1374 | readahead_tree_block(root, block1, blocksize, 0); | ||
1375 | if (block2) | ||
1376 | readahead_tree_block(root, block2, blocksize, 0); | ||
1377 | |||
1378 | if (block1) { | ||
1379 | eb = read_tree_block(root, block1, blocksize, 0); | ||
1380 | free_extent_buffer(eb); | ||
1381 | } | ||
1382 | if (block1) { | ||
1383 | eb = read_tree_block(root, block2, blocksize, 0); | ||
1384 | free_extent_buffer(eb); | ||
1385 | } | ||
1386 | } | ||
1387 | return ret; | ||
1278 | } | 1388 | } |
1279 | 1389 | ||
1390 | |||
1280 | /* | 1391 | /* |
1281 | * when we walk down the tree, it is usually safe to unlock the higher layers | 1392 | * when we walk down the tree, it is usually safe to unlock the higher layers |
1282 | * in the tree. The exceptions are when our path goes through slot 0, because | 1393 | * in the tree. The exceptions are when our path goes through slot 0, because |
@@ -1328,6 +1439,32 @@ static noinline void unlock_up(struct btrfs_path *path, int level, | |||
1328 | } | 1439 | } |
1329 | 1440 | ||
1330 | /* | 1441 | /* |
1442 | * This releases any locks held in the path starting at level and | ||
1443 | * going all the way up to the root. | ||
1444 | * | ||
1445 | * btrfs_search_slot will keep the lock held on higher nodes in a few | ||
1446 | * corner cases, such as COW of the block at slot zero in the node. This | ||
1447 | * ignores those rules, and it should only be called when there are no | ||
1448 | * more updates to be done higher up in the tree. | ||
1449 | */ | ||
1450 | noinline void btrfs_unlock_up_safe(struct btrfs_path *path, int level) | ||
1451 | { | ||
1452 | int i; | ||
1453 | |||
1454 | if (path->keep_locks || path->lowest_level) | ||
1455 | return; | ||
1456 | |||
1457 | for (i = level; i < BTRFS_MAX_LEVEL; i++) { | ||
1458 | if (!path->nodes[i]) | ||
1459 | continue; | ||
1460 | if (!path->locks[i]) | ||
1461 | continue; | ||
1462 | btrfs_tree_unlock(path->nodes[i]); | ||
1463 | path->locks[i] = 0; | ||
1464 | } | ||
1465 | } | ||
1466 | |||
1467 | /* | ||
1331 | * look for key in the tree. path is filled in with nodes along the way | 1468 | * look for key in the tree. path is filled in with nodes along the way |
1332 | * if key is found, we return zero and you can find the item in the leaf | 1469 | * if key is found, we return zero and you can find the item in the leaf |
1333 | * level of the path (level 0) | 1470 | * level of the path (level 0) |
@@ -1387,32 +1524,30 @@ again: | |||
1387 | int wret; | 1524 | int wret; |
1388 | 1525 | ||
1389 | /* is a cow on this block not required */ | 1526 | /* is a cow on this block not required */ |
1390 | spin_lock(&root->fs_info->hash_lock); | ||
1391 | if (btrfs_header_generation(b) == trans->transid && | 1527 | if (btrfs_header_generation(b) == trans->transid && |
1392 | btrfs_header_owner(b) == root->root_key.objectid && | 1528 | btrfs_header_owner(b) == root->root_key.objectid && |
1393 | !btrfs_header_flag(b, BTRFS_HEADER_FLAG_WRITTEN)) { | 1529 | !btrfs_header_flag(b, BTRFS_HEADER_FLAG_WRITTEN)) { |
1394 | spin_unlock(&root->fs_info->hash_lock); | ||
1395 | goto cow_done; | 1530 | goto cow_done; |
1396 | } | 1531 | } |
1397 | spin_unlock(&root->fs_info->hash_lock); | ||
1398 | 1532 | ||
1399 | /* ok, we have to cow, is our old prealloc the right | 1533 | /* ok, we have to cow, is our old prealloc the right |
1400 | * size? | 1534 | * size? |
1401 | */ | 1535 | */ |
1402 | if (prealloc_block.objectid && | 1536 | if (prealloc_block.objectid && |
1403 | prealloc_block.offset != b->len) { | 1537 | prealloc_block.offset != b->len) { |
1538 | btrfs_release_path(root, p); | ||
1404 | btrfs_free_reserved_extent(root, | 1539 | btrfs_free_reserved_extent(root, |
1405 | prealloc_block.objectid, | 1540 | prealloc_block.objectid, |
1406 | prealloc_block.offset); | 1541 | prealloc_block.offset); |
1407 | prealloc_block.objectid = 0; | 1542 | prealloc_block.objectid = 0; |
1543 | goto again; | ||
1408 | } | 1544 | } |
1409 | 1545 | ||
1410 | /* | 1546 | /* |
1411 | * for higher level blocks, try not to allocate blocks | 1547 | * for higher level blocks, try not to allocate blocks |
1412 | * with the block and the parent locks held. | 1548 | * with the block and the parent locks held. |
1413 | */ | 1549 | */ |
1414 | if (level > 1 && !prealloc_block.objectid && | 1550 | if (level > 0 && !prealloc_block.objectid) { |
1415 | btrfs_path_lock_waiting(p, level)) { | ||
1416 | u32 size = b->len; | 1551 | u32 size = b->len; |
1417 | u64 hint = b->start; | 1552 | u64 hint = b->start; |
1418 | 1553 | ||
@@ -1425,6 +1560,8 @@ again: | |||
1425 | goto again; | 1560 | goto again; |
1426 | } | 1561 | } |
1427 | 1562 | ||
1563 | btrfs_set_path_blocking(p); | ||
1564 | |||
1428 | wret = btrfs_cow_block(trans, root, b, | 1565 | wret = btrfs_cow_block(trans, root, b, |
1429 | p->nodes[level + 1], | 1566 | p->nodes[level + 1], |
1430 | p->slots[level + 1], | 1567 | p->slots[level + 1], |
@@ -1446,6 +1583,22 @@ cow_done: | |||
1446 | if (!p->skip_locking) | 1583 | if (!p->skip_locking) |
1447 | p->locks[level] = 1; | 1584 | p->locks[level] = 1; |
1448 | 1585 | ||
1586 | btrfs_clear_path_blocking(p, NULL); | ||
1587 | |||
1588 | /* | ||
1589 | * we have a lock on b and as long as we aren't changing | ||
1590 | * the tree, there is no way to for the items in b to change. | ||
1591 | * It is safe to drop the lock on our parent before we | ||
1592 | * go through the expensive btree search on b. | ||
1593 | * | ||
1594 | * If cow is true, then we might be changing slot zero, | ||
1595 | * which may require changing the parent. So, we can't | ||
1596 | * drop the lock until after we know which slot we're | ||
1597 | * operating on. | ||
1598 | */ | ||
1599 | if (!cow) | ||
1600 | btrfs_unlock_up_safe(p, level + 1); | ||
1601 | |||
1449 | ret = check_block(root, p, level); | 1602 | ret = check_block(root, p, level); |
1450 | if (ret) { | 1603 | if (ret) { |
1451 | ret = -1; | 1604 | ret = -1; |
@@ -1453,6 +1606,7 @@ cow_done: | |||
1453 | } | 1606 | } |
1454 | 1607 | ||
1455 | ret = bin_search(b, key, level, &slot); | 1608 | ret = bin_search(b, key, level, &slot); |
1609 | |||
1456 | if (level != 0) { | 1610 | if (level != 0) { |
1457 | if (ret && slot > 0) | 1611 | if (ret && slot > 0) |
1458 | slot -= 1; | 1612 | slot -= 1; |
@@ -1460,7 +1614,16 @@ cow_done: | |||
1460 | if ((p->search_for_split || ins_len > 0) && | 1614 | if ((p->search_for_split || ins_len > 0) && |
1461 | btrfs_header_nritems(b) >= | 1615 | btrfs_header_nritems(b) >= |
1462 | BTRFS_NODEPTRS_PER_BLOCK(root) - 3) { | 1616 | BTRFS_NODEPTRS_PER_BLOCK(root) - 3) { |
1463 | int sret = split_node(trans, root, p, level); | 1617 | int sret; |
1618 | |||
1619 | sret = reada_for_balance(root, p, level); | ||
1620 | if (sret) | ||
1621 | goto again; | ||
1622 | |||
1623 | btrfs_set_path_blocking(p); | ||
1624 | sret = split_node(trans, root, p, level); | ||
1625 | btrfs_clear_path_blocking(p, NULL); | ||
1626 | |||
1464 | BUG_ON(sret > 0); | 1627 | BUG_ON(sret > 0); |
1465 | if (sret) { | 1628 | if (sret) { |
1466 | ret = sret; | 1629 | ret = sret; |
@@ -1468,9 +1631,19 @@ cow_done: | |||
1468 | } | 1631 | } |
1469 | b = p->nodes[level]; | 1632 | b = p->nodes[level]; |
1470 | slot = p->slots[level]; | 1633 | slot = p->slots[level]; |
1471 | } else if (ins_len < 0) { | 1634 | } else if (ins_len < 0 && |
1472 | int sret = balance_level(trans, root, p, | 1635 | btrfs_header_nritems(b) < |
1473 | level); | 1636 | BTRFS_NODEPTRS_PER_BLOCK(root) / 4) { |
1637 | int sret; | ||
1638 | |||
1639 | sret = reada_for_balance(root, p, level); | ||
1640 | if (sret) | ||
1641 | goto again; | ||
1642 | |||
1643 | btrfs_set_path_blocking(p); | ||
1644 | sret = balance_level(trans, root, p, level); | ||
1645 | btrfs_clear_path_blocking(p, NULL); | ||
1646 | |||
1474 | if (sret) { | 1647 | if (sret) { |
1475 | ret = sret; | 1648 | ret = sret; |
1476 | goto done; | 1649 | goto done; |
@@ -1504,7 +1677,7 @@ cow_done: | |||
1504 | * of the btree by dropping locks before | 1677 | * of the btree by dropping locks before |
1505 | * we read. | 1678 | * we read. |
1506 | */ | 1679 | */ |
1507 | if (level > 1) { | 1680 | if (level > 0) { |
1508 | btrfs_release_path(NULL, p); | 1681 | btrfs_release_path(NULL, p); |
1509 | if (tmp) | 1682 | if (tmp) |
1510 | free_extent_buffer(tmp); | 1683 | free_extent_buffer(tmp); |
@@ -1519,6 +1692,7 @@ cow_done: | |||
1519 | free_extent_buffer(tmp); | 1692 | free_extent_buffer(tmp); |
1520 | goto again; | 1693 | goto again; |
1521 | } else { | 1694 | } else { |
1695 | btrfs_set_path_blocking(p); | ||
1522 | if (tmp) | 1696 | if (tmp) |
1523 | free_extent_buffer(tmp); | 1697 | free_extent_buffer(tmp); |
1524 | if (should_reada) | 1698 | if (should_reada) |
@@ -1528,14 +1702,29 @@ cow_done: | |||
1528 | b = read_node_slot(root, b, slot); | 1702 | b = read_node_slot(root, b, slot); |
1529 | } | 1703 | } |
1530 | } | 1704 | } |
1531 | if (!p->skip_locking) | 1705 | if (!p->skip_locking) { |
1532 | btrfs_tree_lock(b); | 1706 | int lret; |
1707 | |||
1708 | btrfs_clear_path_blocking(p, NULL); | ||
1709 | lret = btrfs_try_spin_lock(b); | ||
1710 | |||
1711 | if (!lret) { | ||
1712 | btrfs_set_path_blocking(p); | ||
1713 | btrfs_tree_lock(b); | ||
1714 | btrfs_clear_path_blocking(p, b); | ||
1715 | } | ||
1716 | } | ||
1533 | } else { | 1717 | } else { |
1534 | p->slots[level] = slot; | 1718 | p->slots[level] = slot; |
1535 | if (ins_len > 0 && | 1719 | if (ins_len > 0 && |
1536 | btrfs_leaf_free_space(root, b) < ins_len) { | 1720 | btrfs_leaf_free_space(root, b) < ins_len) { |
1537 | int sret = split_leaf(trans, root, key, | 1721 | int sret; |
1722 | |||
1723 | btrfs_set_path_blocking(p); | ||
1724 | sret = split_leaf(trans, root, key, | ||
1538 | p, ins_len, ret == 0); | 1725 | p, ins_len, ret == 0); |
1726 | btrfs_clear_path_blocking(p, NULL); | ||
1727 | |||
1539 | BUG_ON(sret > 0); | 1728 | BUG_ON(sret > 0); |
1540 | if (sret) { | 1729 | if (sret) { |
1541 | ret = sret; | 1730 | ret = sret; |
@@ -1549,12 +1738,16 @@ cow_done: | |||
1549 | } | 1738 | } |
1550 | ret = 1; | 1739 | ret = 1; |
1551 | done: | 1740 | done: |
1741 | /* | ||
1742 | * we don't really know what they plan on doing with the path | ||
1743 | * from here on, so for now just mark it as blocking | ||
1744 | */ | ||
1745 | btrfs_set_path_blocking(p); | ||
1552 | if (prealloc_block.objectid) { | 1746 | if (prealloc_block.objectid) { |
1553 | btrfs_free_reserved_extent(root, | 1747 | btrfs_free_reserved_extent(root, |
1554 | prealloc_block.objectid, | 1748 | prealloc_block.objectid, |
1555 | prealloc_block.offset); | 1749 | prealloc_block.offset); |
1556 | } | 1750 | } |
1557 | |||
1558 | return ret; | 1751 | return ret; |
1559 | } | 1752 | } |
1560 | 1753 | ||
@@ -1578,6 +1771,8 @@ int btrfs_merge_path(struct btrfs_trans_handle *trans, | |||
1578 | ret = btrfs_cow_block(trans, root, eb, NULL, 0, &eb, 0); | 1771 | ret = btrfs_cow_block(trans, root, eb, NULL, 0, &eb, 0); |
1579 | BUG_ON(ret); | 1772 | BUG_ON(ret); |
1580 | 1773 | ||
1774 | btrfs_set_lock_blocking(eb); | ||
1775 | |||
1581 | parent = eb; | 1776 | parent = eb; |
1582 | while (1) { | 1777 | while (1) { |
1583 | level = btrfs_header_level(parent); | 1778 | level = btrfs_header_level(parent); |
@@ -1602,6 +1797,7 @@ int btrfs_merge_path(struct btrfs_trans_handle *trans, | |||
1602 | eb = read_tree_block(root, bytenr, blocksize, | 1797 | eb = read_tree_block(root, bytenr, blocksize, |
1603 | generation); | 1798 | generation); |
1604 | btrfs_tree_lock(eb); | 1799 | btrfs_tree_lock(eb); |
1800 | btrfs_set_lock_blocking(eb); | ||
1605 | } | 1801 | } |
1606 | 1802 | ||
1607 | /* | 1803 | /* |
@@ -1626,6 +1822,7 @@ int btrfs_merge_path(struct btrfs_trans_handle *trans, | |||
1626 | eb = read_tree_block(root, bytenr, blocksize, | 1822 | eb = read_tree_block(root, bytenr, blocksize, |
1627 | generation); | 1823 | generation); |
1628 | btrfs_tree_lock(eb); | 1824 | btrfs_tree_lock(eb); |
1825 | btrfs_set_lock_blocking(eb); | ||
1629 | } | 1826 | } |
1630 | 1827 | ||
1631 | ret = btrfs_cow_block(trans, root, eb, parent, slot, | 1828 | ret = btrfs_cow_block(trans, root, eb, parent, slot, |
@@ -2172,6 +2369,8 @@ static int push_leaf_right(struct btrfs_trans_handle *trans, struct btrfs_root | |||
2172 | 2369 | ||
2173 | right = read_node_slot(root, upper, slot + 1); | 2370 | right = read_node_slot(root, upper, slot + 1); |
2174 | btrfs_tree_lock(right); | 2371 | btrfs_tree_lock(right); |
2372 | btrfs_set_lock_blocking(right); | ||
2373 | |||
2175 | free_space = btrfs_leaf_free_space(root, right); | 2374 | free_space = btrfs_leaf_free_space(root, right); |
2176 | if (free_space < data_size) | 2375 | if (free_space < data_size) |
2177 | goto out_unlock; | 2376 | goto out_unlock; |
@@ -2367,6 +2566,8 @@ static int push_leaf_left(struct btrfs_trans_handle *trans, struct btrfs_root | |||
2367 | 2566 | ||
2368 | left = read_node_slot(root, path->nodes[1], slot - 1); | 2567 | left = read_node_slot(root, path->nodes[1], slot - 1); |
2369 | btrfs_tree_lock(left); | 2568 | btrfs_tree_lock(left); |
2569 | btrfs_set_lock_blocking(left); | ||
2570 | |||
2370 | free_space = btrfs_leaf_free_space(root, left); | 2571 | free_space = btrfs_leaf_free_space(root, left); |
2371 | if (free_space < data_size) { | 2572 | if (free_space < data_size) { |
2372 | ret = 1; | 2573 | ret = 1; |
@@ -2825,6 +3026,12 @@ int btrfs_split_item(struct btrfs_trans_handle *trans, | |||
2825 | path->keep_locks = 0; | 3026 | path->keep_locks = 0; |
2826 | BUG_ON(ret); | 3027 | BUG_ON(ret); |
2827 | 3028 | ||
3029 | /* | ||
3030 | * make sure any changes to the path from split_leaf leave it | ||
3031 | * in a blocking state | ||
3032 | */ | ||
3033 | btrfs_set_path_blocking(path); | ||
3034 | |||
2828 | leaf = path->nodes[0]; | 3035 | leaf = path->nodes[0]; |
2829 | BUG_ON(btrfs_leaf_free_space(root, leaf) < sizeof(struct btrfs_item)); | 3036 | BUG_ON(btrfs_leaf_free_space(root, leaf) < sizeof(struct btrfs_item)); |
2830 | 3037 | ||
@@ -3354,6 +3561,7 @@ int btrfs_insert_empty_items(struct btrfs_trans_handle *trans, | |||
3354 | BUG(); | 3561 | BUG(); |
3355 | } | 3562 | } |
3356 | out: | 3563 | out: |
3564 | btrfs_unlock_up_safe(path, 1); | ||
3357 | return ret; | 3565 | return ret; |
3358 | } | 3566 | } |
3359 | 3567 | ||
@@ -3441,15 +3649,22 @@ noinline int btrfs_del_leaf(struct btrfs_trans_handle *trans, | |||
3441 | { | 3649 | { |
3442 | int ret; | 3650 | int ret; |
3443 | u64 root_gen = btrfs_header_generation(path->nodes[1]); | 3651 | u64 root_gen = btrfs_header_generation(path->nodes[1]); |
3652 | u64 parent_start = path->nodes[1]->start; | ||
3653 | u64 parent_owner = btrfs_header_owner(path->nodes[1]); | ||
3444 | 3654 | ||
3445 | ret = del_ptr(trans, root, path, 1, path->slots[1]); | 3655 | ret = del_ptr(trans, root, path, 1, path->slots[1]); |
3446 | if (ret) | 3656 | if (ret) |
3447 | return ret; | 3657 | return ret; |
3448 | 3658 | ||
3659 | /* | ||
3660 | * btrfs_free_extent is expensive, we want to make sure we | ||
3661 | * aren't holding any locks when we call it | ||
3662 | */ | ||
3663 | btrfs_unlock_up_safe(path, 0); | ||
3664 | |||
3449 | ret = btrfs_free_extent(trans, root, bytenr, | 3665 | ret = btrfs_free_extent(trans, root, bytenr, |
3450 | btrfs_level_size(root, 0), | 3666 | btrfs_level_size(root, 0), |
3451 | path->nodes[1]->start, | 3667 | parent_start, parent_owner, |
3452 | btrfs_header_owner(path->nodes[1]), | ||
3453 | root_gen, 0, 1); | 3668 | root_gen, 0, 1); |
3454 | return ret; | 3669 | return ret; |
3455 | } | 3670 | } |
@@ -3721,6 +3936,7 @@ find_next_key: | |||
3721 | */ | 3936 | */ |
3722 | if (slot >= nritems) { | 3937 | if (slot >= nritems) { |
3723 | path->slots[level] = slot; | 3938 | path->slots[level] = slot; |
3939 | btrfs_set_path_blocking(path); | ||
3724 | sret = btrfs_find_next_key(root, path, min_key, level, | 3940 | sret = btrfs_find_next_key(root, path, min_key, level, |
3725 | cache_only, min_trans); | 3941 | cache_only, min_trans); |
3726 | if (sret == 0) { | 3942 | if (sret == 0) { |
@@ -3738,16 +3954,20 @@ find_next_key: | |||
3738 | unlock_up(path, level, 1); | 3954 | unlock_up(path, level, 1); |
3739 | goto out; | 3955 | goto out; |
3740 | } | 3956 | } |
3957 | btrfs_set_path_blocking(path); | ||
3741 | cur = read_node_slot(root, cur, slot); | 3958 | cur = read_node_slot(root, cur, slot); |
3742 | 3959 | ||
3743 | btrfs_tree_lock(cur); | 3960 | btrfs_tree_lock(cur); |
3961 | |||
3744 | path->locks[level - 1] = 1; | 3962 | path->locks[level - 1] = 1; |
3745 | path->nodes[level - 1] = cur; | 3963 | path->nodes[level - 1] = cur; |
3746 | unlock_up(path, level, 1); | 3964 | unlock_up(path, level, 1); |
3965 | btrfs_clear_path_blocking(path, NULL); | ||
3747 | } | 3966 | } |
3748 | out: | 3967 | out: |
3749 | if (ret == 0) | 3968 | if (ret == 0) |
3750 | memcpy(min_key, &found_key, sizeof(found_key)); | 3969 | memcpy(min_key, &found_key, sizeof(found_key)); |
3970 | btrfs_set_path_blocking(path); | ||
3751 | return ret; | 3971 | return ret; |
3752 | } | 3972 | } |
3753 | 3973 | ||
@@ -3843,6 +4063,7 @@ int btrfs_next_leaf(struct btrfs_root *root, struct btrfs_path *path) | |||
3843 | if (ret < 0) | 4063 | if (ret < 0) |
3844 | return ret; | 4064 | return ret; |
3845 | 4065 | ||
4066 | btrfs_set_path_blocking(path); | ||
3846 | nritems = btrfs_header_nritems(path->nodes[0]); | 4067 | nritems = btrfs_header_nritems(path->nodes[0]); |
3847 | /* | 4068 | /* |
3848 | * by releasing the path above we dropped all our locks. A balance | 4069 | * by releasing the path above we dropped all our locks. A balance |
@@ -3873,6 +4094,7 @@ int btrfs_next_leaf(struct btrfs_root *root, struct btrfs_path *path) | |||
3873 | free_extent_buffer(next); | 4094 | free_extent_buffer(next); |
3874 | } | 4095 | } |
3875 | 4096 | ||
4097 | /* the path was set to blocking above */ | ||
3876 | if (level == 1 && (path->locks[1] || path->skip_locking) && | 4098 | if (level == 1 && (path->locks[1] || path->skip_locking) && |
3877 | path->reada) | 4099 | path->reada) |
3878 | reada_for_search(root, path, level, slot, 0); | 4100 | reada_for_search(root, path, level, slot, 0); |
@@ -3881,6 +4103,7 @@ int btrfs_next_leaf(struct btrfs_root *root, struct btrfs_path *path) | |||
3881 | if (!path->skip_locking) { | 4103 | if (!path->skip_locking) { |
3882 | WARN_ON(!btrfs_tree_locked(c)); | 4104 | WARN_ON(!btrfs_tree_locked(c)); |
3883 | btrfs_tree_lock(next); | 4105 | btrfs_tree_lock(next); |
4106 | btrfs_set_lock_blocking(next); | ||
3884 | } | 4107 | } |
3885 | break; | 4108 | break; |
3886 | } | 4109 | } |
@@ -3897,12 +4120,15 @@ int btrfs_next_leaf(struct btrfs_root *root, struct btrfs_path *path) | |||
3897 | path->locks[level] = 1; | 4120 | path->locks[level] = 1; |
3898 | if (!level) | 4121 | if (!level) |
3899 | break; | 4122 | break; |
4123 | |||
4124 | btrfs_set_path_blocking(path); | ||
3900 | if (level == 1 && path->locks[1] && path->reada) | 4125 | if (level == 1 && path->locks[1] && path->reada) |
3901 | reada_for_search(root, path, level, slot, 0); | 4126 | reada_for_search(root, path, level, slot, 0); |
3902 | next = read_node_slot(root, next, 0); | 4127 | next = read_node_slot(root, next, 0); |
3903 | if (!path->skip_locking) { | 4128 | if (!path->skip_locking) { |
3904 | WARN_ON(!btrfs_tree_locked(path->nodes[level])); | 4129 | WARN_ON(!btrfs_tree_locked(path->nodes[level])); |
3905 | btrfs_tree_lock(next); | 4130 | btrfs_tree_lock(next); |
4131 | btrfs_set_lock_blocking(next); | ||
3906 | } | 4132 | } |
3907 | } | 4133 | } |
3908 | done: | 4134 | done: |
@@ -3927,6 +4153,7 @@ int btrfs_previous_item(struct btrfs_root *root, | |||
3927 | 4153 | ||
3928 | while (1) { | 4154 | while (1) { |
3929 | if (path->slots[0] == 0) { | 4155 | if (path->slots[0] == 0) { |
4156 | btrfs_set_path_blocking(path); | ||
3930 | ret = btrfs_prev_leaf(root, path); | 4157 | ret = btrfs_prev_leaf(root, path); |
3931 | if (ret != 0) | 4158 | if (ret != 0) |
3932 | return ret; | 4159 | return ret; |
diff --git a/fs/btrfs/ctree.h b/fs/btrfs/ctree.h index eee060f88113..82491ba8fa40 100644 --- a/fs/btrfs/ctree.h +++ b/fs/btrfs/ctree.h | |||
@@ -43,11 +43,7 @@ struct btrfs_ordered_sum; | |||
43 | 43 | ||
44 | #define BTRFS_ACL_NOT_CACHED ((void *)-1) | 44 | #define BTRFS_ACL_NOT_CACHED ((void *)-1) |
45 | 45 | ||
46 | #ifdef CONFIG_LOCKDEP | 46 | #define BTRFS_MAX_LEVEL 8 |
47 | # define BTRFS_MAX_LEVEL 7 | ||
48 | #else | ||
49 | # define BTRFS_MAX_LEVEL 8 | ||
50 | #endif | ||
51 | 47 | ||
52 | /* holds pointers to all of the tree roots */ | 48 | /* holds pointers to all of the tree roots */ |
53 | #define BTRFS_ROOT_TREE_OBJECTID 1ULL | 49 | #define BTRFS_ROOT_TREE_OBJECTID 1ULL |
@@ -454,17 +450,11 @@ struct btrfs_timespec { | |||
454 | __le32 nsec; | 450 | __le32 nsec; |
455 | } __attribute__ ((__packed__)); | 451 | } __attribute__ ((__packed__)); |
456 | 452 | ||
457 | typedef enum { | 453 | enum btrfs_compression_type { |
458 | BTRFS_COMPRESS_NONE = 0, | 454 | BTRFS_COMPRESS_NONE = 0, |
459 | BTRFS_COMPRESS_ZLIB = 1, | 455 | BTRFS_COMPRESS_ZLIB = 1, |
460 | BTRFS_COMPRESS_LAST = 2, | 456 | BTRFS_COMPRESS_LAST = 2, |
461 | } btrfs_compression_type; | 457 | }; |
462 | |||
463 | /* we don't understand any encryption methods right now */ | ||
464 | typedef enum { | ||
465 | BTRFS_ENCRYPTION_NONE = 0, | ||
466 | BTRFS_ENCRYPTION_LAST = 1, | ||
467 | } btrfs_encryption_type; | ||
468 | 458 | ||
469 | struct btrfs_inode_item { | 459 | struct btrfs_inode_item { |
470 | /* nfs style generation number */ | 460 | /* nfs style generation number */ |
@@ -606,13 +596,27 @@ struct btrfs_block_group_item { | |||
606 | 596 | ||
607 | struct btrfs_space_info { | 597 | struct btrfs_space_info { |
608 | u64 flags; | 598 | u64 flags; |
609 | u64 total_bytes; | 599 | |
610 | u64 bytes_used; | 600 | u64 total_bytes; /* total bytes in the space */ |
611 | u64 bytes_pinned; | 601 | u64 bytes_used; /* total bytes used on disk */ |
612 | u64 bytes_reserved; | 602 | u64 bytes_pinned; /* total bytes pinned, will be freed when the |
613 | u64 bytes_readonly; | 603 | transaction finishes */ |
614 | int full; | 604 | u64 bytes_reserved; /* total bytes the allocator has reserved for |
615 | int force_alloc; | 605 | current allocations */ |
606 | u64 bytes_readonly; /* total bytes that are read only */ | ||
607 | |||
608 | /* delalloc accounting */ | ||
609 | u64 bytes_delalloc; /* number of bytes reserved for allocation, | ||
610 | this space is not necessarily reserved yet | ||
611 | by the allocator */ | ||
612 | u64 bytes_may_use; /* number of bytes that may be used for | ||
613 | delalloc */ | ||
614 | |||
615 | int full; /* indicates that we cannot allocate any more | ||
616 | chunks for this space */ | ||
617 | int force_alloc; /* set if we need to force a chunk alloc for | ||
618 | this space */ | ||
619 | |||
616 | struct list_head list; | 620 | struct list_head list; |
617 | 621 | ||
618 | /* for block groups in our same type */ | 622 | /* for block groups in our same type */ |
@@ -701,9 +705,7 @@ struct btrfs_fs_info { | |||
701 | struct btrfs_transaction *running_transaction; | 705 | struct btrfs_transaction *running_transaction; |
702 | wait_queue_head_t transaction_throttle; | 706 | wait_queue_head_t transaction_throttle; |
703 | wait_queue_head_t transaction_wait; | 707 | wait_queue_head_t transaction_wait; |
704 | |||
705 | wait_queue_head_t async_submit_wait; | 708 | wait_queue_head_t async_submit_wait; |
706 | wait_queue_head_t tree_log_wait; | ||
707 | 709 | ||
708 | struct btrfs_super_block super_copy; | 710 | struct btrfs_super_block super_copy; |
709 | struct btrfs_super_block super_for_commit; | 711 | struct btrfs_super_block super_for_commit; |
@@ -711,7 +713,6 @@ struct btrfs_fs_info { | |||
711 | struct super_block *sb; | 713 | struct super_block *sb; |
712 | struct inode *btree_inode; | 714 | struct inode *btree_inode; |
713 | struct backing_dev_info bdi; | 715 | struct backing_dev_info bdi; |
714 | spinlock_t hash_lock; | ||
715 | struct mutex trans_mutex; | 716 | struct mutex trans_mutex; |
716 | struct mutex tree_log_mutex; | 717 | struct mutex tree_log_mutex; |
717 | struct mutex transaction_kthread_mutex; | 718 | struct mutex transaction_kthread_mutex; |
@@ -730,10 +731,6 @@ struct btrfs_fs_info { | |||
730 | atomic_t async_submit_draining; | 731 | atomic_t async_submit_draining; |
731 | atomic_t nr_async_bios; | 732 | atomic_t nr_async_bios; |
732 | atomic_t async_delalloc_pages; | 733 | atomic_t async_delalloc_pages; |
733 | atomic_t tree_log_writers; | ||
734 | atomic_t tree_log_commit; | ||
735 | unsigned long tree_log_batch; | ||
736 | u64 tree_log_transid; | ||
737 | 734 | ||
738 | /* | 735 | /* |
739 | * this is used by the balancing code to wait for all the pending | 736 | * this is used by the balancing code to wait for all the pending |
@@ -833,7 +830,14 @@ struct btrfs_root { | |||
833 | struct kobject root_kobj; | 830 | struct kobject root_kobj; |
834 | struct completion kobj_unregister; | 831 | struct completion kobj_unregister; |
835 | struct mutex objectid_mutex; | 832 | struct mutex objectid_mutex; |
833 | |||
836 | struct mutex log_mutex; | 834 | struct mutex log_mutex; |
835 | wait_queue_head_t log_writer_wait; | ||
836 | wait_queue_head_t log_commit_wait[2]; | ||
837 | atomic_t log_writers; | ||
838 | atomic_t log_commit[2]; | ||
839 | unsigned long log_transid; | ||
840 | unsigned long log_batch; | ||
837 | 841 | ||
838 | u64 objectid; | 842 | u64 objectid; |
839 | u64 last_trans; | 843 | u64 last_trans; |
@@ -1721,7 +1725,8 @@ struct extent_buffer *btrfs_alloc_free_block(struct btrfs_trans_handle *trans, | |||
1721 | u64 empty_size); | 1725 | u64 empty_size); |
1722 | struct extent_buffer *btrfs_init_new_buffer(struct btrfs_trans_handle *trans, | 1726 | struct extent_buffer *btrfs_init_new_buffer(struct btrfs_trans_handle *trans, |
1723 | struct btrfs_root *root, | 1727 | struct btrfs_root *root, |
1724 | u64 bytenr, u32 blocksize); | 1728 | u64 bytenr, u32 blocksize, |
1729 | int level); | ||
1725 | int btrfs_alloc_extent(struct btrfs_trans_handle *trans, | 1730 | int btrfs_alloc_extent(struct btrfs_trans_handle *trans, |
1726 | struct btrfs_root *root, | 1731 | struct btrfs_root *root, |
1727 | u64 num_bytes, u64 parent, u64 min_bytes, | 1732 | u64 num_bytes, u64 parent, u64 min_bytes, |
@@ -1791,6 +1796,16 @@ int btrfs_add_dead_reloc_root(struct btrfs_root *root); | |||
1791 | int btrfs_cleanup_reloc_trees(struct btrfs_root *root); | 1796 | int btrfs_cleanup_reloc_trees(struct btrfs_root *root); |
1792 | int btrfs_reloc_clone_csums(struct inode *inode, u64 file_pos, u64 len); | 1797 | int btrfs_reloc_clone_csums(struct inode *inode, u64 file_pos, u64 len); |
1793 | u64 btrfs_reduce_alloc_profile(struct btrfs_root *root, u64 flags); | 1798 | u64 btrfs_reduce_alloc_profile(struct btrfs_root *root, u64 flags); |
1799 | void btrfs_set_inode_space_info(struct btrfs_root *root, struct inode *ionde); | ||
1800 | int btrfs_check_metadata_free_space(struct btrfs_root *root); | ||
1801 | int btrfs_check_data_free_space(struct btrfs_root *root, struct inode *inode, | ||
1802 | u64 bytes); | ||
1803 | void btrfs_free_reserved_data_space(struct btrfs_root *root, | ||
1804 | struct inode *inode, u64 bytes); | ||
1805 | void btrfs_delalloc_reserve_space(struct btrfs_root *root, struct inode *inode, | ||
1806 | u64 bytes); | ||
1807 | void btrfs_delalloc_free_space(struct btrfs_root *root, struct inode *inode, | ||
1808 | u64 bytes); | ||
1794 | /* ctree.c */ | 1809 | /* ctree.c */ |
1795 | int btrfs_previous_item(struct btrfs_root *root, | 1810 | int btrfs_previous_item(struct btrfs_root *root, |
1796 | struct btrfs_path *path, u64 min_objectid, | 1811 | struct btrfs_path *path, u64 min_objectid, |
@@ -1840,7 +1855,9 @@ int btrfs_realloc_node(struct btrfs_trans_handle *trans, | |||
1840 | void btrfs_release_path(struct btrfs_root *root, struct btrfs_path *p); | 1855 | void btrfs_release_path(struct btrfs_root *root, struct btrfs_path *p); |
1841 | struct btrfs_path *btrfs_alloc_path(void); | 1856 | struct btrfs_path *btrfs_alloc_path(void); |
1842 | void btrfs_free_path(struct btrfs_path *p); | 1857 | void btrfs_free_path(struct btrfs_path *p); |
1843 | void btrfs_init_path(struct btrfs_path *p); | 1858 | void btrfs_set_path_blocking(struct btrfs_path *p); |
1859 | void btrfs_unlock_up_safe(struct btrfs_path *p, int level); | ||
1860 | |||
1844 | int btrfs_del_items(struct btrfs_trans_handle *trans, struct btrfs_root *root, | 1861 | int btrfs_del_items(struct btrfs_trans_handle *trans, struct btrfs_root *root, |
1845 | struct btrfs_path *path, int slot, int nr); | 1862 | struct btrfs_path *path, int slot, int nr); |
1846 | int btrfs_del_leaf(struct btrfs_trans_handle *trans, | 1863 | int btrfs_del_leaf(struct btrfs_trans_handle *trans, |
@@ -2034,8 +2051,6 @@ int btrfs_merge_bio_hook(struct page *page, unsigned long offset, | |||
2034 | unsigned long btrfs_force_ra(struct address_space *mapping, | 2051 | unsigned long btrfs_force_ra(struct address_space *mapping, |
2035 | struct file_ra_state *ra, struct file *file, | 2052 | struct file_ra_state *ra, struct file *file, |
2036 | pgoff_t offset, pgoff_t last_index); | 2053 | pgoff_t offset, pgoff_t last_index); |
2037 | int btrfs_check_free_space(struct btrfs_root *root, u64 num_required, | ||
2038 | int for_del); | ||
2039 | int btrfs_page_mkwrite(struct vm_area_struct *vma, struct page *page); | 2054 | int btrfs_page_mkwrite(struct vm_area_struct *vma, struct page *page); |
2040 | int btrfs_readpage(struct file *file, struct page *page); | 2055 | int btrfs_readpage(struct file *file, struct page *page); |
2041 | void btrfs_delete_inode(struct inode *inode); | 2056 | void btrfs_delete_inode(struct inode *inode); |
diff --git a/fs/btrfs/disk-io.c b/fs/btrfs/disk-io.c index 81a313874ae5..adda739a0215 100644 --- a/fs/btrfs/disk-io.c +++ b/fs/btrfs/disk-io.c | |||
@@ -16,7 +16,6 @@ | |||
16 | * Boston, MA 021110-1307, USA. | 16 | * Boston, MA 021110-1307, USA. |
17 | */ | 17 | */ |
18 | 18 | ||
19 | #include <linux/version.h> | ||
20 | #include <linux/fs.h> | 19 | #include <linux/fs.h> |
21 | #include <linux/blkdev.h> | 20 | #include <linux/blkdev.h> |
22 | #include <linux/scatterlist.h> | 21 | #include <linux/scatterlist.h> |
@@ -76,6 +75,40 @@ struct async_submit_bio { | |||
76 | struct btrfs_work work; | 75 | struct btrfs_work work; |
77 | }; | 76 | }; |
78 | 77 | ||
78 | /* These are used to set the lockdep class on the extent buffer locks. | ||
79 | * The class is set by the readpage_end_io_hook after the buffer has | ||
80 | * passed csum validation but before the pages are unlocked. | ||
81 | * | ||
82 | * The lockdep class is also set by btrfs_init_new_buffer on freshly | ||
83 | * allocated blocks. | ||
84 | * | ||
85 | * The class is based on the level in the tree block, which allows lockdep | ||
86 | * to know that lower nodes nest inside the locks of higher nodes. | ||
87 | * | ||
88 | * We also add a check to make sure the highest level of the tree is | ||
89 | * the same as our lockdep setup here. If BTRFS_MAX_LEVEL changes, this | ||
90 | * code needs update as well. | ||
91 | */ | ||
92 | #ifdef CONFIG_DEBUG_LOCK_ALLOC | ||
93 | # if BTRFS_MAX_LEVEL != 8 | ||
94 | # error | ||
95 | # endif | ||
96 | static struct lock_class_key btrfs_eb_class[BTRFS_MAX_LEVEL + 1]; | ||
97 | static const char *btrfs_eb_name[BTRFS_MAX_LEVEL + 1] = { | ||
98 | /* leaf */ | ||
99 | "btrfs-extent-00", | ||
100 | "btrfs-extent-01", | ||
101 | "btrfs-extent-02", | ||
102 | "btrfs-extent-03", | ||
103 | "btrfs-extent-04", | ||
104 | "btrfs-extent-05", | ||
105 | "btrfs-extent-06", | ||
106 | "btrfs-extent-07", | ||
107 | /* highest possible level */ | ||
108 | "btrfs-extent-08", | ||
109 | }; | ||
110 | #endif | ||
111 | |||
79 | /* | 112 | /* |
80 | * extents on the btree inode are pretty simple, there's one extent | 113 | * extents on the btree inode are pretty simple, there's one extent |
81 | * that covers the entire device | 114 | * that covers the entire device |
@@ -348,6 +381,15 @@ static int check_tree_block_fsid(struct btrfs_root *root, | |||
348 | return ret; | 381 | return ret; |
349 | } | 382 | } |
350 | 383 | ||
384 | #ifdef CONFIG_DEBUG_LOCK_ALLOC | ||
385 | void btrfs_set_buffer_lockdep_class(struct extent_buffer *eb, int level) | ||
386 | { | ||
387 | lockdep_set_class_and_name(&eb->lock, | ||
388 | &btrfs_eb_class[level], | ||
389 | btrfs_eb_name[level]); | ||
390 | } | ||
391 | #endif | ||
392 | |||
351 | static int btree_readpage_end_io_hook(struct page *page, u64 start, u64 end, | 393 | static int btree_readpage_end_io_hook(struct page *page, u64 start, u64 end, |
352 | struct extent_state *state) | 394 | struct extent_state *state) |
353 | { | 395 | { |
@@ -393,6 +435,8 @@ static int btree_readpage_end_io_hook(struct page *page, u64 start, u64 end, | |||
393 | } | 435 | } |
394 | found_level = btrfs_header_level(eb); | 436 | found_level = btrfs_header_level(eb); |
395 | 437 | ||
438 | btrfs_set_buffer_lockdep_class(eb, found_level); | ||
439 | |||
396 | ret = csum_tree_block(root, eb, 1); | 440 | ret = csum_tree_block(root, eb, 1); |
397 | if (ret) | 441 | if (ret) |
398 | ret = -EIO; | 442 | ret = -EIO; |
@@ -800,7 +844,7 @@ struct extent_buffer *read_tree_block(struct btrfs_root *root, u64 bytenr, | |||
800 | ret = btree_read_extent_buffer_pages(root, buf, 0, parent_transid); | 844 | ret = btree_read_extent_buffer_pages(root, buf, 0, parent_transid); |
801 | 845 | ||
802 | if (ret == 0) | 846 | if (ret == 0) |
803 | buf->flags |= EXTENT_UPTODATE; | 847 | set_bit(EXTENT_BUFFER_UPTODATE, &buf->bflags); |
804 | else | 848 | else |
805 | WARN_ON(1); | 849 | WARN_ON(1); |
806 | return buf; | 850 | return buf; |
@@ -814,6 +858,10 @@ int clean_tree_block(struct btrfs_trans_handle *trans, struct btrfs_root *root, | |||
814 | if (btrfs_header_generation(buf) == | 858 | if (btrfs_header_generation(buf) == |
815 | root->fs_info->running_transaction->transid) { | 859 | root->fs_info->running_transaction->transid) { |
816 | WARN_ON(!btrfs_tree_locked(buf)); | 860 | WARN_ON(!btrfs_tree_locked(buf)); |
861 | |||
862 | /* ugh, clear_extent_buffer_dirty can be expensive */ | ||
863 | btrfs_set_lock_blocking(buf); | ||
864 | |||
817 | clear_extent_buffer_dirty(&BTRFS_I(btree_inode)->io_tree, | 865 | clear_extent_buffer_dirty(&BTRFS_I(btree_inode)->io_tree, |
818 | buf); | 866 | buf); |
819 | } | 867 | } |
@@ -850,6 +898,14 @@ static int __setup_root(u32 nodesize, u32 leafsize, u32 sectorsize, | |||
850 | spin_lock_init(&root->list_lock); | 898 | spin_lock_init(&root->list_lock); |
851 | mutex_init(&root->objectid_mutex); | 899 | mutex_init(&root->objectid_mutex); |
852 | mutex_init(&root->log_mutex); | 900 | mutex_init(&root->log_mutex); |
901 | init_waitqueue_head(&root->log_writer_wait); | ||
902 | init_waitqueue_head(&root->log_commit_wait[0]); | ||
903 | init_waitqueue_head(&root->log_commit_wait[1]); | ||
904 | atomic_set(&root->log_commit[0], 0); | ||
905 | atomic_set(&root->log_commit[1], 0); | ||
906 | atomic_set(&root->log_writers, 0); | ||
907 | root->log_batch = 0; | ||
908 | root->log_transid = 0; | ||
853 | extent_io_tree_init(&root->dirty_log_pages, | 909 | extent_io_tree_init(&root->dirty_log_pages, |
854 | fs_info->btree_inode->i_mapping, GFP_NOFS); | 910 | fs_info->btree_inode->i_mapping, GFP_NOFS); |
855 | 911 | ||
@@ -934,15 +990,16 @@ int btrfs_free_log_root_tree(struct btrfs_trans_handle *trans, | |||
934 | return 0; | 990 | return 0; |
935 | } | 991 | } |
936 | 992 | ||
937 | int btrfs_init_log_root_tree(struct btrfs_trans_handle *trans, | 993 | static struct btrfs_root *alloc_log_tree(struct btrfs_trans_handle *trans, |
938 | struct btrfs_fs_info *fs_info) | 994 | struct btrfs_fs_info *fs_info) |
939 | { | 995 | { |
940 | struct btrfs_root *root; | 996 | struct btrfs_root *root; |
941 | struct btrfs_root *tree_root = fs_info->tree_root; | 997 | struct btrfs_root *tree_root = fs_info->tree_root; |
998 | struct extent_buffer *leaf; | ||
942 | 999 | ||
943 | root = kzalloc(sizeof(*root), GFP_NOFS); | 1000 | root = kzalloc(sizeof(*root), GFP_NOFS); |
944 | if (!root) | 1001 | if (!root) |
945 | return -ENOMEM; | 1002 | return ERR_PTR(-ENOMEM); |
946 | 1003 | ||
947 | __setup_root(tree_root->nodesize, tree_root->leafsize, | 1004 | __setup_root(tree_root->nodesize, tree_root->leafsize, |
948 | tree_root->sectorsize, tree_root->stripesize, | 1005 | tree_root->sectorsize, tree_root->stripesize, |
@@ -951,12 +1008,23 @@ int btrfs_init_log_root_tree(struct btrfs_trans_handle *trans, | |||
951 | root->root_key.objectid = BTRFS_TREE_LOG_OBJECTID; | 1008 | root->root_key.objectid = BTRFS_TREE_LOG_OBJECTID; |
952 | root->root_key.type = BTRFS_ROOT_ITEM_KEY; | 1009 | root->root_key.type = BTRFS_ROOT_ITEM_KEY; |
953 | root->root_key.offset = BTRFS_TREE_LOG_OBJECTID; | 1010 | root->root_key.offset = BTRFS_TREE_LOG_OBJECTID; |
1011 | /* | ||
1012 | * log trees do not get reference counted because they go away | ||
1013 | * before a real commit is actually done. They do store pointers | ||
1014 | * to file data extents, and those reference counts still get | ||
1015 | * updated (along with back refs to the log tree). | ||
1016 | */ | ||
954 | root->ref_cows = 0; | 1017 | root->ref_cows = 0; |
955 | 1018 | ||
956 | root->node = btrfs_alloc_free_block(trans, root, root->leafsize, | 1019 | leaf = btrfs_alloc_free_block(trans, root, root->leafsize, |
957 | 0, BTRFS_TREE_LOG_OBJECTID, | 1020 | 0, BTRFS_TREE_LOG_OBJECTID, |
958 | trans->transid, 0, 0, 0); | 1021 | trans->transid, 0, 0, 0); |
1022 | if (IS_ERR(leaf)) { | ||
1023 | kfree(root); | ||
1024 | return ERR_CAST(leaf); | ||
1025 | } | ||
959 | 1026 | ||
1027 | root->node = leaf; | ||
960 | btrfs_set_header_nritems(root->node, 0); | 1028 | btrfs_set_header_nritems(root->node, 0); |
961 | btrfs_set_header_level(root->node, 0); | 1029 | btrfs_set_header_level(root->node, 0); |
962 | btrfs_set_header_bytenr(root->node, root->node->start); | 1030 | btrfs_set_header_bytenr(root->node, root->node->start); |
@@ -968,7 +1036,48 @@ int btrfs_init_log_root_tree(struct btrfs_trans_handle *trans, | |||
968 | BTRFS_FSID_SIZE); | 1036 | BTRFS_FSID_SIZE); |
969 | btrfs_mark_buffer_dirty(root->node); | 1037 | btrfs_mark_buffer_dirty(root->node); |
970 | btrfs_tree_unlock(root->node); | 1038 | btrfs_tree_unlock(root->node); |
971 | fs_info->log_root_tree = root; | 1039 | return root; |
1040 | } | ||
1041 | |||
1042 | int btrfs_init_log_root_tree(struct btrfs_trans_handle *trans, | ||
1043 | struct btrfs_fs_info *fs_info) | ||
1044 | { | ||
1045 | struct btrfs_root *log_root; | ||
1046 | |||
1047 | log_root = alloc_log_tree(trans, fs_info); | ||
1048 | if (IS_ERR(log_root)) | ||
1049 | return PTR_ERR(log_root); | ||
1050 | WARN_ON(fs_info->log_root_tree); | ||
1051 | fs_info->log_root_tree = log_root; | ||
1052 | return 0; | ||
1053 | } | ||
1054 | |||
1055 | int btrfs_add_log_tree(struct btrfs_trans_handle *trans, | ||
1056 | struct btrfs_root *root) | ||
1057 | { | ||
1058 | struct btrfs_root *log_root; | ||
1059 | struct btrfs_inode_item *inode_item; | ||
1060 | |||
1061 | log_root = alloc_log_tree(trans, root->fs_info); | ||
1062 | if (IS_ERR(log_root)) | ||
1063 | return PTR_ERR(log_root); | ||
1064 | |||
1065 | log_root->last_trans = trans->transid; | ||
1066 | log_root->root_key.offset = root->root_key.objectid; | ||
1067 | |||
1068 | inode_item = &log_root->root_item.inode; | ||
1069 | inode_item->generation = cpu_to_le64(1); | ||
1070 | inode_item->size = cpu_to_le64(3); | ||
1071 | inode_item->nlink = cpu_to_le32(1); | ||
1072 | inode_item->nbytes = cpu_to_le64(root->leafsize); | ||
1073 | inode_item->mode = cpu_to_le32(S_IFDIR | 0755); | ||
1074 | |||
1075 | btrfs_set_root_bytenr(&log_root->root_item, log_root->node->start); | ||
1076 | btrfs_set_root_generation(&log_root->root_item, trans->transid); | ||
1077 | |||
1078 | WARN_ON(root->log_root); | ||
1079 | root->log_root = log_root; | ||
1080 | root->log_transid = 0; | ||
972 | return 0; | 1081 | return 0; |
973 | } | 1082 | } |
974 | 1083 | ||
@@ -1136,7 +1245,6 @@ static int btrfs_congested_fn(void *congested_data, int bdi_bits) | |||
1136 | { | 1245 | { |
1137 | struct btrfs_fs_info *info = (struct btrfs_fs_info *)congested_data; | 1246 | struct btrfs_fs_info *info = (struct btrfs_fs_info *)congested_data; |
1138 | int ret = 0; | 1247 | int ret = 0; |
1139 | struct list_head *cur; | ||
1140 | struct btrfs_device *device; | 1248 | struct btrfs_device *device; |
1141 | struct backing_dev_info *bdi; | 1249 | struct backing_dev_info *bdi; |
1142 | #if 0 | 1250 | #if 0 |
@@ -1144,8 +1252,7 @@ static int btrfs_congested_fn(void *congested_data, int bdi_bits) | |||
1144 | btrfs_congested_async(info, 0)) | 1252 | btrfs_congested_async(info, 0)) |
1145 | return 1; | 1253 | return 1; |
1146 | #endif | 1254 | #endif |
1147 | list_for_each(cur, &info->fs_devices->devices) { | 1255 | list_for_each_entry(device, &info->fs_devices->devices, dev_list) { |
1148 | device = list_entry(cur, struct btrfs_device, dev_list); | ||
1149 | if (!device->bdev) | 1256 | if (!device->bdev) |
1150 | continue; | 1257 | continue; |
1151 | bdi = blk_get_backing_dev_info(device->bdev); | 1258 | bdi = blk_get_backing_dev_info(device->bdev); |
@@ -1163,13 +1270,11 @@ static int btrfs_congested_fn(void *congested_data, int bdi_bits) | |||
1163 | */ | 1270 | */ |
1164 | static void __unplug_io_fn(struct backing_dev_info *bdi, struct page *page) | 1271 | static void __unplug_io_fn(struct backing_dev_info *bdi, struct page *page) |
1165 | { | 1272 | { |
1166 | struct list_head *cur; | ||
1167 | struct btrfs_device *device; | 1273 | struct btrfs_device *device; |
1168 | struct btrfs_fs_info *info; | 1274 | struct btrfs_fs_info *info; |
1169 | 1275 | ||
1170 | info = (struct btrfs_fs_info *)bdi->unplug_io_data; | 1276 | info = (struct btrfs_fs_info *)bdi->unplug_io_data; |
1171 | list_for_each(cur, &info->fs_devices->devices) { | 1277 | list_for_each_entry(device, &info->fs_devices->devices, dev_list) { |
1172 | device = list_entry(cur, struct btrfs_device, dev_list); | ||
1173 | if (!device->bdev) | 1278 | if (!device->bdev) |
1174 | continue; | 1279 | continue; |
1175 | 1280 | ||
@@ -1447,7 +1552,6 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1447 | INIT_LIST_HEAD(&fs_info->dead_roots); | 1552 | INIT_LIST_HEAD(&fs_info->dead_roots); |
1448 | INIT_LIST_HEAD(&fs_info->hashers); | 1553 | INIT_LIST_HEAD(&fs_info->hashers); |
1449 | INIT_LIST_HEAD(&fs_info->delalloc_inodes); | 1554 | INIT_LIST_HEAD(&fs_info->delalloc_inodes); |
1450 | spin_lock_init(&fs_info->hash_lock); | ||
1451 | spin_lock_init(&fs_info->delalloc_lock); | 1555 | spin_lock_init(&fs_info->delalloc_lock); |
1452 | spin_lock_init(&fs_info->new_trans_lock); | 1556 | spin_lock_init(&fs_info->new_trans_lock); |
1453 | spin_lock_init(&fs_info->ref_cache_lock); | 1557 | spin_lock_init(&fs_info->ref_cache_lock); |
@@ -1535,10 +1639,6 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1535 | init_waitqueue_head(&fs_info->transaction_throttle); | 1639 | init_waitqueue_head(&fs_info->transaction_throttle); |
1536 | init_waitqueue_head(&fs_info->transaction_wait); | 1640 | init_waitqueue_head(&fs_info->transaction_wait); |
1537 | init_waitqueue_head(&fs_info->async_submit_wait); | 1641 | init_waitqueue_head(&fs_info->async_submit_wait); |
1538 | init_waitqueue_head(&fs_info->tree_log_wait); | ||
1539 | atomic_set(&fs_info->tree_log_commit, 0); | ||
1540 | atomic_set(&fs_info->tree_log_writers, 0); | ||
1541 | fs_info->tree_log_transid = 0; | ||
1542 | 1642 | ||
1543 | __setup_root(4096, 4096, 4096, 4096, tree_root, | 1643 | __setup_root(4096, 4096, 4096, 4096, tree_root, |
1544 | fs_info, BTRFS_ROOT_TREE_OBJECTID); | 1644 | fs_info, BTRFS_ROOT_TREE_OBJECTID); |
@@ -1627,6 +1727,8 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1627 | * low idle thresh | 1727 | * low idle thresh |
1628 | */ | 1728 | */ |
1629 | fs_info->endio_workers.idle_thresh = 4; | 1729 | fs_info->endio_workers.idle_thresh = 4; |
1730 | fs_info->endio_meta_workers.idle_thresh = 4; | ||
1731 | |||
1630 | fs_info->endio_write_workers.idle_thresh = 64; | 1732 | fs_info->endio_write_workers.idle_thresh = 64; |
1631 | fs_info->endio_meta_write_workers.idle_thresh = 64; | 1733 | fs_info->endio_meta_write_workers.idle_thresh = 64; |
1632 | 1734 | ||
@@ -1720,7 +1822,6 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1720 | ret = find_and_setup_root(tree_root, fs_info, | 1822 | ret = find_and_setup_root(tree_root, fs_info, |
1721 | BTRFS_DEV_TREE_OBJECTID, dev_root); | 1823 | BTRFS_DEV_TREE_OBJECTID, dev_root); |
1722 | dev_root->track_dirty = 1; | 1824 | dev_root->track_dirty = 1; |
1723 | |||
1724 | if (ret) | 1825 | if (ret) |
1725 | goto fail_extent_root; | 1826 | goto fail_extent_root; |
1726 | 1827 | ||
@@ -1740,13 +1841,13 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1740 | fs_info->system_alloc_profile = fs_info->metadata_alloc_profile; | 1841 | fs_info->system_alloc_profile = fs_info->metadata_alloc_profile; |
1741 | fs_info->cleaner_kthread = kthread_run(cleaner_kthread, tree_root, | 1842 | fs_info->cleaner_kthread = kthread_run(cleaner_kthread, tree_root, |
1742 | "btrfs-cleaner"); | 1843 | "btrfs-cleaner"); |
1743 | if (!fs_info->cleaner_kthread) | 1844 | if (IS_ERR(fs_info->cleaner_kthread)) |
1744 | goto fail_csum_root; | 1845 | goto fail_csum_root; |
1745 | 1846 | ||
1746 | fs_info->transaction_kthread = kthread_run(transaction_kthread, | 1847 | fs_info->transaction_kthread = kthread_run(transaction_kthread, |
1747 | tree_root, | 1848 | tree_root, |
1748 | "btrfs-transaction"); | 1849 | "btrfs-transaction"); |
1749 | if (!fs_info->transaction_kthread) | 1850 | if (IS_ERR(fs_info->transaction_kthread)) |
1750 | goto fail_cleaner; | 1851 | goto fail_cleaner; |
1751 | 1852 | ||
1752 | if (btrfs_super_log_root(disk_super) != 0) { | 1853 | if (btrfs_super_log_root(disk_super) != 0) { |
@@ -1828,13 +1929,14 @@ fail_sb_buffer: | |||
1828 | fail_iput: | 1929 | fail_iput: |
1829 | invalidate_inode_pages2(fs_info->btree_inode->i_mapping); | 1930 | invalidate_inode_pages2(fs_info->btree_inode->i_mapping); |
1830 | iput(fs_info->btree_inode); | 1931 | iput(fs_info->btree_inode); |
1831 | fail: | 1932 | |
1832 | btrfs_close_devices(fs_info->fs_devices); | 1933 | btrfs_close_devices(fs_info->fs_devices); |
1833 | btrfs_mapping_tree_free(&fs_info->mapping_tree); | 1934 | btrfs_mapping_tree_free(&fs_info->mapping_tree); |
1935 | bdi_destroy(&fs_info->bdi); | ||
1834 | 1936 | ||
1937 | fail: | ||
1835 | kfree(extent_root); | 1938 | kfree(extent_root); |
1836 | kfree(tree_root); | 1939 | kfree(tree_root); |
1837 | bdi_destroy(&fs_info->bdi); | ||
1838 | kfree(fs_info); | 1940 | kfree(fs_info); |
1839 | kfree(chunk_root); | 1941 | kfree(chunk_root); |
1840 | kfree(dev_root); | 1942 | kfree(dev_root); |
@@ -1995,7 +2097,6 @@ static int write_dev_supers(struct btrfs_device *device, | |||
1995 | 2097 | ||
1996 | int write_all_supers(struct btrfs_root *root, int max_mirrors) | 2098 | int write_all_supers(struct btrfs_root *root, int max_mirrors) |
1997 | { | 2099 | { |
1998 | struct list_head *cur; | ||
1999 | struct list_head *head = &root->fs_info->fs_devices->devices; | 2100 | struct list_head *head = &root->fs_info->fs_devices->devices; |
2000 | struct btrfs_device *dev; | 2101 | struct btrfs_device *dev; |
2001 | struct btrfs_super_block *sb; | 2102 | struct btrfs_super_block *sb; |
@@ -2011,8 +2112,7 @@ int write_all_supers(struct btrfs_root *root, int max_mirrors) | |||
2011 | 2112 | ||
2012 | sb = &root->fs_info->super_for_commit; | 2113 | sb = &root->fs_info->super_for_commit; |
2013 | dev_item = &sb->dev_item; | 2114 | dev_item = &sb->dev_item; |
2014 | list_for_each(cur, head) { | 2115 | list_for_each_entry(dev, head, dev_list) { |
2015 | dev = list_entry(cur, struct btrfs_device, dev_list); | ||
2016 | if (!dev->bdev) { | 2116 | if (!dev->bdev) { |
2017 | total_errors++; | 2117 | total_errors++; |
2018 | continue; | 2118 | continue; |
@@ -2045,8 +2145,7 @@ int write_all_supers(struct btrfs_root *root, int max_mirrors) | |||
2045 | } | 2145 | } |
2046 | 2146 | ||
2047 | total_errors = 0; | 2147 | total_errors = 0; |
2048 | list_for_each(cur, head) { | 2148 | list_for_each_entry(dev, head, dev_list) { |
2049 | dev = list_entry(cur, struct btrfs_device, dev_list); | ||
2050 | if (!dev->bdev) | 2149 | if (!dev->bdev) |
2051 | continue; | 2150 | continue; |
2052 | if (!dev->in_fs_metadata || !dev->writeable) | 2151 | if (!dev->in_fs_metadata || !dev->writeable) |
@@ -2260,6 +2359,8 @@ void btrfs_mark_buffer_dirty(struct extent_buffer *buf) | |||
2260 | u64 transid = btrfs_header_generation(buf); | 2359 | u64 transid = btrfs_header_generation(buf); |
2261 | struct inode *btree_inode = root->fs_info->btree_inode; | 2360 | struct inode *btree_inode = root->fs_info->btree_inode; |
2262 | 2361 | ||
2362 | btrfs_set_lock_blocking(buf); | ||
2363 | |||
2263 | WARN_ON(!btrfs_tree_locked(buf)); | 2364 | WARN_ON(!btrfs_tree_locked(buf)); |
2264 | if (transid != root->fs_info->generation) { | 2365 | if (transid != root->fs_info->generation) { |
2265 | printk(KERN_CRIT "btrfs transid mismatch buffer %llu, " | 2366 | printk(KERN_CRIT "btrfs transid mismatch buffer %llu, " |
@@ -2302,14 +2403,13 @@ int btrfs_read_buffer(struct extent_buffer *buf, u64 parent_transid) | |||
2302 | int ret; | 2403 | int ret; |
2303 | ret = btree_read_extent_buffer_pages(root, buf, 0, parent_transid); | 2404 | ret = btree_read_extent_buffer_pages(root, buf, 0, parent_transid); |
2304 | if (ret == 0) | 2405 | if (ret == 0) |
2305 | buf->flags |= EXTENT_UPTODATE; | 2406 | set_bit(EXTENT_BUFFER_UPTODATE, &buf->bflags); |
2306 | return ret; | 2407 | return ret; |
2307 | } | 2408 | } |
2308 | 2409 | ||
2309 | int btree_lock_page_hook(struct page *page) | 2410 | int btree_lock_page_hook(struct page *page) |
2310 | { | 2411 | { |
2311 | struct inode *inode = page->mapping->host; | 2412 | struct inode *inode = page->mapping->host; |
2312 | struct btrfs_root *root = BTRFS_I(inode)->root; | ||
2313 | struct extent_io_tree *io_tree = &BTRFS_I(inode)->io_tree; | 2413 | struct extent_io_tree *io_tree = &BTRFS_I(inode)->io_tree; |
2314 | struct extent_buffer *eb; | 2414 | struct extent_buffer *eb; |
2315 | unsigned long len; | 2415 | unsigned long len; |
@@ -2324,9 +2424,7 @@ int btree_lock_page_hook(struct page *page) | |||
2324 | goto out; | 2424 | goto out; |
2325 | 2425 | ||
2326 | btrfs_tree_lock(eb); | 2426 | btrfs_tree_lock(eb); |
2327 | spin_lock(&root->fs_info->hash_lock); | ||
2328 | btrfs_set_header_flag(eb, BTRFS_HEADER_FLAG_WRITTEN); | 2427 | btrfs_set_header_flag(eb, BTRFS_HEADER_FLAG_WRITTEN); |
2329 | spin_unlock(&root->fs_info->hash_lock); | ||
2330 | btrfs_tree_unlock(eb); | 2428 | btrfs_tree_unlock(eb); |
2331 | free_extent_buffer(eb); | 2429 | free_extent_buffer(eb); |
2332 | out: | 2430 | out: |
diff --git a/fs/btrfs/disk-io.h b/fs/btrfs/disk-io.h index c0ff404c31b7..95029db227be 100644 --- a/fs/btrfs/disk-io.h +++ b/fs/btrfs/disk-io.h | |||
@@ -98,5 +98,17 @@ int btrfs_free_log_root_tree(struct btrfs_trans_handle *trans, | |||
98 | struct btrfs_fs_info *fs_info); | 98 | struct btrfs_fs_info *fs_info); |
99 | int btrfs_init_log_root_tree(struct btrfs_trans_handle *trans, | 99 | int btrfs_init_log_root_tree(struct btrfs_trans_handle *trans, |
100 | struct btrfs_fs_info *fs_info); | 100 | struct btrfs_fs_info *fs_info); |
101 | int btrfs_add_log_tree(struct btrfs_trans_handle *trans, | ||
102 | struct btrfs_root *root); | ||
101 | int btree_lock_page_hook(struct page *page); | 103 | int btree_lock_page_hook(struct page *page); |
104 | |||
105 | |||
106 | #ifdef CONFIG_DEBUG_LOCK_ALLOC | ||
107 | void btrfs_set_buffer_lockdep_class(struct extent_buffer *eb, int level); | ||
108 | #else | ||
109 | static inline void btrfs_set_buffer_lockdep_class(struct extent_buffer *eb, | ||
110 | int level) | ||
111 | { | ||
112 | } | ||
113 | #endif | ||
102 | #endif | 114 | #endif |
diff --git a/fs/btrfs/extent-tree.c b/fs/btrfs/extent-tree.c index 293da650873f..6b5966aacf44 100644 --- a/fs/btrfs/extent-tree.c +++ b/fs/btrfs/extent-tree.c | |||
@@ -19,7 +19,7 @@ | |||
19 | #include <linux/pagemap.h> | 19 | #include <linux/pagemap.h> |
20 | #include <linux/writeback.h> | 20 | #include <linux/writeback.h> |
21 | #include <linux/blkdev.h> | 21 | #include <linux/blkdev.h> |
22 | #include <linux/version.h> | 22 | #include <linux/sort.h> |
23 | #include "compat.h" | 23 | #include "compat.h" |
24 | #include "hash.h" | 24 | #include "hash.h" |
25 | #include "crc32c.h" | 25 | #include "crc32c.h" |
@@ -30,7 +30,6 @@ | |||
30 | #include "volumes.h" | 30 | #include "volumes.h" |
31 | #include "locking.h" | 31 | #include "locking.h" |
32 | #include "ref-cache.h" | 32 | #include "ref-cache.h" |
33 | #include "compat.h" | ||
34 | 33 | ||
35 | #define PENDING_EXTENT_INSERT 0 | 34 | #define PENDING_EXTENT_INSERT 0 |
36 | #define PENDING_EXTENT_DELETE 1 | 35 | #define PENDING_EXTENT_DELETE 1 |
@@ -61,6 +60,10 @@ static int update_block_group(struct btrfs_trans_handle *trans, | |||
61 | u64 bytenr, u64 num_bytes, int alloc, | 60 | u64 bytenr, u64 num_bytes, int alloc, |
62 | int mark_free); | 61 | int mark_free); |
63 | 62 | ||
63 | static int do_chunk_alloc(struct btrfs_trans_handle *trans, | ||
64 | struct btrfs_root *extent_root, u64 alloc_bytes, | ||
65 | u64 flags, int force); | ||
66 | |||
64 | static int block_group_bits(struct btrfs_block_group_cache *cache, u64 bits) | 67 | static int block_group_bits(struct btrfs_block_group_cache *cache, u64 bits) |
65 | { | 68 | { |
66 | return (cache->flags & bits) == bits; | 69 | return (cache->flags & bits) == bits; |
@@ -326,10 +329,8 @@ static struct btrfs_space_info *__find_space_info(struct btrfs_fs_info *info, | |||
326 | u64 flags) | 329 | u64 flags) |
327 | { | 330 | { |
328 | struct list_head *head = &info->space_info; | 331 | struct list_head *head = &info->space_info; |
329 | struct list_head *cur; | ||
330 | struct btrfs_space_info *found; | 332 | struct btrfs_space_info *found; |
331 | list_for_each(cur, head) { | 333 | list_for_each_entry(found, head, list) { |
332 | found = list_entry(cur, struct btrfs_space_info, list); | ||
333 | if (found->flags == flags) | 334 | if (found->flags == flags) |
334 | return found; | 335 | return found; |
335 | } | 336 | } |
@@ -1326,8 +1327,25 @@ int btrfs_inc_extent_ref(struct btrfs_trans_handle *trans, | |||
1326 | int btrfs_extent_post_op(struct btrfs_trans_handle *trans, | 1327 | int btrfs_extent_post_op(struct btrfs_trans_handle *trans, |
1327 | struct btrfs_root *root) | 1328 | struct btrfs_root *root) |
1328 | { | 1329 | { |
1329 | finish_current_insert(trans, root->fs_info->extent_root, 1); | 1330 | u64 start; |
1330 | del_pending_extents(trans, root->fs_info->extent_root, 1); | 1331 | u64 end; |
1332 | int ret; | ||
1333 | |||
1334 | while(1) { | ||
1335 | finish_current_insert(trans, root->fs_info->extent_root, 1); | ||
1336 | del_pending_extents(trans, root->fs_info->extent_root, 1); | ||
1337 | |||
1338 | /* is there more work to do? */ | ||
1339 | ret = find_first_extent_bit(&root->fs_info->pending_del, | ||
1340 | 0, &start, &end, EXTENT_WRITEBACK); | ||
1341 | if (!ret) | ||
1342 | continue; | ||
1343 | ret = find_first_extent_bit(&root->fs_info->extent_ins, | ||
1344 | 0, &start, &end, EXTENT_WRITEBACK); | ||
1345 | if (!ret) | ||
1346 | continue; | ||
1347 | break; | ||
1348 | } | ||
1331 | return 0; | 1349 | return 0; |
1332 | } | 1350 | } |
1333 | 1351 | ||
@@ -1525,15 +1543,55 @@ out: | |||
1525 | return ret; | 1543 | return ret; |
1526 | } | 1544 | } |
1527 | 1545 | ||
1528 | int btrfs_inc_ref(struct btrfs_trans_handle *trans, struct btrfs_root *root, | 1546 | /* when a block goes through cow, we update the reference counts of |
1529 | struct extent_buffer *orig_buf, struct extent_buffer *buf, | 1547 | * everything that block points to. The internal pointers of the block |
1530 | u32 *nr_extents) | 1548 | * can be in just about any order, and it is likely to have clusters of |
1549 | * things that are close together and clusters of things that are not. | ||
1550 | * | ||
1551 | * To help reduce the seeks that come with updating all of these reference | ||
1552 | * counts, sort them by byte number before actual updates are done. | ||
1553 | * | ||
1554 | * struct refsort is used to match byte number to slot in the btree block. | ||
1555 | * we sort based on the byte number and then use the slot to actually | ||
1556 | * find the item. | ||
1557 | * | ||
1558 | * struct refsort is smaller than strcut btrfs_item and smaller than | ||
1559 | * struct btrfs_key_ptr. Since we're currently limited to the page size | ||
1560 | * for a btree block, there's no way for a kmalloc of refsorts for a | ||
1561 | * single node to be bigger than a page. | ||
1562 | */ | ||
1563 | struct refsort { | ||
1564 | u64 bytenr; | ||
1565 | u32 slot; | ||
1566 | }; | ||
1567 | |||
1568 | /* | ||
1569 | * for passing into sort() | ||
1570 | */ | ||
1571 | static int refsort_cmp(const void *a_void, const void *b_void) | ||
1572 | { | ||
1573 | const struct refsort *a = a_void; | ||
1574 | const struct refsort *b = b_void; | ||
1575 | |||
1576 | if (a->bytenr < b->bytenr) | ||
1577 | return -1; | ||
1578 | if (a->bytenr > b->bytenr) | ||
1579 | return 1; | ||
1580 | return 0; | ||
1581 | } | ||
1582 | |||
1583 | |||
1584 | noinline int btrfs_inc_ref(struct btrfs_trans_handle *trans, | ||
1585 | struct btrfs_root *root, | ||
1586 | struct extent_buffer *orig_buf, | ||
1587 | struct extent_buffer *buf, u32 *nr_extents) | ||
1531 | { | 1588 | { |
1532 | u64 bytenr; | 1589 | u64 bytenr; |
1533 | u64 ref_root; | 1590 | u64 ref_root; |
1534 | u64 orig_root; | 1591 | u64 orig_root; |
1535 | u64 ref_generation; | 1592 | u64 ref_generation; |
1536 | u64 orig_generation; | 1593 | u64 orig_generation; |
1594 | struct refsort *sorted; | ||
1537 | u32 nritems; | 1595 | u32 nritems; |
1538 | u32 nr_file_extents = 0; | 1596 | u32 nr_file_extents = 0; |
1539 | struct btrfs_key key; | 1597 | struct btrfs_key key; |
@@ -1542,6 +1600,8 @@ int btrfs_inc_ref(struct btrfs_trans_handle *trans, struct btrfs_root *root, | |||
1542 | int level; | 1600 | int level; |
1543 | int ret = 0; | 1601 | int ret = 0; |
1544 | int faili = 0; | 1602 | int faili = 0; |
1603 | int refi = 0; | ||
1604 | int slot; | ||
1545 | int (*process_func)(struct btrfs_trans_handle *, struct btrfs_root *, | 1605 | int (*process_func)(struct btrfs_trans_handle *, struct btrfs_root *, |
1546 | u64, u64, u64, u64, u64, u64, u64, u64); | 1606 | u64, u64, u64, u64, u64, u64, u64, u64); |
1547 | 1607 | ||
@@ -1553,6 +1613,9 @@ int btrfs_inc_ref(struct btrfs_trans_handle *trans, struct btrfs_root *root, | |||
1553 | nritems = btrfs_header_nritems(buf); | 1613 | nritems = btrfs_header_nritems(buf); |
1554 | level = btrfs_header_level(buf); | 1614 | level = btrfs_header_level(buf); |
1555 | 1615 | ||
1616 | sorted = kmalloc(sizeof(struct refsort) * nritems, GFP_NOFS); | ||
1617 | BUG_ON(!sorted); | ||
1618 | |||
1556 | if (root->ref_cows) { | 1619 | if (root->ref_cows) { |
1557 | process_func = __btrfs_inc_extent_ref; | 1620 | process_func = __btrfs_inc_extent_ref; |
1558 | } else { | 1621 | } else { |
@@ -1565,6 +1628,11 @@ int btrfs_inc_ref(struct btrfs_trans_handle *trans, struct btrfs_root *root, | |||
1565 | process_func = __btrfs_update_extent_ref; | 1628 | process_func = __btrfs_update_extent_ref; |
1566 | } | 1629 | } |
1567 | 1630 | ||
1631 | /* | ||
1632 | * we make two passes through the items. In the first pass we | ||
1633 | * only record the byte number and slot. Then we sort based on | ||
1634 | * byte number and do the actual work based on the sorted results | ||
1635 | */ | ||
1568 | for (i = 0; i < nritems; i++) { | 1636 | for (i = 0; i < nritems; i++) { |
1569 | cond_resched(); | 1637 | cond_resched(); |
1570 | if (level == 0) { | 1638 | if (level == 0) { |
@@ -1581,6 +1649,32 @@ int btrfs_inc_ref(struct btrfs_trans_handle *trans, struct btrfs_root *root, | |||
1581 | continue; | 1649 | continue; |
1582 | 1650 | ||
1583 | nr_file_extents++; | 1651 | nr_file_extents++; |
1652 | sorted[refi].bytenr = bytenr; | ||
1653 | sorted[refi].slot = i; | ||
1654 | refi++; | ||
1655 | } else { | ||
1656 | bytenr = btrfs_node_blockptr(buf, i); | ||
1657 | sorted[refi].bytenr = bytenr; | ||
1658 | sorted[refi].slot = i; | ||
1659 | refi++; | ||
1660 | } | ||
1661 | } | ||
1662 | /* | ||
1663 | * if refi == 0, we didn't actually put anything into the sorted | ||
1664 | * array and we're done | ||
1665 | */ | ||
1666 | if (refi == 0) | ||
1667 | goto out; | ||
1668 | |||
1669 | sort(sorted, refi, sizeof(struct refsort), refsort_cmp, NULL); | ||
1670 | |||
1671 | for (i = 0; i < refi; i++) { | ||
1672 | cond_resched(); | ||
1673 | slot = sorted[i].slot; | ||
1674 | bytenr = sorted[i].bytenr; | ||
1675 | |||
1676 | if (level == 0) { | ||
1677 | btrfs_item_key_to_cpu(buf, &key, slot); | ||
1584 | 1678 | ||
1585 | ret = process_func(trans, root, bytenr, | 1679 | ret = process_func(trans, root, bytenr, |
1586 | orig_buf->start, buf->start, | 1680 | orig_buf->start, buf->start, |
@@ -1589,25 +1683,25 @@ int btrfs_inc_ref(struct btrfs_trans_handle *trans, struct btrfs_root *root, | |||
1589 | key.objectid); | 1683 | key.objectid); |
1590 | 1684 | ||
1591 | if (ret) { | 1685 | if (ret) { |
1592 | faili = i; | 1686 | faili = slot; |
1593 | WARN_ON(1); | 1687 | WARN_ON(1); |
1594 | goto fail; | 1688 | goto fail; |
1595 | } | 1689 | } |
1596 | } else { | 1690 | } else { |
1597 | bytenr = btrfs_node_blockptr(buf, i); | ||
1598 | ret = process_func(trans, root, bytenr, | 1691 | ret = process_func(trans, root, bytenr, |
1599 | orig_buf->start, buf->start, | 1692 | orig_buf->start, buf->start, |
1600 | orig_root, ref_root, | 1693 | orig_root, ref_root, |
1601 | orig_generation, ref_generation, | 1694 | orig_generation, ref_generation, |
1602 | level - 1); | 1695 | level - 1); |
1603 | if (ret) { | 1696 | if (ret) { |
1604 | faili = i; | 1697 | faili = slot; |
1605 | WARN_ON(1); | 1698 | WARN_ON(1); |
1606 | goto fail; | 1699 | goto fail; |
1607 | } | 1700 | } |
1608 | } | 1701 | } |
1609 | } | 1702 | } |
1610 | out: | 1703 | out: |
1704 | kfree(sorted); | ||
1611 | if (nr_extents) { | 1705 | if (nr_extents) { |
1612 | if (level == 0) | 1706 | if (level == 0) |
1613 | *nr_extents = nr_file_extents; | 1707 | *nr_extents = nr_file_extents; |
@@ -1616,6 +1710,7 @@ out: | |||
1616 | } | 1710 | } |
1617 | return 0; | 1711 | return 0; |
1618 | fail: | 1712 | fail: |
1713 | kfree(sorted); | ||
1619 | WARN_ON(1); | 1714 | WARN_ON(1); |
1620 | return ret; | 1715 | return ret; |
1621 | } | 1716 | } |
@@ -1818,6 +1913,7 @@ static int update_space_info(struct btrfs_fs_info *info, u64 flags, | |||
1818 | found->bytes_pinned = 0; | 1913 | found->bytes_pinned = 0; |
1819 | found->bytes_reserved = 0; | 1914 | found->bytes_reserved = 0; |
1820 | found->bytes_readonly = 0; | 1915 | found->bytes_readonly = 0; |
1916 | found->bytes_delalloc = 0; | ||
1821 | found->full = 0; | 1917 | found->full = 0; |
1822 | found->force_alloc = 0; | 1918 | found->force_alloc = 0; |
1823 | *space_info = found; | 1919 | *space_info = found; |
@@ -1881,6 +1977,233 @@ u64 btrfs_reduce_alloc_profile(struct btrfs_root *root, u64 flags) | |||
1881 | return flags; | 1977 | return flags; |
1882 | } | 1978 | } |
1883 | 1979 | ||
1980 | static u64 btrfs_get_alloc_profile(struct btrfs_root *root, u64 data) | ||
1981 | { | ||
1982 | struct btrfs_fs_info *info = root->fs_info; | ||
1983 | u64 alloc_profile; | ||
1984 | |||
1985 | if (data) { | ||
1986 | alloc_profile = info->avail_data_alloc_bits & | ||
1987 | info->data_alloc_profile; | ||
1988 | data = BTRFS_BLOCK_GROUP_DATA | alloc_profile; | ||
1989 | } else if (root == root->fs_info->chunk_root) { | ||
1990 | alloc_profile = info->avail_system_alloc_bits & | ||
1991 | info->system_alloc_profile; | ||
1992 | data = BTRFS_BLOCK_GROUP_SYSTEM | alloc_profile; | ||
1993 | } else { | ||
1994 | alloc_profile = info->avail_metadata_alloc_bits & | ||
1995 | info->metadata_alloc_profile; | ||
1996 | data = BTRFS_BLOCK_GROUP_METADATA | alloc_profile; | ||
1997 | } | ||
1998 | |||
1999 | return btrfs_reduce_alloc_profile(root, data); | ||
2000 | } | ||
2001 | |||
2002 | void btrfs_set_inode_space_info(struct btrfs_root *root, struct inode *inode) | ||
2003 | { | ||
2004 | u64 alloc_target; | ||
2005 | |||
2006 | alloc_target = btrfs_get_alloc_profile(root, 1); | ||
2007 | BTRFS_I(inode)->space_info = __find_space_info(root->fs_info, | ||
2008 | alloc_target); | ||
2009 | } | ||
2010 | |||
2011 | /* | ||
2012 | * for now this just makes sure we have at least 5% of our metadata space free | ||
2013 | * for use. | ||
2014 | */ | ||
2015 | int btrfs_check_metadata_free_space(struct btrfs_root *root) | ||
2016 | { | ||
2017 | struct btrfs_fs_info *info = root->fs_info; | ||
2018 | struct btrfs_space_info *meta_sinfo; | ||
2019 | u64 alloc_target, thresh; | ||
2020 | int committed = 0, ret; | ||
2021 | |||
2022 | /* get the space info for where the metadata will live */ | ||
2023 | alloc_target = btrfs_get_alloc_profile(root, 0); | ||
2024 | meta_sinfo = __find_space_info(info, alloc_target); | ||
2025 | |||
2026 | again: | ||
2027 | spin_lock(&meta_sinfo->lock); | ||
2028 | if (!meta_sinfo->full) | ||
2029 | thresh = meta_sinfo->total_bytes * 80; | ||
2030 | else | ||
2031 | thresh = meta_sinfo->total_bytes * 95; | ||
2032 | |||
2033 | do_div(thresh, 100); | ||
2034 | |||
2035 | if (meta_sinfo->bytes_used + meta_sinfo->bytes_reserved + | ||
2036 | meta_sinfo->bytes_pinned + meta_sinfo->bytes_readonly > thresh) { | ||
2037 | struct btrfs_trans_handle *trans; | ||
2038 | if (!meta_sinfo->full) { | ||
2039 | meta_sinfo->force_alloc = 1; | ||
2040 | spin_unlock(&meta_sinfo->lock); | ||
2041 | |||
2042 | trans = btrfs_start_transaction(root, 1); | ||
2043 | if (!trans) | ||
2044 | return -ENOMEM; | ||
2045 | |||
2046 | ret = do_chunk_alloc(trans, root->fs_info->extent_root, | ||
2047 | 2 * 1024 * 1024, alloc_target, 0); | ||
2048 | btrfs_end_transaction(trans, root); | ||
2049 | goto again; | ||
2050 | } | ||
2051 | spin_unlock(&meta_sinfo->lock); | ||
2052 | |||
2053 | if (!committed) { | ||
2054 | committed = 1; | ||
2055 | trans = btrfs_join_transaction(root, 1); | ||
2056 | if (!trans) | ||
2057 | return -ENOMEM; | ||
2058 | ret = btrfs_commit_transaction(trans, root); | ||
2059 | if (ret) | ||
2060 | return ret; | ||
2061 | goto again; | ||
2062 | } | ||
2063 | return -ENOSPC; | ||
2064 | } | ||
2065 | spin_unlock(&meta_sinfo->lock); | ||
2066 | |||
2067 | return 0; | ||
2068 | } | ||
2069 | |||
2070 | /* | ||
2071 | * This will check the space that the inode allocates from to make sure we have | ||
2072 | * enough space for bytes. | ||
2073 | */ | ||
2074 | int btrfs_check_data_free_space(struct btrfs_root *root, struct inode *inode, | ||
2075 | u64 bytes) | ||
2076 | { | ||
2077 | struct btrfs_space_info *data_sinfo; | ||
2078 | int ret = 0, committed = 0; | ||
2079 | |||
2080 | /* make sure bytes are sectorsize aligned */ | ||
2081 | bytes = (bytes + root->sectorsize - 1) & ~((u64)root->sectorsize - 1); | ||
2082 | |||
2083 | data_sinfo = BTRFS_I(inode)->space_info; | ||
2084 | again: | ||
2085 | /* make sure we have enough space to handle the data first */ | ||
2086 | spin_lock(&data_sinfo->lock); | ||
2087 | if (data_sinfo->total_bytes - data_sinfo->bytes_used - | ||
2088 | data_sinfo->bytes_delalloc - data_sinfo->bytes_reserved - | ||
2089 | data_sinfo->bytes_pinned - data_sinfo->bytes_readonly - | ||
2090 | data_sinfo->bytes_may_use < bytes) { | ||
2091 | struct btrfs_trans_handle *trans; | ||
2092 | |||
2093 | /* | ||
2094 | * if we don't have enough free bytes in this space then we need | ||
2095 | * to alloc a new chunk. | ||
2096 | */ | ||
2097 | if (!data_sinfo->full) { | ||
2098 | u64 alloc_target; | ||
2099 | |||
2100 | data_sinfo->force_alloc = 1; | ||
2101 | spin_unlock(&data_sinfo->lock); | ||
2102 | |||
2103 | alloc_target = btrfs_get_alloc_profile(root, 1); | ||
2104 | trans = btrfs_start_transaction(root, 1); | ||
2105 | if (!trans) | ||
2106 | return -ENOMEM; | ||
2107 | |||
2108 | ret = do_chunk_alloc(trans, root->fs_info->extent_root, | ||
2109 | bytes + 2 * 1024 * 1024, | ||
2110 | alloc_target, 0); | ||
2111 | btrfs_end_transaction(trans, root); | ||
2112 | if (ret) | ||
2113 | return ret; | ||
2114 | goto again; | ||
2115 | } | ||
2116 | spin_unlock(&data_sinfo->lock); | ||
2117 | |||
2118 | /* commit the current transaction and try again */ | ||
2119 | if (!committed) { | ||
2120 | committed = 1; | ||
2121 | trans = btrfs_join_transaction(root, 1); | ||
2122 | if (!trans) | ||
2123 | return -ENOMEM; | ||
2124 | ret = btrfs_commit_transaction(trans, root); | ||
2125 | if (ret) | ||
2126 | return ret; | ||
2127 | goto again; | ||
2128 | } | ||
2129 | |||
2130 | printk(KERN_ERR "no space left, need %llu, %llu delalloc bytes" | ||
2131 | ", %llu bytes_used, %llu bytes_reserved, " | ||
2132 | "%llu bytes_pinned, %llu bytes_readonly, %llu may use" | ||
2133 | "%llu total\n", bytes, data_sinfo->bytes_delalloc, | ||
2134 | data_sinfo->bytes_used, data_sinfo->bytes_reserved, | ||
2135 | data_sinfo->bytes_pinned, data_sinfo->bytes_readonly, | ||
2136 | data_sinfo->bytes_may_use, data_sinfo->total_bytes); | ||
2137 | return -ENOSPC; | ||
2138 | } | ||
2139 | data_sinfo->bytes_may_use += bytes; | ||
2140 | BTRFS_I(inode)->reserved_bytes += bytes; | ||
2141 | spin_unlock(&data_sinfo->lock); | ||
2142 | |||
2143 | return btrfs_check_metadata_free_space(root); | ||
2144 | } | ||
2145 | |||
2146 | /* | ||
2147 | * if there was an error for whatever reason after calling | ||
2148 | * btrfs_check_data_free_space, call this so we can cleanup the counters. | ||
2149 | */ | ||
2150 | void btrfs_free_reserved_data_space(struct btrfs_root *root, | ||
2151 | struct inode *inode, u64 bytes) | ||
2152 | { | ||
2153 | struct btrfs_space_info *data_sinfo; | ||
2154 | |||
2155 | /* make sure bytes are sectorsize aligned */ | ||
2156 | bytes = (bytes + root->sectorsize - 1) & ~((u64)root->sectorsize - 1); | ||
2157 | |||
2158 | data_sinfo = BTRFS_I(inode)->space_info; | ||
2159 | spin_lock(&data_sinfo->lock); | ||
2160 | data_sinfo->bytes_may_use -= bytes; | ||
2161 | BTRFS_I(inode)->reserved_bytes -= bytes; | ||
2162 | spin_unlock(&data_sinfo->lock); | ||
2163 | } | ||
2164 | |||
2165 | /* called when we are adding a delalloc extent to the inode's io_tree */ | ||
2166 | void btrfs_delalloc_reserve_space(struct btrfs_root *root, struct inode *inode, | ||
2167 | u64 bytes) | ||
2168 | { | ||
2169 | struct btrfs_space_info *data_sinfo; | ||
2170 | |||
2171 | /* get the space info for where this inode will be storing its data */ | ||
2172 | data_sinfo = BTRFS_I(inode)->space_info; | ||
2173 | |||
2174 | /* make sure we have enough space to handle the data first */ | ||
2175 | spin_lock(&data_sinfo->lock); | ||
2176 | data_sinfo->bytes_delalloc += bytes; | ||
2177 | |||
2178 | /* | ||
2179 | * we are adding a delalloc extent without calling | ||
2180 | * btrfs_check_data_free_space first. This happens on a weird | ||
2181 | * writepage condition, but shouldn't hurt our accounting | ||
2182 | */ | ||
2183 | if (unlikely(bytes > BTRFS_I(inode)->reserved_bytes)) { | ||
2184 | data_sinfo->bytes_may_use -= BTRFS_I(inode)->reserved_bytes; | ||
2185 | BTRFS_I(inode)->reserved_bytes = 0; | ||
2186 | } else { | ||
2187 | data_sinfo->bytes_may_use -= bytes; | ||
2188 | BTRFS_I(inode)->reserved_bytes -= bytes; | ||
2189 | } | ||
2190 | |||
2191 | spin_unlock(&data_sinfo->lock); | ||
2192 | } | ||
2193 | |||
2194 | /* called when we are clearing an delalloc extent from the inode's io_tree */ | ||
2195 | void btrfs_delalloc_free_space(struct btrfs_root *root, struct inode *inode, | ||
2196 | u64 bytes) | ||
2197 | { | ||
2198 | struct btrfs_space_info *info; | ||
2199 | |||
2200 | info = BTRFS_I(inode)->space_info; | ||
2201 | |||
2202 | spin_lock(&info->lock); | ||
2203 | info->bytes_delalloc -= bytes; | ||
2204 | spin_unlock(&info->lock); | ||
2205 | } | ||
2206 | |||
1884 | static int do_chunk_alloc(struct btrfs_trans_handle *trans, | 2207 | static int do_chunk_alloc(struct btrfs_trans_handle *trans, |
1885 | struct btrfs_root *extent_root, u64 alloc_bytes, | 2208 | struct btrfs_root *extent_root, u64 alloc_bytes, |
1886 | u64 flags, int force) | 2209 | u64 flags, int force) |
@@ -2137,13 +2460,12 @@ static int finish_current_insert(struct btrfs_trans_handle *trans, | |||
2137 | u64 end; | 2460 | u64 end; |
2138 | u64 priv; | 2461 | u64 priv; |
2139 | u64 search = 0; | 2462 | u64 search = 0; |
2140 | u64 skipped = 0; | ||
2141 | struct btrfs_fs_info *info = extent_root->fs_info; | 2463 | struct btrfs_fs_info *info = extent_root->fs_info; |
2142 | struct btrfs_path *path; | 2464 | struct btrfs_path *path; |
2143 | struct pending_extent_op *extent_op, *tmp; | 2465 | struct pending_extent_op *extent_op, *tmp; |
2144 | struct list_head insert_list, update_list; | 2466 | struct list_head insert_list, update_list; |
2145 | int ret; | 2467 | int ret; |
2146 | int num_inserts = 0, max_inserts; | 2468 | int num_inserts = 0, max_inserts, restart = 0; |
2147 | 2469 | ||
2148 | path = btrfs_alloc_path(); | 2470 | path = btrfs_alloc_path(); |
2149 | INIT_LIST_HEAD(&insert_list); | 2471 | INIT_LIST_HEAD(&insert_list); |
@@ -2159,18 +2481,19 @@ again: | |||
2159 | ret = find_first_extent_bit(&info->extent_ins, search, &start, | 2481 | ret = find_first_extent_bit(&info->extent_ins, search, &start, |
2160 | &end, EXTENT_WRITEBACK); | 2482 | &end, EXTENT_WRITEBACK); |
2161 | if (ret) { | 2483 | if (ret) { |
2162 | if (skipped && all && !num_inserts) { | 2484 | if (restart && !num_inserts && |
2163 | skipped = 0; | 2485 | list_empty(&update_list)) { |
2486 | restart = 0; | ||
2164 | search = 0; | 2487 | search = 0; |
2165 | continue; | 2488 | continue; |
2166 | } | 2489 | } |
2167 | mutex_unlock(&info->extent_ins_mutex); | ||
2168 | break; | 2490 | break; |
2169 | } | 2491 | } |
2170 | 2492 | ||
2171 | ret = try_lock_extent(&info->extent_ins, start, end, GFP_NOFS); | 2493 | ret = try_lock_extent(&info->extent_ins, start, end, GFP_NOFS); |
2172 | if (!ret) { | 2494 | if (!ret) { |
2173 | skipped = 1; | 2495 | if (all) |
2496 | restart = 1; | ||
2174 | search = end + 1; | 2497 | search = end + 1; |
2175 | if (need_resched()) { | 2498 | if (need_resched()) { |
2176 | mutex_unlock(&info->extent_ins_mutex); | 2499 | mutex_unlock(&info->extent_ins_mutex); |
@@ -2189,7 +2512,7 @@ again: | |||
2189 | list_add_tail(&extent_op->list, &insert_list); | 2512 | list_add_tail(&extent_op->list, &insert_list); |
2190 | search = end + 1; | 2513 | search = end + 1; |
2191 | if (num_inserts == max_inserts) { | 2514 | if (num_inserts == max_inserts) { |
2192 | mutex_unlock(&info->extent_ins_mutex); | 2515 | restart = 1; |
2193 | break; | 2516 | break; |
2194 | } | 2517 | } |
2195 | } else if (extent_op->type == PENDING_BACKREF_UPDATE) { | 2518 | } else if (extent_op->type == PENDING_BACKREF_UPDATE) { |
@@ -2205,7 +2528,6 @@ again: | |||
2205 | * somebody marked this thing for deletion then just unlock it and be | 2528 | * somebody marked this thing for deletion then just unlock it and be |
2206 | * done, the free_extents will handle it | 2529 | * done, the free_extents will handle it |
2207 | */ | 2530 | */ |
2208 | mutex_lock(&info->extent_ins_mutex); | ||
2209 | list_for_each_entry_safe(extent_op, tmp, &update_list, list) { | 2531 | list_for_each_entry_safe(extent_op, tmp, &update_list, list) { |
2210 | clear_extent_bits(&info->extent_ins, extent_op->bytenr, | 2532 | clear_extent_bits(&info->extent_ins, extent_op->bytenr, |
2211 | extent_op->bytenr + extent_op->num_bytes - 1, | 2533 | extent_op->bytenr + extent_op->num_bytes - 1, |
@@ -2227,6 +2549,10 @@ again: | |||
2227 | if (!list_empty(&update_list)) { | 2549 | if (!list_empty(&update_list)) { |
2228 | ret = update_backrefs(trans, extent_root, path, &update_list); | 2550 | ret = update_backrefs(trans, extent_root, path, &update_list); |
2229 | BUG_ON(ret); | 2551 | BUG_ON(ret); |
2552 | |||
2553 | /* we may have COW'ed new blocks, so lets start over */ | ||
2554 | if (all) | ||
2555 | restart = 1; | ||
2230 | } | 2556 | } |
2231 | 2557 | ||
2232 | /* | 2558 | /* |
@@ -2234,9 +2560,9 @@ again: | |||
2234 | * need to make sure everything is cleaned then reset everything and | 2560 | * need to make sure everything is cleaned then reset everything and |
2235 | * go back to the beginning | 2561 | * go back to the beginning |
2236 | */ | 2562 | */ |
2237 | if (!num_inserts && all && skipped) { | 2563 | if (!num_inserts && restart) { |
2238 | search = 0; | 2564 | search = 0; |
2239 | skipped = 0; | 2565 | restart = 0; |
2240 | INIT_LIST_HEAD(&update_list); | 2566 | INIT_LIST_HEAD(&update_list); |
2241 | INIT_LIST_HEAD(&insert_list); | 2567 | INIT_LIST_HEAD(&insert_list); |
2242 | goto again; | 2568 | goto again; |
@@ -2293,27 +2619,19 @@ again: | |||
2293 | BUG_ON(ret); | 2619 | BUG_ON(ret); |
2294 | 2620 | ||
2295 | /* | 2621 | /* |
2296 | * if we broke out of the loop in order to insert stuff because we hit | 2622 | * if restart is set for whatever reason we need to go back and start |
2297 | * the maximum number of inserts at a time we can handle, then loop | 2623 | * searching through the pending list again. |
2298 | * back and pick up where we left off | 2624 | * |
2299 | */ | 2625 | * We just inserted some extents, which could have resulted in new |
2300 | if (num_inserts == max_inserts) { | 2626 | * blocks being allocated, which would result in new blocks needing |
2301 | INIT_LIST_HEAD(&insert_list); | 2627 | * updates, so if all is set we _must_ restart to get the updated |
2302 | INIT_LIST_HEAD(&update_list); | 2628 | * blocks. |
2303 | num_inserts = 0; | ||
2304 | goto again; | ||
2305 | } | ||
2306 | |||
2307 | /* | ||
2308 | * again, if we need to make absolutely sure there are no more pending | ||
2309 | * extent operations left and we know that we skipped some, go back to | ||
2310 | * the beginning and do it all again | ||
2311 | */ | 2629 | */ |
2312 | if (all && skipped) { | 2630 | if (restart || all) { |
2313 | INIT_LIST_HEAD(&insert_list); | 2631 | INIT_LIST_HEAD(&insert_list); |
2314 | INIT_LIST_HEAD(&update_list); | 2632 | INIT_LIST_HEAD(&update_list); |
2315 | search = 0; | 2633 | search = 0; |
2316 | skipped = 0; | 2634 | restart = 0; |
2317 | num_inserts = 0; | 2635 | num_inserts = 0; |
2318 | goto again; | 2636 | goto again; |
2319 | } | 2637 | } |
@@ -2547,6 +2865,7 @@ again: | |||
2547 | if (ret) { | 2865 | if (ret) { |
2548 | if (all && skipped && !nr) { | 2866 | if (all && skipped && !nr) { |
2549 | search = 0; | 2867 | search = 0; |
2868 | skipped = 0; | ||
2550 | continue; | 2869 | continue; |
2551 | } | 2870 | } |
2552 | mutex_unlock(&info->extent_ins_mutex); | 2871 | mutex_unlock(&info->extent_ins_mutex); |
@@ -2633,6 +2952,8 @@ again: | |||
2633 | goto again; | 2952 | goto again; |
2634 | } | 2953 | } |
2635 | 2954 | ||
2955 | if (!err) | ||
2956 | finish_current_insert(trans, extent_root, 0); | ||
2636 | return err; | 2957 | return err; |
2637 | } | 2958 | } |
2638 | 2959 | ||
@@ -2700,13 +3021,9 @@ static int __btrfs_free_extent(struct btrfs_trans_handle *trans, | |||
2700 | /* if metadata always pin */ | 3021 | /* if metadata always pin */ |
2701 | if (owner_objectid < BTRFS_FIRST_FREE_OBJECTID) { | 3022 | if (owner_objectid < BTRFS_FIRST_FREE_OBJECTID) { |
2702 | if (root->root_key.objectid == BTRFS_TREE_LOG_OBJECTID) { | 3023 | if (root->root_key.objectid == BTRFS_TREE_LOG_OBJECTID) { |
2703 | struct btrfs_block_group_cache *cache; | 3024 | mutex_lock(&root->fs_info->pinned_mutex); |
2704 | 3025 | btrfs_update_pinned_extents(root, bytenr, num_bytes, 1); | |
2705 | /* btrfs_free_reserved_extent */ | 3026 | mutex_unlock(&root->fs_info->pinned_mutex); |
2706 | cache = btrfs_lookup_block_group(root->fs_info, bytenr); | ||
2707 | BUG_ON(!cache); | ||
2708 | btrfs_add_free_space(cache, bytenr, num_bytes); | ||
2709 | put_block_group(cache); | ||
2710 | update_reserved_extents(root, bytenr, num_bytes, 0); | 3027 | update_reserved_extents(root, bytenr, num_bytes, 0); |
2711 | return 0; | 3028 | return 0; |
2712 | } | 3029 | } |
@@ -2787,7 +3104,8 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans, | |||
2787 | 3104 | ||
2788 | if (data & BTRFS_BLOCK_GROUP_METADATA) { | 3105 | if (data & BTRFS_BLOCK_GROUP_METADATA) { |
2789 | last_ptr = &root->fs_info->last_alloc; | 3106 | last_ptr = &root->fs_info->last_alloc; |
2790 | empty_cluster = 64 * 1024; | 3107 | if (!btrfs_test_opt(root, SSD)) |
3108 | empty_cluster = 64 * 1024; | ||
2791 | } | 3109 | } |
2792 | 3110 | ||
2793 | if ((data & BTRFS_BLOCK_GROUP_DATA) && btrfs_test_opt(root, SSD)) | 3111 | if ((data & BTRFS_BLOCK_GROUP_DATA) && btrfs_test_opt(root, SSD)) |
@@ -3014,16 +3332,18 @@ loop_check: | |||
3014 | static void dump_space_info(struct btrfs_space_info *info, u64 bytes) | 3332 | static void dump_space_info(struct btrfs_space_info *info, u64 bytes) |
3015 | { | 3333 | { |
3016 | struct btrfs_block_group_cache *cache; | 3334 | struct btrfs_block_group_cache *cache; |
3017 | struct list_head *l; | ||
3018 | 3335 | ||
3019 | printk(KERN_INFO "space_info has %llu free, is %sfull\n", | 3336 | printk(KERN_INFO "space_info has %llu free, is %sfull\n", |
3020 | (unsigned long long)(info->total_bytes - info->bytes_used - | 3337 | (unsigned long long)(info->total_bytes - info->bytes_used - |
3021 | info->bytes_pinned - info->bytes_reserved), | 3338 | info->bytes_pinned - info->bytes_reserved), |
3022 | (info->full) ? "" : "not "); | 3339 | (info->full) ? "" : "not "); |
3340 | printk(KERN_INFO "space_info total=%llu, pinned=%llu, delalloc=%llu," | ||
3341 | " may_use=%llu, used=%llu\n", info->total_bytes, | ||
3342 | info->bytes_pinned, info->bytes_delalloc, info->bytes_may_use, | ||
3343 | info->bytes_used); | ||
3023 | 3344 | ||
3024 | down_read(&info->groups_sem); | 3345 | down_read(&info->groups_sem); |
3025 | list_for_each(l, &info->block_groups) { | 3346 | list_for_each_entry(cache, &info->block_groups, list) { |
3026 | cache = list_entry(l, struct btrfs_block_group_cache, list); | ||
3027 | spin_lock(&cache->lock); | 3347 | spin_lock(&cache->lock); |
3028 | printk(KERN_INFO "block group %llu has %llu bytes, %llu used " | 3348 | printk(KERN_INFO "block group %llu has %llu bytes, %llu used " |
3029 | "%llu pinned %llu reserved\n", | 3349 | "%llu pinned %llu reserved\n", |
@@ -3047,24 +3367,10 @@ static int __btrfs_reserve_extent(struct btrfs_trans_handle *trans, | |||
3047 | { | 3367 | { |
3048 | int ret; | 3368 | int ret; |
3049 | u64 search_start = 0; | 3369 | u64 search_start = 0; |
3050 | u64 alloc_profile; | ||
3051 | struct btrfs_fs_info *info = root->fs_info; | 3370 | struct btrfs_fs_info *info = root->fs_info; |
3052 | 3371 | ||
3053 | if (data) { | 3372 | data = btrfs_get_alloc_profile(root, data); |
3054 | alloc_profile = info->avail_data_alloc_bits & | ||
3055 | info->data_alloc_profile; | ||
3056 | data = BTRFS_BLOCK_GROUP_DATA | alloc_profile; | ||
3057 | } else if (root == root->fs_info->chunk_root) { | ||
3058 | alloc_profile = info->avail_system_alloc_bits & | ||
3059 | info->system_alloc_profile; | ||
3060 | data = BTRFS_BLOCK_GROUP_SYSTEM | alloc_profile; | ||
3061 | } else { | ||
3062 | alloc_profile = info->avail_metadata_alloc_bits & | ||
3063 | info->metadata_alloc_profile; | ||
3064 | data = BTRFS_BLOCK_GROUP_METADATA | alloc_profile; | ||
3065 | } | ||
3066 | again: | 3373 | again: |
3067 | data = btrfs_reduce_alloc_profile(root, data); | ||
3068 | /* | 3374 | /* |
3069 | * the only place that sets empty_size is btrfs_realloc_node, which | 3375 | * the only place that sets empty_size is btrfs_realloc_node, which |
3070 | * is not called recursively on allocations | 3376 | * is not called recursively on allocations |
@@ -3332,7 +3638,8 @@ int btrfs_alloc_extent(struct btrfs_trans_handle *trans, | |||
3332 | 3638 | ||
3333 | struct extent_buffer *btrfs_init_new_buffer(struct btrfs_trans_handle *trans, | 3639 | struct extent_buffer *btrfs_init_new_buffer(struct btrfs_trans_handle *trans, |
3334 | struct btrfs_root *root, | 3640 | struct btrfs_root *root, |
3335 | u64 bytenr, u32 blocksize) | 3641 | u64 bytenr, u32 blocksize, |
3642 | int level) | ||
3336 | { | 3643 | { |
3337 | struct extent_buffer *buf; | 3644 | struct extent_buffer *buf; |
3338 | 3645 | ||
@@ -3340,9 +3647,13 @@ struct extent_buffer *btrfs_init_new_buffer(struct btrfs_trans_handle *trans, | |||
3340 | if (!buf) | 3647 | if (!buf) |
3341 | return ERR_PTR(-ENOMEM); | 3648 | return ERR_PTR(-ENOMEM); |
3342 | btrfs_set_header_generation(buf, trans->transid); | 3649 | btrfs_set_header_generation(buf, trans->transid); |
3650 | btrfs_set_buffer_lockdep_class(buf, level); | ||
3343 | btrfs_tree_lock(buf); | 3651 | btrfs_tree_lock(buf); |
3344 | clean_tree_block(trans, root, buf); | 3652 | clean_tree_block(trans, root, buf); |
3653 | |||
3654 | btrfs_set_lock_blocking(buf); | ||
3345 | btrfs_set_buffer_uptodate(buf); | 3655 | btrfs_set_buffer_uptodate(buf); |
3656 | |||
3346 | if (root->root_key.objectid == BTRFS_TREE_LOG_OBJECTID) { | 3657 | if (root->root_key.objectid == BTRFS_TREE_LOG_OBJECTID) { |
3347 | set_extent_dirty(&root->dirty_log_pages, buf->start, | 3658 | set_extent_dirty(&root->dirty_log_pages, buf->start, |
3348 | buf->start + buf->len - 1, GFP_NOFS); | 3659 | buf->start + buf->len - 1, GFP_NOFS); |
@@ -3351,6 +3662,7 @@ struct extent_buffer *btrfs_init_new_buffer(struct btrfs_trans_handle *trans, | |||
3351 | buf->start + buf->len - 1, GFP_NOFS); | 3662 | buf->start + buf->len - 1, GFP_NOFS); |
3352 | } | 3663 | } |
3353 | trans->blocks_used++; | 3664 | trans->blocks_used++; |
3665 | /* this returns a buffer locked for blocking */ | ||
3354 | return buf; | 3666 | return buf; |
3355 | } | 3667 | } |
3356 | 3668 | ||
@@ -3379,7 +3691,8 @@ struct extent_buffer *btrfs_alloc_free_block(struct btrfs_trans_handle *trans, | |||
3379 | return ERR_PTR(ret); | 3691 | return ERR_PTR(ret); |
3380 | } | 3692 | } |
3381 | 3693 | ||
3382 | buf = btrfs_init_new_buffer(trans, root, ins.objectid, blocksize); | 3694 | buf = btrfs_init_new_buffer(trans, root, ins.objectid, |
3695 | blocksize, level); | ||
3383 | return buf; | 3696 | return buf; |
3384 | } | 3697 | } |
3385 | 3698 | ||
@@ -3388,36 +3701,73 @@ int btrfs_drop_leaf_ref(struct btrfs_trans_handle *trans, | |||
3388 | { | 3701 | { |
3389 | u64 leaf_owner; | 3702 | u64 leaf_owner; |
3390 | u64 leaf_generation; | 3703 | u64 leaf_generation; |
3704 | struct refsort *sorted; | ||
3391 | struct btrfs_key key; | 3705 | struct btrfs_key key; |
3392 | struct btrfs_file_extent_item *fi; | 3706 | struct btrfs_file_extent_item *fi; |
3393 | int i; | 3707 | int i; |
3394 | int nritems; | 3708 | int nritems; |
3395 | int ret; | 3709 | int ret; |
3710 | int refi = 0; | ||
3711 | int slot; | ||
3396 | 3712 | ||
3397 | BUG_ON(!btrfs_is_leaf(leaf)); | 3713 | BUG_ON(!btrfs_is_leaf(leaf)); |
3398 | nritems = btrfs_header_nritems(leaf); | 3714 | nritems = btrfs_header_nritems(leaf); |
3399 | leaf_owner = btrfs_header_owner(leaf); | 3715 | leaf_owner = btrfs_header_owner(leaf); |
3400 | leaf_generation = btrfs_header_generation(leaf); | 3716 | leaf_generation = btrfs_header_generation(leaf); |
3401 | 3717 | ||
3718 | sorted = kmalloc(sizeof(*sorted) * nritems, GFP_NOFS); | ||
3719 | /* we do this loop twice. The first time we build a list | ||
3720 | * of the extents we have a reference on, then we sort the list | ||
3721 | * by bytenr. The second time around we actually do the | ||
3722 | * extent freeing. | ||
3723 | */ | ||
3402 | for (i = 0; i < nritems; i++) { | 3724 | for (i = 0; i < nritems; i++) { |
3403 | u64 disk_bytenr; | 3725 | u64 disk_bytenr; |
3404 | cond_resched(); | 3726 | cond_resched(); |
3405 | 3727 | ||
3406 | btrfs_item_key_to_cpu(leaf, &key, i); | 3728 | btrfs_item_key_to_cpu(leaf, &key, i); |
3729 | |||
3730 | /* only extents have references, skip everything else */ | ||
3407 | if (btrfs_key_type(&key) != BTRFS_EXTENT_DATA_KEY) | 3731 | if (btrfs_key_type(&key) != BTRFS_EXTENT_DATA_KEY) |
3408 | continue; | 3732 | continue; |
3733 | |||
3409 | fi = btrfs_item_ptr(leaf, i, struct btrfs_file_extent_item); | 3734 | fi = btrfs_item_ptr(leaf, i, struct btrfs_file_extent_item); |
3735 | |||
3736 | /* inline extents live in the btree, they don't have refs */ | ||
3410 | if (btrfs_file_extent_type(leaf, fi) == | 3737 | if (btrfs_file_extent_type(leaf, fi) == |
3411 | BTRFS_FILE_EXTENT_INLINE) | 3738 | BTRFS_FILE_EXTENT_INLINE) |
3412 | continue; | 3739 | continue; |
3413 | /* | 3740 | |
3414 | * FIXME make sure to insert a trans record that | ||
3415 | * repeats the snapshot del on crash | ||
3416 | */ | ||
3417 | disk_bytenr = btrfs_file_extent_disk_bytenr(leaf, fi); | 3741 | disk_bytenr = btrfs_file_extent_disk_bytenr(leaf, fi); |
3742 | |||
3743 | /* holes don't have refs */ | ||
3418 | if (disk_bytenr == 0) | 3744 | if (disk_bytenr == 0) |
3419 | continue; | 3745 | continue; |
3420 | 3746 | ||
3747 | sorted[refi].bytenr = disk_bytenr; | ||
3748 | sorted[refi].slot = i; | ||
3749 | refi++; | ||
3750 | } | ||
3751 | |||
3752 | if (refi == 0) | ||
3753 | goto out; | ||
3754 | |||
3755 | sort(sorted, refi, sizeof(struct refsort), refsort_cmp, NULL); | ||
3756 | |||
3757 | for (i = 0; i < refi; i++) { | ||
3758 | u64 disk_bytenr; | ||
3759 | |||
3760 | disk_bytenr = sorted[i].bytenr; | ||
3761 | slot = sorted[i].slot; | ||
3762 | |||
3763 | cond_resched(); | ||
3764 | |||
3765 | btrfs_item_key_to_cpu(leaf, &key, slot); | ||
3766 | if (btrfs_key_type(&key) != BTRFS_EXTENT_DATA_KEY) | ||
3767 | continue; | ||
3768 | |||
3769 | fi = btrfs_item_ptr(leaf, slot, struct btrfs_file_extent_item); | ||
3770 | |||
3421 | ret = __btrfs_free_extent(trans, root, disk_bytenr, | 3771 | ret = __btrfs_free_extent(trans, root, disk_bytenr, |
3422 | btrfs_file_extent_disk_num_bytes(leaf, fi), | 3772 | btrfs_file_extent_disk_num_bytes(leaf, fi), |
3423 | leaf->start, leaf_owner, leaf_generation, | 3773 | leaf->start, leaf_owner, leaf_generation, |
@@ -3428,6 +3778,8 @@ int btrfs_drop_leaf_ref(struct btrfs_trans_handle *trans, | |||
3428 | wake_up(&root->fs_info->transaction_throttle); | 3778 | wake_up(&root->fs_info->transaction_throttle); |
3429 | cond_resched(); | 3779 | cond_resched(); |
3430 | } | 3780 | } |
3781 | out: | ||
3782 | kfree(sorted); | ||
3431 | return 0; | 3783 | return 0; |
3432 | } | 3784 | } |
3433 | 3785 | ||
@@ -3437,9 +3789,25 @@ static noinline int cache_drop_leaf_ref(struct btrfs_trans_handle *trans, | |||
3437 | { | 3789 | { |
3438 | int i; | 3790 | int i; |
3439 | int ret; | 3791 | int ret; |
3440 | struct btrfs_extent_info *info = ref->extents; | 3792 | struct btrfs_extent_info *info; |
3793 | struct refsort *sorted; | ||
3794 | |||
3795 | if (ref->nritems == 0) | ||
3796 | return 0; | ||
3441 | 3797 | ||
3798 | sorted = kmalloc(sizeof(*sorted) * ref->nritems, GFP_NOFS); | ||
3442 | for (i = 0; i < ref->nritems; i++) { | 3799 | for (i = 0; i < ref->nritems; i++) { |
3800 | sorted[i].bytenr = ref->extents[i].bytenr; | ||
3801 | sorted[i].slot = i; | ||
3802 | } | ||
3803 | sort(sorted, ref->nritems, sizeof(struct refsort), refsort_cmp, NULL); | ||
3804 | |||
3805 | /* | ||
3806 | * the items in the ref were sorted when the ref was inserted | ||
3807 | * into the ref cache, so this is already in order | ||
3808 | */ | ||
3809 | for (i = 0; i < ref->nritems; i++) { | ||
3810 | info = ref->extents + sorted[i].slot; | ||
3443 | ret = __btrfs_free_extent(trans, root, info->bytenr, | 3811 | ret = __btrfs_free_extent(trans, root, info->bytenr, |
3444 | info->num_bytes, ref->bytenr, | 3812 | info->num_bytes, ref->bytenr, |
3445 | ref->owner, ref->generation, | 3813 | ref->owner, ref->generation, |
@@ -3453,6 +3821,7 @@ static noinline int cache_drop_leaf_ref(struct btrfs_trans_handle *trans, | |||
3453 | info++; | 3821 | info++; |
3454 | } | 3822 | } |
3455 | 3823 | ||
3824 | kfree(sorted); | ||
3456 | return 0; | 3825 | return 0; |
3457 | } | 3826 | } |
3458 | 3827 | ||
@@ -3497,6 +3866,152 @@ static int drop_snap_lookup_refcount(struct btrfs_root *root, u64 start, | |||
3497 | } | 3866 | } |
3498 | 3867 | ||
3499 | /* | 3868 | /* |
3869 | * this is used while deleting old snapshots, and it drops the refs | ||
3870 | * on a whole subtree starting from a level 1 node. | ||
3871 | * | ||
3872 | * The idea is to sort all the leaf pointers, and then drop the | ||
3873 | * ref on all the leaves in order. Most of the time the leaves | ||
3874 | * will have ref cache entries, so no leaf IOs will be required to | ||
3875 | * find the extents they have references on. | ||
3876 | * | ||
3877 | * For each leaf, any references it has are also dropped in order | ||
3878 | * | ||
3879 | * This ends up dropping the references in something close to optimal | ||
3880 | * order for reading and modifying the extent allocation tree. | ||
3881 | */ | ||
3882 | static noinline int drop_level_one_refs(struct btrfs_trans_handle *trans, | ||
3883 | struct btrfs_root *root, | ||
3884 | struct btrfs_path *path) | ||
3885 | { | ||
3886 | u64 bytenr; | ||
3887 | u64 root_owner; | ||
3888 | u64 root_gen; | ||
3889 | struct extent_buffer *eb = path->nodes[1]; | ||
3890 | struct extent_buffer *leaf; | ||
3891 | struct btrfs_leaf_ref *ref; | ||
3892 | struct refsort *sorted = NULL; | ||
3893 | int nritems = btrfs_header_nritems(eb); | ||
3894 | int ret; | ||
3895 | int i; | ||
3896 | int refi = 0; | ||
3897 | int slot = path->slots[1]; | ||
3898 | u32 blocksize = btrfs_level_size(root, 0); | ||
3899 | u32 refs; | ||
3900 | |||
3901 | if (nritems == 0) | ||
3902 | goto out; | ||
3903 | |||
3904 | root_owner = btrfs_header_owner(eb); | ||
3905 | root_gen = btrfs_header_generation(eb); | ||
3906 | sorted = kmalloc(sizeof(*sorted) * nritems, GFP_NOFS); | ||
3907 | |||
3908 | /* | ||
3909 | * step one, sort all the leaf pointers so we don't scribble | ||
3910 | * randomly into the extent allocation tree | ||
3911 | */ | ||
3912 | for (i = slot; i < nritems; i++) { | ||
3913 | sorted[refi].bytenr = btrfs_node_blockptr(eb, i); | ||
3914 | sorted[refi].slot = i; | ||
3915 | refi++; | ||
3916 | } | ||
3917 | |||
3918 | /* | ||
3919 | * nritems won't be zero, but if we're picking up drop_snapshot | ||
3920 | * after a crash, slot might be > 0, so double check things | ||
3921 | * just in case. | ||
3922 | */ | ||
3923 | if (refi == 0) | ||
3924 | goto out; | ||
3925 | |||
3926 | sort(sorted, refi, sizeof(struct refsort), refsort_cmp, NULL); | ||
3927 | |||
3928 | /* | ||
3929 | * the first loop frees everything the leaves point to | ||
3930 | */ | ||
3931 | for (i = 0; i < refi; i++) { | ||
3932 | u64 ptr_gen; | ||
3933 | |||
3934 | bytenr = sorted[i].bytenr; | ||
3935 | |||
3936 | /* | ||
3937 | * check the reference count on this leaf. If it is > 1 | ||
3938 | * we just decrement it below and don't update any | ||
3939 | * of the refs the leaf points to. | ||
3940 | */ | ||
3941 | ret = drop_snap_lookup_refcount(root, bytenr, blocksize, &refs); | ||
3942 | BUG_ON(ret); | ||
3943 | if (refs != 1) | ||
3944 | continue; | ||
3945 | |||
3946 | ptr_gen = btrfs_node_ptr_generation(eb, sorted[i].slot); | ||
3947 | |||
3948 | /* | ||
3949 | * the leaf only had one reference, which means the | ||
3950 | * only thing pointing to this leaf is the snapshot | ||
3951 | * we're deleting. It isn't possible for the reference | ||
3952 | * count to increase again later | ||
3953 | * | ||
3954 | * The reference cache is checked for the leaf, | ||
3955 | * and if found we'll be able to drop any refs held by | ||
3956 | * the leaf without needing to read it in. | ||
3957 | */ | ||
3958 | ref = btrfs_lookup_leaf_ref(root, bytenr); | ||
3959 | if (ref && ref->generation != ptr_gen) { | ||
3960 | btrfs_free_leaf_ref(root, ref); | ||
3961 | ref = NULL; | ||
3962 | } | ||
3963 | if (ref) { | ||
3964 | ret = cache_drop_leaf_ref(trans, root, ref); | ||
3965 | BUG_ON(ret); | ||
3966 | btrfs_remove_leaf_ref(root, ref); | ||
3967 | btrfs_free_leaf_ref(root, ref); | ||
3968 | } else { | ||
3969 | /* | ||
3970 | * the leaf wasn't in the reference cache, so | ||
3971 | * we have to read it. | ||
3972 | */ | ||
3973 | leaf = read_tree_block(root, bytenr, blocksize, | ||
3974 | ptr_gen); | ||
3975 | ret = btrfs_drop_leaf_ref(trans, root, leaf); | ||
3976 | BUG_ON(ret); | ||
3977 | free_extent_buffer(leaf); | ||
3978 | } | ||
3979 | atomic_inc(&root->fs_info->throttle_gen); | ||
3980 | wake_up(&root->fs_info->transaction_throttle); | ||
3981 | cond_resched(); | ||
3982 | } | ||
3983 | |||
3984 | /* | ||
3985 | * run through the loop again to free the refs on the leaves. | ||
3986 | * This is faster than doing it in the loop above because | ||
3987 | * the leaves are likely to be clustered together. We end up | ||
3988 | * working in nice chunks on the extent allocation tree. | ||
3989 | */ | ||
3990 | for (i = 0; i < refi; i++) { | ||
3991 | bytenr = sorted[i].bytenr; | ||
3992 | ret = __btrfs_free_extent(trans, root, bytenr, | ||
3993 | blocksize, eb->start, | ||
3994 | root_owner, root_gen, 0, 1); | ||
3995 | BUG_ON(ret); | ||
3996 | |||
3997 | atomic_inc(&root->fs_info->throttle_gen); | ||
3998 | wake_up(&root->fs_info->transaction_throttle); | ||
3999 | cond_resched(); | ||
4000 | } | ||
4001 | out: | ||
4002 | kfree(sorted); | ||
4003 | |||
4004 | /* | ||
4005 | * update the path to show we've processed the entire level 1 | ||
4006 | * node. This will get saved into the root's drop_snapshot_progress | ||
4007 | * field so these drops are not repeated again if this transaction | ||
4008 | * commits. | ||
4009 | */ | ||
4010 | path->slots[1] = nritems; | ||
4011 | return 0; | ||
4012 | } | ||
4013 | |||
4014 | /* | ||
3500 | * helper function for drop_snapshot, this walks down the tree dropping ref | 4015 | * helper function for drop_snapshot, this walks down the tree dropping ref |
3501 | * counts as it goes. | 4016 | * counts as it goes. |
3502 | */ | 4017 | */ |
@@ -3511,7 +4026,6 @@ static noinline int walk_down_tree(struct btrfs_trans_handle *trans, | |||
3511 | struct extent_buffer *next; | 4026 | struct extent_buffer *next; |
3512 | struct extent_buffer *cur; | 4027 | struct extent_buffer *cur; |
3513 | struct extent_buffer *parent; | 4028 | struct extent_buffer *parent; |
3514 | struct btrfs_leaf_ref *ref; | ||
3515 | u32 blocksize; | 4029 | u32 blocksize; |
3516 | int ret; | 4030 | int ret; |
3517 | u32 refs; | 4031 | u32 refs; |
@@ -3538,17 +4052,46 @@ static noinline int walk_down_tree(struct btrfs_trans_handle *trans, | |||
3538 | if (path->slots[*level] >= | 4052 | if (path->slots[*level] >= |
3539 | btrfs_header_nritems(cur)) | 4053 | btrfs_header_nritems(cur)) |
3540 | break; | 4054 | break; |
4055 | |||
4056 | /* the new code goes down to level 1 and does all the | ||
4057 | * leaves pointed to that node in bulk. So, this check | ||
4058 | * for level 0 will always be false. | ||
4059 | * | ||
4060 | * But, the disk format allows the drop_snapshot_progress | ||
4061 | * field in the root to leave things in a state where | ||
4062 | * a leaf will need cleaning up here. If someone crashes | ||
4063 | * with the old code and then boots with the new code, | ||
4064 | * we might find a leaf here. | ||
4065 | */ | ||
3541 | if (*level == 0) { | 4066 | if (*level == 0) { |
3542 | ret = btrfs_drop_leaf_ref(trans, root, cur); | 4067 | ret = btrfs_drop_leaf_ref(trans, root, cur); |
3543 | BUG_ON(ret); | 4068 | BUG_ON(ret); |
3544 | break; | 4069 | break; |
3545 | } | 4070 | } |
4071 | |||
4072 | /* | ||
4073 | * once we get to level one, process the whole node | ||
4074 | * at once, including everything below it. | ||
4075 | */ | ||
4076 | if (*level == 1) { | ||
4077 | ret = drop_level_one_refs(trans, root, path); | ||
4078 | BUG_ON(ret); | ||
4079 | break; | ||
4080 | } | ||
4081 | |||
3546 | bytenr = btrfs_node_blockptr(cur, path->slots[*level]); | 4082 | bytenr = btrfs_node_blockptr(cur, path->slots[*level]); |
3547 | ptr_gen = btrfs_node_ptr_generation(cur, path->slots[*level]); | 4083 | ptr_gen = btrfs_node_ptr_generation(cur, path->slots[*level]); |
3548 | blocksize = btrfs_level_size(root, *level - 1); | 4084 | blocksize = btrfs_level_size(root, *level - 1); |
3549 | 4085 | ||
3550 | ret = drop_snap_lookup_refcount(root, bytenr, blocksize, &refs); | 4086 | ret = drop_snap_lookup_refcount(root, bytenr, blocksize, &refs); |
3551 | BUG_ON(ret); | 4087 | BUG_ON(ret); |
4088 | |||
4089 | /* | ||
4090 | * if there is more than one reference, we don't need | ||
4091 | * to read that node to drop any references it has. We | ||
4092 | * just drop the ref we hold on that node and move on to the | ||
4093 | * next slot in this level. | ||
4094 | */ | ||
3552 | if (refs != 1) { | 4095 | if (refs != 1) { |
3553 | parent = path->nodes[*level]; | 4096 | parent = path->nodes[*level]; |
3554 | root_owner = btrfs_header_owner(parent); | 4097 | root_owner = btrfs_header_owner(parent); |
@@ -3567,46 +4110,12 @@ static noinline int walk_down_tree(struct btrfs_trans_handle *trans, | |||
3567 | 4110 | ||
3568 | continue; | 4111 | continue; |
3569 | } | 4112 | } |
4113 | |||
3570 | /* | 4114 | /* |
3571 | * at this point, we have a single ref, and since the | 4115 | * we need to keep freeing things in the next level down. |
3572 | * only place referencing this extent is a dead root | 4116 | * read the block and loop around to process it |
3573 | * the reference count should never go higher. | ||
3574 | * So, we don't need to check it again | ||
3575 | */ | 4117 | */ |
3576 | if (*level == 1) { | 4118 | next = read_tree_block(root, bytenr, blocksize, ptr_gen); |
3577 | ref = btrfs_lookup_leaf_ref(root, bytenr); | ||
3578 | if (ref && ref->generation != ptr_gen) { | ||
3579 | btrfs_free_leaf_ref(root, ref); | ||
3580 | ref = NULL; | ||
3581 | } | ||
3582 | if (ref) { | ||
3583 | ret = cache_drop_leaf_ref(trans, root, ref); | ||
3584 | BUG_ON(ret); | ||
3585 | btrfs_remove_leaf_ref(root, ref); | ||
3586 | btrfs_free_leaf_ref(root, ref); | ||
3587 | *level = 0; | ||
3588 | break; | ||
3589 | } | ||
3590 | } | ||
3591 | next = btrfs_find_tree_block(root, bytenr, blocksize); | ||
3592 | if (!next || !btrfs_buffer_uptodate(next, ptr_gen)) { | ||
3593 | free_extent_buffer(next); | ||
3594 | |||
3595 | next = read_tree_block(root, bytenr, blocksize, | ||
3596 | ptr_gen); | ||
3597 | cond_resched(); | ||
3598 | #if 0 | ||
3599 | /* | ||
3600 | * this is a debugging check and can go away | ||
3601 | * the ref should never go all the way down to 1 | ||
3602 | * at this point | ||
3603 | */ | ||
3604 | ret = lookup_extent_ref(NULL, root, bytenr, blocksize, | ||
3605 | &refs); | ||
3606 | BUG_ON(ret); | ||
3607 | WARN_ON(refs != 1); | ||
3608 | #endif | ||
3609 | } | ||
3610 | WARN_ON(*level <= 0); | 4119 | WARN_ON(*level <= 0); |
3611 | if (path->nodes[*level-1]) | 4120 | if (path->nodes[*level-1]) |
3612 | free_extent_buffer(path->nodes[*level-1]); | 4121 | free_extent_buffer(path->nodes[*level-1]); |
@@ -3631,11 +4140,16 @@ out: | |||
3631 | root_owner = btrfs_header_owner(parent); | 4140 | root_owner = btrfs_header_owner(parent); |
3632 | root_gen = btrfs_header_generation(parent); | 4141 | root_gen = btrfs_header_generation(parent); |
3633 | 4142 | ||
4143 | /* | ||
4144 | * cleanup and free the reference on the last node | ||
4145 | * we processed | ||
4146 | */ | ||
3634 | ret = __btrfs_free_extent(trans, root, bytenr, blocksize, | 4147 | ret = __btrfs_free_extent(trans, root, bytenr, blocksize, |
3635 | parent->start, root_owner, root_gen, | 4148 | parent->start, root_owner, root_gen, |
3636 | *level, 1); | 4149 | *level, 1); |
3637 | free_extent_buffer(path->nodes[*level]); | 4150 | free_extent_buffer(path->nodes[*level]); |
3638 | path->nodes[*level] = NULL; | 4151 | path->nodes[*level] = NULL; |
4152 | |||
3639 | *level += 1; | 4153 | *level += 1; |
3640 | BUG_ON(ret); | 4154 | BUG_ON(ret); |
3641 | 4155 | ||
@@ -3687,6 +4201,7 @@ static noinline int walk_down_subtree(struct btrfs_trans_handle *trans, | |||
3687 | 4201 | ||
3688 | next = read_tree_block(root, bytenr, blocksize, ptr_gen); | 4202 | next = read_tree_block(root, bytenr, blocksize, ptr_gen); |
3689 | btrfs_tree_lock(next); | 4203 | btrfs_tree_lock(next); |
4204 | btrfs_set_lock_blocking(next); | ||
3690 | 4205 | ||
3691 | ret = btrfs_lookup_extent_ref(trans, root, bytenr, blocksize, | 4206 | ret = btrfs_lookup_extent_ref(trans, root, bytenr, blocksize, |
3692 | &refs); | 4207 | &refs); |
@@ -3754,6 +4269,13 @@ static noinline int walk_up_tree(struct btrfs_trans_handle *trans, | |||
3754 | if (slot < btrfs_header_nritems(path->nodes[i]) - 1) { | 4269 | if (slot < btrfs_header_nritems(path->nodes[i]) - 1) { |
3755 | struct extent_buffer *node; | 4270 | struct extent_buffer *node; |
3756 | struct btrfs_disk_key disk_key; | 4271 | struct btrfs_disk_key disk_key; |
4272 | |||
4273 | /* | ||
4274 | * there is more work to do in this level. | ||
4275 | * Update the drop_progress marker to reflect | ||
4276 | * the work we've done so far, and then bump | ||
4277 | * the slot number | ||
4278 | */ | ||
3757 | node = path->nodes[i]; | 4279 | node = path->nodes[i]; |
3758 | path->slots[i]++; | 4280 | path->slots[i]++; |
3759 | *level = i; | 4281 | *level = i; |
@@ -3765,6 +4287,11 @@ static noinline int walk_up_tree(struct btrfs_trans_handle *trans, | |||
3765 | return 0; | 4287 | return 0; |
3766 | } else { | 4288 | } else { |
3767 | struct extent_buffer *parent; | 4289 | struct extent_buffer *parent; |
4290 | |||
4291 | /* | ||
4292 | * this whole node is done, free our reference | ||
4293 | * on it and go up one level | ||
4294 | */ | ||
3768 | if (path->nodes[*level] == root->node) | 4295 | if (path->nodes[*level] == root->node) |
3769 | parent = path->nodes[*level]; | 4296 | parent = path->nodes[*level]; |
3770 | else | 4297 | else |
@@ -4444,7 +4971,7 @@ static noinline int replace_one_extent(struct btrfs_trans_handle *trans, | |||
4444 | u64 lock_end = 0; | 4971 | u64 lock_end = 0; |
4445 | u64 num_bytes; | 4972 | u64 num_bytes; |
4446 | u64 ext_offset; | 4973 | u64 ext_offset; |
4447 | u64 first_pos; | 4974 | u64 search_end = (u64)-1; |
4448 | u32 nritems; | 4975 | u32 nritems; |
4449 | int nr_scaned = 0; | 4976 | int nr_scaned = 0; |
4450 | int extent_locked = 0; | 4977 | int extent_locked = 0; |
@@ -4452,7 +4979,6 @@ static noinline int replace_one_extent(struct btrfs_trans_handle *trans, | |||
4452 | int ret; | 4979 | int ret; |
4453 | 4980 | ||
4454 | memcpy(&key, leaf_key, sizeof(key)); | 4981 | memcpy(&key, leaf_key, sizeof(key)); |
4455 | first_pos = INT_LIMIT(loff_t) - extent_key->offset; | ||
4456 | if (ref_path->owner_objectid != BTRFS_MULTIPLE_OBJECTIDS) { | 4982 | if (ref_path->owner_objectid != BTRFS_MULTIPLE_OBJECTIDS) { |
4457 | if (key.objectid < ref_path->owner_objectid || | 4983 | if (key.objectid < ref_path->owner_objectid || |
4458 | (key.objectid == ref_path->owner_objectid && | 4984 | (key.objectid == ref_path->owner_objectid && |
@@ -4501,7 +5027,7 @@ next: | |||
4501 | if ((key.objectid > ref_path->owner_objectid) || | 5027 | if ((key.objectid > ref_path->owner_objectid) || |
4502 | (key.objectid == ref_path->owner_objectid && | 5028 | (key.objectid == ref_path->owner_objectid && |
4503 | key.type > BTRFS_EXTENT_DATA_KEY) || | 5029 | key.type > BTRFS_EXTENT_DATA_KEY) || |
4504 | (key.offset >= first_pos + extent_key->offset)) | 5030 | key.offset >= search_end) |
4505 | break; | 5031 | break; |
4506 | } | 5032 | } |
4507 | 5033 | ||
@@ -4534,8 +5060,10 @@ next: | |||
4534 | num_bytes = btrfs_file_extent_num_bytes(leaf, fi); | 5060 | num_bytes = btrfs_file_extent_num_bytes(leaf, fi); |
4535 | ext_offset = btrfs_file_extent_offset(leaf, fi); | 5061 | ext_offset = btrfs_file_extent_offset(leaf, fi); |
4536 | 5062 | ||
4537 | if (first_pos > key.offset - ext_offset) | 5063 | if (search_end == (u64)-1) { |
4538 | first_pos = key.offset - ext_offset; | 5064 | search_end = key.offset - ext_offset + |
5065 | btrfs_file_extent_ram_bytes(leaf, fi); | ||
5066 | } | ||
4539 | 5067 | ||
4540 | if (!extent_locked) { | 5068 | if (!extent_locked) { |
4541 | lock_start = key.offset; | 5069 | lock_start = key.offset; |
@@ -4724,7 +5252,7 @@ next: | |||
4724 | } | 5252 | } |
4725 | skip: | 5253 | skip: |
4726 | if (ref_path->owner_objectid != BTRFS_MULTIPLE_OBJECTIDS && | 5254 | if (ref_path->owner_objectid != BTRFS_MULTIPLE_OBJECTIDS && |
4727 | key.offset >= first_pos + extent_key->offset) | 5255 | key.offset >= search_end) |
4728 | break; | 5256 | break; |
4729 | 5257 | ||
4730 | cond_resched(); | 5258 | cond_resched(); |
@@ -4778,6 +5306,7 @@ int btrfs_reloc_tree_cache_ref(struct btrfs_trans_handle *trans, | |||
4778 | ref->bytenr = buf->start; | 5306 | ref->bytenr = buf->start; |
4779 | ref->owner = btrfs_header_owner(buf); | 5307 | ref->owner = btrfs_header_owner(buf); |
4780 | ref->generation = btrfs_header_generation(buf); | 5308 | ref->generation = btrfs_header_generation(buf); |
5309 | |||
4781 | ret = btrfs_add_leaf_ref(root, ref, 0); | 5310 | ret = btrfs_add_leaf_ref(root, ref, 0); |
4782 | WARN_ON(ret); | 5311 | WARN_ON(ret); |
4783 | btrfs_free_leaf_ref(root, ref); | 5312 | btrfs_free_leaf_ref(root, ref); |
@@ -5351,7 +5880,9 @@ static noinline int relocate_one_extent(struct btrfs_root *extent_root, | |||
5351 | prev_block = block_start; | 5880 | prev_block = block_start; |
5352 | } | 5881 | } |
5353 | 5882 | ||
5883 | mutex_lock(&extent_root->fs_info->trans_mutex); | ||
5354 | btrfs_record_root_in_trans(found_root); | 5884 | btrfs_record_root_in_trans(found_root); |
5885 | mutex_unlock(&extent_root->fs_info->trans_mutex); | ||
5355 | if (ref_path->owner_objectid >= BTRFS_FIRST_FREE_OBJECTID) { | 5886 | if (ref_path->owner_objectid >= BTRFS_FIRST_FREE_OBJECTID) { |
5356 | /* | 5887 | /* |
5357 | * try to update data extent references while | 5888 | * try to update data extent references while |
@@ -5957,9 +6488,11 @@ int btrfs_remove_block_group(struct btrfs_trans_handle *trans, | |||
5957 | path = btrfs_alloc_path(); | 6488 | path = btrfs_alloc_path(); |
5958 | BUG_ON(!path); | 6489 | BUG_ON(!path); |
5959 | 6490 | ||
5960 | btrfs_remove_free_space_cache(block_group); | 6491 | spin_lock(&root->fs_info->block_group_cache_lock); |
5961 | rb_erase(&block_group->cache_node, | 6492 | rb_erase(&block_group->cache_node, |
5962 | &root->fs_info->block_group_cache_tree); | 6493 | &root->fs_info->block_group_cache_tree); |
6494 | spin_unlock(&root->fs_info->block_group_cache_lock); | ||
6495 | btrfs_remove_free_space_cache(block_group); | ||
5963 | down_write(&block_group->space_info->groups_sem); | 6496 | down_write(&block_group->space_info->groups_sem); |
5964 | list_del(&block_group->list); | 6497 | list_del(&block_group->list); |
5965 | up_write(&block_group->space_info->groups_sem); | 6498 | up_write(&block_group->space_info->groups_sem); |
diff --git a/fs/btrfs/extent_io.c b/fs/btrfs/extent_io.c index e086d407f1fa..ebe6b29e6069 100644 --- a/fs/btrfs/extent_io.c +++ b/fs/btrfs/extent_io.c | |||
@@ -9,7 +9,6 @@ | |||
9 | #include <linux/spinlock.h> | 9 | #include <linux/spinlock.h> |
10 | #include <linux/blkdev.h> | 10 | #include <linux/blkdev.h> |
11 | #include <linux/swap.h> | 11 | #include <linux/swap.h> |
12 | #include <linux/version.h> | ||
13 | #include <linux/writeback.h> | 12 | #include <linux/writeback.h> |
14 | #include <linux/pagevec.h> | 13 | #include <linux/pagevec.h> |
15 | #include "extent_io.h" | 14 | #include "extent_io.h" |
@@ -31,7 +30,7 @@ static LIST_HEAD(buffers); | |||
31 | static LIST_HEAD(states); | 30 | static LIST_HEAD(states); |
32 | 31 | ||
33 | #define LEAK_DEBUG 0 | 32 | #define LEAK_DEBUG 0 |
34 | #ifdef LEAK_DEBUG | 33 | #if LEAK_DEBUG |
35 | static DEFINE_SPINLOCK(leak_lock); | 34 | static DEFINE_SPINLOCK(leak_lock); |
36 | #endif | 35 | #endif |
37 | 36 | ||
@@ -120,7 +119,7 @@ void extent_io_tree_init(struct extent_io_tree *tree, | |||
120 | static struct extent_state *alloc_extent_state(gfp_t mask) | 119 | static struct extent_state *alloc_extent_state(gfp_t mask) |
121 | { | 120 | { |
122 | struct extent_state *state; | 121 | struct extent_state *state; |
123 | #ifdef LEAK_DEBUG | 122 | #if LEAK_DEBUG |
124 | unsigned long flags; | 123 | unsigned long flags; |
125 | #endif | 124 | #endif |
126 | 125 | ||
@@ -130,7 +129,7 @@ static struct extent_state *alloc_extent_state(gfp_t mask) | |||
130 | state->state = 0; | 129 | state->state = 0; |
131 | state->private = 0; | 130 | state->private = 0; |
132 | state->tree = NULL; | 131 | state->tree = NULL; |
133 | #ifdef LEAK_DEBUG | 132 | #if LEAK_DEBUG |
134 | spin_lock_irqsave(&leak_lock, flags); | 133 | spin_lock_irqsave(&leak_lock, flags); |
135 | list_add(&state->leak_list, &states); | 134 | list_add(&state->leak_list, &states); |
136 | spin_unlock_irqrestore(&leak_lock, flags); | 135 | spin_unlock_irqrestore(&leak_lock, flags); |
@@ -145,11 +144,11 @@ static void free_extent_state(struct extent_state *state) | |||
145 | if (!state) | 144 | if (!state) |
146 | return; | 145 | return; |
147 | if (atomic_dec_and_test(&state->refs)) { | 146 | if (atomic_dec_and_test(&state->refs)) { |
148 | #ifdef LEAK_DEBUG | 147 | #if LEAK_DEBUG |
149 | unsigned long flags; | 148 | unsigned long flags; |
150 | #endif | 149 | #endif |
151 | WARN_ON(state->tree); | 150 | WARN_ON(state->tree); |
152 | #ifdef LEAK_DEBUG | 151 | #if LEAK_DEBUG |
153 | spin_lock_irqsave(&leak_lock, flags); | 152 | spin_lock_irqsave(&leak_lock, flags); |
154 | list_del(&state->leak_list); | 153 | list_del(&state->leak_list); |
155 | spin_unlock_irqrestore(&leak_lock, flags); | 154 | spin_unlock_irqrestore(&leak_lock, flags); |
@@ -416,8 +415,6 @@ static int split_state(struct extent_io_tree *tree, struct extent_state *orig, | |||
416 | 415 | ||
417 | node = tree_insert(&tree->state, prealloc->end, &prealloc->rb_node); | 416 | node = tree_insert(&tree->state, prealloc->end, &prealloc->rb_node); |
418 | if (node) { | 417 | if (node) { |
419 | struct extent_state *found; | ||
420 | found = rb_entry(node, struct extent_state, rb_node); | ||
421 | free_extent_state(prealloc); | 418 | free_extent_state(prealloc); |
422 | return -EEXIST; | 419 | return -EEXIST; |
423 | } | 420 | } |
@@ -2378,11 +2375,6 @@ static int extent_write_cache_pages(struct extent_io_tree *tree, | |||
2378 | int scanned = 0; | 2375 | int scanned = 0; |
2379 | int range_whole = 0; | 2376 | int range_whole = 0; |
2380 | 2377 | ||
2381 | if (wbc->nonblocking && bdi_write_congested(bdi)) { | ||
2382 | wbc->encountered_congestion = 1; | ||
2383 | return 0; | ||
2384 | } | ||
2385 | |||
2386 | pagevec_init(&pvec, 0); | 2378 | pagevec_init(&pvec, 0); |
2387 | if (wbc->range_cyclic) { | 2379 | if (wbc->range_cyclic) { |
2388 | index = mapping->writeback_index; /* Start from prev offset */ | 2380 | index = mapping->writeback_index; /* Start from prev offset */ |
@@ -2855,6 +2847,98 @@ out: | |||
2855 | return sector; | 2847 | return sector; |
2856 | } | 2848 | } |
2857 | 2849 | ||
2850 | int extent_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, | ||
2851 | __u64 start, __u64 len, get_extent_t *get_extent) | ||
2852 | { | ||
2853 | int ret; | ||
2854 | u64 off = start; | ||
2855 | u64 max = start + len; | ||
2856 | u32 flags = 0; | ||
2857 | u64 disko = 0; | ||
2858 | struct extent_map *em = NULL; | ||
2859 | int end = 0; | ||
2860 | u64 em_start = 0, em_len = 0; | ||
2861 | unsigned long emflags; | ||
2862 | ret = 0; | ||
2863 | |||
2864 | if (len == 0) | ||
2865 | return -EINVAL; | ||
2866 | |||
2867 | lock_extent(&BTRFS_I(inode)->io_tree, start, start + len, | ||
2868 | GFP_NOFS); | ||
2869 | em = get_extent(inode, NULL, 0, off, max - off, 0); | ||
2870 | if (!em) | ||
2871 | goto out; | ||
2872 | if (IS_ERR(em)) { | ||
2873 | ret = PTR_ERR(em); | ||
2874 | goto out; | ||
2875 | } | ||
2876 | while (!end) { | ||
2877 | off = em->start + em->len; | ||
2878 | if (off >= max) | ||
2879 | end = 1; | ||
2880 | |||
2881 | em_start = em->start; | ||
2882 | em_len = em->len; | ||
2883 | |||
2884 | disko = 0; | ||
2885 | flags = 0; | ||
2886 | |||
2887 | switch (em->block_start) { | ||
2888 | case EXTENT_MAP_LAST_BYTE: | ||
2889 | end = 1; | ||
2890 | flags |= FIEMAP_EXTENT_LAST; | ||
2891 | break; | ||
2892 | case EXTENT_MAP_HOLE: | ||
2893 | flags |= FIEMAP_EXTENT_UNWRITTEN; | ||
2894 | break; | ||
2895 | case EXTENT_MAP_INLINE: | ||
2896 | flags |= (FIEMAP_EXTENT_DATA_INLINE | | ||
2897 | FIEMAP_EXTENT_NOT_ALIGNED); | ||
2898 | break; | ||
2899 | case EXTENT_MAP_DELALLOC: | ||
2900 | flags |= (FIEMAP_EXTENT_DELALLOC | | ||
2901 | FIEMAP_EXTENT_UNKNOWN); | ||
2902 | break; | ||
2903 | default: | ||
2904 | disko = em->block_start; | ||
2905 | break; | ||
2906 | } | ||
2907 | if (test_bit(EXTENT_FLAG_COMPRESSED, &em->flags)) | ||
2908 | flags |= FIEMAP_EXTENT_ENCODED; | ||
2909 | |||
2910 | emflags = em->flags; | ||
2911 | free_extent_map(em); | ||
2912 | em = NULL; | ||
2913 | |||
2914 | if (!end) { | ||
2915 | em = get_extent(inode, NULL, 0, off, max - off, 0); | ||
2916 | if (!em) | ||
2917 | goto out; | ||
2918 | if (IS_ERR(em)) { | ||
2919 | ret = PTR_ERR(em); | ||
2920 | goto out; | ||
2921 | } | ||
2922 | emflags = em->flags; | ||
2923 | } | ||
2924 | if (test_bit(EXTENT_FLAG_VACANCY, &emflags)) { | ||
2925 | flags |= FIEMAP_EXTENT_LAST; | ||
2926 | end = 1; | ||
2927 | } | ||
2928 | |||
2929 | ret = fiemap_fill_next_extent(fieinfo, em_start, disko, | ||
2930 | em_len, flags); | ||
2931 | if (ret) | ||
2932 | goto out_free; | ||
2933 | } | ||
2934 | out_free: | ||
2935 | free_extent_map(em); | ||
2936 | out: | ||
2937 | unlock_extent(&BTRFS_I(inode)->io_tree, start, start + len, | ||
2938 | GFP_NOFS); | ||
2939 | return ret; | ||
2940 | } | ||
2941 | |||
2858 | static inline struct page *extent_buffer_page(struct extent_buffer *eb, | 2942 | static inline struct page *extent_buffer_page(struct extent_buffer *eb, |
2859 | unsigned long i) | 2943 | unsigned long i) |
2860 | { | 2944 | { |
@@ -2892,15 +2976,17 @@ static struct extent_buffer *__alloc_extent_buffer(struct extent_io_tree *tree, | |||
2892 | gfp_t mask) | 2976 | gfp_t mask) |
2893 | { | 2977 | { |
2894 | struct extent_buffer *eb = NULL; | 2978 | struct extent_buffer *eb = NULL; |
2895 | #ifdef LEAK_DEBUG | 2979 | #if LEAK_DEBUG |
2896 | unsigned long flags; | 2980 | unsigned long flags; |
2897 | #endif | 2981 | #endif |
2898 | 2982 | ||
2899 | eb = kmem_cache_zalloc(extent_buffer_cache, mask); | 2983 | eb = kmem_cache_zalloc(extent_buffer_cache, mask); |
2900 | eb->start = start; | 2984 | eb->start = start; |
2901 | eb->len = len; | 2985 | eb->len = len; |
2902 | mutex_init(&eb->mutex); | 2986 | spin_lock_init(&eb->lock); |
2903 | #ifdef LEAK_DEBUG | 2987 | init_waitqueue_head(&eb->lock_wq); |
2988 | |||
2989 | #if LEAK_DEBUG | ||
2904 | spin_lock_irqsave(&leak_lock, flags); | 2990 | spin_lock_irqsave(&leak_lock, flags); |
2905 | list_add(&eb->leak_list, &buffers); | 2991 | list_add(&eb->leak_list, &buffers); |
2906 | spin_unlock_irqrestore(&leak_lock, flags); | 2992 | spin_unlock_irqrestore(&leak_lock, flags); |
@@ -2912,7 +2998,7 @@ static struct extent_buffer *__alloc_extent_buffer(struct extent_io_tree *tree, | |||
2912 | 2998 | ||
2913 | static void __free_extent_buffer(struct extent_buffer *eb) | 2999 | static void __free_extent_buffer(struct extent_buffer *eb) |
2914 | { | 3000 | { |
2915 | #ifdef LEAK_DEBUG | 3001 | #if LEAK_DEBUG |
2916 | unsigned long flags; | 3002 | unsigned long flags; |
2917 | spin_lock_irqsave(&leak_lock, flags); | 3003 | spin_lock_irqsave(&leak_lock, flags); |
2918 | list_del(&eb->leak_list); | 3004 | list_del(&eb->leak_list); |
@@ -2980,8 +3066,7 @@ struct extent_buffer *alloc_extent_buffer(struct extent_io_tree *tree, | |||
2980 | unlock_page(p); | 3066 | unlock_page(p); |
2981 | } | 3067 | } |
2982 | if (uptodate) | 3068 | if (uptodate) |
2983 | eb->flags |= EXTENT_UPTODATE; | 3069 | set_bit(EXTENT_BUFFER_UPTODATE, &eb->bflags); |
2984 | eb->flags |= EXTENT_BUFFER_FILLED; | ||
2985 | 3070 | ||
2986 | spin_lock(&tree->buffer_lock); | 3071 | spin_lock(&tree->buffer_lock); |
2987 | exists = buffer_tree_insert(tree, start, &eb->rb_node); | 3072 | exists = buffer_tree_insert(tree, start, &eb->rb_node); |
@@ -3135,7 +3220,7 @@ int clear_extent_buffer_uptodate(struct extent_io_tree *tree, | |||
3135 | unsigned long num_pages; | 3220 | unsigned long num_pages; |
3136 | 3221 | ||
3137 | num_pages = num_extent_pages(eb->start, eb->len); | 3222 | num_pages = num_extent_pages(eb->start, eb->len); |
3138 | eb->flags &= ~EXTENT_UPTODATE; | 3223 | clear_bit(EXTENT_BUFFER_UPTODATE, &eb->bflags); |
3139 | 3224 | ||
3140 | clear_extent_uptodate(tree, eb->start, eb->start + eb->len - 1, | 3225 | clear_extent_uptodate(tree, eb->start, eb->start + eb->len - 1, |
3141 | GFP_NOFS); | 3226 | GFP_NOFS); |
@@ -3206,7 +3291,7 @@ int extent_buffer_uptodate(struct extent_io_tree *tree, | |||
3206 | struct page *page; | 3291 | struct page *page; |
3207 | int pg_uptodate = 1; | 3292 | int pg_uptodate = 1; |
3208 | 3293 | ||
3209 | if (eb->flags & EXTENT_UPTODATE) | 3294 | if (test_bit(EXTENT_BUFFER_UPTODATE, &eb->bflags)) |
3210 | return 1; | 3295 | return 1; |
3211 | 3296 | ||
3212 | ret = test_range_bit(tree, eb->start, eb->start + eb->len - 1, | 3297 | ret = test_range_bit(tree, eb->start, eb->start + eb->len - 1, |
@@ -3242,7 +3327,7 @@ int read_extent_buffer_pages(struct extent_io_tree *tree, | |||
3242 | struct bio *bio = NULL; | 3327 | struct bio *bio = NULL; |
3243 | unsigned long bio_flags = 0; | 3328 | unsigned long bio_flags = 0; |
3244 | 3329 | ||
3245 | if (eb->flags & EXTENT_UPTODATE) | 3330 | if (test_bit(EXTENT_BUFFER_UPTODATE, &eb->bflags)) |
3246 | return 0; | 3331 | return 0; |
3247 | 3332 | ||
3248 | if (test_range_bit(tree, eb->start, eb->start + eb->len - 1, | 3333 | if (test_range_bit(tree, eb->start, eb->start + eb->len - 1, |
@@ -3273,7 +3358,7 @@ int read_extent_buffer_pages(struct extent_io_tree *tree, | |||
3273 | } | 3358 | } |
3274 | if (all_uptodate) { | 3359 | if (all_uptodate) { |
3275 | if (start_i == 0) | 3360 | if (start_i == 0) |
3276 | eb->flags |= EXTENT_UPTODATE; | 3361 | set_bit(EXTENT_BUFFER_UPTODATE, &eb->bflags); |
3277 | goto unlock_exit; | 3362 | goto unlock_exit; |
3278 | } | 3363 | } |
3279 | 3364 | ||
@@ -3309,7 +3394,7 @@ int read_extent_buffer_pages(struct extent_io_tree *tree, | |||
3309 | } | 3394 | } |
3310 | 3395 | ||
3311 | if (!ret) | 3396 | if (!ret) |
3312 | eb->flags |= EXTENT_UPTODATE; | 3397 | set_bit(EXTENT_BUFFER_UPTODATE, &eb->bflags); |
3313 | return ret; | 3398 | return ret; |
3314 | 3399 | ||
3315 | unlock_exit: | 3400 | unlock_exit: |
@@ -3406,7 +3491,6 @@ int map_extent_buffer(struct extent_buffer *eb, unsigned long start, | |||
3406 | unmap_extent_buffer(eb, eb->map_token, km); | 3491 | unmap_extent_buffer(eb, eb->map_token, km); |
3407 | eb->map_token = NULL; | 3492 | eb->map_token = NULL; |
3408 | save = 1; | 3493 | save = 1; |
3409 | WARN_ON(!mutex_is_locked(&eb->mutex)); | ||
3410 | } | 3494 | } |
3411 | err = map_private_extent_buffer(eb, start, min_len, token, map, | 3495 | err = map_private_extent_buffer(eb, start, min_len, token, map, |
3412 | map_start, map_len, km); | 3496 | map_start, map_len, km); |
diff --git a/fs/btrfs/extent_io.h b/fs/btrfs/extent_io.h index c5b483a79137..1f9df88afbf6 100644 --- a/fs/btrfs/extent_io.h +++ b/fs/btrfs/extent_io.h | |||
@@ -22,6 +22,10 @@ | |||
22 | /* flags for bio submission */ | 22 | /* flags for bio submission */ |
23 | #define EXTENT_BIO_COMPRESSED 1 | 23 | #define EXTENT_BIO_COMPRESSED 1 |
24 | 24 | ||
25 | /* these are bit numbers for test/set bit */ | ||
26 | #define EXTENT_BUFFER_UPTODATE 0 | ||
27 | #define EXTENT_BUFFER_BLOCKING 1 | ||
28 | |||
25 | /* | 29 | /* |
26 | * page->private values. Every page that is controlled by the extent | 30 | * page->private values. Every page that is controlled by the extent |
27 | * map has page->private set to one. | 31 | * map has page->private set to one. |
@@ -95,11 +99,19 @@ struct extent_buffer { | |||
95 | unsigned long map_start; | 99 | unsigned long map_start; |
96 | unsigned long map_len; | 100 | unsigned long map_len; |
97 | struct page *first_page; | 101 | struct page *first_page; |
102 | unsigned long bflags; | ||
98 | atomic_t refs; | 103 | atomic_t refs; |
99 | int flags; | ||
100 | struct list_head leak_list; | 104 | struct list_head leak_list; |
101 | struct rb_node rb_node; | 105 | struct rb_node rb_node; |
102 | struct mutex mutex; | 106 | |
107 | /* the spinlock is used to protect most operations */ | ||
108 | spinlock_t lock; | ||
109 | |||
110 | /* | ||
111 | * when we keep the lock held while blocking, waiters go onto | ||
112 | * the wq | ||
113 | */ | ||
114 | wait_queue_head_t lock_wq; | ||
103 | }; | 115 | }; |
104 | 116 | ||
105 | struct extent_map_tree; | 117 | struct extent_map_tree; |
@@ -193,6 +205,8 @@ int extent_commit_write(struct extent_io_tree *tree, | |||
193 | unsigned from, unsigned to); | 205 | unsigned from, unsigned to); |
194 | sector_t extent_bmap(struct address_space *mapping, sector_t iblock, | 206 | sector_t extent_bmap(struct address_space *mapping, sector_t iblock, |
195 | get_extent_t *get_extent); | 207 | get_extent_t *get_extent); |
208 | int extent_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, | ||
209 | __u64 start, __u64 len, get_extent_t *get_extent); | ||
196 | int set_range_dirty(struct extent_io_tree *tree, u64 start, u64 end); | 210 | int set_range_dirty(struct extent_io_tree *tree, u64 start, u64 end); |
197 | int set_state_private(struct extent_io_tree *tree, u64 start, u64 private); | 211 | int set_state_private(struct extent_io_tree *tree, u64 start, u64 private); |
198 | int get_state_private(struct extent_io_tree *tree, u64 start, u64 *private); | 212 | int get_state_private(struct extent_io_tree *tree, u64 start, u64 *private); |
diff --git a/fs/btrfs/extent_map.c b/fs/btrfs/extent_map.c index 4a83e33ada32..50da69da20ce 100644 --- a/fs/btrfs/extent_map.c +++ b/fs/btrfs/extent_map.c | |||
@@ -3,7 +3,6 @@ | |||
3 | #include <linux/slab.h> | 3 | #include <linux/slab.h> |
4 | #include <linux/module.h> | 4 | #include <linux/module.h> |
5 | #include <linux/spinlock.h> | 5 | #include <linux/spinlock.h> |
6 | #include <linux/version.h> | ||
7 | #include <linux/hardirq.h> | 6 | #include <linux/hardirq.h> |
8 | #include "extent_map.h" | 7 | #include "extent_map.h" |
9 | 8 | ||
diff --git a/fs/btrfs/file.c b/fs/btrfs/file.c index 90268334145e..dc78954861b3 100644 --- a/fs/btrfs/file.c +++ b/fs/btrfs/file.c | |||
@@ -29,7 +29,6 @@ | |||
29 | #include <linux/writeback.h> | 29 | #include <linux/writeback.h> |
30 | #include <linux/statfs.h> | 30 | #include <linux/statfs.h> |
31 | #include <linux/compat.h> | 31 | #include <linux/compat.h> |
32 | #include <linux/version.h> | ||
33 | #include "ctree.h" | 32 | #include "ctree.h" |
34 | #include "disk-io.h" | 33 | #include "disk-io.h" |
35 | #include "transaction.h" | 34 | #include "transaction.h" |
@@ -1092,19 +1091,24 @@ static ssize_t btrfs_file_write(struct file *file, const char __user *buf, | |||
1092 | WARN_ON(num_pages > nrptrs); | 1091 | WARN_ON(num_pages > nrptrs); |
1093 | memset(pages, 0, sizeof(struct page *) * nrptrs); | 1092 | memset(pages, 0, sizeof(struct page *) * nrptrs); |
1094 | 1093 | ||
1095 | ret = btrfs_check_free_space(root, write_bytes, 0); | 1094 | ret = btrfs_check_data_free_space(root, inode, write_bytes); |
1096 | if (ret) | 1095 | if (ret) |
1097 | goto out; | 1096 | goto out; |
1098 | 1097 | ||
1099 | ret = prepare_pages(root, file, pages, num_pages, | 1098 | ret = prepare_pages(root, file, pages, num_pages, |
1100 | pos, first_index, last_index, | 1099 | pos, first_index, last_index, |
1101 | write_bytes); | 1100 | write_bytes); |
1102 | if (ret) | 1101 | if (ret) { |
1102 | btrfs_free_reserved_data_space(root, inode, | ||
1103 | write_bytes); | ||
1103 | goto out; | 1104 | goto out; |
1105 | } | ||
1104 | 1106 | ||
1105 | ret = btrfs_copy_from_user(pos, num_pages, | 1107 | ret = btrfs_copy_from_user(pos, num_pages, |
1106 | write_bytes, pages, buf); | 1108 | write_bytes, pages, buf); |
1107 | if (ret) { | 1109 | if (ret) { |
1110 | btrfs_free_reserved_data_space(root, inode, | ||
1111 | write_bytes); | ||
1108 | btrfs_drop_pages(pages, num_pages); | 1112 | btrfs_drop_pages(pages, num_pages); |
1109 | goto out; | 1113 | goto out; |
1110 | } | 1114 | } |
@@ -1112,8 +1116,11 @@ static ssize_t btrfs_file_write(struct file *file, const char __user *buf, | |||
1112 | ret = dirty_and_release_pages(NULL, root, file, pages, | 1116 | ret = dirty_and_release_pages(NULL, root, file, pages, |
1113 | num_pages, pos, write_bytes); | 1117 | num_pages, pos, write_bytes); |
1114 | btrfs_drop_pages(pages, num_pages); | 1118 | btrfs_drop_pages(pages, num_pages); |
1115 | if (ret) | 1119 | if (ret) { |
1120 | btrfs_free_reserved_data_space(root, inode, | ||
1121 | write_bytes); | ||
1116 | goto out; | 1122 | goto out; |
1123 | } | ||
1117 | 1124 | ||
1118 | if (will_write) { | 1125 | if (will_write) { |
1119 | btrfs_fdatawrite_range(inode->i_mapping, pos, | 1126 | btrfs_fdatawrite_range(inode->i_mapping, pos, |
@@ -1137,6 +1144,8 @@ static ssize_t btrfs_file_write(struct file *file, const char __user *buf, | |||
1137 | } | 1144 | } |
1138 | out: | 1145 | out: |
1139 | mutex_unlock(&inode->i_mutex); | 1146 | mutex_unlock(&inode->i_mutex); |
1147 | if (ret) | ||
1148 | err = ret; | ||
1140 | 1149 | ||
1141 | out_nolock: | 1150 | out_nolock: |
1142 | kfree(pages); | 1151 | kfree(pages); |
@@ -1215,15 +1224,15 @@ int btrfs_sync_file(struct file *file, struct dentry *dentry, int datasync) | |||
1215 | } | 1224 | } |
1216 | mutex_unlock(&root->fs_info->trans_mutex); | 1225 | mutex_unlock(&root->fs_info->trans_mutex); |
1217 | 1226 | ||
1218 | root->fs_info->tree_log_batch++; | 1227 | root->log_batch++; |
1219 | filemap_fdatawrite(inode->i_mapping); | 1228 | filemap_fdatawrite(inode->i_mapping); |
1220 | btrfs_wait_ordered_range(inode, 0, (u64)-1); | 1229 | btrfs_wait_ordered_range(inode, 0, (u64)-1); |
1221 | root->fs_info->tree_log_batch++; | 1230 | root->log_batch++; |
1222 | 1231 | ||
1223 | /* | 1232 | /* |
1224 | * ok we haven't committed the transaction yet, lets do a commit | 1233 | * ok we haven't committed the transaction yet, lets do a commit |
1225 | */ | 1234 | */ |
1226 | if (file->private_data) | 1235 | if (file && file->private_data) |
1227 | btrfs_ioctl_trans_end(file); | 1236 | btrfs_ioctl_trans_end(file); |
1228 | 1237 | ||
1229 | trans = btrfs_start_transaction(root, 1); | 1238 | trans = btrfs_start_transaction(root, 1); |
@@ -1232,7 +1241,7 @@ int btrfs_sync_file(struct file *file, struct dentry *dentry, int datasync) | |||
1232 | goto out; | 1241 | goto out; |
1233 | } | 1242 | } |
1234 | 1243 | ||
1235 | ret = btrfs_log_dentry_safe(trans, root, file->f_dentry); | 1244 | ret = btrfs_log_dentry_safe(trans, root, dentry); |
1236 | if (ret < 0) | 1245 | if (ret < 0) |
1237 | goto out; | 1246 | goto out; |
1238 | 1247 | ||
@@ -1246,7 +1255,7 @@ int btrfs_sync_file(struct file *file, struct dentry *dentry, int datasync) | |||
1246 | * file again, but that will end up using the synchronization | 1255 | * file again, but that will end up using the synchronization |
1247 | * inside btrfs_sync_log to keep things safe. | 1256 | * inside btrfs_sync_log to keep things safe. |
1248 | */ | 1257 | */ |
1249 | mutex_unlock(&file->f_dentry->d_inode->i_mutex); | 1258 | mutex_unlock(&dentry->d_inode->i_mutex); |
1250 | 1259 | ||
1251 | if (ret > 0) { | 1260 | if (ret > 0) { |
1252 | ret = btrfs_commit_transaction(trans, root); | 1261 | ret = btrfs_commit_transaction(trans, root); |
@@ -1254,7 +1263,7 @@ int btrfs_sync_file(struct file *file, struct dentry *dentry, int datasync) | |||
1254 | btrfs_sync_log(trans, root); | 1263 | btrfs_sync_log(trans, root); |
1255 | ret = btrfs_end_transaction(trans, root); | 1264 | ret = btrfs_end_transaction(trans, root); |
1256 | } | 1265 | } |
1257 | mutex_lock(&file->f_dentry->d_inode->i_mutex); | 1266 | mutex_lock(&dentry->d_inode->i_mutex); |
1258 | out: | 1267 | out: |
1259 | return ret > 0 ? EIO : ret; | 1268 | return ret > 0 ? EIO : ret; |
1260 | } | 1269 | } |
diff --git a/fs/btrfs/inode-map.c b/fs/btrfs/inode-map.c index 2aa79873eb46..cc7334d833c9 100644 --- a/fs/btrfs/inode-map.c +++ b/fs/btrfs/inode-map.c | |||
@@ -84,7 +84,6 @@ int btrfs_find_free_objectid(struct btrfs_trans_handle *trans, | |||
84 | search_key.type = 0; | 84 | search_key.type = 0; |
85 | search_key.offset = 0; | 85 | search_key.offset = 0; |
86 | 86 | ||
87 | btrfs_init_path(path); | ||
88 | start_found = 0; | 87 | start_found = 0; |
89 | ret = btrfs_search_slot(trans, root, &search_key, path, 0, 0); | 88 | ret = btrfs_search_slot(trans, root, &search_key, path, 0, 0); |
90 | if (ret < 0) | 89 | if (ret < 0) |
diff --git a/fs/btrfs/inode.c b/fs/btrfs/inode.c index 8adfe059ab41..7d4f948bc22a 100644 --- a/fs/btrfs/inode.c +++ b/fs/btrfs/inode.c | |||
@@ -34,7 +34,6 @@ | |||
34 | #include <linux/statfs.h> | 34 | #include <linux/statfs.h> |
35 | #include <linux/compat.h> | 35 | #include <linux/compat.h> |
36 | #include <linux/bit_spinlock.h> | 36 | #include <linux/bit_spinlock.h> |
37 | #include <linux/version.h> | ||
38 | #include <linux/xattr.h> | 37 | #include <linux/xattr.h> |
39 | #include <linux/posix_acl.h> | 38 | #include <linux/posix_acl.h> |
40 | #include <linux/falloc.h> | 39 | #include <linux/falloc.h> |
@@ -51,6 +50,7 @@ | |||
51 | #include "tree-log.h" | 50 | #include "tree-log.h" |
52 | #include "ref-cache.h" | 51 | #include "ref-cache.h" |
53 | #include "compression.h" | 52 | #include "compression.h" |
53 | #include "locking.h" | ||
54 | 54 | ||
55 | struct btrfs_iget_args { | 55 | struct btrfs_iget_args { |
56 | u64 ino; | 56 | u64 ino; |
@@ -91,32 +91,14 @@ static noinline int cow_file_range(struct inode *inode, | |||
91 | u64 start, u64 end, int *page_started, | 91 | u64 start, u64 end, int *page_started, |
92 | unsigned long *nr_written, int unlock); | 92 | unsigned long *nr_written, int unlock); |
93 | 93 | ||
94 | /* | 94 | static int btrfs_init_inode_security(struct inode *inode, struct inode *dir) |
95 | * a very lame attempt at stopping writes when the FS is 85% full. There | ||
96 | * are countless ways this is incorrect, but it is better than nothing. | ||
97 | */ | ||
98 | int btrfs_check_free_space(struct btrfs_root *root, u64 num_required, | ||
99 | int for_del) | ||
100 | { | 95 | { |
101 | u64 total; | 96 | int err; |
102 | u64 used; | ||
103 | u64 thresh; | ||
104 | int ret = 0; | ||
105 | |||
106 | spin_lock(&root->fs_info->delalloc_lock); | ||
107 | total = btrfs_super_total_bytes(&root->fs_info->super_copy); | ||
108 | used = btrfs_super_bytes_used(&root->fs_info->super_copy); | ||
109 | if (for_del) | ||
110 | thresh = total * 90; | ||
111 | else | ||
112 | thresh = total * 85; | ||
113 | |||
114 | do_div(thresh, 100); | ||
115 | 97 | ||
116 | if (used + root->fs_info->delalloc_bytes + num_required > thresh) | 98 | err = btrfs_init_acl(inode, dir); |
117 | ret = -ENOSPC; | 99 | if (!err) |
118 | spin_unlock(&root->fs_info->delalloc_lock); | 100 | err = btrfs_xattr_security_init(inode, dir); |
119 | return ret; | 101 | return err; |
120 | } | 102 | } |
121 | 103 | ||
122 | /* | 104 | /* |
@@ -350,6 +332,19 @@ again: | |||
350 | nr_pages = (end >> PAGE_CACHE_SHIFT) - (start >> PAGE_CACHE_SHIFT) + 1; | 332 | nr_pages = (end >> PAGE_CACHE_SHIFT) - (start >> PAGE_CACHE_SHIFT) + 1; |
351 | nr_pages = min(nr_pages, (128 * 1024UL) / PAGE_CACHE_SIZE); | 333 | nr_pages = min(nr_pages, (128 * 1024UL) / PAGE_CACHE_SIZE); |
352 | 334 | ||
335 | /* | ||
336 | * we don't want to send crud past the end of i_size through | ||
337 | * compression, that's just a waste of CPU time. So, if the | ||
338 | * end of the file is before the start of our current | ||
339 | * requested range of bytes, we bail out to the uncompressed | ||
340 | * cleanup code that can deal with all of this. | ||
341 | * | ||
342 | * It isn't really the fastest way to fix things, but this is a | ||
343 | * very uncommon corner. | ||
344 | */ | ||
345 | if (actual_end <= start) | ||
346 | goto cleanup_and_bail_uncompressed; | ||
347 | |||
353 | total_compressed = actual_end - start; | 348 | total_compressed = actual_end - start; |
354 | 349 | ||
355 | /* we want to make sure that amount of ram required to uncompress | 350 | /* we want to make sure that amount of ram required to uncompress |
@@ -494,6 +489,7 @@ again: | |||
494 | goto again; | 489 | goto again; |
495 | } | 490 | } |
496 | } else { | 491 | } else { |
492 | cleanup_and_bail_uncompressed: | ||
497 | /* | 493 | /* |
498 | * No compression, but we still need to write the pages in | 494 | * No compression, but we still need to write the pages in |
499 | * the file we've been given so far. redirty the locked | 495 | * the file we've been given so far. redirty the locked |
@@ -1166,6 +1162,7 @@ static int btrfs_set_bit_hook(struct inode *inode, u64 start, u64 end, | |||
1166 | */ | 1162 | */ |
1167 | if (!(old & EXTENT_DELALLOC) && (bits & EXTENT_DELALLOC)) { | 1163 | if (!(old & EXTENT_DELALLOC) && (bits & EXTENT_DELALLOC)) { |
1168 | struct btrfs_root *root = BTRFS_I(inode)->root; | 1164 | struct btrfs_root *root = BTRFS_I(inode)->root; |
1165 | btrfs_delalloc_reserve_space(root, inode, end - start + 1); | ||
1169 | spin_lock(&root->fs_info->delalloc_lock); | 1166 | spin_lock(&root->fs_info->delalloc_lock); |
1170 | BTRFS_I(inode)->delalloc_bytes += end - start + 1; | 1167 | BTRFS_I(inode)->delalloc_bytes += end - start + 1; |
1171 | root->fs_info->delalloc_bytes += end - start + 1; | 1168 | root->fs_info->delalloc_bytes += end - start + 1; |
@@ -1199,9 +1196,12 @@ static int btrfs_clear_bit_hook(struct inode *inode, u64 start, u64 end, | |||
1199 | (unsigned long long)end - start + 1, | 1196 | (unsigned long long)end - start + 1, |
1200 | (unsigned long long) | 1197 | (unsigned long long) |
1201 | root->fs_info->delalloc_bytes); | 1198 | root->fs_info->delalloc_bytes); |
1199 | btrfs_delalloc_free_space(root, inode, (u64)-1); | ||
1202 | root->fs_info->delalloc_bytes = 0; | 1200 | root->fs_info->delalloc_bytes = 0; |
1203 | BTRFS_I(inode)->delalloc_bytes = 0; | 1201 | BTRFS_I(inode)->delalloc_bytes = 0; |
1204 | } else { | 1202 | } else { |
1203 | btrfs_delalloc_free_space(root, inode, | ||
1204 | end - start + 1); | ||
1205 | root->fs_info->delalloc_bytes -= end - start + 1; | 1205 | root->fs_info->delalloc_bytes -= end - start + 1; |
1206 | BTRFS_I(inode)->delalloc_bytes -= end - start + 1; | 1206 | BTRFS_I(inode)->delalloc_bytes -= end - start + 1; |
1207 | } | 1207 | } |
@@ -1324,12 +1324,11 @@ static noinline int add_pending_csums(struct btrfs_trans_handle *trans, | |||
1324 | struct inode *inode, u64 file_offset, | 1324 | struct inode *inode, u64 file_offset, |
1325 | struct list_head *list) | 1325 | struct list_head *list) |
1326 | { | 1326 | { |
1327 | struct list_head *cur; | ||
1328 | struct btrfs_ordered_sum *sum; | 1327 | struct btrfs_ordered_sum *sum; |
1329 | 1328 | ||
1330 | btrfs_set_trans_block_group(trans, inode); | 1329 | btrfs_set_trans_block_group(trans, inode); |
1331 | list_for_each(cur, list) { | 1330 | |
1332 | sum = list_entry(cur, struct btrfs_ordered_sum, list); | 1331 | list_for_each_entry(sum, list, list) { |
1333 | btrfs_csum_file_blocks(trans, | 1332 | btrfs_csum_file_blocks(trans, |
1334 | BTRFS_I(inode)->root->fs_info->csum_root, sum); | 1333 | BTRFS_I(inode)->root->fs_info->csum_root, sum); |
1335 | } | 1334 | } |
@@ -2013,6 +2012,7 @@ void btrfs_read_locked_inode(struct inode *inode) | |||
2013 | BTRFS_I(inode)->flags = btrfs_inode_flags(leaf, inode_item); | 2012 | BTRFS_I(inode)->flags = btrfs_inode_flags(leaf, inode_item); |
2014 | 2013 | ||
2015 | alloc_group_block = btrfs_inode_block_group(leaf, inode_item); | 2014 | alloc_group_block = btrfs_inode_block_group(leaf, inode_item); |
2015 | |||
2016 | BTRFS_I(inode)->block_group = btrfs_find_block_group(root, 0, | 2016 | BTRFS_I(inode)->block_group = btrfs_find_block_group(root, 0, |
2017 | alloc_group_block, 0); | 2017 | alloc_group_block, 0); |
2018 | btrfs_free_path(path); | 2018 | btrfs_free_path(path); |
@@ -2039,6 +2039,7 @@ void btrfs_read_locked_inode(struct inode *inode) | |||
2039 | inode->i_mapping->backing_dev_info = &root->fs_info->bdi; | 2039 | inode->i_mapping->backing_dev_info = &root->fs_info->bdi; |
2040 | break; | 2040 | break; |
2041 | default: | 2041 | default: |
2042 | inode->i_op = &btrfs_special_inode_operations; | ||
2042 | init_special_inode(inode, inode->i_mode, rdev); | 2043 | init_special_inode(inode, inode->i_mode, rdev); |
2043 | break; | 2044 | break; |
2044 | } | 2045 | } |
@@ -2108,6 +2109,7 @@ noinline int btrfs_update_inode(struct btrfs_trans_handle *trans, | |||
2108 | goto failed; | 2109 | goto failed; |
2109 | } | 2110 | } |
2110 | 2111 | ||
2112 | btrfs_unlock_up_safe(path, 1); | ||
2111 | leaf = path->nodes[0]; | 2113 | leaf = path->nodes[0]; |
2112 | inode_item = btrfs_item_ptr(leaf, path->slots[0], | 2114 | inode_item = btrfs_item_ptr(leaf, path->slots[0], |
2113 | struct btrfs_inode_item); | 2115 | struct btrfs_inode_item); |
@@ -2219,10 +2221,6 @@ static int btrfs_unlink(struct inode *dir, struct dentry *dentry) | |||
2219 | 2221 | ||
2220 | root = BTRFS_I(dir)->root; | 2222 | root = BTRFS_I(dir)->root; |
2221 | 2223 | ||
2222 | ret = btrfs_check_free_space(root, 1, 1); | ||
2223 | if (ret) | ||
2224 | goto fail; | ||
2225 | |||
2226 | trans = btrfs_start_transaction(root, 1); | 2224 | trans = btrfs_start_transaction(root, 1); |
2227 | 2225 | ||
2228 | btrfs_set_trans_block_group(trans, dir); | 2226 | btrfs_set_trans_block_group(trans, dir); |
@@ -2235,7 +2233,6 @@ static int btrfs_unlink(struct inode *dir, struct dentry *dentry) | |||
2235 | nr = trans->blocks_used; | 2233 | nr = trans->blocks_used; |
2236 | 2234 | ||
2237 | btrfs_end_transaction_throttle(trans, root); | 2235 | btrfs_end_transaction_throttle(trans, root); |
2238 | fail: | ||
2239 | btrfs_btree_balance_dirty(root, nr); | 2236 | btrfs_btree_balance_dirty(root, nr); |
2240 | return ret; | 2237 | return ret; |
2241 | } | 2238 | } |
@@ -2258,10 +2255,6 @@ static int btrfs_rmdir(struct inode *dir, struct dentry *dentry) | |||
2258 | return -ENOTEMPTY; | 2255 | return -ENOTEMPTY; |
2259 | } | 2256 | } |
2260 | 2257 | ||
2261 | ret = btrfs_check_free_space(root, 1, 1); | ||
2262 | if (ret) | ||
2263 | goto fail; | ||
2264 | |||
2265 | trans = btrfs_start_transaction(root, 1); | 2258 | trans = btrfs_start_transaction(root, 1); |
2266 | btrfs_set_trans_block_group(trans, dir); | 2259 | btrfs_set_trans_block_group(trans, dir); |
2267 | 2260 | ||
@@ -2278,7 +2271,6 @@ static int btrfs_rmdir(struct inode *dir, struct dentry *dentry) | |||
2278 | fail_trans: | 2271 | fail_trans: |
2279 | nr = trans->blocks_used; | 2272 | nr = trans->blocks_used; |
2280 | ret = btrfs_end_transaction_throttle(trans, root); | 2273 | ret = btrfs_end_transaction_throttle(trans, root); |
2281 | fail: | ||
2282 | btrfs_btree_balance_dirty(root, nr); | 2274 | btrfs_btree_balance_dirty(root, nr); |
2283 | 2275 | ||
2284 | if (ret && !err) | 2276 | if (ret && !err) |
@@ -2429,6 +2421,8 @@ next_node: | |||
2429 | ref->generation = leaf_gen; | 2421 | ref->generation = leaf_gen; |
2430 | ref->nritems = 0; | 2422 | ref->nritems = 0; |
2431 | 2423 | ||
2424 | btrfs_sort_leaf_ref(ref); | ||
2425 | |||
2432 | ret = btrfs_add_leaf_ref(root, ref, 0); | 2426 | ret = btrfs_add_leaf_ref(root, ref, 0); |
2433 | WARN_ON(ret); | 2427 | WARN_ON(ret); |
2434 | btrfs_free_leaf_ref(root, ref); | 2428 | btrfs_free_leaf_ref(root, ref); |
@@ -2476,7 +2470,7 @@ noinline int btrfs_truncate_inode_items(struct btrfs_trans_handle *trans, | |||
2476 | struct btrfs_path *path; | 2470 | struct btrfs_path *path; |
2477 | struct btrfs_key key; | 2471 | struct btrfs_key key; |
2478 | struct btrfs_key found_key; | 2472 | struct btrfs_key found_key; |
2479 | u32 found_type; | 2473 | u32 found_type = (u8)-1; |
2480 | struct extent_buffer *leaf; | 2474 | struct extent_buffer *leaf; |
2481 | struct btrfs_file_extent_item *fi; | 2475 | struct btrfs_file_extent_item *fi; |
2482 | u64 extent_start = 0; | 2476 | u64 extent_start = 0; |
@@ -2503,8 +2497,6 @@ noinline int btrfs_truncate_inode_items(struct btrfs_trans_handle *trans, | |||
2503 | key.offset = (u64)-1; | 2497 | key.offset = (u64)-1; |
2504 | key.type = (u8)-1; | 2498 | key.type = (u8)-1; |
2505 | 2499 | ||
2506 | btrfs_init_path(path); | ||
2507 | |||
2508 | search_again: | 2500 | search_again: |
2509 | ret = btrfs_search_slot(trans, root, &key, path, -1, 1); | 2501 | ret = btrfs_search_slot(trans, root, &key, path, -1, 1); |
2510 | if (ret < 0) | 2502 | if (ret < 0) |
@@ -2663,6 +2655,8 @@ next: | |||
2663 | if (pending_del_nr) | 2655 | if (pending_del_nr) |
2664 | goto del_pending; | 2656 | goto del_pending; |
2665 | btrfs_release_path(root, path); | 2657 | btrfs_release_path(root, path); |
2658 | if (found_type == BTRFS_INODE_ITEM_KEY) | ||
2659 | break; | ||
2666 | goto search_again; | 2660 | goto search_again; |
2667 | } | 2661 | } |
2668 | 2662 | ||
@@ -2679,6 +2673,8 @@ del_pending: | |||
2679 | BUG_ON(ret); | 2673 | BUG_ON(ret); |
2680 | pending_del_nr = 0; | 2674 | pending_del_nr = 0; |
2681 | btrfs_release_path(root, path); | 2675 | btrfs_release_path(root, path); |
2676 | if (found_type == BTRFS_INODE_ITEM_KEY) | ||
2677 | break; | ||
2682 | goto search_again; | 2678 | goto search_again; |
2683 | } | 2679 | } |
2684 | } | 2680 | } |
@@ -2788,7 +2784,7 @@ int btrfs_cont_expand(struct inode *inode, loff_t size) | |||
2788 | if (size <= hole_start) | 2784 | if (size <= hole_start) |
2789 | return 0; | 2785 | return 0; |
2790 | 2786 | ||
2791 | err = btrfs_check_free_space(root, 1, 0); | 2787 | err = btrfs_check_metadata_free_space(root); |
2792 | if (err) | 2788 | if (err) |
2793 | return err; | 2789 | return err; |
2794 | 2790 | ||
@@ -2984,6 +2980,7 @@ static noinline void init_btrfs_i(struct inode *inode) | |||
2984 | bi->last_trans = 0; | 2980 | bi->last_trans = 0; |
2985 | bi->logged_trans = 0; | 2981 | bi->logged_trans = 0; |
2986 | bi->delalloc_bytes = 0; | 2982 | bi->delalloc_bytes = 0; |
2983 | bi->reserved_bytes = 0; | ||
2987 | bi->disk_i_size = 0; | 2984 | bi->disk_i_size = 0; |
2988 | bi->flags = 0; | 2985 | bi->flags = 0; |
2989 | bi->index_cnt = (u64)-1; | 2986 | bi->index_cnt = (u64)-1; |
@@ -3005,6 +3002,7 @@ static int btrfs_init_locked_inode(struct inode *inode, void *p) | |||
3005 | inode->i_ino = args->ino; | 3002 | inode->i_ino = args->ino; |
3006 | init_btrfs_i(inode); | 3003 | init_btrfs_i(inode); |
3007 | BTRFS_I(inode)->root = args->root; | 3004 | BTRFS_I(inode)->root = args->root; |
3005 | btrfs_set_inode_space_info(args->root, inode); | ||
3008 | return 0; | 3006 | return 0; |
3009 | } | 3007 | } |
3010 | 3008 | ||
@@ -3265,7 +3263,7 @@ skip: | |||
3265 | 3263 | ||
3266 | /* Reached end of directory/root. Bump pos past the last item. */ | 3264 | /* Reached end of directory/root. Bump pos past the last item. */ |
3267 | if (key_type == BTRFS_DIR_INDEX_KEY) | 3265 | if (key_type == BTRFS_DIR_INDEX_KEY) |
3268 | filp->f_pos = INT_LIMIT(typeof(filp->f_pos)); | 3266 | filp->f_pos = INT_LIMIT(off_t); |
3269 | else | 3267 | else |
3270 | filp->f_pos++; | 3268 | filp->f_pos++; |
3271 | nopos: | 3269 | nopos: |
@@ -3425,6 +3423,7 @@ static struct inode *btrfs_new_inode(struct btrfs_trans_handle *trans, | |||
3425 | BTRFS_I(inode)->index_cnt = 2; | 3423 | BTRFS_I(inode)->index_cnt = 2; |
3426 | BTRFS_I(inode)->root = root; | 3424 | BTRFS_I(inode)->root = root; |
3427 | BTRFS_I(inode)->generation = trans->transid; | 3425 | BTRFS_I(inode)->generation = trans->transid; |
3426 | btrfs_set_inode_space_info(root, inode); | ||
3428 | 3427 | ||
3429 | if (mode & S_IFDIR) | 3428 | if (mode & S_IFDIR) |
3430 | owner = 0; | 3429 | owner = 0; |
@@ -3458,7 +3457,14 @@ static struct inode *btrfs_new_inode(struct btrfs_trans_handle *trans, | |||
3458 | root->highest_inode = objectid; | 3457 | root->highest_inode = objectid; |
3459 | 3458 | ||
3460 | inode->i_uid = current_fsuid(); | 3459 | inode->i_uid = current_fsuid(); |
3461 | inode->i_gid = current_fsgid(); | 3460 | |
3461 | if (dir && (dir->i_mode & S_ISGID)) { | ||
3462 | inode->i_gid = dir->i_gid; | ||
3463 | if (S_ISDIR(mode)) | ||
3464 | mode |= S_ISGID; | ||
3465 | } else | ||
3466 | inode->i_gid = current_fsgid(); | ||
3467 | |||
3462 | inode->i_mode = mode; | 3468 | inode->i_mode = mode; |
3463 | inode->i_ino = objectid; | 3469 | inode->i_ino = objectid; |
3464 | inode_set_bytes(inode, 0); | 3470 | inode_set_bytes(inode, 0); |
@@ -3565,7 +3571,7 @@ static int btrfs_mknod(struct inode *dir, struct dentry *dentry, | |||
3565 | if (!new_valid_dev(rdev)) | 3571 | if (!new_valid_dev(rdev)) |
3566 | return -EINVAL; | 3572 | return -EINVAL; |
3567 | 3573 | ||
3568 | err = btrfs_check_free_space(root, 1, 0); | 3574 | err = btrfs_check_metadata_free_space(root); |
3569 | if (err) | 3575 | if (err) |
3570 | goto fail; | 3576 | goto fail; |
3571 | 3577 | ||
@@ -3586,7 +3592,7 @@ static int btrfs_mknod(struct inode *dir, struct dentry *dentry, | |||
3586 | if (IS_ERR(inode)) | 3592 | if (IS_ERR(inode)) |
3587 | goto out_unlock; | 3593 | goto out_unlock; |
3588 | 3594 | ||
3589 | err = btrfs_init_acl(inode, dir); | 3595 | err = btrfs_init_inode_security(inode, dir); |
3590 | if (err) { | 3596 | if (err) { |
3591 | drop_inode = 1; | 3597 | drop_inode = 1; |
3592 | goto out_unlock; | 3598 | goto out_unlock; |
@@ -3628,7 +3634,7 @@ static int btrfs_create(struct inode *dir, struct dentry *dentry, | |||
3628 | u64 objectid; | 3634 | u64 objectid; |
3629 | u64 index = 0; | 3635 | u64 index = 0; |
3630 | 3636 | ||
3631 | err = btrfs_check_free_space(root, 1, 0); | 3637 | err = btrfs_check_metadata_free_space(root); |
3632 | if (err) | 3638 | if (err) |
3633 | goto fail; | 3639 | goto fail; |
3634 | trans = btrfs_start_transaction(root, 1); | 3640 | trans = btrfs_start_transaction(root, 1); |
@@ -3649,7 +3655,7 @@ static int btrfs_create(struct inode *dir, struct dentry *dentry, | |||
3649 | if (IS_ERR(inode)) | 3655 | if (IS_ERR(inode)) |
3650 | goto out_unlock; | 3656 | goto out_unlock; |
3651 | 3657 | ||
3652 | err = btrfs_init_acl(inode, dir); | 3658 | err = btrfs_init_inode_security(inode, dir); |
3653 | if (err) { | 3659 | if (err) { |
3654 | drop_inode = 1; | 3660 | drop_inode = 1; |
3655 | goto out_unlock; | 3661 | goto out_unlock; |
@@ -3696,7 +3702,7 @@ static int btrfs_link(struct dentry *old_dentry, struct inode *dir, | |||
3696 | return -ENOENT; | 3702 | return -ENOENT; |
3697 | 3703 | ||
3698 | btrfs_inc_nlink(inode); | 3704 | btrfs_inc_nlink(inode); |
3699 | err = btrfs_check_free_space(root, 1, 0); | 3705 | err = btrfs_check_metadata_free_space(root); |
3700 | if (err) | 3706 | if (err) |
3701 | goto fail; | 3707 | goto fail; |
3702 | err = btrfs_set_inode_index(dir, &index); | 3708 | err = btrfs_set_inode_index(dir, &index); |
@@ -3742,7 +3748,7 @@ static int btrfs_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
3742 | u64 index = 0; | 3748 | u64 index = 0; |
3743 | unsigned long nr = 1; | 3749 | unsigned long nr = 1; |
3744 | 3750 | ||
3745 | err = btrfs_check_free_space(root, 1, 0); | 3751 | err = btrfs_check_metadata_free_space(root); |
3746 | if (err) | 3752 | if (err) |
3747 | goto out_unlock; | 3753 | goto out_unlock; |
3748 | 3754 | ||
@@ -3772,7 +3778,7 @@ static int btrfs_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
3772 | 3778 | ||
3773 | drop_on_err = 1; | 3779 | drop_on_err = 1; |
3774 | 3780 | ||
3775 | err = btrfs_init_acl(inode, dir); | 3781 | err = btrfs_init_inode_security(inode, dir); |
3776 | if (err) | 3782 | if (err) |
3777 | goto out_fail; | 3783 | goto out_fail; |
3778 | 3784 | ||
@@ -4158,9 +4164,10 @@ static ssize_t btrfs_direct_IO(int rw, struct kiocb *iocb, | |||
4158 | return -EINVAL; | 4164 | return -EINVAL; |
4159 | } | 4165 | } |
4160 | 4166 | ||
4161 | static sector_t btrfs_bmap(struct address_space *mapping, sector_t iblock) | 4167 | static int btrfs_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, |
4168 | __u64 start, __u64 len) | ||
4162 | { | 4169 | { |
4163 | return extent_bmap(mapping, iblock, btrfs_get_extent); | 4170 | return extent_fiemap(inode, fieinfo, start, len, btrfs_get_extent); |
4164 | } | 4171 | } |
4165 | 4172 | ||
4166 | int btrfs_readpage(struct file *file, struct page *page) | 4173 | int btrfs_readpage(struct file *file, struct page *page) |
@@ -4223,7 +4230,7 @@ static int btrfs_releasepage(struct page *page, gfp_t gfp_flags) | |||
4223 | { | 4230 | { |
4224 | if (PageWriteback(page) || PageDirty(page)) | 4231 | if (PageWriteback(page) || PageDirty(page)) |
4225 | return 0; | 4232 | return 0; |
4226 | return __btrfs_releasepage(page, gfp_flags); | 4233 | return __btrfs_releasepage(page, gfp_flags & GFP_NOFS); |
4227 | } | 4234 | } |
4228 | 4235 | ||
4229 | static void btrfs_invalidatepage(struct page *page, unsigned long offset) | 4236 | static void btrfs_invalidatepage(struct page *page, unsigned long offset) |
@@ -4298,7 +4305,7 @@ int btrfs_page_mkwrite(struct vm_area_struct *vma, struct page *page) | |||
4298 | u64 page_start; | 4305 | u64 page_start; |
4299 | u64 page_end; | 4306 | u64 page_end; |
4300 | 4307 | ||
4301 | ret = btrfs_check_free_space(root, PAGE_CACHE_SIZE, 0); | 4308 | ret = btrfs_check_data_free_space(root, inode, PAGE_CACHE_SIZE); |
4302 | if (ret) | 4309 | if (ret) |
4303 | goto out; | 4310 | goto out; |
4304 | 4311 | ||
@@ -4311,6 +4318,7 @@ again: | |||
4311 | 4318 | ||
4312 | if ((page->mapping != inode->i_mapping) || | 4319 | if ((page->mapping != inode->i_mapping) || |
4313 | (page_start >= size)) { | 4320 | (page_start >= size)) { |
4321 | btrfs_free_reserved_data_space(root, inode, PAGE_CACHE_SIZE); | ||
4314 | /* page got truncated out from underneath us */ | 4322 | /* page got truncated out from underneath us */ |
4315 | goto out_unlock; | 4323 | goto out_unlock; |
4316 | } | 4324 | } |
@@ -4593,7 +4601,7 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
4593 | if (old_inode->i_ino == BTRFS_FIRST_FREE_OBJECTID) | 4601 | if (old_inode->i_ino == BTRFS_FIRST_FREE_OBJECTID) |
4594 | return -EXDEV; | 4602 | return -EXDEV; |
4595 | 4603 | ||
4596 | ret = btrfs_check_free_space(root, 1, 0); | 4604 | ret = btrfs_check_metadata_free_space(root); |
4597 | if (ret) | 4605 | if (ret) |
4598 | goto out_unlock; | 4606 | goto out_unlock; |
4599 | 4607 | ||
@@ -4711,7 +4719,7 @@ static int btrfs_symlink(struct inode *dir, struct dentry *dentry, | |||
4711 | if (name_len > BTRFS_MAX_INLINE_DATA_SIZE(root)) | 4719 | if (name_len > BTRFS_MAX_INLINE_DATA_SIZE(root)) |
4712 | return -ENAMETOOLONG; | 4720 | return -ENAMETOOLONG; |
4713 | 4721 | ||
4714 | err = btrfs_check_free_space(root, 1, 0); | 4722 | err = btrfs_check_metadata_free_space(root); |
4715 | if (err) | 4723 | if (err) |
4716 | goto out_fail; | 4724 | goto out_fail; |
4717 | 4725 | ||
@@ -4733,7 +4741,7 @@ static int btrfs_symlink(struct inode *dir, struct dentry *dentry, | |||
4733 | if (IS_ERR(inode)) | 4741 | if (IS_ERR(inode)) |
4734 | goto out_unlock; | 4742 | goto out_unlock; |
4735 | 4743 | ||
4736 | err = btrfs_init_acl(inode, dir); | 4744 | err = btrfs_init_inode_security(inode, dir); |
4737 | if (err) { | 4745 | if (err) { |
4738 | drop_inode = 1; | 4746 | drop_inode = 1; |
4739 | goto out_unlock; | 4747 | goto out_unlock; |
@@ -4987,13 +4995,24 @@ static struct extent_io_ops btrfs_extent_io_ops = { | |||
4987 | .clear_bit_hook = btrfs_clear_bit_hook, | 4995 | .clear_bit_hook = btrfs_clear_bit_hook, |
4988 | }; | 4996 | }; |
4989 | 4997 | ||
4998 | /* | ||
4999 | * btrfs doesn't support the bmap operation because swapfiles | ||
5000 | * use bmap to make a mapping of extents in the file. They assume | ||
5001 | * these extents won't change over the life of the file and they | ||
5002 | * use the bmap result to do IO directly to the drive. | ||
5003 | * | ||
5004 | * the btrfs bmap call would return logical addresses that aren't | ||
5005 | * suitable for IO and they also will change frequently as COW | ||
5006 | * operations happen. So, swapfile + btrfs == corruption. | ||
5007 | * | ||
5008 | * For now we're avoiding this by dropping bmap. | ||
5009 | */ | ||
4990 | static struct address_space_operations btrfs_aops = { | 5010 | static struct address_space_operations btrfs_aops = { |
4991 | .readpage = btrfs_readpage, | 5011 | .readpage = btrfs_readpage, |
4992 | .writepage = btrfs_writepage, | 5012 | .writepage = btrfs_writepage, |
4993 | .writepages = btrfs_writepages, | 5013 | .writepages = btrfs_writepages, |
4994 | .readpages = btrfs_readpages, | 5014 | .readpages = btrfs_readpages, |
4995 | .sync_page = block_sync_page, | 5015 | .sync_page = block_sync_page, |
4996 | .bmap = btrfs_bmap, | ||
4997 | .direct_IO = btrfs_direct_IO, | 5016 | .direct_IO = btrfs_direct_IO, |
4998 | .invalidatepage = btrfs_invalidatepage, | 5017 | .invalidatepage = btrfs_invalidatepage, |
4999 | .releasepage = btrfs_releasepage, | 5018 | .releasepage = btrfs_releasepage, |
@@ -5017,6 +5036,7 @@ static struct inode_operations btrfs_file_inode_operations = { | |||
5017 | .removexattr = btrfs_removexattr, | 5036 | .removexattr = btrfs_removexattr, |
5018 | .permission = btrfs_permission, | 5037 | .permission = btrfs_permission, |
5019 | .fallocate = btrfs_fallocate, | 5038 | .fallocate = btrfs_fallocate, |
5039 | .fiemap = btrfs_fiemap, | ||
5020 | }; | 5040 | }; |
5021 | static struct inode_operations btrfs_special_inode_operations = { | 5041 | static struct inode_operations btrfs_special_inode_operations = { |
5022 | .getattr = btrfs_getattr, | 5042 | .getattr = btrfs_getattr, |
@@ -5032,4 +5052,8 @@ static struct inode_operations btrfs_symlink_inode_operations = { | |||
5032 | .follow_link = page_follow_link_light, | 5052 | .follow_link = page_follow_link_light, |
5033 | .put_link = page_put_link, | 5053 | .put_link = page_put_link, |
5034 | .permission = btrfs_permission, | 5054 | .permission = btrfs_permission, |
5055 | .setxattr = btrfs_setxattr, | ||
5056 | .getxattr = btrfs_getxattr, | ||
5057 | .listxattr = btrfs_listxattr, | ||
5058 | .removexattr = btrfs_removexattr, | ||
5035 | }; | 5059 | }; |
diff --git a/fs/btrfs/ioctl.c b/fs/btrfs/ioctl.c index c2aa33e3feb5..bca729fc80c8 100644 --- a/fs/btrfs/ioctl.c +++ b/fs/btrfs/ioctl.c | |||
@@ -38,7 +38,6 @@ | |||
38 | #include <linux/compat.h> | 38 | #include <linux/compat.h> |
39 | #include <linux/bit_spinlock.h> | 39 | #include <linux/bit_spinlock.h> |
40 | #include <linux/security.h> | 40 | #include <linux/security.h> |
41 | #include <linux/version.h> | ||
42 | #include <linux/xattr.h> | 41 | #include <linux/xattr.h> |
43 | #include <linux/vmalloc.h> | 42 | #include <linux/vmalloc.h> |
44 | #include "compat.h" | 43 | #include "compat.h" |
@@ -71,7 +70,7 @@ static noinline int create_subvol(struct btrfs_root *root, | |||
71 | u64 index = 0; | 70 | u64 index = 0; |
72 | unsigned long nr = 1; | 71 | unsigned long nr = 1; |
73 | 72 | ||
74 | ret = btrfs_check_free_space(root, 1, 0); | 73 | ret = btrfs_check_metadata_free_space(root); |
75 | if (ret) | 74 | if (ret) |
76 | goto fail_commit; | 75 | goto fail_commit; |
77 | 76 | ||
@@ -204,7 +203,7 @@ static int create_snapshot(struct btrfs_root *root, struct dentry *dentry, | |||
204 | if (!root->ref_cows) | 203 | if (!root->ref_cows) |
205 | return -EINVAL; | 204 | return -EINVAL; |
206 | 205 | ||
207 | ret = btrfs_check_free_space(root, 1, 0); | 206 | ret = btrfs_check_metadata_free_space(root); |
208 | if (ret) | 207 | if (ret) |
209 | goto fail_unlock; | 208 | goto fail_unlock; |
210 | 209 | ||
@@ -375,7 +374,7 @@ static int btrfs_defrag_file(struct file *file) | |||
375 | unsigned long i; | 374 | unsigned long i; |
376 | int ret; | 375 | int ret; |
377 | 376 | ||
378 | ret = btrfs_check_free_space(root, inode->i_size, 0); | 377 | ret = btrfs_check_data_free_space(root, inode, inode->i_size); |
379 | if (ret) | 378 | if (ret) |
380 | return -ENOSPC; | 379 | return -ENOSPC; |
381 | 380 | ||
diff --git a/fs/btrfs/ioctl.h b/fs/btrfs/ioctl.h index 78049ea208db..b320b103fa13 100644 --- a/fs/btrfs/ioctl.h +++ b/fs/btrfs/ioctl.h | |||
@@ -22,13 +22,20 @@ | |||
22 | 22 | ||
23 | #define BTRFS_IOCTL_MAGIC 0x94 | 23 | #define BTRFS_IOCTL_MAGIC 0x94 |
24 | #define BTRFS_VOL_NAME_MAX 255 | 24 | #define BTRFS_VOL_NAME_MAX 255 |
25 | #define BTRFS_PATH_NAME_MAX 3072 | 25 | #define BTRFS_PATH_NAME_MAX 4087 |
26 | 26 | ||
27 | /* this should be 4k */ | ||
27 | struct btrfs_ioctl_vol_args { | 28 | struct btrfs_ioctl_vol_args { |
28 | __s64 fd; | 29 | __s64 fd; |
29 | char name[BTRFS_PATH_NAME_MAX + 1]; | 30 | char name[BTRFS_PATH_NAME_MAX + 1]; |
30 | }; | 31 | }; |
31 | 32 | ||
33 | struct btrfs_ioctl_clone_range_args { | ||
34 | __s64 src_fd; | ||
35 | __u64 src_offset, src_length; | ||
36 | __u64 dest_offset; | ||
37 | }; | ||
38 | |||
32 | #define BTRFS_IOC_SNAP_CREATE _IOW(BTRFS_IOCTL_MAGIC, 1, \ | 39 | #define BTRFS_IOC_SNAP_CREATE _IOW(BTRFS_IOCTL_MAGIC, 1, \ |
33 | struct btrfs_ioctl_vol_args) | 40 | struct btrfs_ioctl_vol_args) |
34 | #define BTRFS_IOC_DEFRAG _IOW(BTRFS_IOCTL_MAGIC, 2, \ | 41 | #define BTRFS_IOC_DEFRAG _IOW(BTRFS_IOCTL_MAGIC, 2, \ |
@@ -52,11 +59,6 @@ struct btrfs_ioctl_vol_args { | |||
52 | struct btrfs_ioctl_vol_args) | 59 | struct btrfs_ioctl_vol_args) |
53 | #define BTRFS_IOC_BALANCE _IOW(BTRFS_IOCTL_MAGIC, 12, \ | 60 | #define BTRFS_IOC_BALANCE _IOW(BTRFS_IOCTL_MAGIC, 12, \ |
54 | struct btrfs_ioctl_vol_args) | 61 | struct btrfs_ioctl_vol_args) |
55 | struct btrfs_ioctl_clone_range_args { | ||
56 | __s64 src_fd; | ||
57 | __u64 src_offset, src_length; | ||
58 | __u64 dest_offset; | ||
59 | }; | ||
60 | 62 | ||
61 | #define BTRFS_IOC_CLONE_RANGE _IOW(BTRFS_IOCTL_MAGIC, 13, \ | 63 | #define BTRFS_IOC_CLONE_RANGE _IOW(BTRFS_IOCTL_MAGIC, 13, \ |
62 | struct btrfs_ioctl_clone_range_args) | 64 | struct btrfs_ioctl_clone_range_args) |
diff --git a/fs/btrfs/locking.c b/fs/btrfs/locking.c index 39bae7761db6..85506c4a3af7 100644 --- a/fs/btrfs/locking.c +++ b/fs/btrfs/locking.c | |||
@@ -25,64 +25,203 @@ | |||
25 | #include "extent_io.h" | 25 | #include "extent_io.h" |
26 | #include "locking.h" | 26 | #include "locking.h" |
27 | 27 | ||
28 | static inline void spin_nested(struct extent_buffer *eb) | ||
29 | { | ||
30 | spin_lock(&eb->lock); | ||
31 | } | ||
32 | |||
28 | /* | 33 | /* |
29 | * locks the per buffer mutex in an extent buffer. This uses adaptive locks | 34 | * Setting a lock to blocking will drop the spinlock and set the |
30 | * and the spin is not tuned very extensively. The spinning does make a big | 35 | * flag that forces other procs who want the lock to wait. After |
31 | * difference in almost every workload, but spinning for the right amount of | 36 | * this you can safely schedule with the lock held. |
32 | * time needs some help. | ||
33 | * | ||
34 | * In general, we want to spin as long as the lock holder is doing btree | ||
35 | * searches, and we should give up if they are in more expensive code. | ||
36 | */ | 37 | */ |
38 | void btrfs_set_lock_blocking(struct extent_buffer *eb) | ||
39 | { | ||
40 | if (!test_bit(EXTENT_BUFFER_BLOCKING, &eb->bflags)) { | ||
41 | set_bit(EXTENT_BUFFER_BLOCKING, &eb->bflags); | ||
42 | spin_unlock(&eb->lock); | ||
43 | } | ||
44 | /* exit with the spin lock released and the bit set */ | ||
45 | } | ||
37 | 46 | ||
38 | int btrfs_tree_lock(struct extent_buffer *eb) | 47 | /* |
48 | * clearing the blocking flag will take the spinlock again. | ||
49 | * After this you can't safely schedule | ||
50 | */ | ||
51 | void btrfs_clear_lock_blocking(struct extent_buffer *eb) | ||
39 | { | 52 | { |
40 | int i; | 53 | if (test_bit(EXTENT_BUFFER_BLOCKING, &eb->bflags)) { |
54 | spin_nested(eb); | ||
55 | clear_bit(EXTENT_BUFFER_BLOCKING, &eb->bflags); | ||
56 | smp_mb__after_clear_bit(); | ||
57 | } | ||
58 | /* exit with the spin lock held */ | ||
59 | } | ||
41 | 60 | ||
42 | if (mutex_trylock(&eb->mutex)) | 61 | /* |
43 | return 0; | 62 | * unfortunately, many of the places that currently set a lock to blocking |
63 | * don't end up blocking for every long, and often they don't block | ||
64 | * at all. For a dbench 50 run, if we don't spin one the blocking bit | ||
65 | * at all, the context switch rate can jump up to 400,000/sec or more. | ||
66 | * | ||
67 | * So, we're still stuck with this crummy spin on the blocking bit, | ||
68 | * at least until the most common causes of the short blocks | ||
69 | * can be dealt with. | ||
70 | */ | ||
71 | static int btrfs_spin_on_block(struct extent_buffer *eb) | ||
72 | { | ||
73 | int i; | ||
44 | for (i = 0; i < 512; i++) { | 74 | for (i = 0; i < 512; i++) { |
45 | cpu_relax(); | 75 | cpu_relax(); |
46 | if (mutex_trylock(&eb->mutex)) | 76 | if (!test_bit(EXTENT_BUFFER_BLOCKING, &eb->bflags)) |
47 | return 0; | 77 | return 1; |
78 | if (need_resched()) | ||
79 | break; | ||
48 | } | 80 | } |
49 | cpu_relax(); | ||
50 | mutex_lock_nested(&eb->mutex, BTRFS_MAX_LEVEL - btrfs_header_level(eb)); | ||
51 | return 0; | 81 | return 0; |
52 | } | 82 | } |
53 | 83 | ||
54 | int btrfs_try_tree_lock(struct extent_buffer *eb) | 84 | /* |
85 | * This is somewhat different from trylock. It will take the | ||
86 | * spinlock but if it finds the lock is set to blocking, it will | ||
87 | * return without the lock held. | ||
88 | * | ||
89 | * returns 1 if it was able to take the lock and zero otherwise | ||
90 | * | ||
91 | * After this call, scheduling is not safe without first calling | ||
92 | * btrfs_set_lock_blocking() | ||
93 | */ | ||
94 | int btrfs_try_spin_lock(struct extent_buffer *eb) | ||
55 | { | 95 | { |
56 | return mutex_trylock(&eb->mutex); | 96 | int i; |
97 | |||
98 | spin_nested(eb); | ||
99 | if (!test_bit(EXTENT_BUFFER_BLOCKING, &eb->bflags)) | ||
100 | return 1; | ||
101 | spin_unlock(&eb->lock); | ||
102 | |||
103 | /* spin for a bit on the BLOCKING flag */ | ||
104 | for (i = 0; i < 2; i++) { | ||
105 | if (!btrfs_spin_on_block(eb)) | ||
106 | break; | ||
107 | |||
108 | spin_nested(eb); | ||
109 | if (!test_bit(EXTENT_BUFFER_BLOCKING, &eb->bflags)) | ||
110 | return 1; | ||
111 | spin_unlock(&eb->lock); | ||
112 | } | ||
113 | return 0; | ||
57 | } | 114 | } |
58 | 115 | ||
59 | int btrfs_tree_unlock(struct extent_buffer *eb) | 116 | /* |
117 | * the autoremove wake function will return 0 if it tried to wake up | ||
118 | * a process that was already awake, which means that process won't | ||
119 | * count as an exclusive wakeup. The waitq code will continue waking | ||
120 | * procs until it finds one that was actually sleeping. | ||
121 | * | ||
122 | * For btrfs, this isn't quite what we want. We want a single proc | ||
123 | * to be notified that the lock is ready for taking. If that proc | ||
124 | * already happen to be awake, great, it will loop around and try for | ||
125 | * the lock. | ||
126 | * | ||
127 | * So, btrfs_wake_function always returns 1, even when the proc that we | ||
128 | * tried to wake up was already awake. | ||
129 | */ | ||
130 | static int btrfs_wake_function(wait_queue_t *wait, unsigned mode, | ||
131 | int sync, void *key) | ||
60 | { | 132 | { |
61 | mutex_unlock(&eb->mutex); | 133 | autoremove_wake_function(wait, mode, sync, key); |
62 | return 0; | 134 | return 1; |
63 | } | 135 | } |
64 | 136 | ||
65 | int btrfs_tree_locked(struct extent_buffer *eb) | 137 | /* |
138 | * returns with the extent buffer spinlocked. | ||
139 | * | ||
140 | * This will spin and/or wait as required to take the lock, and then | ||
141 | * return with the spinlock held. | ||
142 | * | ||
143 | * After this call, scheduling is not safe without first calling | ||
144 | * btrfs_set_lock_blocking() | ||
145 | */ | ||
146 | int btrfs_tree_lock(struct extent_buffer *eb) | ||
66 | { | 147 | { |
67 | return mutex_is_locked(&eb->mutex); | 148 | DEFINE_WAIT(wait); |
149 | wait.func = btrfs_wake_function; | ||
150 | |||
151 | while(1) { | ||
152 | spin_nested(eb); | ||
153 | |||
154 | /* nobody is blocking, exit with the spinlock held */ | ||
155 | if (!test_bit(EXTENT_BUFFER_BLOCKING, &eb->bflags)) | ||
156 | return 0; | ||
157 | |||
158 | /* | ||
159 | * we have the spinlock, but the real owner is blocking. | ||
160 | * wait for them | ||
161 | */ | ||
162 | spin_unlock(&eb->lock); | ||
163 | |||
164 | /* | ||
165 | * spin for a bit, and if the blocking flag goes away, | ||
166 | * loop around | ||
167 | */ | ||
168 | if (btrfs_spin_on_block(eb)) | ||
169 | continue; | ||
170 | |||
171 | prepare_to_wait_exclusive(&eb->lock_wq, &wait, | ||
172 | TASK_UNINTERRUPTIBLE); | ||
173 | |||
174 | if (test_bit(EXTENT_BUFFER_BLOCKING, &eb->bflags)) | ||
175 | schedule(); | ||
176 | |||
177 | finish_wait(&eb->lock_wq, &wait); | ||
178 | } | ||
179 | return 0; | ||
68 | } | 180 | } |
69 | 181 | ||
70 | /* | 182 | /* |
71 | * btrfs_search_slot uses this to decide if it should drop its locks | 183 | * Very quick trylock, this does not spin or schedule. It returns |
72 | * before doing something expensive like allocating free blocks for cow. | 184 | * 1 with the spinlock held if it was able to take the lock, or it |
185 | * returns zero if it was unable to take the lock. | ||
186 | * | ||
187 | * After this call, scheduling is not safe without first calling | ||
188 | * btrfs_set_lock_blocking() | ||
73 | */ | 189 | */ |
74 | int btrfs_path_lock_waiting(struct btrfs_path *path, int level) | 190 | int btrfs_try_tree_lock(struct extent_buffer *eb) |
75 | { | 191 | { |
76 | int i; | 192 | if (spin_trylock(&eb->lock)) { |
77 | struct extent_buffer *eb; | 193 | if (test_bit(EXTENT_BUFFER_BLOCKING, &eb->bflags)) { |
78 | for (i = level; i <= level + 1 && i < BTRFS_MAX_LEVEL; i++) { | 194 | /* |
79 | eb = path->nodes[i]; | 195 | * we've got the spinlock, but the real owner is |
80 | if (!eb) | 196 | * blocking. Drop the spinlock and return failure |
81 | break; | 197 | */ |
82 | smp_mb(); | 198 | spin_unlock(&eb->lock); |
83 | if (!list_empty(&eb->mutex.wait_list)) | 199 | return 0; |
84 | return 1; | 200 | } |
201 | return 1; | ||
85 | } | 202 | } |
203 | /* someone else has the spinlock giveup */ | ||
86 | return 0; | 204 | return 0; |
87 | } | 205 | } |
88 | 206 | ||
207 | int btrfs_tree_unlock(struct extent_buffer *eb) | ||
208 | { | ||
209 | /* | ||
210 | * if we were a blocking owner, we don't have the spinlock held | ||
211 | * just clear the bit and look for waiters | ||
212 | */ | ||
213 | if (test_and_clear_bit(EXTENT_BUFFER_BLOCKING, &eb->bflags)) | ||
214 | smp_mb__after_clear_bit(); | ||
215 | else | ||
216 | spin_unlock(&eb->lock); | ||
217 | |||
218 | if (waitqueue_active(&eb->lock_wq)) | ||
219 | wake_up(&eb->lock_wq); | ||
220 | return 0; | ||
221 | } | ||
222 | |||
223 | int btrfs_tree_locked(struct extent_buffer *eb) | ||
224 | { | ||
225 | return test_bit(EXTENT_BUFFER_BLOCKING, &eb->bflags) || | ||
226 | spin_is_locked(&eb->lock); | ||
227 | } | ||
diff --git a/fs/btrfs/locking.h b/fs/btrfs/locking.h index bc1faef12519..6bb0afbff928 100644 --- a/fs/btrfs/locking.h +++ b/fs/btrfs/locking.h | |||
@@ -22,6 +22,10 @@ | |||
22 | int btrfs_tree_lock(struct extent_buffer *eb); | 22 | int btrfs_tree_lock(struct extent_buffer *eb); |
23 | int btrfs_tree_unlock(struct extent_buffer *eb); | 23 | int btrfs_tree_unlock(struct extent_buffer *eb); |
24 | int btrfs_tree_locked(struct extent_buffer *eb); | 24 | int btrfs_tree_locked(struct extent_buffer *eb); |
25 | |||
25 | int btrfs_try_tree_lock(struct extent_buffer *eb); | 26 | int btrfs_try_tree_lock(struct extent_buffer *eb); |
26 | int btrfs_path_lock_waiting(struct btrfs_path *path, int level); | 27 | int btrfs_try_spin_lock(struct extent_buffer *eb); |
28 | |||
29 | void btrfs_set_lock_blocking(struct extent_buffer *eb); | ||
30 | void btrfs_clear_lock_blocking(struct extent_buffer *eb); | ||
27 | #endif | 31 | #endif |
diff --git a/fs/btrfs/ordered-data.c b/fs/btrfs/ordered-data.c index a20940170274..77c2411a5f0f 100644 --- a/fs/btrfs/ordered-data.c +++ b/fs/btrfs/ordered-data.c | |||
@@ -613,7 +613,6 @@ int btrfs_find_ordered_sum(struct inode *inode, u64 offset, u64 disk_bytenr, | |||
613 | struct btrfs_sector_sum *sector_sums; | 613 | struct btrfs_sector_sum *sector_sums; |
614 | struct btrfs_ordered_extent *ordered; | 614 | struct btrfs_ordered_extent *ordered; |
615 | struct btrfs_ordered_inode_tree *tree = &BTRFS_I(inode)->ordered_tree; | 615 | struct btrfs_ordered_inode_tree *tree = &BTRFS_I(inode)->ordered_tree; |
616 | struct list_head *cur; | ||
617 | unsigned long num_sectors; | 616 | unsigned long num_sectors; |
618 | unsigned long i; | 617 | unsigned long i; |
619 | u32 sectorsize = BTRFS_I(inode)->root->sectorsize; | 618 | u32 sectorsize = BTRFS_I(inode)->root->sectorsize; |
@@ -624,8 +623,7 @@ int btrfs_find_ordered_sum(struct inode *inode, u64 offset, u64 disk_bytenr, | |||
624 | return 1; | 623 | return 1; |
625 | 624 | ||
626 | mutex_lock(&tree->mutex); | 625 | mutex_lock(&tree->mutex); |
627 | list_for_each_prev(cur, &ordered->list) { | 626 | list_for_each_entry_reverse(ordered_sum, &ordered->list, list) { |
628 | ordered_sum = list_entry(cur, struct btrfs_ordered_sum, list); | ||
629 | if (disk_bytenr >= ordered_sum->bytenr) { | 627 | if (disk_bytenr >= ordered_sum->bytenr) { |
630 | num_sectors = ordered_sum->len / sectorsize; | 628 | num_sectors = ordered_sum->len / sectorsize; |
631 | sector_sums = ordered_sum->sums; | 629 | sector_sums = ordered_sum->sums; |
diff --git a/fs/btrfs/ref-cache.c b/fs/btrfs/ref-cache.c index 6f0acc4c9eab..d0cc62bccb94 100644 --- a/fs/btrfs/ref-cache.c +++ b/fs/btrfs/ref-cache.c | |||
@@ -17,6 +17,7 @@ | |||
17 | */ | 17 | */ |
18 | 18 | ||
19 | #include <linux/sched.h> | 19 | #include <linux/sched.h> |
20 | #include <linux/sort.h> | ||
20 | #include "ctree.h" | 21 | #include "ctree.h" |
21 | #include "ref-cache.h" | 22 | #include "ref-cache.h" |
22 | #include "transaction.h" | 23 | #include "transaction.h" |
diff --git a/fs/btrfs/ref-cache.h b/fs/btrfs/ref-cache.h index 16f3183d7c59..bc283ad2db73 100644 --- a/fs/btrfs/ref-cache.h +++ b/fs/btrfs/ref-cache.h | |||
@@ -73,5 +73,4 @@ int btrfs_add_leaf_ref(struct btrfs_root *root, struct btrfs_leaf_ref *ref, | |||
73 | int btrfs_remove_leaf_refs(struct btrfs_root *root, u64 max_root_gen, | 73 | int btrfs_remove_leaf_refs(struct btrfs_root *root, u64 max_root_gen, |
74 | int shared); | 74 | int shared); |
75 | int btrfs_remove_leaf_ref(struct btrfs_root *root, struct btrfs_leaf_ref *ref); | 75 | int btrfs_remove_leaf_ref(struct btrfs_root *root, struct btrfs_leaf_ref *ref); |
76 | |||
77 | #endif | 76 | #endif |
diff --git a/fs/btrfs/super.c b/fs/btrfs/super.c index 0a14b495532f..19a4daf03ccb 100644 --- a/fs/btrfs/super.c +++ b/fs/btrfs/super.c | |||
@@ -37,7 +37,7 @@ | |||
37 | #include <linux/ctype.h> | 37 | #include <linux/ctype.h> |
38 | #include <linux/namei.h> | 38 | #include <linux/namei.h> |
39 | #include <linux/miscdevice.h> | 39 | #include <linux/miscdevice.h> |
40 | #include <linux/version.h> | 40 | #include <linux/magic.h> |
41 | #include "compat.h" | 41 | #include "compat.h" |
42 | #include "ctree.h" | 42 | #include "ctree.h" |
43 | #include "disk-io.h" | 43 | #include "disk-io.h" |
@@ -51,7 +51,6 @@ | |||
51 | #include "export.h" | 51 | #include "export.h" |
52 | #include "compression.h" | 52 | #include "compression.h" |
53 | 53 | ||
54 | #define BTRFS_SUPER_MAGIC 0x9123683E | ||
55 | 54 | ||
56 | static struct super_operations btrfs_super_ops; | 55 | static struct super_operations btrfs_super_ops; |
57 | 56 | ||
@@ -380,7 +379,6 @@ int btrfs_sync_fs(struct super_block *sb, int wait) | |||
380 | btrfs_start_delalloc_inodes(root); | 379 | btrfs_start_delalloc_inodes(root); |
381 | btrfs_wait_ordered_extents(root, 0); | 380 | btrfs_wait_ordered_extents(root, 0); |
382 | 381 | ||
383 | btrfs_clean_old_snapshots(root); | ||
384 | trans = btrfs_start_transaction(root, 1); | 382 | trans = btrfs_start_transaction(root, 1); |
385 | ret = btrfs_commit_transaction(trans, root); | 383 | ret = btrfs_commit_transaction(trans, root); |
386 | sb->s_dirt = 0; | 384 | sb->s_dirt = 0; |
@@ -512,6 +510,10 @@ static int btrfs_remount(struct super_block *sb, int *flags, char *data) | |||
512 | struct btrfs_root *root = btrfs_sb(sb); | 510 | struct btrfs_root *root = btrfs_sb(sb); |
513 | int ret; | 511 | int ret; |
514 | 512 | ||
513 | ret = btrfs_parse_options(root, data); | ||
514 | if (ret) | ||
515 | return -EINVAL; | ||
516 | |||
515 | if ((*flags & MS_RDONLY) == (sb->s_flags & MS_RDONLY)) | 517 | if ((*flags & MS_RDONLY) == (sb->s_flags & MS_RDONLY)) |
516 | return 0; | 518 | return 0; |
517 | 519 | ||
@@ -582,18 +584,20 @@ static long btrfs_control_ioctl(struct file *file, unsigned int cmd, | |||
582 | { | 584 | { |
583 | struct btrfs_ioctl_vol_args *vol; | 585 | struct btrfs_ioctl_vol_args *vol; |
584 | struct btrfs_fs_devices *fs_devices; | 586 | struct btrfs_fs_devices *fs_devices; |
585 | int ret = 0; | 587 | int ret = -ENOTTY; |
586 | int len; | ||
587 | 588 | ||
588 | if (!capable(CAP_SYS_ADMIN)) | 589 | if (!capable(CAP_SYS_ADMIN)) |
589 | return -EPERM; | 590 | return -EPERM; |
590 | 591 | ||
591 | vol = kmalloc(sizeof(*vol), GFP_KERNEL); | 592 | vol = kmalloc(sizeof(*vol), GFP_KERNEL); |
593 | if (!vol) | ||
594 | return -ENOMEM; | ||
595 | |||
592 | if (copy_from_user(vol, (void __user *)arg, sizeof(*vol))) { | 596 | if (copy_from_user(vol, (void __user *)arg, sizeof(*vol))) { |
593 | ret = -EFAULT; | 597 | ret = -EFAULT; |
594 | goto out; | 598 | goto out; |
595 | } | 599 | } |
596 | len = strnlen(vol->name, BTRFS_PATH_NAME_MAX); | 600 | |
597 | switch (cmd) { | 601 | switch (cmd) { |
598 | case BTRFS_IOC_SCAN_DEV: | 602 | case BTRFS_IOC_SCAN_DEV: |
599 | ret = btrfs_scan_one_device(vol->name, FMODE_READ, | 603 | ret = btrfs_scan_one_device(vol->name, FMODE_READ, |
diff --git a/fs/btrfs/transaction.c b/fs/btrfs/transaction.c index 8a08f9443340..4112d53d4f4d 100644 --- a/fs/btrfs/transaction.c +++ b/fs/btrfs/transaction.c | |||
@@ -688,7 +688,9 @@ static noinline int drop_dirty_roots(struct btrfs_root *tree_root, | |||
688 | num_bytes -= btrfs_root_used(&dirty->root->root_item); | 688 | num_bytes -= btrfs_root_used(&dirty->root->root_item); |
689 | bytes_used = btrfs_root_used(&root->root_item); | 689 | bytes_used = btrfs_root_used(&root->root_item); |
690 | if (num_bytes) { | 690 | if (num_bytes) { |
691 | mutex_lock(&root->fs_info->trans_mutex); | ||
691 | btrfs_record_root_in_trans(root); | 692 | btrfs_record_root_in_trans(root); |
693 | mutex_unlock(&root->fs_info->trans_mutex); | ||
692 | btrfs_set_root_used(&root->root_item, | 694 | btrfs_set_root_used(&root->root_item, |
693 | bytes_used - num_bytes); | 695 | bytes_used - num_bytes); |
694 | } | 696 | } |
@@ -852,11 +854,9 @@ static noinline int create_pending_snapshots(struct btrfs_trans_handle *trans, | |||
852 | { | 854 | { |
853 | struct btrfs_pending_snapshot *pending; | 855 | struct btrfs_pending_snapshot *pending; |
854 | struct list_head *head = &trans->transaction->pending_snapshots; | 856 | struct list_head *head = &trans->transaction->pending_snapshots; |
855 | struct list_head *cur; | ||
856 | int ret; | 857 | int ret; |
857 | 858 | ||
858 | list_for_each(cur, head) { | 859 | list_for_each_entry(pending, head, list) { |
859 | pending = list_entry(cur, struct btrfs_pending_snapshot, list); | ||
860 | ret = create_pending_snapshot(trans, fs_info, pending); | 860 | ret = create_pending_snapshot(trans, fs_info, pending); |
861 | BUG_ON(ret); | 861 | BUG_ON(ret); |
862 | } | 862 | } |
diff --git a/fs/btrfs/tree-defrag.c b/fs/btrfs/tree-defrag.c index 3e8358c36165..98d25fa4570e 100644 --- a/fs/btrfs/tree-defrag.c +++ b/fs/btrfs/tree-defrag.c | |||
@@ -74,6 +74,7 @@ int btrfs_defrag_leaves(struct btrfs_trans_handle *trans, | |||
74 | u32 nritems; | 74 | u32 nritems; |
75 | 75 | ||
76 | root_node = btrfs_lock_root_node(root); | 76 | root_node = btrfs_lock_root_node(root); |
77 | btrfs_set_lock_blocking(root_node); | ||
77 | nritems = btrfs_header_nritems(root_node); | 78 | nritems = btrfs_header_nritems(root_node); |
78 | root->defrag_max.objectid = 0; | 79 | root->defrag_max.objectid = 0; |
79 | /* from above we know this is not a leaf */ | 80 | /* from above we know this is not a leaf */ |
diff --git a/fs/btrfs/tree-log.c b/fs/btrfs/tree-log.c index d81cda2e077c..9c462fbd60fa 100644 --- a/fs/btrfs/tree-log.c +++ b/fs/btrfs/tree-log.c | |||
@@ -78,104 +78,6 @@ static int link_to_fixup_dir(struct btrfs_trans_handle *trans, | |||
78 | */ | 78 | */ |
79 | 79 | ||
80 | /* | 80 | /* |
81 | * btrfs_add_log_tree adds a new per-subvolume log tree into the | ||
82 | * tree of log tree roots. This must be called with a tree log transaction | ||
83 | * running (see start_log_trans). | ||
84 | */ | ||
85 | static int btrfs_add_log_tree(struct btrfs_trans_handle *trans, | ||
86 | struct btrfs_root *root) | ||
87 | { | ||
88 | struct btrfs_key key; | ||
89 | struct btrfs_root_item root_item; | ||
90 | struct btrfs_inode_item *inode_item; | ||
91 | struct extent_buffer *leaf; | ||
92 | struct btrfs_root *new_root = root; | ||
93 | int ret; | ||
94 | u64 objectid = root->root_key.objectid; | ||
95 | |||
96 | leaf = btrfs_alloc_free_block(trans, root, root->leafsize, 0, | ||
97 | BTRFS_TREE_LOG_OBJECTID, | ||
98 | trans->transid, 0, 0, 0); | ||
99 | if (IS_ERR(leaf)) { | ||
100 | ret = PTR_ERR(leaf); | ||
101 | return ret; | ||
102 | } | ||
103 | |||
104 | btrfs_set_header_nritems(leaf, 0); | ||
105 | btrfs_set_header_level(leaf, 0); | ||
106 | btrfs_set_header_bytenr(leaf, leaf->start); | ||
107 | btrfs_set_header_generation(leaf, trans->transid); | ||
108 | btrfs_set_header_owner(leaf, BTRFS_TREE_LOG_OBJECTID); | ||
109 | |||
110 | write_extent_buffer(leaf, root->fs_info->fsid, | ||
111 | (unsigned long)btrfs_header_fsid(leaf), | ||
112 | BTRFS_FSID_SIZE); | ||
113 | btrfs_mark_buffer_dirty(leaf); | ||
114 | |||
115 | inode_item = &root_item.inode; | ||
116 | memset(inode_item, 0, sizeof(*inode_item)); | ||
117 | inode_item->generation = cpu_to_le64(1); | ||
118 | inode_item->size = cpu_to_le64(3); | ||
119 | inode_item->nlink = cpu_to_le32(1); | ||
120 | inode_item->nbytes = cpu_to_le64(root->leafsize); | ||
121 | inode_item->mode = cpu_to_le32(S_IFDIR | 0755); | ||
122 | |||
123 | btrfs_set_root_bytenr(&root_item, leaf->start); | ||
124 | btrfs_set_root_generation(&root_item, trans->transid); | ||
125 | btrfs_set_root_level(&root_item, 0); | ||
126 | btrfs_set_root_refs(&root_item, 0); | ||
127 | btrfs_set_root_used(&root_item, 0); | ||
128 | |||
129 | memset(&root_item.drop_progress, 0, sizeof(root_item.drop_progress)); | ||
130 | root_item.drop_level = 0; | ||
131 | |||
132 | btrfs_tree_unlock(leaf); | ||
133 | free_extent_buffer(leaf); | ||
134 | leaf = NULL; | ||
135 | |||
136 | btrfs_set_root_dirid(&root_item, 0); | ||
137 | |||
138 | key.objectid = BTRFS_TREE_LOG_OBJECTID; | ||
139 | key.offset = objectid; | ||
140 | btrfs_set_key_type(&key, BTRFS_ROOT_ITEM_KEY); | ||
141 | ret = btrfs_insert_root(trans, root->fs_info->log_root_tree, &key, | ||
142 | &root_item); | ||
143 | if (ret) | ||
144 | goto fail; | ||
145 | |||
146 | new_root = btrfs_read_fs_root_no_radix(root->fs_info->log_root_tree, | ||
147 | &key); | ||
148 | BUG_ON(!new_root); | ||
149 | |||
150 | WARN_ON(root->log_root); | ||
151 | root->log_root = new_root; | ||
152 | |||
153 | /* | ||
154 | * log trees do not get reference counted because they go away | ||
155 | * before a real commit is actually done. They do store pointers | ||
156 | * to file data extents, and those reference counts still get | ||
157 | * updated (along with back refs to the log tree). | ||
158 | */ | ||
159 | new_root->ref_cows = 0; | ||
160 | new_root->last_trans = trans->transid; | ||
161 | |||
162 | /* | ||
163 | * we need to make sure the root block for this new tree | ||
164 | * is marked as dirty in the dirty_log_pages tree. This | ||
165 | * is how it gets flushed down to disk at tree log commit time. | ||
166 | * | ||
167 | * the tree logging mutex keeps others from coming in and changing | ||
168 | * the new_root->node, so we can safely access it here | ||
169 | */ | ||
170 | set_extent_dirty(&new_root->dirty_log_pages, new_root->node->start, | ||
171 | new_root->node->start + new_root->node->len - 1, | ||
172 | GFP_NOFS); | ||
173 | |||
174 | fail: | ||
175 | return ret; | ||
176 | } | ||
177 | |||
178 | /* | ||
179 | * start a sub transaction and setup the log tree | 81 | * start a sub transaction and setup the log tree |
180 | * this increments the log tree writer count to make the people | 82 | * this increments the log tree writer count to make the people |
181 | * syncing the tree wait for us to finish | 83 | * syncing the tree wait for us to finish |
@@ -184,6 +86,14 @@ static int start_log_trans(struct btrfs_trans_handle *trans, | |||
184 | struct btrfs_root *root) | 86 | struct btrfs_root *root) |
185 | { | 87 | { |
186 | int ret; | 88 | int ret; |
89 | |||
90 | mutex_lock(&root->log_mutex); | ||
91 | if (root->log_root) { | ||
92 | root->log_batch++; | ||
93 | atomic_inc(&root->log_writers); | ||
94 | mutex_unlock(&root->log_mutex); | ||
95 | return 0; | ||
96 | } | ||
187 | mutex_lock(&root->fs_info->tree_log_mutex); | 97 | mutex_lock(&root->fs_info->tree_log_mutex); |
188 | if (!root->fs_info->log_root_tree) { | 98 | if (!root->fs_info->log_root_tree) { |
189 | ret = btrfs_init_log_root_tree(trans, root->fs_info); | 99 | ret = btrfs_init_log_root_tree(trans, root->fs_info); |
@@ -193,9 +103,10 @@ static int start_log_trans(struct btrfs_trans_handle *trans, | |||
193 | ret = btrfs_add_log_tree(trans, root); | 103 | ret = btrfs_add_log_tree(trans, root); |
194 | BUG_ON(ret); | 104 | BUG_ON(ret); |
195 | } | 105 | } |
196 | atomic_inc(&root->fs_info->tree_log_writers); | ||
197 | root->fs_info->tree_log_batch++; | ||
198 | mutex_unlock(&root->fs_info->tree_log_mutex); | 106 | mutex_unlock(&root->fs_info->tree_log_mutex); |
107 | root->log_batch++; | ||
108 | atomic_inc(&root->log_writers); | ||
109 | mutex_unlock(&root->log_mutex); | ||
199 | return 0; | 110 | return 0; |
200 | } | 111 | } |
201 | 112 | ||
@@ -212,13 +123,12 @@ static int join_running_log_trans(struct btrfs_root *root) | |||
212 | if (!root->log_root) | 123 | if (!root->log_root) |
213 | return -ENOENT; | 124 | return -ENOENT; |
214 | 125 | ||
215 | mutex_lock(&root->fs_info->tree_log_mutex); | 126 | mutex_lock(&root->log_mutex); |
216 | if (root->log_root) { | 127 | if (root->log_root) { |
217 | ret = 0; | 128 | ret = 0; |
218 | atomic_inc(&root->fs_info->tree_log_writers); | 129 | atomic_inc(&root->log_writers); |
219 | root->fs_info->tree_log_batch++; | ||
220 | } | 130 | } |
221 | mutex_unlock(&root->fs_info->tree_log_mutex); | 131 | mutex_unlock(&root->log_mutex); |
222 | return ret; | 132 | return ret; |
223 | } | 133 | } |
224 | 134 | ||
@@ -228,10 +138,11 @@ static int join_running_log_trans(struct btrfs_root *root) | |||
228 | */ | 138 | */ |
229 | static int end_log_trans(struct btrfs_root *root) | 139 | static int end_log_trans(struct btrfs_root *root) |
230 | { | 140 | { |
231 | atomic_dec(&root->fs_info->tree_log_writers); | 141 | if (atomic_dec_and_test(&root->log_writers)) { |
232 | smp_mb(); | 142 | smp_mb(); |
233 | if (waitqueue_active(&root->fs_info->tree_log_wait)) | 143 | if (waitqueue_active(&root->log_writer_wait)) |
234 | wake_up(&root->fs_info->tree_log_wait); | 144 | wake_up(&root->log_writer_wait); |
145 | } | ||
235 | return 0; | 146 | return 0; |
236 | } | 147 | } |
237 | 148 | ||
@@ -1704,6 +1615,7 @@ static noinline int walk_down_log_tree(struct btrfs_trans_handle *trans, | |||
1704 | 1615 | ||
1705 | btrfs_tree_lock(next); | 1616 | btrfs_tree_lock(next); |
1706 | clean_tree_block(trans, root, next); | 1617 | clean_tree_block(trans, root, next); |
1618 | btrfs_set_lock_blocking(next); | ||
1707 | btrfs_wait_tree_block_writeback(next); | 1619 | btrfs_wait_tree_block_writeback(next); |
1708 | btrfs_tree_unlock(next); | 1620 | btrfs_tree_unlock(next); |
1709 | 1621 | ||
@@ -1750,6 +1662,7 @@ static noinline int walk_down_log_tree(struct btrfs_trans_handle *trans, | |||
1750 | next = path->nodes[*level]; | 1662 | next = path->nodes[*level]; |
1751 | btrfs_tree_lock(next); | 1663 | btrfs_tree_lock(next); |
1752 | clean_tree_block(trans, root, next); | 1664 | clean_tree_block(trans, root, next); |
1665 | btrfs_set_lock_blocking(next); | ||
1753 | btrfs_wait_tree_block_writeback(next); | 1666 | btrfs_wait_tree_block_writeback(next); |
1754 | btrfs_tree_unlock(next); | 1667 | btrfs_tree_unlock(next); |
1755 | 1668 | ||
@@ -1807,6 +1720,7 @@ static noinline int walk_up_log_tree(struct btrfs_trans_handle *trans, | |||
1807 | 1720 | ||
1808 | btrfs_tree_lock(next); | 1721 | btrfs_tree_lock(next); |
1809 | clean_tree_block(trans, root, next); | 1722 | clean_tree_block(trans, root, next); |
1723 | btrfs_set_lock_blocking(next); | ||
1810 | btrfs_wait_tree_block_writeback(next); | 1724 | btrfs_wait_tree_block_writeback(next); |
1811 | btrfs_tree_unlock(next); | 1725 | btrfs_tree_unlock(next); |
1812 | 1726 | ||
@@ -1879,6 +1793,7 @@ static int walk_log_tree(struct btrfs_trans_handle *trans, | |||
1879 | 1793 | ||
1880 | btrfs_tree_lock(next); | 1794 | btrfs_tree_lock(next); |
1881 | clean_tree_block(trans, log, next); | 1795 | clean_tree_block(trans, log, next); |
1796 | btrfs_set_lock_blocking(next); | ||
1882 | btrfs_wait_tree_block_writeback(next); | 1797 | btrfs_wait_tree_block_writeback(next); |
1883 | btrfs_tree_unlock(next); | 1798 | btrfs_tree_unlock(next); |
1884 | 1799 | ||
@@ -1902,26 +1817,65 @@ static int walk_log_tree(struct btrfs_trans_handle *trans, | |||
1902 | } | 1817 | } |
1903 | } | 1818 | } |
1904 | btrfs_free_path(path); | 1819 | btrfs_free_path(path); |
1905 | if (wc->free) | ||
1906 | free_extent_buffer(log->node); | ||
1907 | return ret; | 1820 | return ret; |
1908 | } | 1821 | } |
1909 | 1822 | ||
1910 | static int wait_log_commit(struct btrfs_root *log) | 1823 | /* |
1824 | * helper function to update the item for a given subvolumes log root | ||
1825 | * in the tree of log roots | ||
1826 | */ | ||
1827 | static int update_log_root(struct btrfs_trans_handle *trans, | ||
1828 | struct btrfs_root *log) | ||
1829 | { | ||
1830 | int ret; | ||
1831 | |||
1832 | if (log->log_transid == 1) { | ||
1833 | /* insert root item on the first sync */ | ||
1834 | ret = btrfs_insert_root(trans, log->fs_info->log_root_tree, | ||
1835 | &log->root_key, &log->root_item); | ||
1836 | } else { | ||
1837 | ret = btrfs_update_root(trans, log->fs_info->log_root_tree, | ||
1838 | &log->root_key, &log->root_item); | ||
1839 | } | ||
1840 | return ret; | ||
1841 | } | ||
1842 | |||
1843 | static int wait_log_commit(struct btrfs_root *root, unsigned long transid) | ||
1911 | { | 1844 | { |
1912 | DEFINE_WAIT(wait); | 1845 | DEFINE_WAIT(wait); |
1913 | u64 transid = log->fs_info->tree_log_transid; | 1846 | int index = transid % 2; |
1914 | 1847 | ||
1848 | /* | ||
1849 | * we only allow two pending log transactions at a time, | ||
1850 | * so we know that if ours is more than 2 older than the | ||
1851 | * current transaction, we're done | ||
1852 | */ | ||
1915 | do { | 1853 | do { |
1916 | prepare_to_wait(&log->fs_info->tree_log_wait, &wait, | 1854 | prepare_to_wait(&root->log_commit_wait[index], |
1917 | TASK_UNINTERRUPTIBLE); | 1855 | &wait, TASK_UNINTERRUPTIBLE); |
1918 | mutex_unlock(&log->fs_info->tree_log_mutex); | 1856 | mutex_unlock(&root->log_mutex); |
1919 | if (atomic_read(&log->fs_info->tree_log_commit)) | 1857 | if (root->log_transid < transid + 2 && |
1858 | atomic_read(&root->log_commit[index])) | ||
1920 | schedule(); | 1859 | schedule(); |
1921 | finish_wait(&log->fs_info->tree_log_wait, &wait); | 1860 | finish_wait(&root->log_commit_wait[index], &wait); |
1922 | mutex_lock(&log->fs_info->tree_log_mutex); | 1861 | mutex_lock(&root->log_mutex); |
1923 | } while (transid == log->fs_info->tree_log_transid && | 1862 | } while (root->log_transid < transid + 2 && |
1924 | atomic_read(&log->fs_info->tree_log_commit)); | 1863 | atomic_read(&root->log_commit[index])); |
1864 | return 0; | ||
1865 | } | ||
1866 | |||
1867 | static int wait_for_writer(struct btrfs_root *root) | ||
1868 | { | ||
1869 | DEFINE_WAIT(wait); | ||
1870 | while (atomic_read(&root->log_writers)) { | ||
1871 | prepare_to_wait(&root->log_writer_wait, | ||
1872 | &wait, TASK_UNINTERRUPTIBLE); | ||
1873 | mutex_unlock(&root->log_mutex); | ||
1874 | if (atomic_read(&root->log_writers)) | ||
1875 | schedule(); | ||
1876 | mutex_lock(&root->log_mutex); | ||
1877 | finish_wait(&root->log_writer_wait, &wait); | ||
1878 | } | ||
1925 | return 0; | 1879 | return 0; |
1926 | } | 1880 | } |
1927 | 1881 | ||
@@ -1933,57 +1887,114 @@ static int wait_log_commit(struct btrfs_root *log) | |||
1933 | int btrfs_sync_log(struct btrfs_trans_handle *trans, | 1887 | int btrfs_sync_log(struct btrfs_trans_handle *trans, |
1934 | struct btrfs_root *root) | 1888 | struct btrfs_root *root) |
1935 | { | 1889 | { |
1890 | int index1; | ||
1891 | int index2; | ||
1936 | int ret; | 1892 | int ret; |
1937 | unsigned long batch; | ||
1938 | struct btrfs_root *log = root->log_root; | 1893 | struct btrfs_root *log = root->log_root; |
1894 | struct btrfs_root *log_root_tree = root->fs_info->log_root_tree; | ||
1939 | 1895 | ||
1940 | mutex_lock(&log->fs_info->tree_log_mutex); | 1896 | mutex_lock(&root->log_mutex); |
1941 | if (atomic_read(&log->fs_info->tree_log_commit)) { | 1897 | index1 = root->log_transid % 2; |
1942 | wait_log_commit(log); | 1898 | if (atomic_read(&root->log_commit[index1])) { |
1943 | goto out; | 1899 | wait_log_commit(root, root->log_transid); |
1900 | mutex_unlock(&root->log_mutex); | ||
1901 | return 0; | ||
1944 | } | 1902 | } |
1945 | atomic_set(&log->fs_info->tree_log_commit, 1); | 1903 | atomic_set(&root->log_commit[index1], 1); |
1904 | |||
1905 | /* wait for previous tree log sync to complete */ | ||
1906 | if (atomic_read(&root->log_commit[(index1 + 1) % 2])) | ||
1907 | wait_log_commit(root, root->log_transid - 1); | ||
1946 | 1908 | ||
1947 | while (1) { | 1909 | while (1) { |
1948 | batch = log->fs_info->tree_log_batch; | 1910 | unsigned long batch = root->log_batch; |
1949 | mutex_unlock(&log->fs_info->tree_log_mutex); | 1911 | mutex_unlock(&root->log_mutex); |
1950 | schedule_timeout_uninterruptible(1); | 1912 | schedule_timeout_uninterruptible(1); |
1951 | mutex_lock(&log->fs_info->tree_log_mutex); | 1913 | mutex_lock(&root->log_mutex); |
1952 | 1914 | wait_for_writer(root); | |
1953 | while (atomic_read(&log->fs_info->tree_log_writers)) { | 1915 | if (batch == root->log_batch) |
1954 | DEFINE_WAIT(wait); | ||
1955 | prepare_to_wait(&log->fs_info->tree_log_wait, &wait, | ||
1956 | TASK_UNINTERRUPTIBLE); | ||
1957 | mutex_unlock(&log->fs_info->tree_log_mutex); | ||
1958 | if (atomic_read(&log->fs_info->tree_log_writers)) | ||
1959 | schedule(); | ||
1960 | mutex_lock(&log->fs_info->tree_log_mutex); | ||
1961 | finish_wait(&log->fs_info->tree_log_wait, &wait); | ||
1962 | } | ||
1963 | if (batch == log->fs_info->tree_log_batch) | ||
1964 | break; | 1916 | break; |
1965 | } | 1917 | } |
1966 | 1918 | ||
1967 | ret = btrfs_write_and_wait_marked_extents(log, &log->dirty_log_pages); | 1919 | ret = btrfs_write_and_wait_marked_extents(log, &log->dirty_log_pages); |
1968 | BUG_ON(ret); | 1920 | BUG_ON(ret); |
1969 | ret = btrfs_write_and_wait_marked_extents(root->fs_info->log_root_tree, | 1921 | |
1970 | &root->fs_info->log_root_tree->dirty_log_pages); | 1922 | btrfs_set_root_bytenr(&log->root_item, log->node->start); |
1923 | btrfs_set_root_generation(&log->root_item, trans->transid); | ||
1924 | btrfs_set_root_level(&log->root_item, btrfs_header_level(log->node)); | ||
1925 | |||
1926 | root->log_batch = 0; | ||
1927 | root->log_transid++; | ||
1928 | log->log_transid = root->log_transid; | ||
1929 | smp_mb(); | ||
1930 | /* | ||
1931 | * log tree has been flushed to disk, new modifications of | ||
1932 | * the log will be written to new positions. so it's safe to | ||
1933 | * allow log writers to go in. | ||
1934 | */ | ||
1935 | mutex_unlock(&root->log_mutex); | ||
1936 | |||
1937 | mutex_lock(&log_root_tree->log_mutex); | ||
1938 | log_root_tree->log_batch++; | ||
1939 | atomic_inc(&log_root_tree->log_writers); | ||
1940 | mutex_unlock(&log_root_tree->log_mutex); | ||
1941 | |||
1942 | ret = update_log_root(trans, log); | ||
1943 | BUG_ON(ret); | ||
1944 | |||
1945 | mutex_lock(&log_root_tree->log_mutex); | ||
1946 | if (atomic_dec_and_test(&log_root_tree->log_writers)) { | ||
1947 | smp_mb(); | ||
1948 | if (waitqueue_active(&log_root_tree->log_writer_wait)) | ||
1949 | wake_up(&log_root_tree->log_writer_wait); | ||
1950 | } | ||
1951 | |||
1952 | index2 = log_root_tree->log_transid % 2; | ||
1953 | if (atomic_read(&log_root_tree->log_commit[index2])) { | ||
1954 | wait_log_commit(log_root_tree, log_root_tree->log_transid); | ||
1955 | mutex_unlock(&log_root_tree->log_mutex); | ||
1956 | goto out; | ||
1957 | } | ||
1958 | atomic_set(&log_root_tree->log_commit[index2], 1); | ||
1959 | |||
1960 | if (atomic_read(&log_root_tree->log_commit[(index2 + 1) % 2])) | ||
1961 | wait_log_commit(log_root_tree, log_root_tree->log_transid - 1); | ||
1962 | |||
1963 | wait_for_writer(log_root_tree); | ||
1964 | |||
1965 | ret = btrfs_write_and_wait_marked_extents(log_root_tree, | ||
1966 | &log_root_tree->dirty_log_pages); | ||
1971 | BUG_ON(ret); | 1967 | BUG_ON(ret); |
1972 | 1968 | ||
1973 | btrfs_set_super_log_root(&root->fs_info->super_for_commit, | 1969 | btrfs_set_super_log_root(&root->fs_info->super_for_commit, |
1974 | log->fs_info->log_root_tree->node->start); | 1970 | log_root_tree->node->start); |
1975 | btrfs_set_super_log_root_level(&root->fs_info->super_for_commit, | 1971 | btrfs_set_super_log_root_level(&root->fs_info->super_for_commit, |
1976 | btrfs_header_level(log->fs_info->log_root_tree->node)); | 1972 | btrfs_header_level(log_root_tree->node)); |
1973 | |||
1974 | log_root_tree->log_batch = 0; | ||
1975 | log_root_tree->log_transid++; | ||
1976 | smp_mb(); | ||
1977 | |||
1978 | mutex_unlock(&log_root_tree->log_mutex); | ||
1979 | |||
1980 | /* | ||
1981 | * nobody else is going to jump in and write the the ctree | ||
1982 | * super here because the log_commit atomic below is protecting | ||
1983 | * us. We must be called with a transaction handle pinning | ||
1984 | * the running transaction open, so a full commit can't hop | ||
1985 | * in and cause problems either. | ||
1986 | */ | ||
1987 | write_ctree_super(trans, root->fs_info->tree_root, 2); | ||
1977 | 1988 | ||
1978 | write_ctree_super(trans, log->fs_info->tree_root, 2); | 1989 | atomic_set(&log_root_tree->log_commit[index2], 0); |
1979 | log->fs_info->tree_log_transid++; | ||
1980 | log->fs_info->tree_log_batch = 0; | ||
1981 | atomic_set(&log->fs_info->tree_log_commit, 0); | ||
1982 | smp_mb(); | 1990 | smp_mb(); |
1983 | if (waitqueue_active(&log->fs_info->tree_log_wait)) | 1991 | if (waitqueue_active(&log_root_tree->log_commit_wait[index2])) |
1984 | wake_up(&log->fs_info->tree_log_wait); | 1992 | wake_up(&log_root_tree->log_commit_wait[index2]); |
1985 | out: | 1993 | out: |
1986 | mutex_unlock(&log->fs_info->tree_log_mutex); | 1994 | atomic_set(&root->log_commit[index1], 0); |
1995 | smp_mb(); | ||
1996 | if (waitqueue_active(&root->log_commit_wait[index1])) | ||
1997 | wake_up(&root->log_commit_wait[index1]); | ||
1987 | return 0; | 1998 | return 0; |
1988 | } | 1999 | } |
1989 | 2000 | ||
@@ -2019,38 +2030,18 @@ int btrfs_free_log(struct btrfs_trans_handle *trans, struct btrfs_root *root) | |||
2019 | start, end, GFP_NOFS); | 2030 | start, end, GFP_NOFS); |
2020 | } | 2031 | } |
2021 | 2032 | ||
2022 | log = root->log_root; | 2033 | if (log->log_transid > 0) { |
2023 | ret = btrfs_del_root(trans, root->fs_info->log_root_tree, | 2034 | ret = btrfs_del_root(trans, root->fs_info->log_root_tree, |
2024 | &log->root_key); | 2035 | &log->root_key); |
2025 | BUG_ON(ret); | 2036 | BUG_ON(ret); |
2037 | } | ||
2026 | root->log_root = NULL; | 2038 | root->log_root = NULL; |
2027 | kfree(root->log_root); | 2039 | free_extent_buffer(log->node); |
2040 | kfree(log); | ||
2028 | return 0; | 2041 | return 0; |
2029 | } | 2042 | } |
2030 | 2043 | ||
2031 | /* | 2044 | /* |
2032 | * helper function to update the item for a given subvolumes log root | ||
2033 | * in the tree of log roots | ||
2034 | */ | ||
2035 | static int update_log_root(struct btrfs_trans_handle *trans, | ||
2036 | struct btrfs_root *log) | ||
2037 | { | ||
2038 | u64 bytenr = btrfs_root_bytenr(&log->root_item); | ||
2039 | int ret; | ||
2040 | |||
2041 | if (log->node->start == bytenr) | ||
2042 | return 0; | ||
2043 | |||
2044 | btrfs_set_root_bytenr(&log->root_item, log->node->start); | ||
2045 | btrfs_set_root_generation(&log->root_item, trans->transid); | ||
2046 | btrfs_set_root_level(&log->root_item, btrfs_header_level(log->node)); | ||
2047 | ret = btrfs_update_root(trans, log->fs_info->log_root_tree, | ||
2048 | &log->root_key, &log->root_item); | ||
2049 | BUG_ON(ret); | ||
2050 | return ret; | ||
2051 | } | ||
2052 | |||
2053 | /* | ||
2054 | * If both a file and directory are logged, and unlinks or renames are | 2045 | * If both a file and directory are logged, and unlinks or renames are |
2055 | * mixed in, we have a few interesting corners: | 2046 | * mixed in, we have a few interesting corners: |
2056 | * | 2047 | * |
@@ -2711,11 +2702,6 @@ next_slot: | |||
2711 | 2702 | ||
2712 | btrfs_free_path(path); | 2703 | btrfs_free_path(path); |
2713 | btrfs_free_path(dst_path); | 2704 | btrfs_free_path(dst_path); |
2714 | |||
2715 | mutex_lock(&root->fs_info->tree_log_mutex); | ||
2716 | ret = update_log_root(trans, log); | ||
2717 | BUG_ON(ret); | ||
2718 | mutex_unlock(&root->fs_info->tree_log_mutex); | ||
2719 | out: | 2705 | out: |
2720 | return 0; | 2706 | return 0; |
2721 | } | 2707 | } |
@@ -2846,7 +2832,9 @@ again: | |||
2846 | BUG_ON(!wc.replay_dest); | 2832 | BUG_ON(!wc.replay_dest); |
2847 | 2833 | ||
2848 | wc.replay_dest->log_root = log; | 2834 | wc.replay_dest->log_root = log; |
2835 | mutex_lock(&fs_info->trans_mutex); | ||
2849 | btrfs_record_root_in_trans(wc.replay_dest); | 2836 | btrfs_record_root_in_trans(wc.replay_dest); |
2837 | mutex_unlock(&fs_info->trans_mutex); | ||
2850 | ret = walk_log_tree(trans, log, &wc); | 2838 | ret = walk_log_tree(trans, log, &wc); |
2851 | BUG_ON(ret); | 2839 | BUG_ON(ret); |
2852 | 2840 | ||
diff --git a/fs/btrfs/volumes.c b/fs/btrfs/volumes.c index b187b537888e..1316139bf9e8 100644 --- a/fs/btrfs/volumes.c +++ b/fs/btrfs/volumes.c | |||
@@ -20,7 +20,6 @@ | |||
20 | #include <linux/buffer_head.h> | 20 | #include <linux/buffer_head.h> |
21 | #include <linux/blkdev.h> | 21 | #include <linux/blkdev.h> |
22 | #include <linux/random.h> | 22 | #include <linux/random.h> |
23 | #include <linux/version.h> | ||
24 | #include <asm/div64.h> | 23 | #include <asm/div64.h> |
25 | #include "compat.h" | 24 | #include "compat.h" |
26 | #include "ctree.h" | 25 | #include "ctree.h" |
@@ -104,10 +103,8 @@ static noinline struct btrfs_device *__find_device(struct list_head *head, | |||
104 | u64 devid, u8 *uuid) | 103 | u64 devid, u8 *uuid) |
105 | { | 104 | { |
106 | struct btrfs_device *dev; | 105 | struct btrfs_device *dev; |
107 | struct list_head *cur; | ||
108 | 106 | ||
109 | list_for_each(cur, head) { | 107 | list_for_each_entry(dev, head, dev_list) { |
110 | dev = list_entry(cur, struct btrfs_device, dev_list); | ||
111 | if (dev->devid == devid && | 108 | if (dev->devid == devid && |
112 | (!uuid || !memcmp(dev->uuid, uuid, BTRFS_UUID_SIZE))) { | 109 | (!uuid || !memcmp(dev->uuid, uuid, BTRFS_UUID_SIZE))) { |
113 | return dev; | 110 | return dev; |
@@ -118,11 +115,9 @@ static noinline struct btrfs_device *__find_device(struct list_head *head, | |||
118 | 115 | ||
119 | static noinline struct btrfs_fs_devices *find_fsid(u8 *fsid) | 116 | static noinline struct btrfs_fs_devices *find_fsid(u8 *fsid) |
120 | { | 117 | { |
121 | struct list_head *cur; | ||
122 | struct btrfs_fs_devices *fs_devices; | 118 | struct btrfs_fs_devices *fs_devices; |
123 | 119 | ||
124 | list_for_each(cur, &fs_uuids) { | 120 | list_for_each_entry(fs_devices, &fs_uuids, list) { |
125 | fs_devices = list_entry(cur, struct btrfs_fs_devices, list); | ||
126 | if (memcmp(fsid, fs_devices->fsid, BTRFS_FSID_SIZE) == 0) | 121 | if (memcmp(fsid, fs_devices->fsid, BTRFS_FSID_SIZE) == 0) |
127 | return fs_devices; | 122 | return fs_devices; |
128 | } | 123 | } |
@@ -159,6 +154,7 @@ static noinline int run_scheduled_bios(struct btrfs_device *device) | |||
159 | loop: | 154 | loop: |
160 | spin_lock(&device->io_lock); | 155 | spin_lock(&device->io_lock); |
161 | 156 | ||
157 | loop_lock: | ||
162 | /* take all the bios off the list at once and process them | 158 | /* take all the bios off the list at once and process them |
163 | * later on (without the lock held). But, remember the | 159 | * later on (without the lock held). But, remember the |
164 | * tail and other pointers so the bios can be properly reinserted | 160 | * tail and other pointers so the bios can be properly reinserted |
@@ -208,7 +204,7 @@ loop: | |||
208 | * is now congested. Back off and let other work structs | 204 | * is now congested. Back off and let other work structs |
209 | * run instead | 205 | * run instead |
210 | */ | 206 | */ |
211 | if (pending && bdi_write_congested(bdi) && | 207 | if (pending && bdi_write_congested(bdi) && num_run > 16 && |
212 | fs_info->fs_devices->open_devices > 1) { | 208 | fs_info->fs_devices->open_devices > 1) { |
213 | struct bio *old_head; | 209 | struct bio *old_head; |
214 | 210 | ||
@@ -221,6 +217,8 @@ loop: | |||
221 | else | 217 | else |
222 | device->pending_bio_tail = tail; | 218 | device->pending_bio_tail = tail; |
223 | 219 | ||
220 | device->running_pending = 1; | ||
221 | |||
224 | spin_unlock(&device->io_lock); | 222 | spin_unlock(&device->io_lock); |
225 | btrfs_requeue_work(&device->work); | 223 | btrfs_requeue_work(&device->work); |
226 | goto done; | 224 | goto done; |
@@ -228,6 +226,11 @@ loop: | |||
228 | } | 226 | } |
229 | if (again) | 227 | if (again) |
230 | goto loop; | 228 | goto loop; |
229 | |||
230 | spin_lock(&device->io_lock); | ||
231 | if (device->pending_bios) | ||
232 | goto loop_lock; | ||
233 | spin_unlock(&device->io_lock); | ||
231 | done: | 234 | done: |
232 | return 0; | 235 | return 0; |
233 | } | 236 | } |
@@ -344,14 +347,11 @@ error: | |||
344 | 347 | ||
345 | int btrfs_close_extra_devices(struct btrfs_fs_devices *fs_devices) | 348 | int btrfs_close_extra_devices(struct btrfs_fs_devices *fs_devices) |
346 | { | 349 | { |
347 | struct list_head *tmp; | 350 | struct btrfs_device *device, *next; |
348 | struct list_head *cur; | ||
349 | struct btrfs_device *device; | ||
350 | 351 | ||
351 | mutex_lock(&uuid_mutex); | 352 | mutex_lock(&uuid_mutex); |
352 | again: | 353 | again: |
353 | list_for_each_safe(cur, tmp, &fs_devices->devices) { | 354 | list_for_each_entry_safe(device, next, &fs_devices->devices, dev_list) { |
354 | device = list_entry(cur, struct btrfs_device, dev_list); | ||
355 | if (device->in_fs_metadata) | 355 | if (device->in_fs_metadata) |
356 | continue; | 356 | continue; |
357 | 357 | ||
@@ -382,14 +382,12 @@ again: | |||
382 | 382 | ||
383 | static int __btrfs_close_devices(struct btrfs_fs_devices *fs_devices) | 383 | static int __btrfs_close_devices(struct btrfs_fs_devices *fs_devices) |
384 | { | 384 | { |
385 | struct list_head *cur; | ||
386 | struct btrfs_device *device; | 385 | struct btrfs_device *device; |
387 | 386 | ||
388 | if (--fs_devices->opened > 0) | 387 | if (--fs_devices->opened > 0) |
389 | return 0; | 388 | return 0; |
390 | 389 | ||
391 | list_for_each(cur, &fs_devices->devices) { | 390 | list_for_each_entry(device, &fs_devices->devices, dev_list) { |
392 | device = list_entry(cur, struct btrfs_device, dev_list); | ||
393 | if (device->bdev) { | 391 | if (device->bdev) { |
394 | close_bdev_exclusive(device->bdev, device->mode); | 392 | close_bdev_exclusive(device->bdev, device->mode); |
395 | fs_devices->open_devices--; | 393 | fs_devices->open_devices--; |
@@ -438,7 +436,6 @@ static int __btrfs_open_devices(struct btrfs_fs_devices *fs_devices, | |||
438 | { | 436 | { |
439 | struct block_device *bdev; | 437 | struct block_device *bdev; |
440 | struct list_head *head = &fs_devices->devices; | 438 | struct list_head *head = &fs_devices->devices; |
441 | struct list_head *cur; | ||
442 | struct btrfs_device *device; | 439 | struct btrfs_device *device; |
443 | struct block_device *latest_bdev = NULL; | 440 | struct block_device *latest_bdev = NULL; |
444 | struct buffer_head *bh; | 441 | struct buffer_head *bh; |
@@ -449,8 +446,7 @@ static int __btrfs_open_devices(struct btrfs_fs_devices *fs_devices, | |||
449 | int seeding = 1; | 446 | int seeding = 1; |
450 | int ret = 0; | 447 | int ret = 0; |
451 | 448 | ||
452 | list_for_each(cur, head) { | 449 | list_for_each_entry(device, head, dev_list) { |
453 | device = list_entry(cur, struct btrfs_device, dev_list); | ||
454 | if (device->bdev) | 450 | if (device->bdev) |
455 | continue; | 451 | continue; |
456 | if (!device->name) | 452 | if (!device->name) |
@@ -577,7 +573,7 @@ int btrfs_scan_one_device(const char *path, fmode_t flags, void *holder, | |||
577 | *(unsigned long long *)disk_super->fsid, | 573 | *(unsigned long long *)disk_super->fsid, |
578 | *(unsigned long long *)(disk_super->fsid + 8)); | 574 | *(unsigned long long *)(disk_super->fsid + 8)); |
579 | } | 575 | } |
580 | printk(KERN_INFO "devid %llu transid %llu %s\n", | 576 | printk(KERN_CONT "devid %llu transid %llu %s\n", |
581 | (unsigned long long)devid, (unsigned long long)transid, path); | 577 | (unsigned long long)devid, (unsigned long long)transid, path); |
582 | ret = device_list_add(path, disk_super, devid, fs_devices_ret); | 578 | ret = device_list_add(path, disk_super, devid, fs_devices_ret); |
583 | 579 | ||
@@ -1016,14 +1012,12 @@ int btrfs_rm_device(struct btrfs_root *root, char *device_path) | |||
1016 | } | 1012 | } |
1017 | 1013 | ||
1018 | if (strcmp(device_path, "missing") == 0) { | 1014 | if (strcmp(device_path, "missing") == 0) { |
1019 | struct list_head *cur; | ||
1020 | struct list_head *devices; | 1015 | struct list_head *devices; |
1021 | struct btrfs_device *tmp; | 1016 | struct btrfs_device *tmp; |
1022 | 1017 | ||
1023 | device = NULL; | 1018 | device = NULL; |
1024 | devices = &root->fs_info->fs_devices->devices; | 1019 | devices = &root->fs_info->fs_devices->devices; |
1025 | list_for_each(cur, devices) { | 1020 | list_for_each_entry(tmp, devices, dev_list) { |
1026 | tmp = list_entry(cur, struct btrfs_device, dev_list); | ||
1027 | if (tmp->in_fs_metadata && !tmp->bdev) { | 1021 | if (tmp->in_fs_metadata && !tmp->bdev) { |
1028 | device = tmp; | 1022 | device = tmp; |
1029 | break; | 1023 | break; |
@@ -1279,7 +1273,6 @@ int btrfs_init_new_device(struct btrfs_root *root, char *device_path) | |||
1279 | struct btrfs_trans_handle *trans; | 1273 | struct btrfs_trans_handle *trans; |
1280 | struct btrfs_device *device; | 1274 | struct btrfs_device *device; |
1281 | struct block_device *bdev; | 1275 | struct block_device *bdev; |
1282 | struct list_head *cur; | ||
1283 | struct list_head *devices; | 1276 | struct list_head *devices; |
1284 | struct super_block *sb = root->fs_info->sb; | 1277 | struct super_block *sb = root->fs_info->sb; |
1285 | u64 total_bytes; | 1278 | u64 total_bytes; |
@@ -1303,8 +1296,7 @@ int btrfs_init_new_device(struct btrfs_root *root, char *device_path) | |||
1303 | mutex_lock(&root->fs_info->volume_mutex); | 1296 | mutex_lock(&root->fs_info->volume_mutex); |
1304 | 1297 | ||
1305 | devices = &root->fs_info->fs_devices->devices; | 1298 | devices = &root->fs_info->fs_devices->devices; |
1306 | list_for_each(cur, devices) { | 1299 | list_for_each_entry(device, devices, dev_list) { |
1307 | device = list_entry(cur, struct btrfs_device, dev_list); | ||
1308 | if (device->bdev == bdev) { | 1300 | if (device->bdev == bdev) { |
1309 | ret = -EEXIST; | 1301 | ret = -EEXIST; |
1310 | goto error; | 1302 | goto error; |
@@ -1703,7 +1695,6 @@ static u64 div_factor(u64 num, int factor) | |||
1703 | int btrfs_balance(struct btrfs_root *dev_root) | 1695 | int btrfs_balance(struct btrfs_root *dev_root) |
1704 | { | 1696 | { |
1705 | int ret; | 1697 | int ret; |
1706 | struct list_head *cur; | ||
1707 | struct list_head *devices = &dev_root->fs_info->fs_devices->devices; | 1698 | struct list_head *devices = &dev_root->fs_info->fs_devices->devices; |
1708 | struct btrfs_device *device; | 1699 | struct btrfs_device *device; |
1709 | u64 old_size; | 1700 | u64 old_size; |
@@ -1722,8 +1713,7 @@ int btrfs_balance(struct btrfs_root *dev_root) | |||
1722 | dev_root = dev_root->fs_info->dev_root; | 1713 | dev_root = dev_root->fs_info->dev_root; |
1723 | 1714 | ||
1724 | /* step one make some room on all the devices */ | 1715 | /* step one make some room on all the devices */ |
1725 | list_for_each(cur, devices) { | 1716 | list_for_each_entry(device, devices, dev_list) { |
1726 | device = list_entry(cur, struct btrfs_device, dev_list); | ||
1727 | old_size = device->total_bytes; | 1717 | old_size = device->total_bytes; |
1728 | size_to_free = div_factor(old_size, 1); | 1718 | size_to_free = div_factor(old_size, 1); |
1729 | size_to_free = min(size_to_free, (u64)1 * 1024 * 1024); | 1719 | size_to_free = min(size_to_free, (u64)1 * 1024 * 1024); |
@@ -2904,10 +2894,6 @@ static int read_one_chunk(struct btrfs_root *root, struct btrfs_key *key, | |||
2904 | free_extent_map(em); | 2894 | free_extent_map(em); |
2905 | } | 2895 | } |
2906 | 2896 | ||
2907 | map = kzalloc(sizeof(*map), GFP_NOFS); | ||
2908 | if (!map) | ||
2909 | return -ENOMEM; | ||
2910 | |||
2911 | em = alloc_extent_map(GFP_NOFS); | 2897 | em = alloc_extent_map(GFP_NOFS); |
2912 | if (!em) | 2898 | if (!em) |
2913 | return -ENOMEM; | 2899 | return -ENOMEM; |
@@ -3116,6 +3102,8 @@ int btrfs_read_sys_array(struct btrfs_root *root) | |||
3116 | if (!sb) | 3102 | if (!sb) |
3117 | return -ENOMEM; | 3103 | return -ENOMEM; |
3118 | btrfs_set_buffer_uptodate(sb); | 3104 | btrfs_set_buffer_uptodate(sb); |
3105 | btrfs_set_buffer_lockdep_class(sb, 0); | ||
3106 | |||
3119 | write_extent_buffer(sb, super_copy, 0, BTRFS_SUPER_INFO_SIZE); | 3107 | write_extent_buffer(sb, super_copy, 0, BTRFS_SUPER_INFO_SIZE); |
3120 | array_size = btrfs_super_sys_array_size(super_copy); | 3108 | array_size = btrfs_super_sys_array_size(super_copy); |
3121 | 3109 | ||
diff --git a/fs/btrfs/xattr.c b/fs/btrfs/xattr.c index 7f332e270894..a9d3bf4d2689 100644 --- a/fs/btrfs/xattr.c +++ b/fs/btrfs/xattr.c | |||
@@ -21,6 +21,7 @@ | |||
21 | #include <linux/slab.h> | 21 | #include <linux/slab.h> |
22 | #include <linux/rwsem.h> | 22 | #include <linux/rwsem.h> |
23 | #include <linux/xattr.h> | 23 | #include <linux/xattr.h> |
24 | #include <linux/security.h> | ||
24 | #include "ctree.h" | 25 | #include "ctree.h" |
25 | #include "btrfs_inode.h" | 26 | #include "btrfs_inode.h" |
26 | #include "transaction.h" | 27 | #include "transaction.h" |
@@ -45,9 +46,12 @@ ssize_t __btrfs_getxattr(struct inode *inode, const char *name, | |||
45 | /* lookup the xattr by name */ | 46 | /* lookup the xattr by name */ |
46 | di = btrfs_lookup_xattr(NULL, root, path, inode->i_ino, name, | 47 | di = btrfs_lookup_xattr(NULL, root, path, inode->i_ino, name, |
47 | strlen(name), 0); | 48 | strlen(name), 0); |
48 | if (!di || IS_ERR(di)) { | 49 | if (!di) { |
49 | ret = -ENODATA; | 50 | ret = -ENODATA; |
50 | goto out; | 51 | goto out; |
52 | } else if (IS_ERR(di)) { | ||
53 | ret = PTR_ERR(di); | ||
54 | goto out; | ||
51 | } | 55 | } |
52 | 56 | ||
53 | leaf = path->nodes[0]; | 57 | leaf = path->nodes[0]; |
@@ -62,6 +66,14 @@ ssize_t __btrfs_getxattr(struct inode *inode, const char *name, | |||
62 | ret = -ERANGE; | 66 | ret = -ERANGE; |
63 | goto out; | 67 | goto out; |
64 | } | 68 | } |
69 | |||
70 | /* | ||
71 | * The way things are packed into the leaf is like this | ||
72 | * |struct btrfs_dir_item|name|data| | ||
73 | * where name is the xattr name, so security.foo, and data is the | ||
74 | * content of the xattr. data_ptr points to the location in memory | ||
75 | * where the data starts in the in memory leaf | ||
76 | */ | ||
65 | data_ptr = (unsigned long)((char *)(di + 1) + | 77 | data_ptr = (unsigned long)((char *)(di + 1) + |
66 | btrfs_dir_name_len(leaf, di)); | 78 | btrfs_dir_name_len(leaf, di)); |
67 | read_extent_buffer(leaf, buffer, data_ptr, | 79 | read_extent_buffer(leaf, buffer, data_ptr, |
@@ -86,7 +98,7 @@ int __btrfs_setxattr(struct inode *inode, const char *name, | |||
86 | if (!path) | 98 | if (!path) |
87 | return -ENOMEM; | 99 | return -ENOMEM; |
88 | 100 | ||
89 | trans = btrfs_start_transaction(root, 1); | 101 | trans = btrfs_join_transaction(root, 1); |
90 | btrfs_set_trans_block_group(trans, inode); | 102 | btrfs_set_trans_block_group(trans, inode); |
91 | 103 | ||
92 | /* first lets see if we already have this xattr */ | 104 | /* first lets see if we already have this xattr */ |
@@ -176,7 +188,6 @@ ssize_t btrfs_listxattr(struct dentry *dentry, char *buffer, size_t size) | |||
176 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); | 188 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); |
177 | if (ret < 0) | 189 | if (ret < 0) |
178 | goto err; | 190 | goto err; |
179 | ret = 0; | ||
180 | advance = 0; | 191 | advance = 0; |
181 | while (1) { | 192 | while (1) { |
182 | leaf = path->nodes[0]; | 193 | leaf = path->nodes[0]; |
@@ -320,3 +331,34 @@ int btrfs_removexattr(struct dentry *dentry, const char *name) | |||
320 | return -EOPNOTSUPP; | 331 | return -EOPNOTSUPP; |
321 | return __btrfs_setxattr(dentry->d_inode, name, NULL, 0, XATTR_REPLACE); | 332 | return __btrfs_setxattr(dentry->d_inode, name, NULL, 0, XATTR_REPLACE); |
322 | } | 333 | } |
334 | |||
335 | int btrfs_xattr_security_init(struct inode *inode, struct inode *dir) | ||
336 | { | ||
337 | int err; | ||
338 | size_t len; | ||
339 | void *value; | ||
340 | char *suffix; | ||
341 | char *name; | ||
342 | |||
343 | err = security_inode_init_security(inode, dir, &suffix, &value, &len); | ||
344 | if (err) { | ||
345 | if (err == -EOPNOTSUPP) | ||
346 | return 0; | ||
347 | return err; | ||
348 | } | ||
349 | |||
350 | name = kmalloc(XATTR_SECURITY_PREFIX_LEN + strlen(suffix) + 1, | ||
351 | GFP_NOFS); | ||
352 | if (!name) { | ||
353 | err = -ENOMEM; | ||
354 | } else { | ||
355 | strcpy(name, XATTR_SECURITY_PREFIX); | ||
356 | strcpy(name + XATTR_SECURITY_PREFIX_LEN, suffix); | ||
357 | err = __btrfs_setxattr(inode, name, value, len, 0); | ||
358 | kfree(name); | ||
359 | } | ||
360 | |||
361 | kfree(suffix); | ||
362 | kfree(value); | ||
363 | return err; | ||
364 | } | ||
diff --git a/fs/btrfs/xattr.h b/fs/btrfs/xattr.h index 5b1d08f8e68d..c71e9c3cf3f7 100644 --- a/fs/btrfs/xattr.h +++ b/fs/btrfs/xattr.h | |||
@@ -36,4 +36,6 @@ extern int btrfs_setxattr(struct dentry *dentry, const char *name, | |||
36 | const void *value, size_t size, int flags); | 36 | const void *value, size_t size, int flags); |
37 | extern int btrfs_removexattr(struct dentry *dentry, const char *name); | 37 | extern int btrfs_removexattr(struct dentry *dentry, const char *name); |
38 | 38 | ||
39 | extern int btrfs_xattr_security_init(struct inode *inode, struct inode *dir); | ||
40 | |||
39 | #endif /* __XATTR__ */ | 41 | #endif /* __XATTR__ */ |
diff --git a/fs/buffer.c b/fs/buffer.c index b6e8b8632e2f..9f697419ed8e 100644 --- a/fs/buffer.c +++ b/fs/buffer.c | |||
@@ -777,6 +777,7 @@ static int __set_page_dirty(struct page *page, | |||
777 | __inc_zone_page_state(page, NR_FILE_DIRTY); | 777 | __inc_zone_page_state(page, NR_FILE_DIRTY); |
778 | __inc_bdi_stat(mapping->backing_dev_info, | 778 | __inc_bdi_stat(mapping->backing_dev_info, |
779 | BDI_RECLAIMABLE); | 779 | BDI_RECLAIMABLE); |
780 | task_dirty_inc(current); | ||
780 | task_io_account_write(PAGE_CACHE_SIZE); | 781 | task_io_account_write(PAGE_CACHE_SIZE); |
781 | } | 782 | } |
782 | radix_tree_tag_set(&mapping->page_tree, | 783 | radix_tree_tag_set(&mapping->page_tree, |
@@ -2688,7 +2689,7 @@ int nobh_write_end(struct file *file, struct address_space *mapping, | |||
2688 | struct buffer_head *bh; | 2689 | struct buffer_head *bh; |
2689 | BUG_ON(fsdata != NULL && page_has_buffers(page)); | 2690 | BUG_ON(fsdata != NULL && page_has_buffers(page)); |
2690 | 2691 | ||
2691 | if (unlikely(copied < len) && !page_has_buffers(page)) | 2692 | if (unlikely(copied < len) && head) |
2692 | attach_nobh_buffers(page, head); | 2693 | attach_nobh_buffers(page, head); |
2693 | if (page_has_buffers(page)) | 2694 | if (page_has_buffers(page)) |
2694 | return generic_write_end(file, mapping, pos, len, | 2695 | return generic_write_end(file, mapping, pos, len, |
@@ -3108,7 +3109,7 @@ int sync_dirty_buffer(struct buffer_head *bh) | |||
3108 | if (test_clear_buffer_dirty(bh)) { | 3109 | if (test_clear_buffer_dirty(bh)) { |
3109 | get_bh(bh); | 3110 | get_bh(bh); |
3110 | bh->b_end_io = end_buffer_write_sync; | 3111 | bh->b_end_io = end_buffer_write_sync; |
3111 | ret = submit_bh(WRITE_SYNC, bh); | 3112 | ret = submit_bh(WRITE, bh); |
3112 | wait_on_buffer(bh); | 3113 | wait_on_buffer(bh); |
3113 | if (buffer_eopnotsupp(bh)) { | 3114 | if (buffer_eopnotsupp(bh)) { |
3114 | clear_buffer_eopnotsupp(bh); | 3115 | clear_buffer_eopnotsupp(bh); |
@@ -3243,7 +3244,7 @@ void block_sync_page(struct page *page) | |||
3243 | * Use of bdflush() is deprecated and will be removed in a future kernel. | 3244 | * Use of bdflush() is deprecated and will be removed in a future kernel. |
3244 | * The `pdflush' kernel threads fully replace bdflush daemons and this call. | 3245 | * The `pdflush' kernel threads fully replace bdflush daemons and this call. |
3245 | */ | 3246 | */ |
3246 | asmlinkage long sys_bdflush(int func, long data) | 3247 | SYSCALL_DEFINE2(bdflush, int, func, long, data) |
3247 | { | 3248 | { |
3248 | static int msg_count; | 3249 | static int msg_count; |
3249 | 3250 | ||
diff --git a/fs/cifs/CHANGES b/fs/cifs/CHANGES index 080703a15f44..851388fafc73 100644 --- a/fs/cifs/CHANGES +++ b/fs/cifs/CHANGES | |||
@@ -1,3 +1,13 @@ | |||
1 | Version 1.57 | ||
2 | ------------ | ||
3 | Improve support for multiple security contexts to the same server. We | ||
4 | used to use the same "vcnumber" for all connections which could cause | ||
5 | the server to treat subsequent connections, especially those that | ||
6 | are authenticated as guest, as reconnections, invalidating the earlier | ||
7 | user's smb session. This fix allows cifs to mount multiple times to the | ||
8 | same server with different userids without risking invalidating earlier | ||
9 | established security contexts. | ||
10 | |||
1 | Version 1.56 | 11 | Version 1.56 |
2 | ------------ | 12 | ------------ |
3 | Add "forcemandatorylock" mount option to allow user to use mandatory | 13 | Add "forcemandatorylock" mount option to allow user to use mandatory |
@@ -5,7 +15,12 @@ rather than posix (advisory) byte range locks, even though server would | |||
5 | support posix byte range locks. Fix query of root inode when prefixpath | 15 | support posix byte range locks. Fix query of root inode when prefixpath |
6 | specified and user does not have access to query information about the | 16 | specified and user does not have access to query information about the |
7 | top of the share. Fix problem in 2.6.28 resolving DFS paths to | 17 | top of the share. Fix problem in 2.6.28 resolving DFS paths to |
8 | Samba servers (worked to Windows). | 18 | Samba servers (worked to Windows). Fix rmdir so that pending search |
19 | (readdir) requests do not get invalid results which include the now | ||
20 | removed directory. Fix oops in cifs_dfs_ref.c when prefixpath is not reachable | ||
21 | when using DFS. Add better file create support to servers which support | ||
22 | the CIFS POSIX protocol extensions (this adds support for new flags | ||
23 | on create, and improves semantics for write of locked ranges). | ||
9 | 24 | ||
10 | Version 1.55 | 25 | Version 1.55 |
11 | ------------ | 26 | ------------ |
diff --git a/fs/cifs/cifsencrypt.c b/fs/cifs/cifsencrypt.c index d4839cf0cb2c..7c9809523f42 100644 --- a/fs/cifs/cifsencrypt.c +++ b/fs/cifs/cifsencrypt.c | |||
@@ -48,11 +48,11 @@ static int cifs_calculate_signature(const struct smb_hdr *cifs_pdu, | |||
48 | if ((cifs_pdu == NULL) || (signature == NULL) || (key == NULL)) | 48 | if ((cifs_pdu == NULL) || (signature == NULL) || (key == NULL)) |
49 | return -EINVAL; | 49 | return -EINVAL; |
50 | 50 | ||
51 | MD5Init(&context); | 51 | cifs_MD5_init(&context); |
52 | MD5Update(&context, (char *)&key->data, key->len); | 52 | cifs_MD5_update(&context, (char *)&key->data, key->len); |
53 | MD5Update(&context, cifs_pdu->Protocol, cifs_pdu->smb_buf_length); | 53 | cifs_MD5_update(&context, cifs_pdu->Protocol, cifs_pdu->smb_buf_length); |
54 | 54 | ||
55 | MD5Final(signature, &context); | 55 | cifs_MD5_final(signature, &context); |
56 | return 0; | 56 | return 0; |
57 | } | 57 | } |
58 | 58 | ||
@@ -96,8 +96,8 @@ static int cifs_calc_signature2(const struct kvec *iov, int n_vec, | |||
96 | if ((iov == NULL) || (signature == NULL) || (key == NULL)) | 96 | if ((iov == NULL) || (signature == NULL) || (key == NULL)) |
97 | return -EINVAL; | 97 | return -EINVAL; |
98 | 98 | ||
99 | MD5Init(&context); | 99 | cifs_MD5_init(&context); |
100 | MD5Update(&context, (char *)&key->data, key->len); | 100 | cifs_MD5_update(&context, (char *)&key->data, key->len); |
101 | for (i = 0; i < n_vec; i++) { | 101 | for (i = 0; i < n_vec; i++) { |
102 | if (iov[i].iov_len == 0) | 102 | if (iov[i].iov_len == 0) |
103 | continue; | 103 | continue; |
@@ -110,13 +110,13 @@ static int cifs_calc_signature2(const struct kvec *iov, int n_vec, | |||
110 | if (i == 0) { | 110 | if (i == 0) { |
111 | if (iov[0].iov_len <= 8) /* cmd field at offset 9 */ | 111 | if (iov[0].iov_len <= 8) /* cmd field at offset 9 */ |
112 | break; /* nothing to sign or corrupt header */ | 112 | break; /* nothing to sign or corrupt header */ |
113 | MD5Update(&context, iov[0].iov_base+4, | 113 | cifs_MD5_update(&context, iov[0].iov_base+4, |
114 | iov[0].iov_len-4); | 114 | iov[0].iov_len-4); |
115 | } else | 115 | } else |
116 | MD5Update(&context, iov[i].iov_base, iov[i].iov_len); | 116 | cifs_MD5_update(&context, iov[i].iov_base, iov[i].iov_len); |
117 | } | 117 | } |
118 | 118 | ||
119 | MD5Final(signature, &context); | 119 | cifs_MD5_final(signature, &context); |
120 | 120 | ||
121 | return 0; | 121 | return 0; |
122 | } | 122 | } |
diff --git a/fs/cifs/cifsfs.h b/fs/cifs/cifsfs.h index 7ac481841f87..2b1d28a9ee28 100644 --- a/fs/cifs/cifsfs.h +++ b/fs/cifs/cifsfs.h | |||
@@ -100,5 +100,5 @@ extern long cifs_ioctl(struct file *filep, unsigned int cmd, unsigned long arg); | |||
100 | extern const struct export_operations cifs_export_ops; | 100 | extern const struct export_operations cifs_export_ops; |
101 | #endif /* EXPERIMENTAL */ | 101 | #endif /* EXPERIMENTAL */ |
102 | 102 | ||
103 | #define CIFS_VERSION "1.56" | 103 | #define CIFS_VERSION "1.57" |
104 | #endif /* _CIFSFS_H */ | 104 | #endif /* _CIFSFS_H */ |
diff --git a/fs/cifs/cifsglob.h b/fs/cifs/cifsglob.h index 94c1ca0ec953..e004f6db5fc8 100644 --- a/fs/cifs/cifsglob.h +++ b/fs/cifs/cifsglob.h | |||
@@ -164,9 +164,12 @@ struct TCP_Server_Info { | |||
164 | /* multiplexed reads or writes */ | 164 | /* multiplexed reads or writes */ |
165 | unsigned int maxBuf; /* maxBuf specifies the maximum */ | 165 | unsigned int maxBuf; /* maxBuf specifies the maximum */ |
166 | /* message size the server can send or receive for non-raw SMBs */ | 166 | /* message size the server can send or receive for non-raw SMBs */ |
167 | unsigned int maxRw; /* maxRw specifies the maximum */ | 167 | unsigned int max_rw; /* maxRw specifies the maximum */ |
168 | /* message size the server can send or receive for */ | 168 | /* message size the server can send or receive for */ |
169 | /* SMB_COM_WRITE_RAW or SMB_COM_READ_RAW. */ | 169 | /* SMB_COM_WRITE_RAW or SMB_COM_READ_RAW. */ |
170 | unsigned int max_vcs; /* maximum number of smb sessions, at least | ||
171 | those that can be specified uniquely with | ||
172 | vcnumbers */ | ||
170 | char sessid[4]; /* unique token id for this session */ | 173 | char sessid[4]; /* unique token id for this session */ |
171 | /* (returned on Negotiate */ | 174 | /* (returned on Negotiate */ |
172 | int capabilities; /* allow selective disabling of caps by smb sess */ | 175 | int capabilities; /* allow selective disabling of caps by smb sess */ |
@@ -210,6 +213,7 @@ struct cifsSesInfo { | |||
210 | unsigned overrideSecFlg; /* if non-zero override global sec flags */ | 213 | unsigned overrideSecFlg; /* if non-zero override global sec flags */ |
211 | __u16 ipc_tid; /* special tid for connection to IPC share */ | 214 | __u16 ipc_tid; /* special tid for connection to IPC share */ |
212 | __u16 flags; | 215 | __u16 flags; |
216 | __u16 vcnum; | ||
213 | char *serverOS; /* name of operating system underlying server */ | 217 | char *serverOS; /* name of operating system underlying server */ |
214 | char *serverNOS; /* name of network operating system of server */ | 218 | char *serverNOS; /* name of network operating system of server */ |
215 | char *serverDomain; /* security realm of server */ | 219 | char *serverDomain; /* security realm of server */ |
diff --git a/fs/cifs/cifsproto.h b/fs/cifs/cifsproto.h index 06f6779988bf..083dfc57c7a3 100644 --- a/fs/cifs/cifsproto.h +++ b/fs/cifs/cifsproto.h | |||
@@ -35,13 +35,14 @@ extern struct smb_hdr *cifs_buf_get(void); | |||
35 | extern void cifs_buf_release(void *); | 35 | extern void cifs_buf_release(void *); |
36 | extern struct smb_hdr *cifs_small_buf_get(void); | 36 | extern struct smb_hdr *cifs_small_buf_get(void); |
37 | extern void cifs_small_buf_release(void *); | 37 | extern void cifs_small_buf_release(void *); |
38 | extern int smb_send(struct socket *, struct smb_hdr *, | 38 | extern int smb_send(struct TCP_Server_Info *, struct smb_hdr *, |
39 | unsigned int /* length */ , struct sockaddr *, bool); | 39 | unsigned int /* length */); |
40 | extern unsigned int _GetXid(void); | 40 | extern unsigned int _GetXid(void); |
41 | extern void _FreeXid(unsigned int); | 41 | extern void _FreeXid(unsigned int); |
42 | #define GetXid() (int)_GetXid(); cFYI(1,("CIFS VFS: in %s as Xid: %d with uid: %d",__func__, xid,current_fsuid())); | 42 | #define GetXid() (int)_GetXid(); cFYI(1,("CIFS VFS: in %s as Xid: %d with uid: %d",__func__, xid,current_fsuid())); |
43 | #define FreeXid(curr_xid) {_FreeXid(curr_xid); cFYI(1,("CIFS VFS: leaving %s (xid = %d) rc = %d",__func__,curr_xid,(int)rc));} | 43 | #define FreeXid(curr_xid) {_FreeXid(curr_xid); cFYI(1,("CIFS VFS: leaving %s (xid = %d) rc = %d",__func__,curr_xid,(int)rc));} |
44 | extern char *build_path_from_dentry(struct dentry *); | 44 | extern char *build_path_from_dentry(struct dentry *); |
45 | extern char *cifs_build_path_to_root(struct cifs_sb_info *cifs_sb); | ||
45 | extern char *build_wildcard_path_from_dentry(struct dentry *direntry); | 46 | extern char *build_wildcard_path_from_dentry(struct dentry *direntry); |
46 | /* extern void renew_parental_timestamps(struct dentry *direntry);*/ | 47 | /* extern void renew_parental_timestamps(struct dentry *direntry);*/ |
47 | extern int SendReceive(const unsigned int /* xid */ , struct cifsSesInfo *, | 48 | extern int SendReceive(const unsigned int /* xid */ , struct cifsSesInfo *, |
@@ -91,6 +92,9 @@ extern u64 cifs_UnixTimeToNT(struct timespec); | |||
91 | extern __le64 cnvrtDosCifsTm(__u16 date, __u16 time); | 92 | extern __le64 cnvrtDosCifsTm(__u16 date, __u16 time); |
92 | extern struct timespec cnvrtDosUnixTm(__u16 date, __u16 time); | 93 | extern struct timespec cnvrtDosUnixTm(__u16 date, __u16 time); |
93 | 94 | ||
95 | extern void posix_fill_in_inode(struct inode *tmp_inode, | ||
96 | FILE_UNIX_BASIC_INFO *pData, int isNewInode); | ||
97 | extern struct inode *cifs_new_inode(struct super_block *sb, __u64 *inum); | ||
94 | extern int cifs_get_inode_info(struct inode **pinode, | 98 | extern int cifs_get_inode_info(struct inode **pinode, |
95 | const unsigned char *search_path, | 99 | const unsigned char *search_path, |
96 | FILE_ALL_INFO *pfile_info, | 100 | FILE_ALL_INFO *pfile_info, |
diff --git a/fs/cifs/cifssmb.c b/fs/cifs/cifssmb.c index 552642a507c4..939e2f76b959 100644 --- a/fs/cifs/cifssmb.c +++ b/fs/cifs/cifssmb.c | |||
@@ -528,14 +528,15 @@ CIFSSMBNegotiate(unsigned int xid, struct cifsSesInfo *ses) | |||
528 | server->maxReq = le16_to_cpu(rsp->MaxMpxCount); | 528 | server->maxReq = le16_to_cpu(rsp->MaxMpxCount); |
529 | server->maxBuf = min((__u32)le16_to_cpu(rsp->MaxBufSize), | 529 | server->maxBuf = min((__u32)le16_to_cpu(rsp->MaxBufSize), |
530 | (__u32)CIFSMaxBufSize + MAX_CIFS_HDR_SIZE); | 530 | (__u32)CIFSMaxBufSize + MAX_CIFS_HDR_SIZE); |
531 | server->max_vcs = le16_to_cpu(rsp->MaxNumberVcs); | ||
531 | GETU32(server->sessid) = le32_to_cpu(rsp->SessionKey); | 532 | GETU32(server->sessid) = le32_to_cpu(rsp->SessionKey); |
532 | /* even though we do not use raw we might as well set this | 533 | /* even though we do not use raw we might as well set this |
533 | accurately, in case we ever find a need for it */ | 534 | accurately, in case we ever find a need for it */ |
534 | if ((le16_to_cpu(rsp->RawMode) & RAW_ENABLE) == RAW_ENABLE) { | 535 | if ((le16_to_cpu(rsp->RawMode) & RAW_ENABLE) == RAW_ENABLE) { |
535 | server->maxRw = 0xFF00; | 536 | server->max_rw = 0xFF00; |
536 | server->capabilities = CAP_MPX_MODE | CAP_RAW_MODE; | 537 | server->capabilities = CAP_MPX_MODE | CAP_RAW_MODE; |
537 | } else { | 538 | } else { |
538 | server->maxRw = 0;/* we do not need to use raw anyway */ | 539 | server->max_rw = 0;/* do not need to use raw anyway */ |
539 | server->capabilities = CAP_MPX_MODE; | 540 | server->capabilities = CAP_MPX_MODE; |
540 | } | 541 | } |
541 | tmp = (__s16)le16_to_cpu(rsp->ServerTimeZone); | 542 | tmp = (__s16)le16_to_cpu(rsp->ServerTimeZone); |
@@ -638,7 +639,7 @@ CIFSSMBNegotiate(unsigned int xid, struct cifsSesInfo *ses) | |||
638 | /* probably no need to store and check maxvcs */ | 639 | /* probably no need to store and check maxvcs */ |
639 | server->maxBuf = min(le32_to_cpu(pSMBr->MaxBufferSize), | 640 | server->maxBuf = min(le32_to_cpu(pSMBr->MaxBufferSize), |
640 | (__u32) CIFSMaxBufSize + MAX_CIFS_HDR_SIZE); | 641 | (__u32) CIFSMaxBufSize + MAX_CIFS_HDR_SIZE); |
641 | server->maxRw = le32_to_cpu(pSMBr->MaxRawSize); | 642 | server->max_rw = le32_to_cpu(pSMBr->MaxRawSize); |
642 | cFYI(DBG2, ("Max buf = %d", ses->server->maxBuf)); | 643 | cFYI(DBG2, ("Max buf = %d", ses->server->maxBuf)); |
643 | GETU32(ses->server->sessid) = le32_to_cpu(pSMBr->SessionKey); | 644 | GETU32(ses->server->sessid) = le32_to_cpu(pSMBr->SessionKey); |
644 | server->capabilities = le32_to_cpu(pSMBr->Capabilities); | 645 | server->capabilities = le32_to_cpu(pSMBr->Capabilities); |
diff --git a/fs/cifs/connect.c b/fs/cifs/connect.c index e9ea394ee075..da0f4ffa0613 100644 --- a/fs/cifs/connect.c +++ b/fs/cifs/connect.c | |||
@@ -23,7 +23,6 @@ | |||
23 | #include <linux/string.h> | 23 | #include <linux/string.h> |
24 | #include <linux/list.h> | 24 | #include <linux/list.h> |
25 | #include <linux/wait.h> | 25 | #include <linux/wait.h> |
26 | #include <linux/ipv6.h> | ||
27 | #include <linux/pagemap.h> | 26 | #include <linux/pagemap.h> |
28 | #include <linux/ctype.h> | 27 | #include <linux/ctype.h> |
29 | #include <linux/utsname.h> | 28 | #include <linux/utsname.h> |
@@ -35,6 +34,7 @@ | |||
35 | #include <linux/freezer.h> | 34 | #include <linux/freezer.h> |
36 | #include <asm/uaccess.h> | 35 | #include <asm/uaccess.h> |
37 | #include <asm/processor.h> | 36 | #include <asm/processor.h> |
37 | #include <net/ipv6.h> | ||
38 | #include "cifspdu.h" | 38 | #include "cifspdu.h" |
39 | #include "cifsglob.h" | 39 | #include "cifsglob.h" |
40 | #include "cifsproto.h" | 40 | #include "cifsproto.h" |
@@ -1354,7 +1354,7 @@ cifs_parse_mount_options(char *options, const char *devname, | |||
1354 | } | 1354 | } |
1355 | 1355 | ||
1356 | static struct TCP_Server_Info * | 1356 | static struct TCP_Server_Info * |
1357 | cifs_find_tcp_session(struct sockaddr *addr) | 1357 | cifs_find_tcp_session(struct sockaddr_storage *addr) |
1358 | { | 1358 | { |
1359 | struct list_head *tmp; | 1359 | struct list_head *tmp; |
1360 | struct TCP_Server_Info *server; | 1360 | struct TCP_Server_Info *server; |
@@ -1374,13 +1374,13 @@ cifs_find_tcp_session(struct sockaddr *addr) | |||
1374 | if (server->tcpStatus == CifsNew) | 1374 | if (server->tcpStatus == CifsNew) |
1375 | continue; | 1375 | continue; |
1376 | 1376 | ||
1377 | if (addr->sa_family == AF_INET && | 1377 | if (addr->ss_family == AF_INET && |
1378 | (addr4->sin_addr.s_addr != | 1378 | (addr4->sin_addr.s_addr != |
1379 | server->addr.sockAddr.sin_addr.s_addr)) | 1379 | server->addr.sockAddr.sin_addr.s_addr)) |
1380 | continue; | 1380 | continue; |
1381 | else if (addr->sa_family == AF_INET6 && | 1381 | else if (addr->ss_family == AF_INET6 && |
1382 | memcmp(&server->addr.sockAddr6.sin6_addr, | 1382 | !ipv6_addr_equal(&server->addr.sockAddr6.sin6_addr, |
1383 | &addr6->sin6_addr, sizeof(addr6->sin6_addr))) | 1383 | &addr6->sin6_addr)) |
1384 | continue; | 1384 | continue; |
1385 | 1385 | ||
1386 | ++server->srv_count; | 1386 | ++server->srv_count; |
@@ -1419,12 +1419,12 @@ static struct TCP_Server_Info * | |||
1419 | cifs_get_tcp_session(struct smb_vol *volume_info) | 1419 | cifs_get_tcp_session(struct smb_vol *volume_info) |
1420 | { | 1420 | { |
1421 | struct TCP_Server_Info *tcp_ses = NULL; | 1421 | struct TCP_Server_Info *tcp_ses = NULL; |
1422 | struct sockaddr addr; | 1422 | struct sockaddr_storage addr; |
1423 | struct sockaddr_in *sin_server = (struct sockaddr_in *) &addr; | 1423 | struct sockaddr_in *sin_server = (struct sockaddr_in *) &addr; |
1424 | struct sockaddr_in6 *sin_server6 = (struct sockaddr_in6 *) &addr; | 1424 | struct sockaddr_in6 *sin_server6 = (struct sockaddr_in6 *) &addr; |
1425 | int rc; | 1425 | int rc; |
1426 | 1426 | ||
1427 | memset(&addr, 0, sizeof(struct sockaddr)); | 1427 | memset(&addr, 0, sizeof(struct sockaddr_storage)); |
1428 | 1428 | ||
1429 | if (volume_info->UNCip && volume_info->UNC) { | 1429 | if (volume_info->UNCip && volume_info->UNC) { |
1430 | rc = cifs_inet_pton(AF_INET, volume_info->UNCip, | 1430 | rc = cifs_inet_pton(AF_INET, volume_info->UNCip, |
@@ -1435,9 +1435,9 @@ cifs_get_tcp_session(struct smb_vol *volume_info) | |||
1435 | rc = cifs_inet_pton(AF_INET6, volume_info->UNCip, | 1435 | rc = cifs_inet_pton(AF_INET6, volume_info->UNCip, |
1436 | &sin_server6->sin6_addr.in6_u); | 1436 | &sin_server6->sin6_addr.in6_u); |
1437 | if (rc > 0) | 1437 | if (rc > 0) |
1438 | addr.sa_family = AF_INET6; | 1438 | addr.ss_family = AF_INET6; |
1439 | } else { | 1439 | } else { |
1440 | addr.sa_family = AF_INET; | 1440 | addr.ss_family = AF_INET; |
1441 | } | 1441 | } |
1442 | 1442 | ||
1443 | if (rc <= 0) { | 1443 | if (rc <= 0) { |
@@ -1502,7 +1502,7 @@ cifs_get_tcp_session(struct smb_vol *volume_info) | |||
1502 | tcp_ses->tcpStatus = CifsNew; | 1502 | tcp_ses->tcpStatus = CifsNew; |
1503 | ++tcp_ses->srv_count; | 1503 | ++tcp_ses->srv_count; |
1504 | 1504 | ||
1505 | if (addr.sa_family == AF_INET6) { | 1505 | if (addr.ss_family == AF_INET6) { |
1506 | cFYI(1, ("attempting ipv6 connect")); | 1506 | cFYI(1, ("attempting ipv6 connect")); |
1507 | /* BB should we allow ipv6 on port 139? */ | 1507 | /* BB should we allow ipv6 on port 139? */ |
1508 | /* other OS never observed in Wild doing 139 with v6 */ | 1508 | /* other OS never observed in Wild doing 139 with v6 */ |
@@ -1802,7 +1802,7 @@ ipv4_connect(struct TCP_Server_Info *server) | |||
1802 | * user space buffer | 1802 | * user space buffer |
1803 | */ | 1803 | */ |
1804 | socket->sk->sk_rcvtimeo = 7 * HZ; | 1804 | socket->sk->sk_rcvtimeo = 7 * HZ; |
1805 | socket->sk->sk_sndtimeo = 3 * HZ; | 1805 | socket->sk->sk_sndtimeo = 5 * HZ; |
1806 | 1806 | ||
1807 | /* make the bufsizes depend on wsize/rsize and max requests */ | 1807 | /* make the bufsizes depend on wsize/rsize and max requests */ |
1808 | if (server->noautotune) { | 1808 | if (server->noautotune) { |
@@ -1860,9 +1860,7 @@ ipv4_connect(struct TCP_Server_Info *server) | |||
1860 | smb_buf = (struct smb_hdr *)ses_init_buf; | 1860 | smb_buf = (struct smb_hdr *)ses_init_buf; |
1861 | /* sizeof RFC1002_SESSION_REQUEST with no scope */ | 1861 | /* sizeof RFC1002_SESSION_REQUEST with no scope */ |
1862 | smb_buf->smb_buf_length = 0x81000044; | 1862 | smb_buf->smb_buf_length = 0x81000044; |
1863 | rc = smb_send(socket, smb_buf, 0x44, | 1863 | rc = smb_send(server, smb_buf, 0x44); |
1864 | (struct sockaddr *) &server->addr.sockAddr, | ||
1865 | server->noblocksnd); | ||
1866 | kfree(ses_init_buf); | 1864 | kfree(ses_init_buf); |
1867 | msleep(1); /* RFC1001 layer in at least one server | 1865 | msleep(1); /* RFC1001 layer in at least one server |
1868 | requires very short break before negprot | 1866 | requires very short break before negprot |
@@ -1955,7 +1953,7 @@ ipv6_connect(struct TCP_Server_Info *server) | |||
1955 | * user space buffer | 1953 | * user space buffer |
1956 | */ | 1954 | */ |
1957 | socket->sk->sk_rcvtimeo = 7 * HZ; | 1955 | socket->sk->sk_rcvtimeo = 7 * HZ; |
1958 | socket->sk->sk_sndtimeo = 3 * HZ; | 1956 | socket->sk->sk_sndtimeo = 5 * HZ; |
1959 | server->ssocket = socket; | 1957 | server->ssocket = socket; |
1960 | 1958 | ||
1961 | return rc; | 1959 | return rc; |
@@ -2182,6 +2180,33 @@ static void setup_cifs_sb(struct smb_vol *pvolume_info, | |||
2182 | "mount option supported")); | 2180 | "mount option supported")); |
2183 | } | 2181 | } |
2184 | 2182 | ||
2183 | static int | ||
2184 | is_path_accessible(int xid, struct cifsTconInfo *tcon, | ||
2185 | struct cifs_sb_info *cifs_sb, const char *full_path) | ||
2186 | { | ||
2187 | int rc; | ||
2188 | __u64 inode_num; | ||
2189 | FILE_ALL_INFO *pfile_info; | ||
2190 | |||
2191 | rc = CIFSGetSrvInodeNumber(xid, tcon, full_path, &inode_num, | ||
2192 | cifs_sb->local_nls, | ||
2193 | cifs_sb->mnt_cifs_flags & | ||
2194 | CIFS_MOUNT_MAP_SPECIAL_CHR); | ||
2195 | if (rc != -EOPNOTSUPP) | ||
2196 | return rc; | ||
2197 | |||
2198 | pfile_info = kmalloc(sizeof(FILE_ALL_INFO), GFP_KERNEL); | ||
2199 | if (pfile_info == NULL) | ||
2200 | return -ENOMEM; | ||
2201 | |||
2202 | rc = CIFSSMBQPathInfo(xid, tcon, full_path, pfile_info, | ||
2203 | 0 /* not legacy */, cifs_sb->local_nls, | ||
2204 | cifs_sb->mnt_cifs_flags & | ||
2205 | CIFS_MOUNT_MAP_SPECIAL_CHR); | ||
2206 | kfree(pfile_info); | ||
2207 | return rc; | ||
2208 | } | ||
2209 | |||
2185 | int | 2210 | int |
2186 | cifs_mount(struct super_block *sb, struct cifs_sb_info *cifs_sb, | 2211 | cifs_mount(struct super_block *sb, struct cifs_sb_info *cifs_sb, |
2187 | char *mount_data, const char *devname) | 2212 | char *mount_data, const char *devname) |
@@ -2192,6 +2217,7 @@ cifs_mount(struct super_block *sb, struct cifs_sb_info *cifs_sb, | |||
2192 | struct cifsSesInfo *pSesInfo = NULL; | 2217 | struct cifsSesInfo *pSesInfo = NULL; |
2193 | struct cifsTconInfo *tcon = NULL; | 2218 | struct cifsTconInfo *tcon = NULL; |
2194 | struct TCP_Server_Info *srvTcp = NULL; | 2219 | struct TCP_Server_Info *srvTcp = NULL; |
2220 | char *full_path; | ||
2195 | 2221 | ||
2196 | xid = GetXid(); | 2222 | xid = GetXid(); |
2197 | 2223 | ||
@@ -2428,6 +2454,23 @@ mount_fail_check: | |||
2428 | cifs_sb->rsize = min(cifs_sb->rsize, | 2454 | cifs_sb->rsize = min(cifs_sb->rsize, |
2429 | (tcon->ses->server->maxBuf - MAX_CIFS_HDR_SIZE)); | 2455 | (tcon->ses->server->maxBuf - MAX_CIFS_HDR_SIZE)); |
2430 | 2456 | ||
2457 | if (!rc && cifs_sb->prepathlen) { | ||
2458 | /* build_path_to_root works only when we have a valid tcon */ | ||
2459 | full_path = cifs_build_path_to_root(cifs_sb); | ||
2460 | if (full_path == NULL) { | ||
2461 | rc = -ENOMEM; | ||
2462 | goto mount_fail_check; | ||
2463 | } | ||
2464 | rc = is_path_accessible(xid, tcon, cifs_sb, full_path); | ||
2465 | if (rc) { | ||
2466 | cERROR(1, ("Path %s in not accessible: %d", | ||
2467 | full_path, rc)); | ||
2468 | kfree(full_path); | ||
2469 | goto mount_fail_check; | ||
2470 | } | ||
2471 | kfree(full_path); | ||
2472 | } | ||
2473 | |||
2431 | /* volume_info->password is freed above when existing session found | 2474 | /* volume_info->password is freed above when existing session found |
2432 | (in which case it is not needed anymore) but when new sesion is created | 2475 | (in which case it is not needed anymore) but when new sesion is created |
2433 | the password ptr is put in the new session structure (in which case the | 2476 | the password ptr is put in the new session structure (in which case the |
diff --git a/fs/cifs/dir.c b/fs/cifs/dir.c index 838d9c720a5c..89fb72832652 100644 --- a/fs/cifs/dir.c +++ b/fs/cifs/dir.c | |||
@@ -3,7 +3,7 @@ | |||
3 | * | 3 | * |
4 | * vfs operations that deal with dentries | 4 | * vfs operations that deal with dentries |
5 | * | 5 | * |
6 | * Copyright (C) International Business Machines Corp., 2002,2008 | 6 | * Copyright (C) International Business Machines Corp., 2002,2009 |
7 | * Author(s): Steve French (sfrench@us.ibm.com) | 7 | * Author(s): Steve French (sfrench@us.ibm.com) |
8 | * | 8 | * |
9 | * This library is free software; you can redistribute it and/or modify | 9 | * This library is free software; you can redistribute it and/or modify |
@@ -129,6 +129,89 @@ cifs_bp_rename_retry: | |||
129 | return full_path; | 129 | return full_path; |
130 | } | 130 | } |
131 | 131 | ||
132 | static int cifs_posix_open(char *full_path, struct inode **pinode, | ||
133 | struct super_block *sb, int mode, int oflags, | ||
134 | int *poplock, __u16 *pnetfid, int xid) | ||
135 | { | ||
136 | int rc; | ||
137 | __u32 oplock; | ||
138 | FILE_UNIX_BASIC_INFO *presp_data; | ||
139 | __u32 posix_flags = 0; | ||
140 | struct cifs_sb_info *cifs_sb = CIFS_SB(sb); | ||
141 | |||
142 | cFYI(1, ("posix open %s", full_path)); | ||
143 | |||
144 | presp_data = kzalloc(sizeof(FILE_UNIX_BASIC_INFO), GFP_KERNEL); | ||
145 | if (presp_data == NULL) | ||
146 | return -ENOMEM; | ||
147 | |||
148 | /* So far cifs posix extensions can only map the following flags. | ||
149 | There are other valid fmode oflags such as FMODE_LSEEK, FMODE_PREAD, but | ||
150 | so far we do not seem to need them, and we can treat them as local only */ | ||
151 | if ((oflags & (FMODE_READ | FMODE_WRITE)) == | ||
152 | (FMODE_READ | FMODE_WRITE)) | ||
153 | posix_flags = SMB_O_RDWR; | ||
154 | else if (oflags & FMODE_READ) | ||
155 | posix_flags = SMB_O_RDONLY; | ||
156 | else if (oflags & FMODE_WRITE) | ||
157 | posix_flags = SMB_O_WRONLY; | ||
158 | if (oflags & O_CREAT) | ||
159 | posix_flags |= SMB_O_CREAT; | ||
160 | if (oflags & O_EXCL) | ||
161 | posix_flags |= SMB_O_EXCL; | ||
162 | if (oflags & O_TRUNC) | ||
163 | posix_flags |= SMB_O_TRUNC; | ||
164 | if (oflags & O_APPEND) | ||
165 | posix_flags |= SMB_O_APPEND; | ||
166 | if (oflags & O_SYNC) | ||
167 | posix_flags |= SMB_O_SYNC; | ||
168 | if (oflags & O_DIRECTORY) | ||
169 | posix_flags |= SMB_O_DIRECTORY; | ||
170 | if (oflags & O_NOFOLLOW) | ||
171 | posix_flags |= SMB_O_NOFOLLOW; | ||
172 | if (oflags & O_DIRECT) | ||
173 | posix_flags |= SMB_O_DIRECT; | ||
174 | |||
175 | |||
176 | rc = CIFSPOSIXCreate(xid, cifs_sb->tcon, posix_flags, mode, | ||
177 | pnetfid, presp_data, &oplock, full_path, | ||
178 | cifs_sb->local_nls, cifs_sb->mnt_cifs_flags & | ||
179 | CIFS_MOUNT_MAP_SPECIAL_CHR); | ||
180 | if (rc) | ||
181 | goto posix_open_ret; | ||
182 | |||
183 | if (presp_data->Type == cpu_to_le32(-1)) | ||
184 | goto posix_open_ret; /* open ok, caller does qpathinfo */ | ||
185 | |||
186 | /* get new inode and set it up */ | ||
187 | if (!pinode) | ||
188 | goto posix_open_ret; /* caller does not need info */ | ||
189 | |||
190 | *pinode = cifs_new_inode(sb, &presp_data->UniqueId); | ||
191 | |||
192 | /* We do not need to close the file if new_inode fails since | ||
193 | the caller will retry qpathinfo as long as inode is null */ | ||
194 | if (*pinode == NULL) | ||
195 | goto posix_open_ret; | ||
196 | |||
197 | posix_fill_in_inode(*pinode, presp_data, 1); | ||
198 | |||
199 | posix_open_ret: | ||
200 | kfree(presp_data); | ||
201 | return rc; | ||
202 | } | ||
203 | |||
204 | static void setup_cifs_dentry(struct cifsTconInfo *tcon, | ||
205 | struct dentry *direntry, | ||
206 | struct inode *newinode) | ||
207 | { | ||
208 | if (tcon->nocase) | ||
209 | direntry->d_op = &cifs_ci_dentry_ops; | ||
210 | else | ||
211 | direntry->d_op = &cifs_dentry_ops; | ||
212 | d_instantiate(direntry, newinode); | ||
213 | } | ||
214 | |||
132 | /* Inode operations in similar order to how they appear in Linux file fs.h */ | 215 | /* Inode operations in similar order to how they appear in Linux file fs.h */ |
133 | 216 | ||
134 | int | 217 | int |
@@ -139,14 +222,21 @@ cifs_create(struct inode *inode, struct dentry *direntry, int mode, | |||
139 | int xid; | 222 | int xid; |
140 | int create_options = CREATE_NOT_DIR; | 223 | int create_options = CREATE_NOT_DIR; |
141 | int oplock = 0; | 224 | int oplock = 0; |
225 | int oflags; | ||
226 | /* | ||
227 | * BB below access is probably too much for mknod to request | ||
228 | * but we have to do query and setpathinfo so requesting | ||
229 | * less could fail (unless we want to request getatr and setatr | ||
230 | * permissions (only). At least for POSIX we do not have to | ||
231 | * request so much. | ||
232 | */ | ||
142 | int desiredAccess = GENERIC_READ | GENERIC_WRITE; | 233 | int desiredAccess = GENERIC_READ | GENERIC_WRITE; |
143 | __u16 fileHandle; | 234 | __u16 fileHandle; |
144 | struct cifs_sb_info *cifs_sb; | 235 | struct cifs_sb_info *cifs_sb; |
145 | struct cifsTconInfo *pTcon; | 236 | struct cifsTconInfo *tcon; |
146 | char *full_path = NULL; | 237 | char *full_path = NULL; |
147 | FILE_ALL_INFO *buf = NULL; | 238 | FILE_ALL_INFO *buf = NULL; |
148 | struct inode *newinode = NULL; | 239 | struct inode *newinode = NULL; |
149 | struct cifsFileInfo *pCifsFile = NULL; | ||
150 | struct cifsInodeInfo *pCifsInode; | 240 | struct cifsInodeInfo *pCifsInode; |
151 | int disposition = FILE_OVERWRITE_IF; | 241 | int disposition = FILE_OVERWRITE_IF; |
152 | bool write_only = false; | 242 | bool write_only = false; |
@@ -154,7 +244,7 @@ cifs_create(struct inode *inode, struct dentry *direntry, int mode, | |||
154 | xid = GetXid(); | 244 | xid = GetXid(); |
155 | 245 | ||
156 | cifs_sb = CIFS_SB(inode->i_sb); | 246 | cifs_sb = CIFS_SB(inode->i_sb); |
157 | pTcon = cifs_sb->tcon; | 247 | tcon = cifs_sb->tcon; |
158 | 248 | ||
159 | full_path = build_path_from_dentry(direntry); | 249 | full_path = build_path_from_dentry(direntry); |
160 | if (full_path == NULL) { | 250 | if (full_path == NULL) { |
@@ -162,12 +252,44 @@ cifs_create(struct inode *inode, struct dentry *direntry, int mode, | |||
162 | return -ENOMEM; | 252 | return -ENOMEM; |
163 | } | 253 | } |
164 | 254 | ||
165 | if (nd && (nd->flags & LOOKUP_OPEN)) { | 255 | mode &= ~current->fs->umask; |
166 | int oflags = nd->intent.open.flags; | 256 | if (oplockEnabled) |
257 | oplock = REQ_OPLOCK; | ||
258 | |||
259 | if (nd && (nd->flags & LOOKUP_OPEN)) | ||
260 | oflags = nd->intent.open.flags; | ||
261 | else | ||
262 | oflags = FMODE_READ; | ||
263 | |||
264 | if (tcon->unix_ext && (tcon->ses->capabilities & CAP_UNIX) && | ||
265 | (CIFS_UNIX_POSIX_PATH_OPS_CAP & | ||
266 | le64_to_cpu(tcon->fsUnixInfo.Capability))) { | ||
267 | rc = cifs_posix_open(full_path, &newinode, inode->i_sb, | ||
268 | mode, oflags, &oplock, &fileHandle, xid); | ||
269 | /* EIO could indicate that (posix open) operation is not | ||
270 | supported, despite what server claimed in capability | ||
271 | negotation. EREMOTE indicates DFS junction, which is not | ||
272 | handled in posix open */ | ||
273 | |||
274 | if ((rc == 0) && (newinode == NULL)) | ||
275 | goto cifs_create_get_file_info; /* query inode info */ | ||
276 | else if (rc == 0) /* success, no need to query */ | ||
277 | goto cifs_create_set_dentry; | ||
278 | else if ((rc != -EIO) && (rc != -EREMOTE) && | ||
279 | (rc != -EOPNOTSUPP)) /* path not found or net err */ | ||
280 | goto cifs_create_out; | ||
281 | /* else fallthrough to retry, using older open call, this is | ||
282 | case where server does not support this SMB level, and | ||
283 | falsely claims capability (also get here for DFS case | ||
284 | which should be rare for path not covered on files) */ | ||
285 | } | ||
167 | 286 | ||
287 | if (nd && (nd->flags & LOOKUP_OPEN)) { | ||
288 | /* if the file is going to stay open, then we | ||
289 | need to set the desired access properly */ | ||
168 | desiredAccess = 0; | 290 | desiredAccess = 0; |
169 | if (oflags & FMODE_READ) | 291 | if (oflags & FMODE_READ) |
170 | desiredAccess |= GENERIC_READ; | 292 | desiredAccess |= GENERIC_READ; /* is this too little? */ |
171 | if (oflags & FMODE_WRITE) { | 293 | if (oflags & FMODE_WRITE) { |
172 | desiredAccess |= GENERIC_WRITE; | 294 | desiredAccess |= GENERIC_WRITE; |
173 | if (!(oflags & FMODE_READ)) | 295 | if (!(oflags & FMODE_READ)) |
@@ -186,8 +308,6 @@ cifs_create(struct inode *inode, struct dentry *direntry, int mode, | |||
186 | 308 | ||
187 | /* BB add processing to set equivalent of mode - e.g. via CreateX with | 309 | /* BB add processing to set equivalent of mode - e.g. via CreateX with |
188 | ACLs */ | 310 | ACLs */ |
189 | if (oplockEnabled) | ||
190 | oplock = REQ_OPLOCK; | ||
191 | 311 | ||
192 | buf = kmalloc(sizeof(FILE_ALL_INFO), GFP_KERNEL); | 312 | buf = kmalloc(sizeof(FILE_ALL_INFO), GFP_KERNEL); |
193 | if (buf == NULL) { | 313 | if (buf == NULL) { |
@@ -196,17 +316,15 @@ cifs_create(struct inode *inode, struct dentry *direntry, int mode, | |||
196 | return -ENOMEM; | 316 | return -ENOMEM; |
197 | } | 317 | } |
198 | 318 | ||
199 | mode &= ~current->fs->umask; | ||
200 | |||
201 | /* | 319 | /* |
202 | * if we're not using unix extensions, see if we need to set | 320 | * if we're not using unix extensions, see if we need to set |
203 | * ATTR_READONLY on the create call | 321 | * ATTR_READONLY on the create call |
204 | */ | 322 | */ |
205 | if (!pTcon->unix_ext && (mode & S_IWUGO) == 0) | 323 | if (!tcon->unix_ext && (mode & S_IWUGO) == 0) |
206 | create_options |= CREATE_OPTION_READONLY; | 324 | create_options |= CREATE_OPTION_READONLY; |
207 | 325 | ||
208 | if (cifs_sb->tcon->ses->capabilities & CAP_NT_SMBS) | 326 | if (cifs_sb->tcon->ses->capabilities & CAP_NT_SMBS) |
209 | rc = CIFSSMBOpen(xid, pTcon, full_path, disposition, | 327 | rc = CIFSSMBOpen(xid, tcon, full_path, disposition, |
210 | desiredAccess, create_options, | 328 | desiredAccess, create_options, |
211 | &fileHandle, &oplock, buf, cifs_sb->local_nls, | 329 | &fileHandle, &oplock, buf, cifs_sb->local_nls, |
212 | cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MAP_SPECIAL_CHR); | 330 | cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MAP_SPECIAL_CHR); |
@@ -215,128 +333,119 @@ cifs_create(struct inode *inode, struct dentry *direntry, int mode, | |||
215 | 333 | ||
216 | if (rc == -EIO) { | 334 | if (rc == -EIO) { |
217 | /* old server, retry the open legacy style */ | 335 | /* old server, retry the open legacy style */ |
218 | rc = SMBLegacyOpen(xid, pTcon, full_path, disposition, | 336 | rc = SMBLegacyOpen(xid, tcon, full_path, disposition, |
219 | desiredAccess, create_options, | 337 | desiredAccess, create_options, |
220 | &fileHandle, &oplock, buf, cifs_sb->local_nls, | 338 | &fileHandle, &oplock, buf, cifs_sb->local_nls, |
221 | cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MAP_SPECIAL_CHR); | 339 | cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MAP_SPECIAL_CHR); |
222 | } | 340 | } |
223 | if (rc) { | 341 | if (rc) { |
224 | cFYI(1, ("cifs_create returned 0x%x", rc)); | 342 | cFYI(1, ("cifs_create returned 0x%x", rc)); |
225 | } else { | 343 | goto cifs_create_out; |
226 | /* If Open reported that we actually created a file | 344 | } |
227 | then we now have to set the mode if possible */ | 345 | |
228 | if ((pTcon->unix_ext) && (oplock & CIFS_CREATE_ACTION)) { | 346 | /* If Open reported that we actually created a file |
229 | struct cifs_unix_set_info_args args = { | 347 | then we now have to set the mode if possible */ |
348 | if ((tcon->unix_ext) && (oplock & CIFS_CREATE_ACTION)) { | ||
349 | struct cifs_unix_set_info_args args = { | ||
230 | .mode = mode, | 350 | .mode = mode, |
231 | .ctime = NO_CHANGE_64, | 351 | .ctime = NO_CHANGE_64, |
232 | .atime = NO_CHANGE_64, | 352 | .atime = NO_CHANGE_64, |
233 | .mtime = NO_CHANGE_64, | 353 | .mtime = NO_CHANGE_64, |
234 | .device = 0, | 354 | .device = 0, |
235 | }; | 355 | }; |
236 | 356 | ||
237 | if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_SET_UID) { | 357 | if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_SET_UID) { |
238 | args.uid = (__u64) current_fsuid(); | 358 | args.uid = (__u64) current_fsuid(); |
239 | if (inode->i_mode & S_ISGID) | 359 | if (inode->i_mode & S_ISGID) |
240 | args.gid = (__u64) inode->i_gid; | 360 | args.gid = (__u64) inode->i_gid; |
241 | else | 361 | else |
242 | args.gid = (__u64) current_fsgid(); | 362 | args.gid = (__u64) current_fsgid(); |
243 | } else { | ||
244 | args.uid = NO_CHANGE_64; | ||
245 | args.gid = NO_CHANGE_64; | ||
246 | } | ||
247 | CIFSSMBUnixSetInfo(xid, pTcon, full_path, &args, | ||
248 | cifs_sb->local_nls, | ||
249 | cifs_sb->mnt_cifs_flags & | ||
250 | CIFS_MOUNT_MAP_SPECIAL_CHR); | ||
251 | } else { | 363 | } else { |
252 | /* BB implement mode setting via Windows security | 364 | args.uid = NO_CHANGE_64; |
253 | descriptors e.g. */ | 365 | args.gid = NO_CHANGE_64; |
254 | /* CIFSSMBWinSetPerms(xid,pTcon,path,mode,-1,-1,nls);*/ | ||
255 | |||
256 | /* Could set r/o dos attribute if mode & 0222 == 0 */ | ||
257 | } | 366 | } |
367 | CIFSSMBUnixSetInfo(xid, tcon, full_path, &args, | ||
368 | cifs_sb->local_nls, | ||
369 | cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MAP_SPECIAL_CHR); | ||
370 | } else { | ||
371 | /* BB implement mode setting via Windows security | ||
372 | descriptors e.g. */ | ||
373 | /* CIFSSMBWinSetPerms(xid,tcon,path,mode,-1,-1,nls);*/ | ||
258 | 374 | ||
259 | /* server might mask mode so we have to query for it */ | 375 | /* Could set r/o dos attribute if mode & 0222 == 0 */ |
260 | if (pTcon->unix_ext) | 376 | } |
261 | rc = cifs_get_inode_info_unix(&newinode, full_path, | 377 | |
262 | inode->i_sb, xid); | 378 | cifs_create_get_file_info: |
263 | else { | 379 | /* server might mask mode so we have to query for it */ |
264 | rc = cifs_get_inode_info(&newinode, full_path, | 380 | if (tcon->unix_ext) |
265 | buf, inode->i_sb, xid, | 381 | rc = cifs_get_inode_info_unix(&newinode, full_path, |
266 | &fileHandle); | 382 | inode->i_sb, xid); |
267 | if (newinode) { | 383 | else { |
268 | if (cifs_sb->mnt_cifs_flags & | 384 | rc = cifs_get_inode_info(&newinode, full_path, buf, |
269 | CIFS_MOUNT_DYNPERM) | 385 | inode->i_sb, xid, &fileHandle); |
270 | newinode->i_mode = mode; | 386 | if (newinode) { |
271 | if ((oplock & CIFS_CREATE_ACTION) && | 387 | if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_DYNPERM) |
272 | (cifs_sb->mnt_cifs_flags & | 388 | newinode->i_mode = mode; |
273 | CIFS_MOUNT_SET_UID)) { | 389 | if ((oplock & CIFS_CREATE_ACTION) && |
274 | newinode->i_uid = current_fsuid(); | 390 | (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_SET_UID)) { |
275 | if (inode->i_mode & S_ISGID) | 391 | newinode->i_uid = current_fsuid(); |
276 | newinode->i_gid = | 392 | if (inode->i_mode & S_ISGID) |
277 | inode->i_gid; | 393 | newinode->i_gid = inode->i_gid; |
278 | else | 394 | else |
279 | newinode->i_gid = | 395 | newinode->i_gid = current_fsgid(); |
280 | current_fsgid(); | ||
281 | } | ||
282 | } | 396 | } |
283 | } | 397 | } |
398 | } | ||
284 | 399 | ||
285 | if (rc != 0) { | 400 | cifs_create_set_dentry: |
286 | cFYI(1, | 401 | if (rc == 0) |
287 | ("Create worked but get_inode_info failed rc = %d", | 402 | setup_cifs_dentry(tcon, direntry, newinode); |
288 | rc)); | 403 | else |
289 | } else { | 404 | cFYI(1, ("Create worked, get_inode_info failed rc = %d", rc)); |
290 | if (pTcon->nocase) | 405 | |
291 | direntry->d_op = &cifs_ci_dentry_ops; | 406 | /* nfsd case - nfs srv does not set nd */ |
292 | else | 407 | if ((nd == NULL) || (!(nd->flags & LOOKUP_OPEN))) { |
293 | direntry->d_op = &cifs_dentry_ops; | 408 | /* mknod case - do not leave file open */ |
294 | d_instantiate(direntry, newinode); | 409 | CIFSSMBClose(xid, tcon, fileHandle); |
295 | } | 410 | } else if (newinode) { |
296 | if ((nd == NULL /* nfsd case - nfs srv does not set nd */) || | 411 | struct cifsFileInfo *pCifsFile = |
297 | (!(nd->flags & LOOKUP_OPEN))) { | 412 | kzalloc(sizeof(struct cifsFileInfo), GFP_KERNEL); |
298 | /* mknod case - do not leave file open */ | 413 | |
299 | CIFSSMBClose(xid, pTcon, fileHandle); | 414 | if (pCifsFile == NULL) |
300 | } else if (newinode) { | 415 | goto cifs_create_out; |
301 | pCifsFile = | 416 | pCifsFile->netfid = fileHandle; |
302 | kzalloc(sizeof(struct cifsFileInfo), GFP_KERNEL); | 417 | pCifsFile->pid = current->tgid; |
303 | 418 | pCifsFile->pInode = newinode; | |
304 | if (pCifsFile == NULL) | 419 | pCifsFile->invalidHandle = false; |
305 | goto cifs_create_out; | 420 | pCifsFile->closePend = false; |
306 | pCifsFile->netfid = fileHandle; | 421 | init_MUTEX(&pCifsFile->fh_sem); |
307 | pCifsFile->pid = current->tgid; | 422 | mutex_init(&pCifsFile->lock_mutex); |
308 | pCifsFile->pInode = newinode; | 423 | INIT_LIST_HEAD(&pCifsFile->llist); |
309 | pCifsFile->invalidHandle = false; | 424 | atomic_set(&pCifsFile->wrtPending, 0); |
310 | pCifsFile->closePend = false; | 425 | |
311 | init_MUTEX(&pCifsFile->fh_sem); | 426 | /* set the following in open now |
312 | mutex_init(&pCifsFile->lock_mutex); | ||
313 | INIT_LIST_HEAD(&pCifsFile->llist); | ||
314 | atomic_set(&pCifsFile->wrtPending, 0); | ||
315 | |||
316 | /* set the following in open now | ||
317 | pCifsFile->pfile = file; */ | 427 | pCifsFile->pfile = file; */ |
318 | write_lock(&GlobalSMBSeslock); | 428 | write_lock(&GlobalSMBSeslock); |
319 | list_add(&pCifsFile->tlist, &pTcon->openFileList); | 429 | list_add(&pCifsFile->tlist, &tcon->openFileList); |
320 | pCifsInode = CIFS_I(newinode); | 430 | pCifsInode = CIFS_I(newinode); |
321 | if (pCifsInode) { | 431 | if (pCifsInode) { |
322 | /* if readable file instance put first in list*/ | 432 | /* if readable file instance put first in list*/ |
323 | if (write_only) { | 433 | if (write_only) { |
324 | list_add_tail(&pCifsFile->flist, | 434 | list_add_tail(&pCifsFile->flist, |
325 | &pCifsInode->openFileList); | 435 | &pCifsInode->openFileList); |
326 | } else { | 436 | } else { |
327 | list_add(&pCifsFile->flist, | 437 | list_add(&pCifsFile->flist, |
328 | &pCifsInode->openFileList); | 438 | &pCifsInode->openFileList); |
329 | } | ||
330 | if ((oplock & 0xF) == OPLOCK_EXCLUSIVE) { | ||
331 | pCifsInode->clientCanCacheAll = true; | ||
332 | pCifsInode->clientCanCacheRead = true; | ||
333 | cFYI(1, ("Exclusive Oplock inode %p", | ||
334 | newinode)); | ||
335 | } else if ((oplock & 0xF) == OPLOCK_READ) | ||
336 | pCifsInode->clientCanCacheRead = true; | ||
337 | } | 439 | } |
338 | write_unlock(&GlobalSMBSeslock); | 440 | if ((oplock & 0xF) == OPLOCK_EXCLUSIVE) { |
441 | pCifsInode->clientCanCacheAll = true; | ||
442 | pCifsInode->clientCanCacheRead = true; | ||
443 | cFYI(1, ("Exclusive Oplock inode %p", | ||
444 | newinode)); | ||
445 | } else if ((oplock & 0xF) == OPLOCK_READ) | ||
446 | pCifsInode->clientCanCacheRead = true; | ||
339 | } | 447 | } |
448 | write_unlock(&GlobalSMBSeslock); | ||
340 | } | 449 | } |
341 | cifs_create_out: | 450 | cifs_create_out: |
342 | kfree(buf); | 451 | kfree(buf); |
diff --git a/fs/cifs/inode.c b/fs/cifs/inode.c index 5ab9896fdcb2..4690a360c855 100644 --- a/fs/cifs/inode.c +++ b/fs/cifs/inode.c | |||
@@ -199,6 +199,49 @@ static void fill_fake_finddataunix(FILE_UNIX_BASIC_INFO *pfnd_dat, | |||
199 | pfnd_dat->Gid = cpu_to_le64(pinode->i_gid); | 199 | pfnd_dat->Gid = cpu_to_le64(pinode->i_gid); |
200 | } | 200 | } |
201 | 201 | ||
202 | /** | ||
203 | * cifs_new inode - create new inode, initialize, and hash it | ||
204 | * @sb - pointer to superblock | ||
205 | * @inum - if valid pointer and serverino is enabled, replace i_ino with val | ||
206 | * | ||
207 | * Create a new inode, initialize it for CIFS and hash it. Returns the new | ||
208 | * inode or NULL if one couldn't be allocated. | ||
209 | * | ||
210 | * If the share isn't mounted with "serverino" or inum is a NULL pointer then | ||
211 | * we'll just use the inode number assigned by new_inode(). Note that this can | ||
212 | * mean i_ino collisions since the i_ino assigned by new_inode is not | ||
213 | * guaranteed to be unique. | ||
214 | */ | ||
215 | struct inode * | ||
216 | cifs_new_inode(struct super_block *sb, __u64 *inum) | ||
217 | { | ||
218 | struct inode *inode; | ||
219 | |||
220 | inode = new_inode(sb); | ||
221 | if (inode == NULL) | ||
222 | return NULL; | ||
223 | |||
224 | /* | ||
225 | * BB: Is i_ino == 0 legal? Here, we assume that it is. If it isn't we | ||
226 | * stop passing inum as ptr. Are there sanity checks we can use to | ||
227 | * ensure that the server is really filling in that field? Also, | ||
228 | * if serverino is disabled, perhaps we should be using iunique()? | ||
229 | */ | ||
230 | if (inum && (CIFS_SB(sb)->mnt_cifs_flags & CIFS_MOUNT_SERVER_INUM)) | ||
231 | inode->i_ino = (unsigned long) *inum; | ||
232 | |||
233 | /* | ||
234 | * must set this here instead of cifs_alloc_inode since VFS will | ||
235 | * clobber i_flags | ||
236 | */ | ||
237 | if (sb->s_flags & MS_NOATIME) | ||
238 | inode->i_flags |= S_NOATIME | S_NOCMTIME; | ||
239 | |||
240 | insert_inode_hash(inode); | ||
241 | |||
242 | return inode; | ||
243 | } | ||
244 | |||
202 | int cifs_get_inode_info_unix(struct inode **pinode, | 245 | int cifs_get_inode_info_unix(struct inode **pinode, |
203 | const unsigned char *full_path, struct super_block *sb, int xid) | 246 | const unsigned char *full_path, struct super_block *sb, int xid) |
204 | { | 247 | { |
@@ -233,22 +276,11 @@ int cifs_get_inode_info_unix(struct inode **pinode, | |||
233 | 276 | ||
234 | /* get new inode */ | 277 | /* get new inode */ |
235 | if (*pinode == NULL) { | 278 | if (*pinode == NULL) { |
236 | *pinode = new_inode(sb); | 279 | *pinode = cifs_new_inode(sb, &find_data.UniqueId); |
237 | if (*pinode == NULL) { | 280 | if (*pinode == NULL) { |
238 | rc = -ENOMEM; | 281 | rc = -ENOMEM; |
239 | goto cgiiu_exit; | 282 | goto cgiiu_exit; |
240 | } | 283 | } |
241 | /* Is an i_ino of zero legal? */ | ||
242 | /* note ino incremented to unique num in new_inode */ | ||
243 | /* Are there sanity checks we can use to ensure that | ||
244 | the server is really filling in that field? */ | ||
245 | if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_SERVER_INUM) | ||
246 | (*pinode)->i_ino = (unsigned long)find_data.UniqueId; | ||
247 | |||
248 | if (sb->s_flags & MS_NOATIME) | ||
249 | (*pinode)->i_flags |= S_NOATIME | S_NOCMTIME; | ||
250 | |||
251 | insert_inode_hash(*pinode); | ||
252 | } | 284 | } |
253 | 285 | ||
254 | inode = *pinode; | 286 | inode = *pinode; |
@@ -465,11 +497,9 @@ int cifs_get_inode_info(struct inode **pinode, | |||
465 | 497 | ||
466 | /* get new inode */ | 498 | /* get new inode */ |
467 | if (*pinode == NULL) { | 499 | if (*pinode == NULL) { |
468 | *pinode = new_inode(sb); | 500 | __u64 inode_num; |
469 | if (*pinode == NULL) { | 501 | __u64 *pinum = &inode_num; |
470 | rc = -ENOMEM; | 502 | |
471 | goto cgii_exit; | ||
472 | } | ||
473 | /* Is an i_ino of zero legal? Can we use that to check | 503 | /* Is an i_ino of zero legal? Can we use that to check |
474 | if the server supports returning inode numbers? Are | 504 | if the server supports returning inode numbers? Are |
475 | there other sanity checks we can use to ensure that | 505 | there other sanity checks we can use to ensure that |
@@ -486,22 +516,26 @@ int cifs_get_inode_info(struct inode **pinode, | |||
486 | 516 | ||
487 | if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_SERVER_INUM) { | 517 | if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_SERVER_INUM) { |
488 | int rc1 = 0; | 518 | int rc1 = 0; |
489 | __u64 inode_num; | ||
490 | 519 | ||
491 | rc1 = CIFSGetSrvInodeNumber(xid, pTcon, | 520 | rc1 = CIFSGetSrvInodeNumber(xid, pTcon, |
492 | full_path, &inode_num, | 521 | full_path, pinum, |
493 | cifs_sb->local_nls, | 522 | cifs_sb->local_nls, |
494 | cifs_sb->mnt_cifs_flags & | 523 | cifs_sb->mnt_cifs_flags & |
495 | CIFS_MOUNT_MAP_SPECIAL_CHR); | 524 | CIFS_MOUNT_MAP_SPECIAL_CHR); |
496 | if (rc1) { | 525 | if (rc1) { |
497 | cFYI(1, ("GetSrvInodeNum rc %d", rc1)); | 526 | cFYI(1, ("GetSrvInodeNum rc %d", rc1)); |
527 | pinum = NULL; | ||
498 | /* BB EOPNOSUPP disable SERVER_INUM? */ | 528 | /* BB EOPNOSUPP disable SERVER_INUM? */ |
499 | } else /* do we need cast or hash to ino? */ | 529 | } |
500 | (*pinode)->i_ino = inode_num; | 530 | } else { |
501 | } /* else ino incremented to unique num in new_inode*/ | 531 | pinum = NULL; |
502 | if (sb->s_flags & MS_NOATIME) | 532 | } |
503 | (*pinode)->i_flags |= S_NOATIME | S_NOCMTIME; | 533 | |
504 | insert_inode_hash(*pinode); | 534 | *pinode = cifs_new_inode(sb, pinum); |
535 | if (*pinode == NULL) { | ||
536 | rc = -ENOMEM; | ||
537 | goto cgii_exit; | ||
538 | } | ||
505 | } | 539 | } |
506 | inode = *pinode; | 540 | inode = *pinode; |
507 | cifsInfo = CIFS_I(inode); | 541 | cifsInfo = CIFS_I(inode); |
@@ -621,7 +655,7 @@ static const struct inode_operations cifs_ipc_inode_ops = { | |||
621 | .lookup = cifs_lookup, | 655 | .lookup = cifs_lookup, |
622 | }; | 656 | }; |
623 | 657 | ||
624 | static char *build_path_to_root(struct cifs_sb_info *cifs_sb) | 658 | char *cifs_build_path_to_root(struct cifs_sb_info *cifs_sb) |
625 | { | 659 | { |
626 | int pplen = cifs_sb->prepathlen; | 660 | int pplen = cifs_sb->prepathlen; |
627 | int dfsplen; | 661 | int dfsplen; |
@@ -678,7 +712,7 @@ struct inode *cifs_iget(struct super_block *sb, unsigned long ino) | |||
678 | return inode; | 712 | return inode; |
679 | 713 | ||
680 | cifs_sb = CIFS_SB(inode->i_sb); | 714 | cifs_sb = CIFS_SB(inode->i_sb); |
681 | full_path = build_path_to_root(cifs_sb); | 715 | full_path = cifs_build_path_to_root(cifs_sb); |
682 | if (full_path == NULL) | 716 | if (full_path == NULL) |
683 | return ERR_PTR(-ENOMEM); | 717 | return ERR_PTR(-ENOMEM); |
684 | 718 | ||
@@ -1017,7 +1051,7 @@ out_reval: | |||
1017 | return rc; | 1051 | return rc; |
1018 | } | 1052 | } |
1019 | 1053 | ||
1020 | static void posix_fill_in_inode(struct inode *tmp_inode, | 1054 | void posix_fill_in_inode(struct inode *tmp_inode, |
1021 | FILE_UNIX_BASIC_INFO *pData, int isNewInode) | 1055 | FILE_UNIX_BASIC_INFO *pData, int isNewInode) |
1022 | { | 1056 | { |
1023 | struct cifsInodeInfo *cifsInfo = CIFS_I(tmp_inode); | 1057 | struct cifsInodeInfo *cifsInfo = CIFS_I(tmp_inode); |
@@ -1114,24 +1148,14 @@ int cifs_mkdir(struct inode *inode, struct dentry *direntry, int mode) | |||
1114 | else | 1148 | else |
1115 | direntry->d_op = &cifs_dentry_ops; | 1149 | direntry->d_op = &cifs_dentry_ops; |
1116 | 1150 | ||
1117 | newinode = new_inode(inode->i_sb); | 1151 | newinode = cifs_new_inode(inode->i_sb, |
1152 | &pInfo->UniqueId); | ||
1118 | if (newinode == NULL) { | 1153 | if (newinode == NULL) { |
1119 | kfree(pInfo); | 1154 | kfree(pInfo); |
1120 | goto mkdir_get_info; | 1155 | goto mkdir_get_info; |
1121 | } | 1156 | } |
1122 | 1157 | ||
1123 | /* Is an i_ino of zero legal? */ | ||
1124 | /* Are there sanity checks we can use to ensure that | ||
1125 | the server is really filling in that field? */ | ||
1126 | if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_SERVER_INUM) { | ||
1127 | newinode->i_ino = | ||
1128 | (unsigned long)pInfo->UniqueId; | ||
1129 | } /* note ino incremented to unique num in new_inode */ | ||
1130 | if (inode->i_sb->s_flags & MS_NOATIME) | ||
1131 | newinode->i_flags |= S_NOATIME | S_NOCMTIME; | ||
1132 | newinode->i_nlink = 2; | 1158 | newinode->i_nlink = 2; |
1133 | |||
1134 | insert_inode_hash(newinode); | ||
1135 | d_instantiate(direntry, newinode); | 1159 | d_instantiate(direntry, newinode); |
1136 | 1160 | ||
1137 | /* we already checked in POSIXCreate whether | 1161 | /* we already checked in POSIXCreate whether |
@@ -1285,6 +1309,11 @@ int cifs_rmdir(struct inode *inode, struct dentry *direntry) | |||
1285 | cifsInode = CIFS_I(direntry->d_inode); | 1309 | cifsInode = CIFS_I(direntry->d_inode); |
1286 | cifsInode->time = 0; /* force revalidate to go get info when | 1310 | cifsInode->time = 0; /* force revalidate to go get info when |
1287 | needed */ | 1311 | needed */ |
1312 | |||
1313 | cifsInode = CIFS_I(inode); | ||
1314 | cifsInode->time = 0; /* force revalidate to get parent dir info | ||
1315 | since cached search results now invalid */ | ||
1316 | |||
1288 | direntry->d_inode->i_ctime = inode->i_ctime = inode->i_mtime = | 1317 | direntry->d_inode->i_ctime = inode->i_ctime = inode->i_mtime = |
1289 | current_fs_time(inode->i_sb); | 1318 | current_fs_time(inode->i_sb); |
1290 | 1319 | ||
diff --git a/fs/cifs/md5.c b/fs/cifs/md5.c index 462bbfefd4b6..98b66a54c319 100644 --- a/fs/cifs/md5.c +++ b/fs/cifs/md5.c | |||
@@ -10,8 +10,8 @@ | |||
10 | * with every copy. | 10 | * with every copy. |
11 | * | 11 | * |
12 | * To compute the message digest of a chunk of bytes, declare an | 12 | * To compute the message digest of a chunk of bytes, declare an |
13 | * MD5Context structure, pass it to MD5Init, call MD5Update as | 13 | * MD5Context structure, pass it to cifs_MD5_init, call cifs_MD5_update as |
14 | * needed on buffers full of bytes, and then call MD5Final, which | 14 | * needed on buffers full of bytes, and then call cifs_MD5_final, which |
15 | * will fill a supplied 16-byte array with the digest. | 15 | * will fill a supplied 16-byte array with the digest. |
16 | */ | 16 | */ |
17 | 17 | ||
@@ -45,7 +45,7 @@ byteReverse(unsigned char *buf, unsigned longs) | |||
45 | * initialization constants. | 45 | * initialization constants. |
46 | */ | 46 | */ |
47 | void | 47 | void |
48 | MD5Init(struct MD5Context *ctx) | 48 | cifs_MD5_init(struct MD5Context *ctx) |
49 | { | 49 | { |
50 | ctx->buf[0] = 0x67452301; | 50 | ctx->buf[0] = 0x67452301; |
51 | ctx->buf[1] = 0xefcdab89; | 51 | ctx->buf[1] = 0xefcdab89; |
@@ -61,7 +61,7 @@ MD5Init(struct MD5Context *ctx) | |||
61 | * of bytes. | 61 | * of bytes. |
62 | */ | 62 | */ |
63 | void | 63 | void |
64 | MD5Update(struct MD5Context *ctx, unsigned char const *buf, unsigned len) | 64 | cifs_MD5_update(struct MD5Context *ctx, unsigned char const *buf, unsigned len) |
65 | { | 65 | { |
66 | register __u32 t; | 66 | register __u32 t; |
67 | 67 | ||
@@ -110,7 +110,7 @@ MD5Update(struct MD5Context *ctx, unsigned char const *buf, unsigned len) | |||
110 | * 1 0* (64-bit count of bits processed, MSB-first) | 110 | * 1 0* (64-bit count of bits processed, MSB-first) |
111 | */ | 111 | */ |
112 | void | 112 | void |
113 | MD5Final(unsigned char digest[16], struct MD5Context *ctx) | 113 | cifs_MD5_final(unsigned char digest[16], struct MD5Context *ctx) |
114 | { | 114 | { |
115 | unsigned int count; | 115 | unsigned int count; |
116 | unsigned char *p; | 116 | unsigned char *p; |
@@ -165,7 +165,7 @@ MD5Final(unsigned char digest[16], struct MD5Context *ctx) | |||
165 | 165 | ||
166 | /* | 166 | /* |
167 | * The core of the MD5 algorithm, this alters an existing MD5 hash to | 167 | * The core of the MD5 algorithm, this alters an existing MD5 hash to |
168 | * reflect the addition of 16 longwords of new data. MD5Update blocks | 168 | * reflect the addition of 16 longwords of new data. cifs_MD5_update blocks |
169 | * the data and converts bytes into longwords for this routine. | 169 | * the data and converts bytes into longwords for this routine. |
170 | */ | 170 | */ |
171 | static void | 171 | static void |
@@ -267,9 +267,9 @@ hmac_md5_init_rfc2104(unsigned char *key, int key_len, | |||
267 | unsigned char tk[16]; | 267 | unsigned char tk[16]; |
268 | struct MD5Context tctx; | 268 | struct MD5Context tctx; |
269 | 269 | ||
270 | MD5Init(&tctx); | 270 | cifs_MD5_init(&tctx); |
271 | MD5Update(&tctx, key, key_len); | 271 | cifs_MD5_update(&tctx, key, key_len); |
272 | MD5Final(tk, &tctx); | 272 | cifs_MD5_final(tk, &tctx); |
273 | 273 | ||
274 | key = tk; | 274 | key = tk; |
275 | key_len = 16; | 275 | key_len = 16; |
@@ -287,8 +287,8 @@ hmac_md5_init_rfc2104(unsigned char *key, int key_len, | |||
287 | ctx->k_opad[i] ^= 0x5c; | 287 | ctx->k_opad[i] ^= 0x5c; |
288 | } | 288 | } |
289 | 289 | ||
290 | MD5Init(&ctx->ctx); | 290 | cifs_MD5_init(&ctx->ctx); |
291 | MD5Update(&ctx->ctx, ctx->k_ipad, 64); | 291 | cifs_MD5_update(&ctx->ctx, ctx->k_ipad, 64); |
292 | } | 292 | } |
293 | #endif | 293 | #endif |
294 | 294 | ||
@@ -317,8 +317,8 @@ hmac_md5_init_limK_to_64(const unsigned char *key, int key_len, | |||
317 | ctx->k_opad[i] ^= 0x5c; | 317 | ctx->k_opad[i] ^= 0x5c; |
318 | } | 318 | } |
319 | 319 | ||
320 | MD5Init(&ctx->ctx); | 320 | cifs_MD5_init(&ctx->ctx); |
321 | MD5Update(&ctx->ctx, ctx->k_ipad, 64); | 321 | cifs_MD5_update(&ctx->ctx, ctx->k_ipad, 64); |
322 | } | 322 | } |
323 | 323 | ||
324 | /*********************************************************************** | 324 | /*********************************************************************** |
@@ -328,7 +328,7 @@ void | |||
328 | hmac_md5_update(const unsigned char *text, int text_len, | 328 | hmac_md5_update(const unsigned char *text, int text_len, |
329 | struct HMACMD5Context *ctx) | 329 | struct HMACMD5Context *ctx) |
330 | { | 330 | { |
331 | MD5Update(&ctx->ctx, text, text_len); /* then text of datagram */ | 331 | cifs_MD5_update(&ctx->ctx, text, text_len); /* then text of datagram */ |
332 | } | 332 | } |
333 | 333 | ||
334 | /*********************************************************************** | 334 | /*********************************************************************** |
@@ -339,12 +339,12 @@ hmac_md5_final(unsigned char *digest, struct HMACMD5Context *ctx) | |||
339 | { | 339 | { |
340 | struct MD5Context ctx_o; | 340 | struct MD5Context ctx_o; |
341 | 341 | ||
342 | MD5Final(digest, &ctx->ctx); | 342 | cifs_MD5_final(digest, &ctx->ctx); |
343 | 343 | ||
344 | MD5Init(&ctx_o); | 344 | cifs_MD5_init(&ctx_o); |
345 | MD5Update(&ctx_o, ctx->k_opad, 64); | 345 | cifs_MD5_update(&ctx_o, ctx->k_opad, 64); |
346 | MD5Update(&ctx_o, digest, 16); | 346 | cifs_MD5_update(&ctx_o, digest, 16); |
347 | MD5Final(digest, &ctx_o); | 347 | cifs_MD5_final(digest, &ctx_o); |
348 | } | 348 | } |
349 | 349 | ||
350 | /*********************************************************** | 350 | /*********************************************************** |
diff --git a/fs/cifs/md5.h b/fs/cifs/md5.h index f7d4f4197bac..6fba8cb402fd 100644 --- a/fs/cifs/md5.h +++ b/fs/cifs/md5.h | |||
@@ -20,10 +20,10 @@ struct HMACMD5Context { | |||
20 | }; | 20 | }; |
21 | #endif /* _HMAC_MD5_H */ | 21 | #endif /* _HMAC_MD5_H */ |
22 | 22 | ||
23 | void MD5Init(struct MD5Context *context); | 23 | void cifs_MD5_init(struct MD5Context *context); |
24 | void MD5Update(struct MD5Context *context, unsigned char const *buf, | 24 | void cifs_MD5_update(struct MD5Context *context, unsigned char const *buf, |
25 | unsigned len); | 25 | unsigned len); |
26 | void MD5Final(unsigned char digest[16], struct MD5Context *context); | 26 | void cifs_MD5_final(unsigned char digest[16], struct MD5Context *context); |
27 | 27 | ||
28 | /* The following definitions come from lib/hmacmd5.c */ | 28 | /* The following definitions come from lib/hmacmd5.c */ |
29 | 29 | ||
diff --git a/fs/cifs/readdir.c b/fs/cifs/readdir.c index 9f51f9bf0292..c2c01ff4c32c 100644 --- a/fs/cifs/readdir.c +++ b/fs/cifs/readdir.c | |||
@@ -56,35 +56,34 @@ static inline void dump_cifs_file_struct(struct file *file, char *label) | |||
56 | } | 56 | } |
57 | #endif /* DEBUG2 */ | 57 | #endif /* DEBUG2 */ |
58 | 58 | ||
59 | /* Returns one if new inode created (which therefore needs to be hashed) */ | 59 | /* Returns 1 if new inode created, 2 if both dentry and inode were */ |
60 | /* Might check in the future if inode number changed so we can rehash inode */ | 60 | /* Might check in the future if inode number changed so we can rehash inode */ |
61 | static int construct_dentry(struct qstr *qstring, struct file *file, | 61 | static int |
62 | struct inode **ptmp_inode, struct dentry **pnew_dentry) | 62 | construct_dentry(struct qstr *qstring, struct file *file, |
63 | struct inode **ptmp_inode, struct dentry **pnew_dentry, | ||
64 | __u64 *inum) | ||
63 | { | 65 | { |
64 | struct dentry *tmp_dentry; | 66 | struct dentry *tmp_dentry = NULL; |
65 | struct cifs_sb_info *cifs_sb; | 67 | struct super_block *sb = file->f_path.dentry->d_sb; |
66 | struct cifsTconInfo *pTcon; | ||
67 | int rc = 0; | 68 | int rc = 0; |
68 | 69 | ||
69 | cFYI(1, ("For %s", qstring->name)); | 70 | cFYI(1, ("For %s", qstring->name)); |
70 | cifs_sb = CIFS_SB(file->f_path.dentry->d_sb); | ||
71 | pTcon = cifs_sb->tcon; | ||
72 | 71 | ||
73 | qstring->hash = full_name_hash(qstring->name, qstring->len); | 72 | qstring->hash = full_name_hash(qstring->name, qstring->len); |
74 | tmp_dentry = d_lookup(file->f_path.dentry, qstring); | 73 | tmp_dentry = d_lookup(file->f_path.dentry, qstring); |
75 | if (tmp_dentry) { | 74 | if (tmp_dentry) { |
75 | /* BB: overwrite old name? i.e. tmp_dentry->d_name and | ||
76 | * tmp_dentry->d_name.len?? | ||
77 | */ | ||
76 | cFYI(0, ("existing dentry with inode 0x%p", | 78 | cFYI(0, ("existing dentry with inode 0x%p", |
77 | tmp_dentry->d_inode)); | 79 | tmp_dentry->d_inode)); |
78 | *ptmp_inode = tmp_dentry->d_inode; | 80 | *ptmp_inode = tmp_dentry->d_inode; |
79 | /* BB overwrite old name? i.e. tmp_dentry->d_name and tmp_dentry->d_name.len??*/ | ||
80 | if (*ptmp_inode == NULL) { | 81 | if (*ptmp_inode == NULL) { |
81 | *ptmp_inode = new_inode(file->f_path.dentry->d_sb); | 82 | *ptmp_inode = cifs_new_inode(sb, inum); |
82 | if (*ptmp_inode == NULL) | 83 | if (*ptmp_inode == NULL) |
83 | return rc; | 84 | return rc; |
84 | rc = 1; | 85 | rc = 1; |
85 | } | 86 | } |
86 | if (file->f_path.dentry->d_sb->s_flags & MS_NOATIME) | ||
87 | (*ptmp_inode)->i_flags |= S_NOATIME | S_NOCMTIME; | ||
88 | } else { | 87 | } else { |
89 | tmp_dentry = d_alloc(file->f_path.dentry, qstring); | 88 | tmp_dentry = d_alloc(file->f_path.dentry, qstring); |
90 | if (tmp_dentry == NULL) { | 89 | if (tmp_dentry == NULL) { |
@@ -93,15 +92,14 @@ static int construct_dentry(struct qstr *qstring, struct file *file, | |||
93 | return rc; | 92 | return rc; |
94 | } | 93 | } |
95 | 94 | ||
96 | *ptmp_inode = new_inode(file->f_path.dentry->d_sb); | 95 | if (CIFS_SB(sb)->tcon->nocase) |
97 | if (pTcon->nocase) | ||
98 | tmp_dentry->d_op = &cifs_ci_dentry_ops; | 96 | tmp_dentry->d_op = &cifs_ci_dentry_ops; |
99 | else | 97 | else |
100 | tmp_dentry->d_op = &cifs_dentry_ops; | 98 | tmp_dentry->d_op = &cifs_dentry_ops; |
99 | |||
100 | *ptmp_inode = cifs_new_inode(sb, inum); | ||
101 | if (*ptmp_inode == NULL) | 101 | if (*ptmp_inode == NULL) |
102 | return rc; | 102 | return rc; |
103 | if (file->f_path.dentry->d_sb->s_flags & MS_NOATIME) | ||
104 | (*ptmp_inode)->i_flags |= S_NOATIME | S_NOCMTIME; | ||
105 | rc = 2; | 103 | rc = 2; |
106 | } | 104 | } |
107 | 105 | ||
@@ -822,7 +820,7 @@ static int find_cifs_entry(const int xid, struct cifsTconInfo *pTcon, | |||
822 | /* inode num, inode type and filename returned */ | 820 | /* inode num, inode type and filename returned */ |
823 | static int cifs_get_name_from_search_buf(struct qstr *pqst, | 821 | static int cifs_get_name_from_search_buf(struct qstr *pqst, |
824 | char *current_entry, __u16 level, unsigned int unicode, | 822 | char *current_entry, __u16 level, unsigned int unicode, |
825 | struct cifs_sb_info *cifs_sb, int max_len, ino_t *pinum) | 823 | struct cifs_sb_info *cifs_sb, int max_len, __u64 *pinum) |
826 | { | 824 | { |
827 | int rc = 0; | 825 | int rc = 0; |
828 | unsigned int len = 0; | 826 | unsigned int len = 0; |
@@ -842,9 +840,7 @@ static int cifs_get_name_from_search_buf(struct qstr *pqst, | |||
842 | len = strnlen(filename, PATH_MAX); | 840 | len = strnlen(filename, PATH_MAX); |
843 | } | 841 | } |
844 | 842 | ||
845 | /* BB fixme - hash low and high 32 bits if not 64 bit arch BB */ | 843 | *pinum = pFindData->UniqueId; |
846 | if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_SERVER_INUM) | ||
847 | *pinum = pFindData->UniqueId; | ||
848 | } else if (level == SMB_FIND_FILE_DIRECTORY_INFO) { | 844 | } else if (level == SMB_FIND_FILE_DIRECTORY_INFO) { |
849 | FILE_DIRECTORY_INFO *pFindData = | 845 | FILE_DIRECTORY_INFO *pFindData = |
850 | (FILE_DIRECTORY_INFO *)current_entry; | 846 | (FILE_DIRECTORY_INFO *)current_entry; |
@@ -907,7 +903,7 @@ static int cifs_filldir(char *pfindEntry, struct file *file, | |||
907 | struct qstr qstring; | 903 | struct qstr qstring; |
908 | struct cifsFileInfo *pCifsF; | 904 | struct cifsFileInfo *pCifsF; |
909 | unsigned int obj_type; | 905 | unsigned int obj_type; |
910 | ino_t inum; | 906 | __u64 inum; |
911 | struct cifs_sb_info *cifs_sb; | 907 | struct cifs_sb_info *cifs_sb; |
912 | struct inode *tmp_inode; | 908 | struct inode *tmp_inode; |
913 | struct dentry *tmp_dentry; | 909 | struct dentry *tmp_dentry; |
@@ -940,20 +936,18 @@ static int cifs_filldir(char *pfindEntry, struct file *file, | |||
940 | if (rc) | 936 | if (rc) |
941 | return rc; | 937 | return rc; |
942 | 938 | ||
943 | rc = construct_dentry(&qstring, file, &tmp_inode, &tmp_dentry); | 939 | /* only these two infolevels return valid inode numbers */ |
940 | if (pCifsF->srch_inf.info_level == SMB_FIND_FILE_UNIX || | ||
941 | pCifsF->srch_inf.info_level == SMB_FIND_FILE_ID_FULL_DIR_INFO) | ||
942 | rc = construct_dentry(&qstring, file, &tmp_inode, &tmp_dentry, | ||
943 | &inum); | ||
944 | else | ||
945 | rc = construct_dentry(&qstring, file, &tmp_inode, &tmp_dentry, | ||
946 | NULL); | ||
947 | |||
944 | if ((tmp_inode == NULL) || (tmp_dentry == NULL)) | 948 | if ((tmp_inode == NULL) || (tmp_dentry == NULL)) |
945 | return -ENOMEM; | 949 | return -ENOMEM; |
946 | 950 | ||
947 | if (rc) { | ||
948 | /* inode created, we need to hash it with right inode number */ | ||
949 | if (inum != 0) { | ||
950 | /* BB fixme - hash the 2 32 quantities bits together if | ||
951 | * necessary BB */ | ||
952 | tmp_inode->i_ino = inum; | ||
953 | } | ||
954 | insert_inode_hash(tmp_inode); | ||
955 | } | ||
956 | |||
957 | /* we pass in rc below, indicating whether it is a new inode, | 951 | /* we pass in rc below, indicating whether it is a new inode, |
958 | so we can figure out whether to invalidate the inode cached | 952 | so we can figure out whether to invalidate the inode cached |
959 | data if the file has changed */ | 953 | data if the file has changed */ |
diff --git a/fs/cifs/sess.c b/fs/cifs/sess.c index 5f22de7b79a9..5c68b4282be9 100644 --- a/fs/cifs/sess.c +++ b/fs/cifs/sess.c | |||
@@ -34,15 +34,99 @@ | |||
34 | extern void SMBNTencrypt(unsigned char *passwd, unsigned char *c8, | 34 | extern void SMBNTencrypt(unsigned char *passwd, unsigned char *c8, |
35 | unsigned char *p24); | 35 | unsigned char *p24); |
36 | 36 | ||
37 | /* Checks if this is the first smb session to be reconnected after | ||
38 | the socket has been reestablished (so we know whether to use vc 0). | ||
39 | Called while holding the cifs_tcp_ses_lock, so do not block */ | ||
40 | static bool is_first_ses_reconnect(struct cifsSesInfo *ses) | ||
41 | { | ||
42 | struct list_head *tmp; | ||
43 | struct cifsSesInfo *tmp_ses; | ||
44 | |||
45 | list_for_each(tmp, &ses->server->smb_ses_list) { | ||
46 | tmp_ses = list_entry(tmp, struct cifsSesInfo, | ||
47 | smb_ses_list); | ||
48 | if (tmp_ses->need_reconnect == false) | ||
49 | return false; | ||
50 | } | ||
51 | /* could not find a session that was already connected, | ||
52 | this must be the first one we are reconnecting */ | ||
53 | return true; | ||
54 | } | ||
55 | |||
56 | /* | ||
57 | * vc number 0 is treated specially by some servers, and should be the | ||
58 | * first one we request. After that we can use vcnumbers up to maxvcs, | ||
59 | * one for each smb session (some Windows versions set maxvcs incorrectly | ||
60 | * so maxvc=1 can be ignored). If we have too many vcs, we can reuse | ||
61 | * any vc but zero (some servers reset the connection on vcnum zero) | ||
62 | * | ||
63 | */ | ||
64 | static __le16 get_next_vcnum(struct cifsSesInfo *ses) | ||
65 | { | ||
66 | __u16 vcnum = 0; | ||
67 | struct list_head *tmp; | ||
68 | struct cifsSesInfo *tmp_ses; | ||
69 | __u16 max_vcs = ses->server->max_vcs; | ||
70 | __u16 i; | ||
71 | int free_vc_found = 0; | ||
72 | |||
73 | /* Quoting the MS-SMB specification: "Windows-based SMB servers set this | ||
74 | field to one but do not enforce this limit, which allows an SMB client | ||
75 | to establish more virtual circuits than allowed by this value ... but | ||
76 | other server implementations can enforce this limit." */ | ||
77 | if (max_vcs < 2) | ||
78 | max_vcs = 0xFFFF; | ||
79 | |||
80 | write_lock(&cifs_tcp_ses_lock); | ||
81 | if ((ses->need_reconnect) && is_first_ses_reconnect(ses)) | ||
82 | goto get_vc_num_exit; /* vcnum will be zero */ | ||
83 | for (i = ses->server->srv_count - 1; i < max_vcs; i++) { | ||
84 | if (i == 0) /* this is the only connection, use vc 0 */ | ||
85 | break; | ||
86 | |||
87 | free_vc_found = 1; | ||
88 | |||
89 | list_for_each(tmp, &ses->server->smb_ses_list) { | ||
90 | tmp_ses = list_entry(tmp, struct cifsSesInfo, | ||
91 | smb_ses_list); | ||
92 | if (tmp_ses->vcnum == i) { | ||
93 | free_vc_found = 0; | ||
94 | break; /* found duplicate, try next vcnum */ | ||
95 | } | ||
96 | } | ||
97 | if (free_vc_found) | ||
98 | break; /* we found a vcnumber that will work - use it */ | ||
99 | } | ||
100 | |||
101 | if (i == 0) | ||
102 | vcnum = 0; /* for most common case, ie if one smb session, use | ||
103 | vc zero. Also for case when no free vcnum, zero | ||
104 | is safest to send (some clients only send zero) */ | ||
105 | else if (free_vc_found == 0) | ||
106 | vcnum = 1; /* we can not reuse vc=0 safely, since some servers | ||
107 | reset all uids on that, but 1 is ok. */ | ||
108 | else | ||
109 | vcnum = i; | ||
110 | ses->vcnum = vcnum; | ||
111 | get_vc_num_exit: | ||
112 | write_unlock(&cifs_tcp_ses_lock); | ||
113 | |||
114 | return le16_to_cpu(vcnum); | ||
115 | } | ||
116 | |||
37 | static __u32 cifs_ssetup_hdr(struct cifsSesInfo *ses, SESSION_SETUP_ANDX *pSMB) | 117 | static __u32 cifs_ssetup_hdr(struct cifsSesInfo *ses, SESSION_SETUP_ANDX *pSMB) |
38 | { | 118 | { |
39 | __u32 capabilities = 0; | 119 | __u32 capabilities = 0; |
40 | 120 | ||
41 | /* init fields common to all four types of SessSetup */ | 121 | /* init fields common to all four types of SessSetup */ |
42 | /* note that header is initialized to zero in header_assemble */ | 122 | /* Note that offsets for first seven fields in req struct are same */ |
123 | /* in CIFS Specs so does not matter which of 3 forms of struct */ | ||
124 | /* that we use in next few lines */ | ||
125 | /* Note that header is initialized to zero in header_assemble */ | ||
43 | pSMB->req.AndXCommand = 0xFF; | 126 | pSMB->req.AndXCommand = 0xFF; |
44 | pSMB->req.MaxBufferSize = cpu_to_le16(ses->server->maxBuf); | 127 | pSMB->req.MaxBufferSize = cpu_to_le16(ses->server->maxBuf); |
45 | pSMB->req.MaxMpxCount = cpu_to_le16(ses->server->maxReq); | 128 | pSMB->req.MaxMpxCount = cpu_to_le16(ses->server->maxReq); |
129 | pSMB->req.VcNumber = get_next_vcnum(ses); | ||
46 | 130 | ||
47 | /* Now no need to set SMBFLG_CASELESS or obsolete CANONICAL PATH */ | 131 | /* Now no need to set SMBFLG_CASELESS or obsolete CANONICAL PATH */ |
48 | 132 | ||
@@ -71,7 +155,6 @@ static __u32 cifs_ssetup_hdr(struct cifsSesInfo *ses, SESSION_SETUP_ANDX *pSMB) | |||
71 | if (ses->capabilities & CAP_UNIX) | 155 | if (ses->capabilities & CAP_UNIX) |
72 | capabilities |= CAP_UNIX; | 156 | capabilities |= CAP_UNIX; |
73 | 157 | ||
74 | /* BB check whether to init vcnum BB */ | ||
75 | return capabilities; | 158 | return capabilities; |
76 | } | 159 | } |
77 | 160 | ||
@@ -228,7 +311,7 @@ static int decode_unicode_ssetup(char **pbcc_area, int bleft, | |||
228 | 311 | ||
229 | kfree(ses->serverOS); | 312 | kfree(ses->serverOS); |
230 | /* UTF-8 string will not grow more than four times as big as UCS-16 */ | 313 | /* UTF-8 string will not grow more than four times as big as UCS-16 */ |
231 | ses->serverOS = kzalloc(4 * len, GFP_KERNEL); | 314 | ses->serverOS = kzalloc((4 * len) + 2 /* trailing null */, GFP_KERNEL); |
232 | if (ses->serverOS != NULL) | 315 | if (ses->serverOS != NULL) |
233 | cifs_strfromUCS_le(ses->serverOS, (__le16 *)data, len, nls_cp); | 316 | cifs_strfromUCS_le(ses->serverOS, (__le16 *)data, len, nls_cp); |
234 | data += 2 * (len + 1); | 317 | data += 2 * (len + 1); |
@@ -241,7 +324,7 @@ static int decode_unicode_ssetup(char **pbcc_area, int bleft, | |||
241 | return rc; | 324 | return rc; |
242 | 325 | ||
243 | kfree(ses->serverNOS); | 326 | kfree(ses->serverNOS); |
244 | ses->serverNOS = kzalloc(4 * len, GFP_KERNEL); /* BB this is wrong length FIXME BB */ | 327 | ses->serverNOS = kzalloc((4 * len) + 2 /* trailing null */, GFP_KERNEL); |
245 | if (ses->serverNOS != NULL) { | 328 | if (ses->serverNOS != NULL) { |
246 | cifs_strfromUCS_le(ses->serverNOS, (__le16 *)data, len, | 329 | cifs_strfromUCS_le(ses->serverNOS, (__le16 *)data, len, |
247 | nls_cp); | 330 | nls_cp); |
diff --git a/fs/cifs/transport.c b/fs/cifs/transport.c index 7ebe6599ed3a..0ad3e2d116a6 100644 --- a/fs/cifs/transport.c +++ b/fs/cifs/transport.c | |||
@@ -154,81 +154,8 @@ void DeleteTconOplockQEntries(struct cifsTconInfo *tcon) | |||
154 | spin_unlock(&GlobalMid_Lock); | 154 | spin_unlock(&GlobalMid_Lock); |
155 | } | 155 | } |
156 | 156 | ||
157 | int | ||
158 | smb_send(struct socket *ssocket, struct smb_hdr *smb_buffer, | ||
159 | unsigned int smb_buf_length, struct sockaddr *sin, bool noblocksnd) | ||
160 | { | ||
161 | int rc = 0; | ||
162 | int i = 0; | ||
163 | struct msghdr smb_msg; | ||
164 | struct kvec iov; | ||
165 | unsigned len = smb_buf_length + 4; | ||
166 | |||
167 | if (ssocket == NULL) | ||
168 | return -ENOTSOCK; /* BB eventually add reconnect code here */ | ||
169 | iov.iov_base = smb_buffer; | ||
170 | iov.iov_len = len; | ||
171 | |||
172 | smb_msg.msg_name = sin; | ||
173 | smb_msg.msg_namelen = sizeof(struct sockaddr); | ||
174 | smb_msg.msg_control = NULL; | ||
175 | smb_msg.msg_controllen = 0; | ||
176 | if (noblocksnd) | ||
177 | smb_msg.msg_flags = MSG_DONTWAIT + MSG_NOSIGNAL; | ||
178 | else | ||
179 | smb_msg.msg_flags = MSG_NOSIGNAL; | ||
180 | |||
181 | /* smb header is converted in header_assemble. bcc and rest of SMB word | ||
182 | area, and byte area if necessary, is converted to littleendian in | ||
183 | cifssmb.c and RFC1001 len is converted to bigendian in smb_send | ||
184 | Flags2 is converted in SendReceive */ | ||
185 | |||
186 | smb_buffer->smb_buf_length = cpu_to_be32(smb_buffer->smb_buf_length); | ||
187 | cFYI(1, ("Sending smb of length %d", smb_buf_length)); | ||
188 | dump_smb(smb_buffer, len); | ||
189 | |||
190 | while (len > 0) { | ||
191 | rc = kernel_sendmsg(ssocket, &smb_msg, &iov, 1, len); | ||
192 | if ((rc == -ENOSPC) || (rc == -EAGAIN)) { | ||
193 | i++; | ||
194 | /* smaller timeout here than send2 since smaller size */ | ||
195 | /* Although it may not be required, this also is smaller | ||
196 | oplock break time */ | ||
197 | if (i > 12) { | ||
198 | cERROR(1, | ||
199 | ("sends on sock %p stuck for 7 seconds", | ||
200 | ssocket)); | ||
201 | rc = -EAGAIN; | ||
202 | break; | ||
203 | } | ||
204 | msleep(1 << i); | ||
205 | continue; | ||
206 | } | ||
207 | if (rc < 0) | ||
208 | break; | ||
209 | else | ||
210 | i = 0; /* reset i after each successful send */ | ||
211 | iov.iov_base += rc; | ||
212 | iov.iov_len -= rc; | ||
213 | len -= rc; | ||
214 | } | ||
215 | |||
216 | if (rc < 0) { | ||
217 | cERROR(1, ("Error %d sending data on socket to server", rc)); | ||
218 | } else { | ||
219 | rc = 0; | ||
220 | } | ||
221 | |||
222 | /* Don't want to modify the buffer as a | ||
223 | side effect of this call. */ | ||
224 | smb_buffer->smb_buf_length = smb_buf_length; | ||
225 | |||
226 | return rc; | ||
227 | } | ||
228 | |||
229 | static int | 157 | static int |
230 | smb_send2(struct TCP_Server_Info *server, struct kvec *iov, int n_vec, | 158 | smb_sendv(struct TCP_Server_Info *server, struct kvec *iov, int n_vec) |
231 | struct sockaddr *sin, bool noblocksnd) | ||
232 | { | 159 | { |
233 | int rc = 0; | 160 | int rc = 0; |
234 | int i = 0; | 161 | int i = 0; |
@@ -243,11 +170,11 @@ smb_send2(struct TCP_Server_Info *server, struct kvec *iov, int n_vec, | |||
243 | if (ssocket == NULL) | 170 | if (ssocket == NULL) |
244 | return -ENOTSOCK; /* BB eventually add reconnect code here */ | 171 | return -ENOTSOCK; /* BB eventually add reconnect code here */ |
245 | 172 | ||
246 | smb_msg.msg_name = sin; | 173 | smb_msg.msg_name = (struct sockaddr *) &server->addr.sockAddr; |
247 | smb_msg.msg_namelen = sizeof(struct sockaddr); | 174 | smb_msg.msg_namelen = sizeof(struct sockaddr); |
248 | smb_msg.msg_control = NULL; | 175 | smb_msg.msg_control = NULL; |
249 | smb_msg.msg_controllen = 0; | 176 | smb_msg.msg_controllen = 0; |
250 | if (noblocksnd) | 177 | if (server->noblocksnd) |
251 | smb_msg.msg_flags = MSG_DONTWAIT + MSG_NOSIGNAL; | 178 | smb_msg.msg_flags = MSG_DONTWAIT + MSG_NOSIGNAL; |
252 | else | 179 | else |
253 | smb_msg.msg_flags = MSG_NOSIGNAL; | 180 | smb_msg.msg_flags = MSG_NOSIGNAL; |
@@ -272,7 +199,25 @@ smb_send2(struct TCP_Server_Info *server, struct kvec *iov, int n_vec, | |||
272 | n_vec - first_vec, total_len); | 199 | n_vec - first_vec, total_len); |
273 | if ((rc == -ENOSPC) || (rc == -EAGAIN)) { | 200 | if ((rc == -ENOSPC) || (rc == -EAGAIN)) { |
274 | i++; | 201 | i++; |
275 | if (i >= 14) { | 202 | /* if blocking send we try 3 times, since each can block |
203 | for 5 seconds. For nonblocking we have to try more | ||
204 | but wait increasing amounts of time allowing time for | ||
205 | socket to clear. The overall time we wait in either | ||
206 | case to send on the socket is about 15 seconds. | ||
207 | Similarly we wait for 15 seconds for | ||
208 | a response from the server in SendReceive[2] | ||
209 | for the server to send a response back for | ||
210 | most types of requests (except SMB Write | ||
211 | past end of file which can be slow, and | ||
212 | blocking lock operations). NFS waits slightly longer | ||
213 | than CIFS, but this can make it take longer for | ||
214 | nonresponsive servers to be detected and 15 seconds | ||
215 | is more than enough time for modern networks to | ||
216 | send a packet. In most cases if we fail to send | ||
217 | after the retries we will kill the socket and | ||
218 | reconnect which may clear the network problem. | ||
219 | */ | ||
220 | if ((i >= 14) || (!server->noblocksnd && (i > 2))) { | ||
276 | cERROR(1, | 221 | cERROR(1, |
277 | ("sends on sock %p stuck for 15 seconds", | 222 | ("sends on sock %p stuck for 15 seconds", |
278 | ssocket)); | 223 | ssocket)); |
@@ -339,6 +284,18 @@ smb_send2(struct TCP_Server_Info *server, struct kvec *iov, int n_vec, | |||
339 | return rc; | 284 | return rc; |
340 | } | 285 | } |
341 | 286 | ||
287 | int | ||
288 | smb_send(struct TCP_Server_Info *server, struct smb_hdr *smb_buffer, | ||
289 | unsigned int smb_buf_length) | ||
290 | { | ||
291 | struct kvec iov; | ||
292 | |||
293 | iov.iov_base = smb_buffer; | ||
294 | iov.iov_len = smb_buf_length + 4; | ||
295 | |||
296 | return smb_sendv(server, &iov, 1); | ||
297 | } | ||
298 | |||
342 | static int wait_for_free_request(struct cifsSesInfo *ses, const int long_op) | 299 | static int wait_for_free_request(struct cifsSesInfo *ses, const int long_op) |
343 | { | 300 | { |
344 | if (long_op == CIFS_ASYNC_OP) { | 301 | if (long_op == CIFS_ASYNC_OP) { |
@@ -540,9 +497,7 @@ SendReceive2(const unsigned int xid, struct cifsSesInfo *ses, | |||
540 | #ifdef CONFIG_CIFS_STATS2 | 497 | #ifdef CONFIG_CIFS_STATS2 |
541 | atomic_inc(&ses->server->inSend); | 498 | atomic_inc(&ses->server->inSend); |
542 | #endif | 499 | #endif |
543 | rc = smb_send2(ses->server, iov, n_vec, | 500 | rc = smb_sendv(ses->server, iov, n_vec); |
544 | (struct sockaddr *) &(ses->server->addr.sockAddr), | ||
545 | ses->server->noblocksnd); | ||
546 | #ifdef CONFIG_CIFS_STATS2 | 501 | #ifdef CONFIG_CIFS_STATS2 |
547 | atomic_dec(&ses->server->inSend); | 502 | atomic_dec(&ses->server->inSend); |
548 | midQ->when_sent = jiffies; | 503 | midQ->when_sent = jiffies; |
@@ -736,9 +691,7 @@ SendReceive(const unsigned int xid, struct cifsSesInfo *ses, | |||
736 | #ifdef CONFIG_CIFS_STATS2 | 691 | #ifdef CONFIG_CIFS_STATS2 |
737 | atomic_inc(&ses->server->inSend); | 692 | atomic_inc(&ses->server->inSend); |
738 | #endif | 693 | #endif |
739 | rc = smb_send(ses->server->ssocket, in_buf, in_buf->smb_buf_length, | 694 | rc = smb_send(ses->server, in_buf, in_buf->smb_buf_length); |
740 | (struct sockaddr *) &(ses->server->addr.sockAddr), | ||
741 | ses->server->noblocksnd); | ||
742 | #ifdef CONFIG_CIFS_STATS2 | 695 | #ifdef CONFIG_CIFS_STATS2 |
743 | atomic_dec(&ses->server->inSend); | 696 | atomic_dec(&ses->server->inSend); |
744 | midQ->when_sent = jiffies; | 697 | midQ->when_sent = jiffies; |
@@ -879,9 +832,7 @@ send_nt_cancel(struct cifsTconInfo *tcon, struct smb_hdr *in_buf, | |||
879 | mutex_unlock(&ses->server->srv_mutex); | 832 | mutex_unlock(&ses->server->srv_mutex); |
880 | return rc; | 833 | return rc; |
881 | } | 834 | } |
882 | rc = smb_send(ses->server->ssocket, in_buf, in_buf->smb_buf_length, | 835 | rc = smb_send(ses->server, in_buf, in_buf->smb_buf_length); |
883 | (struct sockaddr *) &(ses->server->addr.sockAddr), | ||
884 | ses->server->noblocksnd); | ||
885 | mutex_unlock(&ses->server->srv_mutex); | 836 | mutex_unlock(&ses->server->srv_mutex); |
886 | return rc; | 837 | return rc; |
887 | } | 838 | } |
@@ -973,9 +924,7 @@ SendReceiveBlockingLock(const unsigned int xid, struct cifsTconInfo *tcon, | |||
973 | #ifdef CONFIG_CIFS_STATS2 | 924 | #ifdef CONFIG_CIFS_STATS2 |
974 | atomic_inc(&ses->server->inSend); | 925 | atomic_inc(&ses->server->inSend); |
975 | #endif | 926 | #endif |
976 | rc = smb_send(ses->server->ssocket, in_buf, in_buf->smb_buf_length, | 927 | rc = smb_send(ses->server, in_buf, in_buf->smb_buf_length); |
977 | (struct sockaddr *) &(ses->server->addr.sockAddr), | ||
978 | ses->server->noblocksnd); | ||
979 | #ifdef CONFIG_CIFS_STATS2 | 928 | #ifdef CONFIG_CIFS_STATS2 |
980 | atomic_dec(&ses->server->inSend); | 929 | atomic_dec(&ses->server->inSend); |
981 | midQ->when_sent = jiffies; | 930 | midQ->when_sent = jiffies; |
diff --git a/fs/coda/Kconfig b/fs/coda/Kconfig new file mode 100644 index 000000000000..c0e5a7fad06d --- /dev/null +++ b/fs/coda/Kconfig | |||
@@ -0,0 +1,21 @@ | |||
1 | config CODA_FS | ||
2 | tristate "Coda file system support (advanced network fs)" | ||
3 | depends on INET | ||
4 | help | ||
5 | Coda is an advanced network file system, similar to NFS in that it | ||
6 | enables you to mount file systems of a remote server and access them | ||
7 | with regular Unix commands as if they were sitting on your hard | ||
8 | disk. Coda has several advantages over NFS: support for | ||
9 | disconnected operation (e.g. for laptops), read/write server | ||
10 | replication, security model for authentication and encryption, | ||
11 | persistent client caches and write back caching. | ||
12 | |||
13 | If you say Y here, your Linux box will be able to act as a Coda | ||
14 | *client*. You will need user level code as well, both for the | ||
15 | client and server. Servers are currently user level, i.e. they need | ||
16 | no kernel support. Please read | ||
17 | <file:Documentation/filesystems/coda.txt> and check out the Coda | ||
18 | home page <http://www.coda.cs.cmu.edu/>. | ||
19 | |||
20 | To compile the coda client support as a module, choose M here: the | ||
21 | module will be called coda. | ||
diff --git a/fs/compat.c b/fs/compat.c index 30f2faa22f5c..d0145ca27572 100644 --- a/fs/compat.c +++ b/fs/compat.c | |||
@@ -1407,7 +1407,7 @@ int compat_do_execve(char * filename, | |||
1407 | bprm->cred = prepare_exec_creds(); | 1407 | bprm->cred = prepare_exec_creds(); |
1408 | if (!bprm->cred) | 1408 | if (!bprm->cred) |
1409 | goto out_unlock; | 1409 | goto out_unlock; |
1410 | check_unsafe_exec(bprm); | 1410 | check_unsafe_exec(bprm, current->files); |
1411 | 1411 | ||
1412 | file = open_exec(filename); | 1412 | file = open_exec(filename); |
1413 | retval = PTR_ERR(file); | 1413 | retval = PTR_ERR(file); |
@@ -1709,7 +1709,7 @@ asmlinkage long compat_sys_select(int n, compat_ulong_t __user *inp, | |||
1709 | } | 1709 | } |
1710 | 1710 | ||
1711 | #ifdef HAVE_SET_RESTORE_SIGMASK | 1711 | #ifdef HAVE_SET_RESTORE_SIGMASK |
1712 | asmlinkage long compat_sys_pselect7(int n, compat_ulong_t __user *inp, | 1712 | static long do_compat_pselect(int n, compat_ulong_t __user *inp, |
1713 | compat_ulong_t __user *outp, compat_ulong_t __user *exp, | 1713 | compat_ulong_t __user *outp, compat_ulong_t __user *exp, |
1714 | struct compat_timespec __user *tsp, compat_sigset_t __user *sigmask, | 1714 | struct compat_timespec __user *tsp, compat_sigset_t __user *sigmask, |
1715 | compat_size_t sigsetsize) | 1715 | compat_size_t sigsetsize) |
@@ -1775,8 +1775,8 @@ asmlinkage long compat_sys_pselect6(int n, compat_ulong_t __user *inp, | |||
1775 | (compat_size_t __user *)(sig+sizeof(up)))) | 1775 | (compat_size_t __user *)(sig+sizeof(up)))) |
1776 | return -EFAULT; | 1776 | return -EFAULT; |
1777 | } | 1777 | } |
1778 | return compat_sys_pselect7(n, inp, outp, exp, tsp, compat_ptr(up), | 1778 | return do_compat_pselect(n, inp, outp, exp, tsp, compat_ptr(up), |
1779 | sigsetsize); | 1779 | sigsetsize); |
1780 | } | 1780 | } |
1781 | 1781 | ||
1782 | asmlinkage long compat_sys_ppoll(struct pollfd __user *ufds, | 1782 | asmlinkage long compat_sys_ppoll(struct pollfd __user *ufds, |
diff --git a/fs/compat_ioctl.c b/fs/compat_ioctl.c index 5235c67e7594..39bd4d38e889 100644 --- a/fs/compat_ioctl.c +++ b/fs/compat_ioctl.c | |||
@@ -538,6 +538,7 @@ static int dev_ifsioc(unsigned int fd, unsigned int cmd, unsigned long arg) | |||
538 | * cannot be fixed without breaking all existing apps. | 538 | * cannot be fixed without breaking all existing apps. |
539 | */ | 539 | */ |
540 | case TUNSETIFF: | 540 | case TUNSETIFF: |
541 | case TUNGETIFF: | ||
541 | case SIOCGIFFLAGS: | 542 | case SIOCGIFFLAGS: |
542 | case SIOCGIFMETRIC: | 543 | case SIOCGIFMETRIC: |
543 | case SIOCGIFMTU: | 544 | case SIOCGIFMTU: |
@@ -784,7 +785,7 @@ static int sg_ioctl_trans(unsigned int fd, unsigned int cmd, unsigned long arg) | |||
784 | 785 | ||
785 | if (copy_in_user(&sgio->status, &sgio32->status, | 786 | if (copy_in_user(&sgio->status, &sgio32->status, |
786 | (4 * sizeof(unsigned char)) + | 787 | (4 * sizeof(unsigned char)) + |
787 | (2 * sizeof(unsigned (short))) + | 788 | (2 * sizeof(unsigned short)) + |
788 | (3 * sizeof(int)))) | 789 | (3 * sizeof(int)))) |
789 | return -EFAULT; | 790 | return -EFAULT; |
790 | 791 | ||
@@ -1937,6 +1938,8 @@ ULONG_IOCTL(SET_BITMAP_FILE) | |||
1937 | /* Big K */ | 1938 | /* Big K */ |
1938 | COMPATIBLE_IOCTL(PIO_FONT) | 1939 | COMPATIBLE_IOCTL(PIO_FONT) |
1939 | COMPATIBLE_IOCTL(GIO_FONT) | 1940 | COMPATIBLE_IOCTL(GIO_FONT) |
1941 | COMPATIBLE_IOCTL(PIO_CMAP) | ||
1942 | COMPATIBLE_IOCTL(GIO_CMAP) | ||
1940 | ULONG_IOCTL(KDSIGACCEPT) | 1943 | ULONG_IOCTL(KDSIGACCEPT) |
1941 | COMPATIBLE_IOCTL(KDGETKEYCODE) | 1944 | COMPATIBLE_IOCTL(KDGETKEYCODE) |
1942 | COMPATIBLE_IOCTL(KDSETKEYCODE) | 1945 | COMPATIBLE_IOCTL(KDSETKEYCODE) |
@@ -1982,6 +1985,11 @@ COMPATIBLE_IOCTL(TUNSETNOCSUM) | |||
1982 | COMPATIBLE_IOCTL(TUNSETDEBUG) | 1985 | COMPATIBLE_IOCTL(TUNSETDEBUG) |
1983 | COMPATIBLE_IOCTL(TUNSETPERSIST) | 1986 | COMPATIBLE_IOCTL(TUNSETPERSIST) |
1984 | COMPATIBLE_IOCTL(TUNSETOWNER) | 1987 | COMPATIBLE_IOCTL(TUNSETOWNER) |
1988 | COMPATIBLE_IOCTL(TUNSETLINK) | ||
1989 | COMPATIBLE_IOCTL(TUNSETGROUP) | ||
1990 | COMPATIBLE_IOCTL(TUNGETFEATURES) | ||
1991 | COMPATIBLE_IOCTL(TUNSETOFFLOAD) | ||
1992 | COMPATIBLE_IOCTL(TUNSETTXFILTER) | ||
1985 | /* Big V */ | 1993 | /* Big V */ |
1986 | COMPATIBLE_IOCTL(VT_SETMODE) | 1994 | COMPATIBLE_IOCTL(VT_SETMODE) |
1987 | COMPATIBLE_IOCTL(VT_GETMODE) | 1995 | COMPATIBLE_IOCTL(VT_GETMODE) |
@@ -2573,6 +2581,7 @@ HANDLE_IOCTL(SIOCGIFPFLAGS, dev_ifsioc) | |||
2573 | HANDLE_IOCTL(SIOCGIFTXQLEN, dev_ifsioc) | 2581 | HANDLE_IOCTL(SIOCGIFTXQLEN, dev_ifsioc) |
2574 | HANDLE_IOCTL(SIOCSIFTXQLEN, dev_ifsioc) | 2582 | HANDLE_IOCTL(SIOCSIFTXQLEN, dev_ifsioc) |
2575 | HANDLE_IOCTL(TUNSETIFF, dev_ifsioc) | 2583 | HANDLE_IOCTL(TUNSETIFF, dev_ifsioc) |
2584 | HANDLE_IOCTL(TUNGETIFF, dev_ifsioc) | ||
2576 | HANDLE_IOCTL(SIOCETHTOOL, ethtool_ioctl) | 2585 | HANDLE_IOCTL(SIOCETHTOOL, ethtool_ioctl) |
2577 | HANDLE_IOCTL(SIOCBONDENSLAVE, bond_ioctl) | 2586 | HANDLE_IOCTL(SIOCBONDENSLAVE, bond_ioctl) |
2578 | HANDLE_IOCTL(SIOCBONDRELEASE, bond_ioctl) | 2587 | HANDLE_IOCTL(SIOCBONDRELEASE, bond_ioctl) |
diff --git a/fs/configfs/Kconfig b/fs/configfs/Kconfig new file mode 100644 index 000000000000..13587cc97a0b --- /dev/null +++ b/fs/configfs/Kconfig | |||
@@ -0,0 +1,11 @@ | |||
1 | config CONFIGFS_FS | ||
2 | tristate "Userspace-driven configuration filesystem" | ||
3 | depends on SYSFS | ||
4 | help | ||
5 | configfs is a ram-based filesystem that provides the converse | ||
6 | of sysfs's functionality. Where sysfs is a filesystem-based | ||
7 | view of kernel objects, configfs is a filesystem-based manager | ||
8 | of kernel objects, or config_items. | ||
9 | |||
10 | Both sysfs and configfs can and should exist together on the | ||
11 | same system. One is not a replacement for the other. | ||
diff --git a/fs/cramfs/Kconfig b/fs/cramfs/Kconfig new file mode 100644 index 000000000000..cd06466f365e --- /dev/null +++ b/fs/cramfs/Kconfig | |||
@@ -0,0 +1,19 @@ | |||
1 | config CRAMFS | ||
2 | tristate "Compressed ROM file system support (cramfs)" | ||
3 | depends on BLOCK | ||
4 | select ZLIB_INFLATE | ||
5 | help | ||
6 | Saying Y here includes support for CramFs (Compressed ROM File | ||
7 | System). CramFs is designed to be a simple, small, and compressed | ||
8 | file system for ROM based embedded systems. CramFs is read-only, | ||
9 | limited to 256MB file systems (with 16MB files), and doesn't support | ||
10 | 16/32 bits uid/gid, hard links and timestamps. | ||
11 | |||
12 | See <file:Documentation/filesystems/cramfs.txt> and | ||
13 | <file:fs/cramfs/README> for further information. | ||
14 | |||
15 | To compile this as a module, choose M here: the module will be called | ||
16 | cramfs. Note that the root file system (the one containing the | ||
17 | directory /) cannot be compiled as a module. | ||
18 | |||
19 | If unsure, say N. | ||
diff --git a/fs/dcache.c b/fs/dcache.c index 4547f66884a0..937df0fb0da5 100644 --- a/fs/dcache.c +++ b/fs/dcache.c | |||
@@ -2092,7 +2092,7 @@ Elong: | |||
2092 | * return NULL; | 2092 | * return NULL; |
2093 | * } | 2093 | * } |
2094 | */ | 2094 | */ |
2095 | asmlinkage long sys_getcwd(char __user *buf, unsigned long size) | 2095 | SYSCALL_DEFINE2(getcwd, char __user *, buf, unsigned long, size) |
2096 | { | 2096 | { |
2097 | int error; | 2097 | int error; |
2098 | struct path pwd, root; | 2098 | struct path pwd, root; |
diff --git a/fs/dcookies.c b/fs/dcookies.c index 180e9fec4ad8..a21cabdbd87b 100644 --- a/fs/dcookies.c +++ b/fs/dcookies.c | |||
@@ -145,7 +145,7 @@ out: | |||
145 | /* And here is where the userspace process can look up the cookie value | 145 | /* And here is where the userspace process can look up the cookie value |
146 | * to retrieve the path. | 146 | * to retrieve the path. |
147 | */ | 147 | */ |
148 | asmlinkage long sys_lookup_dcookie(u64 cookie64, char __user * buf, size_t len) | 148 | SYSCALL_DEFINE(lookup_dcookie)(u64 cookie64, char __user * buf, size_t len) |
149 | { | 149 | { |
150 | unsigned long cookie = (unsigned long)cookie64; | 150 | unsigned long cookie = (unsigned long)cookie64; |
151 | int err = -EINVAL; | 151 | int err = -EINVAL; |
@@ -198,7 +198,13 @@ out: | |||
198 | mutex_unlock(&dcookie_mutex); | 198 | mutex_unlock(&dcookie_mutex); |
199 | return err; | 199 | return err; |
200 | } | 200 | } |
201 | 201 | #ifdef CONFIG_HAVE_SYSCALL_WRAPPERS | |
202 | asmlinkage long SyS_lookup_dcookie(u64 cookie64, long buf, long len) | ||
203 | { | ||
204 | return SYSC_lookup_dcookie(cookie64, (char __user *) buf, (size_t) len); | ||
205 | } | ||
206 | SYSCALL_ALIAS(sys_lookup_dcookie, SyS_lookup_dcookie); | ||
207 | #endif | ||
202 | 208 | ||
203 | static int dcookie_init(void) | 209 | static int dcookie_init(void) |
204 | { | 210 | { |
diff --git a/fs/dlm/debug_fs.c b/fs/dlm/debug_fs.c index 2f107d1a6a45..1d1d27442235 100644 --- a/fs/dlm/debug_fs.c +++ b/fs/dlm/debug_fs.c | |||
@@ -1,7 +1,7 @@ | |||
1 | /****************************************************************************** | 1 | /****************************************************************************** |
2 | ******************************************************************************* | 2 | ******************************************************************************* |
3 | ** | 3 | ** |
4 | ** Copyright (C) 2005-2008 Red Hat, Inc. All rights reserved. | 4 | ** Copyright (C) 2005-2009 Red Hat, Inc. All rights reserved. |
5 | ** | 5 | ** |
6 | ** This copyrighted material is made available to anyone wishing to use, | 6 | ** This copyrighted material is made available to anyone wishing to use, |
7 | ** modify, copy, or redistribute it subject to the terms and conditions | 7 | ** modify, copy, or redistribute it subject to the terms and conditions |
@@ -25,19 +25,6 @@ static struct mutex debug_buf_lock; | |||
25 | 25 | ||
26 | static struct dentry *dlm_root; | 26 | static struct dentry *dlm_root; |
27 | 27 | ||
28 | struct rsb_iter { | ||
29 | int entry; | ||
30 | int format; | ||
31 | int header; | ||
32 | struct dlm_ls *ls; | ||
33 | struct list_head *next; | ||
34 | struct dlm_rsb *rsb; | ||
35 | }; | ||
36 | |||
37 | /* | ||
38 | * dump all rsb's in the lockspace hash table | ||
39 | */ | ||
40 | |||
41 | static char *print_lockmode(int mode) | 28 | static char *print_lockmode(int mode) |
42 | { | 29 | { |
43 | switch (mode) { | 30 | switch (mode) { |
@@ -60,13 +47,13 @@ static char *print_lockmode(int mode) | |||
60 | } | 47 | } |
61 | } | 48 | } |
62 | 49 | ||
63 | static void print_format1_lock(struct seq_file *s, struct dlm_lkb *lkb, | 50 | static int print_format1_lock(struct seq_file *s, struct dlm_lkb *lkb, |
64 | struct dlm_rsb *res) | 51 | struct dlm_rsb *res) |
65 | { | 52 | { |
66 | seq_printf(s, "%08x %s", lkb->lkb_id, print_lockmode(lkb->lkb_grmode)); | 53 | seq_printf(s, "%08x %s", lkb->lkb_id, print_lockmode(lkb->lkb_grmode)); |
67 | 54 | ||
68 | if (lkb->lkb_status == DLM_LKSTS_CONVERT | 55 | if (lkb->lkb_status == DLM_LKSTS_CONVERT || |
69 | || lkb->lkb_status == DLM_LKSTS_WAITING) | 56 | lkb->lkb_status == DLM_LKSTS_WAITING) |
70 | seq_printf(s, " (%s)", print_lockmode(lkb->lkb_rqmode)); | 57 | seq_printf(s, " (%s)", print_lockmode(lkb->lkb_rqmode)); |
71 | 58 | ||
72 | if (lkb->lkb_nodeid) { | 59 | if (lkb->lkb_nodeid) { |
@@ -80,33 +67,42 @@ static void print_format1_lock(struct seq_file *s, struct dlm_lkb *lkb, | |||
80 | if (lkb->lkb_wait_type) | 67 | if (lkb->lkb_wait_type) |
81 | seq_printf(s, " wait_type: %d", lkb->lkb_wait_type); | 68 | seq_printf(s, " wait_type: %d", lkb->lkb_wait_type); |
82 | 69 | ||
83 | seq_printf(s, "\n"); | 70 | return seq_printf(s, "\n"); |
84 | } | 71 | } |
85 | 72 | ||
86 | static int print_format1(struct dlm_rsb *res, struct seq_file *s) | 73 | static int print_format1(struct dlm_rsb *res, struct seq_file *s) |
87 | { | 74 | { |
88 | struct dlm_lkb *lkb; | 75 | struct dlm_lkb *lkb; |
89 | int i, lvblen = res->res_ls->ls_lvblen, recover_list, root_list; | 76 | int i, lvblen = res->res_ls->ls_lvblen, recover_list, root_list; |
77 | int rv; | ||
90 | 78 | ||
91 | lock_rsb(res); | 79 | lock_rsb(res); |
92 | 80 | ||
93 | seq_printf(s, "\nResource %p Name (len=%d) \"", res, res->res_length); | 81 | rv = seq_printf(s, "\nResource %p Name (len=%d) \"", |
82 | res, res->res_length); | ||
83 | if (rv) | ||
84 | goto out; | ||
85 | |||
94 | for (i = 0; i < res->res_length; i++) { | 86 | for (i = 0; i < res->res_length; i++) { |
95 | if (isprint(res->res_name[i])) | 87 | if (isprint(res->res_name[i])) |
96 | seq_printf(s, "%c", res->res_name[i]); | 88 | seq_printf(s, "%c", res->res_name[i]); |
97 | else | 89 | else |
98 | seq_printf(s, "%c", '.'); | 90 | seq_printf(s, "%c", '.'); |
99 | } | 91 | } |
92 | |||
100 | if (res->res_nodeid > 0) | 93 | if (res->res_nodeid > 0) |
101 | seq_printf(s, "\" \nLocal Copy, Master is node %d\n", | 94 | rv = seq_printf(s, "\" \nLocal Copy, Master is node %d\n", |
102 | res->res_nodeid); | 95 | res->res_nodeid); |
103 | else if (res->res_nodeid == 0) | 96 | else if (res->res_nodeid == 0) |
104 | seq_printf(s, "\" \nMaster Copy\n"); | 97 | rv = seq_printf(s, "\" \nMaster Copy\n"); |
105 | else if (res->res_nodeid == -1) | 98 | else if (res->res_nodeid == -1) |
106 | seq_printf(s, "\" \nLooking up master (lkid %x)\n", | 99 | rv = seq_printf(s, "\" \nLooking up master (lkid %x)\n", |
107 | res->res_first_lkid); | 100 | res->res_first_lkid); |
108 | else | 101 | else |
109 | seq_printf(s, "\" \nInvalid master %d\n", res->res_nodeid); | 102 | rv = seq_printf(s, "\" \nInvalid master %d\n", |
103 | res->res_nodeid); | ||
104 | if (rv) | ||
105 | goto out; | ||
110 | 106 | ||
111 | /* Print the LVB: */ | 107 | /* Print the LVB: */ |
112 | if (res->res_lvbptr) { | 108 | if (res->res_lvbptr) { |
@@ -119,52 +115,66 @@ static int print_format1(struct dlm_rsb *res, struct seq_file *s) | |||
119 | } | 115 | } |
120 | if (rsb_flag(res, RSB_VALNOTVALID)) | 116 | if (rsb_flag(res, RSB_VALNOTVALID)) |
121 | seq_printf(s, " (INVALID)"); | 117 | seq_printf(s, " (INVALID)"); |
122 | seq_printf(s, "\n"); | 118 | rv = seq_printf(s, "\n"); |
119 | if (rv) | ||
120 | goto out; | ||
123 | } | 121 | } |
124 | 122 | ||
125 | root_list = !list_empty(&res->res_root_list); | 123 | root_list = !list_empty(&res->res_root_list); |
126 | recover_list = !list_empty(&res->res_recover_list); | 124 | recover_list = !list_empty(&res->res_recover_list); |
127 | 125 | ||
128 | if (root_list || recover_list) { | 126 | if (root_list || recover_list) { |
129 | seq_printf(s, "Recovery: root %d recover %d flags %lx " | 127 | rv = seq_printf(s, "Recovery: root %d recover %d flags %lx " |
130 | "count %d\n", root_list, recover_list, | 128 | "count %d\n", root_list, recover_list, |
131 | res->res_flags, res->res_recover_locks_count); | 129 | res->res_flags, res->res_recover_locks_count); |
130 | if (rv) | ||
131 | goto out; | ||
132 | } | 132 | } |
133 | 133 | ||
134 | /* Print the locks attached to this resource */ | 134 | /* Print the locks attached to this resource */ |
135 | seq_printf(s, "Granted Queue\n"); | 135 | seq_printf(s, "Granted Queue\n"); |
136 | list_for_each_entry(lkb, &res->res_grantqueue, lkb_statequeue) | 136 | list_for_each_entry(lkb, &res->res_grantqueue, lkb_statequeue) { |
137 | print_format1_lock(s, lkb, res); | 137 | rv = print_format1_lock(s, lkb, res); |
138 | if (rv) | ||
139 | goto out; | ||
140 | } | ||
138 | 141 | ||
139 | seq_printf(s, "Conversion Queue\n"); | 142 | seq_printf(s, "Conversion Queue\n"); |
140 | list_for_each_entry(lkb, &res->res_convertqueue, lkb_statequeue) | 143 | list_for_each_entry(lkb, &res->res_convertqueue, lkb_statequeue) { |
141 | print_format1_lock(s, lkb, res); | 144 | rv = print_format1_lock(s, lkb, res); |
145 | if (rv) | ||
146 | goto out; | ||
147 | } | ||
142 | 148 | ||
143 | seq_printf(s, "Waiting Queue\n"); | 149 | seq_printf(s, "Waiting Queue\n"); |
144 | list_for_each_entry(lkb, &res->res_waitqueue, lkb_statequeue) | 150 | list_for_each_entry(lkb, &res->res_waitqueue, lkb_statequeue) { |
145 | print_format1_lock(s, lkb, res); | 151 | rv = print_format1_lock(s, lkb, res); |
152 | if (rv) | ||
153 | goto out; | ||
154 | } | ||
146 | 155 | ||
147 | if (list_empty(&res->res_lookup)) | 156 | if (list_empty(&res->res_lookup)) |
148 | goto out; | 157 | goto out; |
149 | 158 | ||
150 | seq_printf(s, "Lookup Queue\n"); | 159 | seq_printf(s, "Lookup Queue\n"); |
151 | list_for_each_entry(lkb, &res->res_lookup, lkb_rsb_lookup) { | 160 | list_for_each_entry(lkb, &res->res_lookup, lkb_rsb_lookup) { |
152 | seq_printf(s, "%08x %s", lkb->lkb_id, | 161 | rv = seq_printf(s, "%08x %s", lkb->lkb_id, |
153 | print_lockmode(lkb->lkb_rqmode)); | 162 | print_lockmode(lkb->lkb_rqmode)); |
154 | if (lkb->lkb_wait_type) | 163 | if (lkb->lkb_wait_type) |
155 | seq_printf(s, " wait_type: %d", lkb->lkb_wait_type); | 164 | seq_printf(s, " wait_type: %d", lkb->lkb_wait_type); |
156 | seq_printf(s, "\n"); | 165 | rv = seq_printf(s, "\n"); |
157 | } | 166 | } |
158 | out: | 167 | out: |
159 | unlock_rsb(res); | 168 | unlock_rsb(res); |
160 | return 0; | 169 | return rv; |
161 | } | 170 | } |
162 | 171 | ||
163 | static void print_format2_lock(struct seq_file *s, struct dlm_lkb *lkb, | 172 | static int print_format2_lock(struct seq_file *s, struct dlm_lkb *lkb, |
164 | struct dlm_rsb *r) | 173 | struct dlm_rsb *r) |
165 | { | 174 | { |
166 | u64 xid = 0; | 175 | u64 xid = 0; |
167 | u64 us; | 176 | u64 us; |
177 | int rv; | ||
168 | 178 | ||
169 | if (lkb->lkb_flags & DLM_IFL_USER) { | 179 | if (lkb->lkb_flags & DLM_IFL_USER) { |
170 | if (lkb->lkb_ua) | 180 | if (lkb->lkb_ua) |
@@ -177,69 +187,82 @@ static void print_format2_lock(struct seq_file *s, struct dlm_lkb *lkb, | |||
177 | /* id nodeid remid pid xid exflags flags sts grmode rqmode time_us | 187 | /* id nodeid remid pid xid exflags flags sts grmode rqmode time_us |
178 | r_nodeid r_len r_name */ | 188 | r_nodeid r_len r_name */ |
179 | 189 | ||
180 | seq_printf(s, "%x %d %x %u %llu %x %x %d %d %d %llu %u %d \"%s\"\n", | 190 | rv = seq_printf(s, "%x %d %x %u %llu %x %x %d %d %d %llu %u %d \"%s\"\n", |
181 | lkb->lkb_id, | 191 | lkb->lkb_id, |
182 | lkb->lkb_nodeid, | 192 | lkb->lkb_nodeid, |
183 | lkb->lkb_remid, | 193 | lkb->lkb_remid, |
184 | lkb->lkb_ownpid, | 194 | lkb->lkb_ownpid, |
185 | (unsigned long long)xid, | 195 | (unsigned long long)xid, |
186 | lkb->lkb_exflags, | 196 | lkb->lkb_exflags, |
187 | lkb->lkb_flags, | 197 | lkb->lkb_flags, |
188 | lkb->lkb_status, | 198 | lkb->lkb_status, |
189 | lkb->lkb_grmode, | 199 | lkb->lkb_grmode, |
190 | lkb->lkb_rqmode, | 200 | lkb->lkb_rqmode, |
191 | (unsigned long long)us, | 201 | (unsigned long long)us, |
192 | r->res_nodeid, | 202 | r->res_nodeid, |
193 | r->res_length, | 203 | r->res_length, |
194 | r->res_name); | 204 | r->res_name); |
205 | return rv; | ||
195 | } | 206 | } |
196 | 207 | ||
197 | static int print_format2(struct dlm_rsb *r, struct seq_file *s) | 208 | static int print_format2(struct dlm_rsb *r, struct seq_file *s) |
198 | { | 209 | { |
199 | struct dlm_lkb *lkb; | 210 | struct dlm_lkb *lkb; |
211 | int rv = 0; | ||
200 | 212 | ||
201 | lock_rsb(r); | 213 | lock_rsb(r); |
202 | 214 | ||
203 | list_for_each_entry(lkb, &r->res_grantqueue, lkb_statequeue) | 215 | list_for_each_entry(lkb, &r->res_grantqueue, lkb_statequeue) { |
204 | print_format2_lock(s, lkb, r); | 216 | rv = print_format2_lock(s, lkb, r); |
205 | 217 | if (rv) | |
206 | list_for_each_entry(lkb, &r->res_convertqueue, lkb_statequeue) | 218 | goto out; |
207 | print_format2_lock(s, lkb, r); | 219 | } |
208 | 220 | ||
209 | list_for_each_entry(lkb, &r->res_waitqueue, lkb_statequeue) | 221 | list_for_each_entry(lkb, &r->res_convertqueue, lkb_statequeue) { |
210 | print_format2_lock(s, lkb, r); | 222 | rv = print_format2_lock(s, lkb, r); |
223 | if (rv) | ||
224 | goto out; | ||
225 | } | ||
211 | 226 | ||
227 | list_for_each_entry(lkb, &r->res_waitqueue, lkb_statequeue) { | ||
228 | rv = print_format2_lock(s, lkb, r); | ||
229 | if (rv) | ||
230 | goto out; | ||
231 | } | ||
232 | out: | ||
212 | unlock_rsb(r); | 233 | unlock_rsb(r); |
213 | return 0; | 234 | return rv; |
214 | } | 235 | } |
215 | 236 | ||
216 | static void print_format3_lock(struct seq_file *s, struct dlm_lkb *lkb, | 237 | static int print_format3_lock(struct seq_file *s, struct dlm_lkb *lkb, |
217 | int rsb_lookup) | 238 | int rsb_lookup) |
218 | { | 239 | { |
219 | u64 xid = 0; | 240 | u64 xid = 0; |
241 | int rv; | ||
220 | 242 | ||
221 | if (lkb->lkb_flags & DLM_IFL_USER) { | 243 | if (lkb->lkb_flags & DLM_IFL_USER) { |
222 | if (lkb->lkb_ua) | 244 | if (lkb->lkb_ua) |
223 | xid = lkb->lkb_ua->xid; | 245 | xid = lkb->lkb_ua->xid; |
224 | } | 246 | } |
225 | 247 | ||
226 | seq_printf(s, "lkb %x %d %x %u %llu %x %x %d %d %d %d %d %d %u %llu %llu\n", | 248 | rv = seq_printf(s, "lkb %x %d %x %u %llu %x %x %d %d %d %d %d %d %u %llu %llu\n", |
227 | lkb->lkb_id, | 249 | lkb->lkb_id, |
228 | lkb->lkb_nodeid, | 250 | lkb->lkb_nodeid, |
229 | lkb->lkb_remid, | 251 | lkb->lkb_remid, |
230 | lkb->lkb_ownpid, | 252 | lkb->lkb_ownpid, |
231 | (unsigned long long)xid, | 253 | (unsigned long long)xid, |
232 | lkb->lkb_exflags, | 254 | lkb->lkb_exflags, |
233 | lkb->lkb_flags, | 255 | lkb->lkb_flags, |
234 | lkb->lkb_status, | 256 | lkb->lkb_status, |
235 | lkb->lkb_grmode, | 257 | lkb->lkb_grmode, |
236 | lkb->lkb_rqmode, | 258 | lkb->lkb_rqmode, |
237 | lkb->lkb_highbast, | 259 | lkb->lkb_highbast, |
238 | rsb_lookup, | 260 | rsb_lookup, |
239 | lkb->lkb_wait_type, | 261 | lkb->lkb_wait_type, |
240 | lkb->lkb_lvbseq, | 262 | lkb->lkb_lvbseq, |
241 | (unsigned long long)ktime_to_ns(lkb->lkb_timestamp), | 263 | (unsigned long long)ktime_to_ns(lkb->lkb_timestamp), |
242 | (unsigned long long)ktime_to_ns(lkb->lkb_time_bast)); | 264 | (unsigned long long)ktime_to_ns(lkb->lkb_time_bast)); |
265 | return rv; | ||
243 | } | 266 | } |
244 | 267 | ||
245 | static int print_format3(struct dlm_rsb *r, struct seq_file *s) | 268 | static int print_format3(struct dlm_rsb *r, struct seq_file *s) |
@@ -247,18 +270,21 @@ static int print_format3(struct dlm_rsb *r, struct seq_file *s) | |||
247 | struct dlm_lkb *lkb; | 270 | struct dlm_lkb *lkb; |
248 | int i, lvblen = r->res_ls->ls_lvblen; | 271 | int i, lvblen = r->res_ls->ls_lvblen; |
249 | int print_name = 1; | 272 | int print_name = 1; |
273 | int rv; | ||
250 | 274 | ||
251 | lock_rsb(r); | 275 | lock_rsb(r); |
252 | 276 | ||
253 | seq_printf(s, "rsb %p %d %x %lx %d %d %u %d ", | 277 | rv = seq_printf(s, "rsb %p %d %x %lx %d %d %u %d ", |
254 | r, | 278 | r, |
255 | r->res_nodeid, | 279 | r->res_nodeid, |
256 | r->res_first_lkid, | 280 | r->res_first_lkid, |
257 | r->res_flags, | 281 | r->res_flags, |
258 | !list_empty(&r->res_root_list), | 282 | !list_empty(&r->res_root_list), |
259 | !list_empty(&r->res_recover_list), | 283 | !list_empty(&r->res_recover_list), |
260 | r->res_recover_locks_count, | 284 | r->res_recover_locks_count, |
261 | r->res_length); | 285 | r->res_length); |
286 | if (rv) | ||
287 | goto out; | ||
262 | 288 | ||
263 | for (i = 0; i < r->res_length; i++) { | 289 | for (i = 0; i < r->res_length; i++) { |
264 | if (!isascii(r->res_name[i]) || !isprint(r->res_name[i])) | 290 | if (!isascii(r->res_name[i]) || !isprint(r->res_name[i])) |
@@ -273,7 +299,9 @@ static int print_format3(struct dlm_rsb *r, struct seq_file *s) | |||
273 | else | 299 | else |
274 | seq_printf(s, " %02x", (unsigned char)r->res_name[i]); | 300 | seq_printf(s, " %02x", (unsigned char)r->res_name[i]); |
275 | } | 301 | } |
276 | seq_printf(s, "\n"); | 302 | rv = seq_printf(s, "\n"); |
303 | if (rv) | ||
304 | goto out; | ||
277 | 305 | ||
278 | if (!r->res_lvbptr) | 306 | if (!r->res_lvbptr) |
279 | goto do_locks; | 307 | goto do_locks; |
@@ -282,344 +310,294 @@ static int print_format3(struct dlm_rsb *r, struct seq_file *s) | |||
282 | 310 | ||
283 | for (i = 0; i < lvblen; i++) | 311 | for (i = 0; i < lvblen; i++) |
284 | seq_printf(s, " %02x", (unsigned char)r->res_lvbptr[i]); | 312 | seq_printf(s, " %02x", (unsigned char)r->res_lvbptr[i]); |
285 | seq_printf(s, "\n"); | 313 | rv = seq_printf(s, "\n"); |
314 | if (rv) | ||
315 | goto out; | ||
286 | 316 | ||
287 | do_locks: | 317 | do_locks: |
288 | list_for_each_entry(lkb, &r->res_grantqueue, lkb_statequeue) | 318 | list_for_each_entry(lkb, &r->res_grantqueue, lkb_statequeue) { |
289 | print_format3_lock(s, lkb, 0); | 319 | rv = print_format3_lock(s, lkb, 0); |
290 | 320 | if (rv) | |
291 | list_for_each_entry(lkb, &r->res_convertqueue, lkb_statequeue) | 321 | goto out; |
292 | print_format3_lock(s, lkb, 0); | ||
293 | |||
294 | list_for_each_entry(lkb, &r->res_waitqueue, lkb_statequeue) | ||
295 | print_format3_lock(s, lkb, 0); | ||
296 | |||
297 | list_for_each_entry(lkb, &r->res_lookup, lkb_rsb_lookup) | ||
298 | print_format3_lock(s, lkb, 1); | ||
299 | |||
300 | unlock_rsb(r); | ||
301 | return 0; | ||
302 | } | ||
303 | |||
304 | static int rsb_iter_next(struct rsb_iter *ri) | ||
305 | { | ||
306 | struct dlm_ls *ls = ri->ls; | ||
307 | int i; | ||
308 | |||
309 | if (!ri->next) { | ||
310 | top: | ||
311 | /* Find the next non-empty hash bucket */ | ||
312 | for (i = ri->entry; i < ls->ls_rsbtbl_size; i++) { | ||
313 | read_lock(&ls->ls_rsbtbl[i].lock); | ||
314 | if (!list_empty(&ls->ls_rsbtbl[i].list)) { | ||
315 | ri->next = ls->ls_rsbtbl[i].list.next; | ||
316 | ri->rsb = list_entry(ri->next, struct dlm_rsb, | ||
317 | res_hashchain); | ||
318 | dlm_hold_rsb(ri->rsb); | ||
319 | read_unlock(&ls->ls_rsbtbl[i].lock); | ||
320 | break; | ||
321 | } | ||
322 | read_unlock(&ls->ls_rsbtbl[i].lock); | ||
323 | } | ||
324 | ri->entry = i; | ||
325 | |||
326 | if (ri->entry >= ls->ls_rsbtbl_size) | ||
327 | return 1; | ||
328 | } else { | ||
329 | struct dlm_rsb *old = ri->rsb; | ||
330 | i = ri->entry; | ||
331 | read_lock(&ls->ls_rsbtbl[i].lock); | ||
332 | ri->next = ri->next->next; | ||
333 | if (ri->next->next == ls->ls_rsbtbl[i].list.next) { | ||
334 | /* End of list - move to next bucket */ | ||
335 | ri->next = NULL; | ||
336 | ri->entry++; | ||
337 | read_unlock(&ls->ls_rsbtbl[i].lock); | ||
338 | dlm_put_rsb(old); | ||
339 | goto top; | ||
340 | } | ||
341 | ri->rsb = list_entry(ri->next, struct dlm_rsb, res_hashchain); | ||
342 | dlm_hold_rsb(ri->rsb); | ||
343 | read_unlock(&ls->ls_rsbtbl[i].lock); | ||
344 | dlm_put_rsb(old); | ||
345 | } | 322 | } |
346 | 323 | ||
347 | return 0; | 324 | list_for_each_entry(lkb, &r->res_convertqueue, lkb_statequeue) { |
348 | } | 325 | rv = print_format3_lock(s, lkb, 0); |
349 | 326 | if (rv) | |
350 | static void rsb_iter_free(struct rsb_iter *ri) | 327 | goto out; |
351 | { | ||
352 | kfree(ri); | ||
353 | } | ||
354 | |||
355 | static struct rsb_iter *rsb_iter_init(struct dlm_ls *ls) | ||
356 | { | ||
357 | struct rsb_iter *ri; | ||
358 | |||
359 | ri = kzalloc(sizeof *ri, GFP_KERNEL); | ||
360 | if (!ri) | ||
361 | return NULL; | ||
362 | |||
363 | ri->ls = ls; | ||
364 | ri->entry = 0; | ||
365 | ri->next = NULL; | ||
366 | ri->format = 1; | ||
367 | |||
368 | if (rsb_iter_next(ri)) { | ||
369 | rsb_iter_free(ri); | ||
370 | return NULL; | ||
371 | } | 328 | } |
372 | 329 | ||
373 | return ri; | 330 | list_for_each_entry(lkb, &r->res_waitqueue, lkb_statequeue) { |
374 | } | 331 | rv = print_format3_lock(s, lkb, 0); |
375 | 332 | if (rv) | |
376 | static void *rsb_seq_start(struct seq_file *file, loff_t *pos) | 333 | goto out; |
377 | { | ||
378 | struct rsb_iter *ri; | ||
379 | loff_t n = *pos; | ||
380 | |||
381 | ri = rsb_iter_init(file->private); | ||
382 | if (!ri) | ||
383 | return NULL; | ||
384 | |||
385 | while (n--) { | ||
386 | if (rsb_iter_next(ri)) { | ||
387 | rsb_iter_free(ri); | ||
388 | return NULL; | ||
389 | } | ||
390 | } | 334 | } |
391 | 335 | ||
392 | return ri; | 336 | list_for_each_entry(lkb, &r->res_lookup, lkb_rsb_lookup) { |
393 | } | 337 | rv = print_format3_lock(s, lkb, 1); |
394 | 338 | if (rv) | |
395 | static void *rsb_seq_next(struct seq_file *file, void *iter_ptr, loff_t *pos) | 339 | goto out; |
396 | { | ||
397 | struct rsb_iter *ri = iter_ptr; | ||
398 | |||
399 | (*pos)++; | ||
400 | |||
401 | if (rsb_iter_next(ri)) { | ||
402 | rsb_iter_free(ri); | ||
403 | return NULL; | ||
404 | } | 340 | } |
405 | 341 | out: | |
406 | return ri; | 342 | unlock_rsb(r); |
343 | return rv; | ||
407 | } | 344 | } |
408 | 345 | ||
409 | static void rsb_seq_stop(struct seq_file *file, void *iter_ptr) | 346 | struct rsbtbl_iter { |
410 | { | 347 | struct dlm_rsb *rsb; |
411 | /* nothing for now */ | 348 | unsigned bucket; |
412 | } | 349 | int format; |
350 | int header; | ||
351 | }; | ||
413 | 352 | ||
414 | static int rsb_seq_show(struct seq_file *file, void *iter_ptr) | 353 | /* seq_printf returns -1 if the buffer is full, and 0 otherwise. |
354 | If the buffer is full, seq_printf can be called again, but it | ||
355 | does nothing and just returns -1. So, the these printing routines | ||
356 | periodically check the return value to avoid wasting too much time | ||
357 | trying to print to a full buffer. */ | ||
358 | |||
359 | static int table_seq_show(struct seq_file *seq, void *iter_ptr) | ||
415 | { | 360 | { |
416 | struct rsb_iter *ri = iter_ptr; | 361 | struct rsbtbl_iter *ri = iter_ptr; |
362 | int rv = 0; | ||
417 | 363 | ||
418 | switch (ri->format) { | 364 | switch (ri->format) { |
419 | case 1: | 365 | case 1: |
420 | print_format1(ri->rsb, file); | 366 | rv = print_format1(ri->rsb, seq); |
421 | break; | 367 | break; |
422 | case 2: | 368 | case 2: |
423 | if (ri->header) { | 369 | if (ri->header) { |
424 | seq_printf(file, "id nodeid remid pid xid exflags " | 370 | seq_printf(seq, "id nodeid remid pid xid exflags " |
425 | "flags sts grmode rqmode time_ms " | 371 | "flags sts grmode rqmode time_ms " |
426 | "r_nodeid r_len r_name\n"); | 372 | "r_nodeid r_len r_name\n"); |
427 | ri->header = 0; | 373 | ri->header = 0; |
428 | } | 374 | } |
429 | print_format2(ri->rsb, file); | 375 | rv = print_format2(ri->rsb, seq); |
430 | break; | 376 | break; |
431 | case 3: | 377 | case 3: |
432 | if (ri->header) { | 378 | if (ri->header) { |
433 | seq_printf(file, "version rsb 1.1 lvb 1.1 lkb 1.1\n"); | 379 | seq_printf(seq, "version rsb 1.1 lvb 1.1 lkb 1.1\n"); |
434 | ri->header = 0; | 380 | ri->header = 0; |
435 | } | 381 | } |
436 | print_format3(ri->rsb, file); | 382 | rv = print_format3(ri->rsb, seq); |
437 | break; | 383 | break; |
438 | } | 384 | } |
439 | 385 | ||
440 | return 0; | 386 | return rv; |
441 | } | 387 | } |
442 | 388 | ||
443 | static struct seq_operations rsb_seq_ops = { | 389 | static struct seq_operations format1_seq_ops; |
444 | .start = rsb_seq_start, | 390 | static struct seq_operations format2_seq_ops; |
445 | .next = rsb_seq_next, | 391 | static struct seq_operations format3_seq_ops; |
446 | .stop = rsb_seq_stop, | ||
447 | .show = rsb_seq_show, | ||
448 | }; | ||
449 | 392 | ||
450 | static int rsb_open(struct inode *inode, struct file *file) | 393 | static void *table_seq_start(struct seq_file *seq, loff_t *pos) |
451 | { | 394 | { |
452 | struct seq_file *seq; | 395 | struct dlm_ls *ls = seq->private; |
453 | int ret; | 396 | struct rsbtbl_iter *ri; |
454 | 397 | struct dlm_rsb *r; | |
455 | ret = seq_open(file, &rsb_seq_ops); | 398 | loff_t n = *pos; |
456 | if (ret) | 399 | unsigned bucket, entry; |
457 | return ret; | ||
458 | |||
459 | seq = file->private_data; | ||
460 | seq->private = inode->i_private; | ||
461 | |||
462 | return 0; | ||
463 | } | ||
464 | |||
465 | static const struct file_operations rsb_fops = { | ||
466 | .owner = THIS_MODULE, | ||
467 | .open = rsb_open, | ||
468 | .read = seq_read, | ||
469 | .llseek = seq_lseek, | ||
470 | .release = seq_release | ||
471 | }; | ||
472 | 400 | ||
473 | /* | 401 | bucket = n >> 32; |
474 | * Dump state in compact per-lock listing | 402 | entry = n & ((1LL << 32) - 1); |
475 | */ | ||
476 | 403 | ||
477 | static struct rsb_iter *locks_iter_init(struct dlm_ls *ls, loff_t *pos) | 404 | if (bucket >= ls->ls_rsbtbl_size) |
478 | { | 405 | return NULL; |
479 | struct rsb_iter *ri; | ||
480 | 406 | ||
481 | ri = kzalloc(sizeof *ri, GFP_KERNEL); | 407 | ri = kzalloc(sizeof(struct rsbtbl_iter), GFP_KERNEL); |
482 | if (!ri) | 408 | if (!ri) |
483 | return NULL; | 409 | return NULL; |
484 | 410 | if (n == 0) | |
485 | ri->ls = ls; | ||
486 | ri->entry = 0; | ||
487 | ri->next = NULL; | ||
488 | ri->format = 2; | ||
489 | |||
490 | if (*pos == 0) | ||
491 | ri->header = 1; | 411 | ri->header = 1; |
492 | 412 | if (seq->op == &format1_seq_ops) | |
493 | if (rsb_iter_next(ri)) { | 413 | ri->format = 1; |
494 | rsb_iter_free(ri); | 414 | if (seq->op == &format2_seq_ops) |
495 | return NULL; | 415 | ri->format = 2; |
416 | if (seq->op == &format3_seq_ops) | ||
417 | ri->format = 3; | ||
418 | |||
419 | spin_lock(&ls->ls_rsbtbl[bucket].lock); | ||
420 | if (!list_empty(&ls->ls_rsbtbl[bucket].list)) { | ||
421 | list_for_each_entry(r, &ls->ls_rsbtbl[bucket].list, | ||
422 | res_hashchain) { | ||
423 | if (!entry--) { | ||
424 | dlm_hold_rsb(r); | ||
425 | ri->rsb = r; | ||
426 | ri->bucket = bucket; | ||
427 | spin_unlock(&ls->ls_rsbtbl[bucket].lock); | ||
428 | return ri; | ||
429 | } | ||
430 | } | ||
496 | } | 431 | } |
432 | spin_unlock(&ls->ls_rsbtbl[bucket].lock); | ||
497 | 433 | ||
498 | return ri; | 434 | /* |
499 | } | 435 | * move to the first rsb in the next non-empty bucket |
436 | */ | ||
500 | 437 | ||
501 | static void *locks_seq_start(struct seq_file *file, loff_t *pos) | 438 | /* zero the entry */ |
502 | { | 439 | n &= ~((1LL << 32) - 1); |
503 | struct rsb_iter *ri; | ||
504 | loff_t n = *pos; | ||
505 | 440 | ||
506 | ri = locks_iter_init(file->private, pos); | 441 | while (1) { |
507 | if (!ri) | 442 | bucket++; |
508 | return NULL; | 443 | n += 1LL << 32; |
509 | 444 | ||
510 | while (n--) { | 445 | if (bucket >= ls->ls_rsbtbl_size) { |
511 | if (rsb_iter_next(ri)) { | 446 | kfree(ri); |
512 | rsb_iter_free(ri); | ||
513 | return NULL; | 447 | return NULL; |
514 | } | 448 | } |
515 | } | ||
516 | 449 | ||
517 | return ri; | 450 | spin_lock(&ls->ls_rsbtbl[bucket].lock); |
451 | if (!list_empty(&ls->ls_rsbtbl[bucket].list)) { | ||
452 | r = list_first_entry(&ls->ls_rsbtbl[bucket].list, | ||
453 | struct dlm_rsb, res_hashchain); | ||
454 | dlm_hold_rsb(r); | ||
455 | ri->rsb = r; | ||
456 | ri->bucket = bucket; | ||
457 | spin_unlock(&ls->ls_rsbtbl[bucket].lock); | ||
458 | *pos = n; | ||
459 | return ri; | ||
460 | } | ||
461 | spin_unlock(&ls->ls_rsbtbl[bucket].lock); | ||
462 | } | ||
518 | } | 463 | } |
519 | 464 | ||
520 | static struct seq_operations locks_seq_ops = { | 465 | static void *table_seq_next(struct seq_file *seq, void *iter_ptr, loff_t *pos) |
521 | .start = locks_seq_start, | ||
522 | .next = rsb_seq_next, | ||
523 | .stop = rsb_seq_stop, | ||
524 | .show = rsb_seq_show, | ||
525 | }; | ||
526 | |||
527 | static int locks_open(struct inode *inode, struct file *file) | ||
528 | { | 466 | { |
529 | struct seq_file *seq; | 467 | struct dlm_ls *ls = seq->private; |
530 | int ret; | 468 | struct rsbtbl_iter *ri = iter_ptr; |
531 | 469 | struct list_head *next; | |
532 | ret = seq_open(file, &locks_seq_ops); | 470 | struct dlm_rsb *r, *rp; |
533 | if (ret) | 471 | loff_t n = *pos; |
534 | return ret; | 472 | unsigned bucket; |
535 | 473 | ||
536 | seq = file->private_data; | 474 | bucket = n >> 32; |
537 | seq->private = inode->i_private; | 475 | |
538 | 476 | /* | |
539 | return 0; | 477 | * move to the next rsb in the same bucket |
540 | } | 478 | */ |
541 | 479 | ||
542 | static const struct file_operations locks_fops = { | 480 | spin_lock(&ls->ls_rsbtbl[bucket].lock); |
543 | .owner = THIS_MODULE, | 481 | rp = ri->rsb; |
544 | .open = locks_open, | 482 | next = rp->res_hashchain.next; |
545 | .read = seq_read, | 483 | |
546 | .llseek = seq_lseek, | 484 | if (next != &ls->ls_rsbtbl[bucket].list) { |
547 | .release = seq_release | 485 | r = list_entry(next, struct dlm_rsb, res_hashchain); |
548 | }; | 486 | dlm_hold_rsb(r); |
549 | 487 | ri->rsb = r; | |
550 | /* | 488 | spin_unlock(&ls->ls_rsbtbl[bucket].lock); |
551 | * Dump all rsb/lvb/lkb state in compact listing, more complete than _locks | 489 | dlm_put_rsb(rp); |
552 | * This can replace both formats 1 and 2 eventually. | 490 | ++*pos; |
553 | */ | 491 | return ri; |
492 | } | ||
493 | spin_unlock(&ls->ls_rsbtbl[bucket].lock); | ||
494 | dlm_put_rsb(rp); | ||
554 | 495 | ||
555 | static struct rsb_iter *all_iter_init(struct dlm_ls *ls, loff_t *pos) | 496 | /* |
556 | { | 497 | * move to the first rsb in the next non-empty bucket |
557 | struct rsb_iter *ri; | 498 | */ |
558 | 499 | ||
559 | ri = kzalloc(sizeof *ri, GFP_KERNEL); | 500 | /* zero the entry */ |
560 | if (!ri) | 501 | n &= ~((1LL << 32) - 1); |
561 | return NULL; | ||
562 | 502 | ||
563 | ri->ls = ls; | 503 | while (1) { |
564 | ri->entry = 0; | 504 | bucket++; |
565 | ri->next = NULL; | 505 | n += 1LL << 32; |
566 | ri->format = 3; | ||
567 | 506 | ||
568 | if (*pos == 0) | 507 | if (bucket >= ls->ls_rsbtbl_size) { |
569 | ri->header = 1; | 508 | kfree(ri); |
509 | return NULL; | ||
510 | } | ||
570 | 511 | ||
571 | if (rsb_iter_next(ri)) { | 512 | spin_lock(&ls->ls_rsbtbl[bucket].lock); |
572 | rsb_iter_free(ri); | 513 | if (!list_empty(&ls->ls_rsbtbl[bucket].list)) { |
573 | return NULL; | 514 | r = list_first_entry(&ls->ls_rsbtbl[bucket].list, |
515 | struct dlm_rsb, res_hashchain); | ||
516 | dlm_hold_rsb(r); | ||
517 | ri->rsb = r; | ||
518 | ri->bucket = bucket; | ||
519 | spin_unlock(&ls->ls_rsbtbl[bucket].lock); | ||
520 | *pos = n; | ||
521 | return ri; | ||
522 | } | ||
523 | spin_unlock(&ls->ls_rsbtbl[bucket].lock); | ||
574 | } | 524 | } |
575 | |||
576 | return ri; | ||
577 | } | 525 | } |
578 | 526 | ||
579 | static void *all_seq_start(struct seq_file *file, loff_t *pos) | 527 | static void table_seq_stop(struct seq_file *seq, void *iter_ptr) |
580 | { | 528 | { |
581 | struct rsb_iter *ri; | 529 | struct rsbtbl_iter *ri = iter_ptr; |
582 | loff_t n = *pos; | ||
583 | |||
584 | ri = all_iter_init(file->private, pos); | ||
585 | if (!ri) | ||
586 | return NULL; | ||
587 | 530 | ||
588 | while (n--) { | 531 | if (ri) { |
589 | if (rsb_iter_next(ri)) { | 532 | dlm_put_rsb(ri->rsb); |
590 | rsb_iter_free(ri); | 533 | kfree(ri); |
591 | return NULL; | ||
592 | } | ||
593 | } | 534 | } |
594 | |||
595 | return ri; | ||
596 | } | 535 | } |
597 | 536 | ||
598 | static struct seq_operations all_seq_ops = { | 537 | static struct seq_operations format1_seq_ops = { |
599 | .start = all_seq_start, | 538 | .start = table_seq_start, |
600 | .next = rsb_seq_next, | 539 | .next = table_seq_next, |
601 | .stop = rsb_seq_stop, | 540 | .stop = table_seq_stop, |
602 | .show = rsb_seq_show, | 541 | .show = table_seq_show, |
603 | }; | 542 | }; |
604 | 543 | ||
605 | static int all_open(struct inode *inode, struct file *file) | 544 | static struct seq_operations format2_seq_ops = { |
545 | .start = table_seq_start, | ||
546 | .next = table_seq_next, | ||
547 | .stop = table_seq_stop, | ||
548 | .show = table_seq_show, | ||
549 | }; | ||
550 | |||
551 | static struct seq_operations format3_seq_ops = { | ||
552 | .start = table_seq_start, | ||
553 | .next = table_seq_next, | ||
554 | .stop = table_seq_stop, | ||
555 | .show = table_seq_show, | ||
556 | }; | ||
557 | |||
558 | static const struct file_operations format1_fops; | ||
559 | static const struct file_operations format2_fops; | ||
560 | static const struct file_operations format3_fops; | ||
561 | |||
562 | static int table_open(struct inode *inode, struct file *file) | ||
606 | { | 563 | { |
607 | struct seq_file *seq; | 564 | struct seq_file *seq; |
608 | int ret; | 565 | int ret = -1; |
566 | |||
567 | if (file->f_op == &format1_fops) | ||
568 | ret = seq_open(file, &format1_seq_ops); | ||
569 | else if (file->f_op == &format2_fops) | ||
570 | ret = seq_open(file, &format2_seq_ops); | ||
571 | else if (file->f_op == &format3_fops) | ||
572 | ret = seq_open(file, &format3_seq_ops); | ||
609 | 573 | ||
610 | ret = seq_open(file, &all_seq_ops); | ||
611 | if (ret) | 574 | if (ret) |
612 | return ret; | 575 | return ret; |
613 | 576 | ||
614 | seq = file->private_data; | 577 | seq = file->private_data; |
615 | seq->private = inode->i_private; | 578 | seq->private = inode->i_private; /* the dlm_ls */ |
616 | |||
617 | return 0; | 579 | return 0; |
618 | } | 580 | } |
619 | 581 | ||
620 | static const struct file_operations all_fops = { | 582 | static const struct file_operations format1_fops = { |
583 | .owner = THIS_MODULE, | ||
584 | .open = table_open, | ||
585 | .read = seq_read, | ||
586 | .llseek = seq_lseek, | ||
587 | .release = seq_release | ||
588 | }; | ||
589 | |||
590 | static const struct file_operations format2_fops = { | ||
591 | .owner = THIS_MODULE, | ||
592 | .open = table_open, | ||
593 | .read = seq_read, | ||
594 | .llseek = seq_lseek, | ||
595 | .release = seq_release | ||
596 | }; | ||
597 | |||
598 | static const struct file_operations format3_fops = { | ||
621 | .owner = THIS_MODULE, | 599 | .owner = THIS_MODULE, |
622 | .open = all_open, | 600 | .open = table_open, |
623 | .read = seq_read, | 601 | .read = seq_read, |
624 | .llseek = seq_lseek, | 602 | .llseek = seq_lseek, |
625 | .release = seq_release | 603 | .release = seq_release |
@@ -689,7 +667,7 @@ int dlm_create_debug_file(struct dlm_ls *ls) | |||
689 | S_IFREG | S_IRUGO, | 667 | S_IFREG | S_IRUGO, |
690 | dlm_root, | 668 | dlm_root, |
691 | ls, | 669 | ls, |
692 | &rsb_fops); | 670 | &format1_fops); |
693 | if (!ls->ls_debug_rsb_dentry) | 671 | if (!ls->ls_debug_rsb_dentry) |
694 | goto fail; | 672 | goto fail; |
695 | 673 | ||
@@ -702,7 +680,7 @@ int dlm_create_debug_file(struct dlm_ls *ls) | |||
702 | S_IFREG | S_IRUGO, | 680 | S_IFREG | S_IRUGO, |
703 | dlm_root, | 681 | dlm_root, |
704 | ls, | 682 | ls, |
705 | &locks_fops); | 683 | &format2_fops); |
706 | if (!ls->ls_debug_locks_dentry) | 684 | if (!ls->ls_debug_locks_dentry) |
707 | goto fail; | 685 | goto fail; |
708 | 686 | ||
@@ -715,7 +693,7 @@ int dlm_create_debug_file(struct dlm_ls *ls) | |||
715 | S_IFREG | S_IRUGO, | 693 | S_IFREG | S_IRUGO, |
716 | dlm_root, | 694 | dlm_root, |
717 | ls, | 695 | ls, |
718 | &all_fops); | 696 | &format3_fops); |
719 | if (!ls->ls_debug_all_dentry) | 697 | if (!ls->ls_debug_all_dentry) |
720 | goto fail; | 698 | goto fail; |
721 | 699 | ||
diff --git a/fs/dlm/dlm_internal.h b/fs/dlm/dlm_internal.h index ef2f1e353966..076e86f38bc8 100644 --- a/fs/dlm/dlm_internal.h +++ b/fs/dlm/dlm_internal.h | |||
@@ -105,7 +105,7 @@ struct dlm_dirtable { | |||
105 | struct dlm_rsbtable { | 105 | struct dlm_rsbtable { |
106 | struct list_head list; | 106 | struct list_head list; |
107 | struct list_head toss; | 107 | struct list_head toss; |
108 | rwlock_t lock; | 108 | spinlock_t lock; |
109 | }; | 109 | }; |
110 | 110 | ||
111 | struct dlm_lkbtable { | 111 | struct dlm_lkbtable { |
diff --git a/fs/dlm/lock.c b/fs/dlm/lock.c index 6cfe65bbf4a2..01e7d39c5fba 100644 --- a/fs/dlm/lock.c +++ b/fs/dlm/lock.c | |||
@@ -412,9 +412,9 @@ static int search_rsb(struct dlm_ls *ls, char *name, int len, int b, | |||
412 | unsigned int flags, struct dlm_rsb **r_ret) | 412 | unsigned int flags, struct dlm_rsb **r_ret) |
413 | { | 413 | { |
414 | int error; | 414 | int error; |
415 | write_lock(&ls->ls_rsbtbl[b].lock); | 415 | spin_lock(&ls->ls_rsbtbl[b].lock); |
416 | error = _search_rsb(ls, name, len, b, flags, r_ret); | 416 | error = _search_rsb(ls, name, len, b, flags, r_ret); |
417 | write_unlock(&ls->ls_rsbtbl[b].lock); | 417 | spin_unlock(&ls->ls_rsbtbl[b].lock); |
418 | return error; | 418 | return error; |
419 | } | 419 | } |
420 | 420 | ||
@@ -478,16 +478,16 @@ static int find_rsb(struct dlm_ls *ls, char *name, int namelen, | |||
478 | r->res_nodeid = nodeid; | 478 | r->res_nodeid = nodeid; |
479 | } | 479 | } |
480 | 480 | ||
481 | write_lock(&ls->ls_rsbtbl[bucket].lock); | 481 | spin_lock(&ls->ls_rsbtbl[bucket].lock); |
482 | error = _search_rsb(ls, name, namelen, bucket, 0, &tmp); | 482 | error = _search_rsb(ls, name, namelen, bucket, 0, &tmp); |
483 | if (!error) { | 483 | if (!error) { |
484 | write_unlock(&ls->ls_rsbtbl[bucket].lock); | 484 | spin_unlock(&ls->ls_rsbtbl[bucket].lock); |
485 | dlm_free_rsb(r); | 485 | dlm_free_rsb(r); |
486 | r = tmp; | 486 | r = tmp; |
487 | goto out; | 487 | goto out; |
488 | } | 488 | } |
489 | list_add(&r->res_hashchain, &ls->ls_rsbtbl[bucket].list); | 489 | list_add(&r->res_hashchain, &ls->ls_rsbtbl[bucket].list); |
490 | write_unlock(&ls->ls_rsbtbl[bucket].lock); | 490 | spin_unlock(&ls->ls_rsbtbl[bucket].lock); |
491 | error = 0; | 491 | error = 0; |
492 | out: | 492 | out: |
493 | *r_ret = r; | 493 | *r_ret = r; |
@@ -530,9 +530,9 @@ static void put_rsb(struct dlm_rsb *r) | |||
530 | struct dlm_ls *ls = r->res_ls; | 530 | struct dlm_ls *ls = r->res_ls; |
531 | uint32_t bucket = r->res_bucket; | 531 | uint32_t bucket = r->res_bucket; |
532 | 532 | ||
533 | write_lock(&ls->ls_rsbtbl[bucket].lock); | 533 | spin_lock(&ls->ls_rsbtbl[bucket].lock); |
534 | kref_put(&r->res_ref, toss_rsb); | 534 | kref_put(&r->res_ref, toss_rsb); |
535 | write_unlock(&ls->ls_rsbtbl[bucket].lock); | 535 | spin_unlock(&ls->ls_rsbtbl[bucket].lock); |
536 | } | 536 | } |
537 | 537 | ||
538 | void dlm_put_rsb(struct dlm_rsb *r) | 538 | void dlm_put_rsb(struct dlm_rsb *r) |
@@ -967,7 +967,7 @@ static int shrink_bucket(struct dlm_ls *ls, int b) | |||
967 | 967 | ||
968 | for (;;) { | 968 | for (;;) { |
969 | found = 0; | 969 | found = 0; |
970 | write_lock(&ls->ls_rsbtbl[b].lock); | 970 | spin_lock(&ls->ls_rsbtbl[b].lock); |
971 | list_for_each_entry_reverse(r, &ls->ls_rsbtbl[b].toss, | 971 | list_for_each_entry_reverse(r, &ls->ls_rsbtbl[b].toss, |
972 | res_hashchain) { | 972 | res_hashchain) { |
973 | if (!time_after_eq(jiffies, r->res_toss_time + | 973 | if (!time_after_eq(jiffies, r->res_toss_time + |
@@ -978,20 +978,20 @@ static int shrink_bucket(struct dlm_ls *ls, int b) | |||
978 | } | 978 | } |
979 | 979 | ||
980 | if (!found) { | 980 | if (!found) { |
981 | write_unlock(&ls->ls_rsbtbl[b].lock); | 981 | spin_unlock(&ls->ls_rsbtbl[b].lock); |
982 | break; | 982 | break; |
983 | } | 983 | } |
984 | 984 | ||
985 | if (kref_put(&r->res_ref, kill_rsb)) { | 985 | if (kref_put(&r->res_ref, kill_rsb)) { |
986 | list_del(&r->res_hashchain); | 986 | list_del(&r->res_hashchain); |
987 | write_unlock(&ls->ls_rsbtbl[b].lock); | 987 | spin_unlock(&ls->ls_rsbtbl[b].lock); |
988 | 988 | ||
989 | if (is_master(r)) | 989 | if (is_master(r)) |
990 | dir_remove(r); | 990 | dir_remove(r); |
991 | dlm_free_rsb(r); | 991 | dlm_free_rsb(r); |
992 | count++; | 992 | count++; |
993 | } else { | 993 | } else { |
994 | write_unlock(&ls->ls_rsbtbl[b].lock); | 994 | spin_unlock(&ls->ls_rsbtbl[b].lock); |
995 | log_error(ls, "tossed rsb in use %s", r->res_name); | 995 | log_error(ls, "tossed rsb in use %s", r->res_name); |
996 | } | 996 | } |
997 | } | 997 | } |
@@ -4224,7 +4224,7 @@ static struct dlm_rsb *find_purged_rsb(struct dlm_ls *ls, int bucket) | |||
4224 | { | 4224 | { |
4225 | struct dlm_rsb *r, *r_ret = NULL; | 4225 | struct dlm_rsb *r, *r_ret = NULL; |
4226 | 4226 | ||
4227 | read_lock(&ls->ls_rsbtbl[bucket].lock); | 4227 | spin_lock(&ls->ls_rsbtbl[bucket].lock); |
4228 | list_for_each_entry(r, &ls->ls_rsbtbl[bucket].list, res_hashchain) { | 4228 | list_for_each_entry(r, &ls->ls_rsbtbl[bucket].list, res_hashchain) { |
4229 | if (!rsb_flag(r, RSB_LOCKS_PURGED)) | 4229 | if (!rsb_flag(r, RSB_LOCKS_PURGED)) |
4230 | continue; | 4230 | continue; |
@@ -4233,7 +4233,7 @@ static struct dlm_rsb *find_purged_rsb(struct dlm_ls *ls, int bucket) | |||
4233 | r_ret = r; | 4233 | r_ret = r; |
4234 | break; | 4234 | break; |
4235 | } | 4235 | } |
4236 | read_unlock(&ls->ls_rsbtbl[bucket].lock); | 4236 | spin_unlock(&ls->ls_rsbtbl[bucket].lock); |
4237 | return r_ret; | 4237 | return r_ret; |
4238 | } | 4238 | } |
4239 | 4239 | ||
diff --git a/fs/dlm/lockspace.c b/fs/dlm/lockspace.c index 8d86b7960f0d..aa32e5f02493 100644 --- a/fs/dlm/lockspace.c +++ b/fs/dlm/lockspace.c | |||
@@ -464,7 +464,7 @@ static int new_lockspace(char *name, int namelen, void **lockspace, | |||
464 | for (i = 0; i < size; i++) { | 464 | for (i = 0; i < size; i++) { |
465 | INIT_LIST_HEAD(&ls->ls_rsbtbl[i].list); | 465 | INIT_LIST_HEAD(&ls->ls_rsbtbl[i].list); |
466 | INIT_LIST_HEAD(&ls->ls_rsbtbl[i].toss); | 466 | INIT_LIST_HEAD(&ls->ls_rsbtbl[i].toss); |
467 | rwlock_init(&ls->ls_rsbtbl[i].lock); | 467 | spin_lock_init(&ls->ls_rsbtbl[i].lock); |
468 | } | 468 | } |
469 | 469 | ||
470 | size = dlm_config.ci_lkbtbl_size; | 470 | size = dlm_config.ci_lkbtbl_size; |
diff --git a/fs/dlm/plock.c b/fs/dlm/plock.c index eba87ff3177b..894a32d438d5 100644 --- a/fs/dlm/plock.c +++ b/fs/dlm/plock.c | |||
@@ -168,7 +168,7 @@ static int dlm_plock_callback(struct plock_op *op) | |||
168 | notify = xop->callback; | 168 | notify = xop->callback; |
169 | 169 | ||
170 | if (op->info.rv) { | 170 | if (op->info.rv) { |
171 | notify(flc, NULL, op->info.rv); | 171 | notify(fl, NULL, op->info.rv); |
172 | goto out; | 172 | goto out; |
173 | } | 173 | } |
174 | 174 | ||
@@ -187,7 +187,7 @@ static int dlm_plock_callback(struct plock_op *op) | |||
187 | (unsigned long long)op->info.number, file, fl); | 187 | (unsigned long long)op->info.number, file, fl); |
188 | } | 188 | } |
189 | 189 | ||
190 | rv = notify(flc, NULL, 0); | 190 | rv = notify(fl, NULL, 0); |
191 | if (rv) { | 191 | if (rv) { |
192 | /* XXX: We need to cancel the fs lock here: */ | 192 | /* XXX: We need to cancel the fs lock here: */ |
193 | log_print("dlm_plock_callback: lock granted after lock request " | 193 | log_print("dlm_plock_callback: lock granted after lock request " |
@@ -304,7 +304,9 @@ int dlm_posix_get(dlm_lockspace_t *lockspace, u64 number, struct file *file, | |||
304 | if (rv == -ENOENT) | 304 | if (rv == -ENOENT) |
305 | rv = 0; | 305 | rv = 0; |
306 | else if (rv > 0) { | 306 | else if (rv > 0) { |
307 | locks_init_lock(fl); | ||
307 | fl->fl_type = (op->info.ex) ? F_WRLCK : F_RDLCK; | 308 | fl->fl_type = (op->info.ex) ? F_WRLCK : F_RDLCK; |
309 | fl->fl_flags = FL_POSIX; | ||
308 | fl->fl_pid = op->info.pid; | 310 | fl->fl_pid = op->info.pid; |
309 | fl->fl_start = op->info.start; | 311 | fl->fl_start = op->info.start; |
310 | fl->fl_end = op->info.end; | 312 | fl->fl_end = op->info.end; |
diff --git a/fs/dlm/recover.c b/fs/dlm/recover.c index 80aba5bdd4a4..eda43f362616 100644 --- a/fs/dlm/recover.c +++ b/fs/dlm/recover.c | |||
@@ -726,7 +726,7 @@ int dlm_create_root_list(struct dlm_ls *ls) | |||
726 | } | 726 | } |
727 | 727 | ||
728 | for (i = 0; i < ls->ls_rsbtbl_size; i++) { | 728 | for (i = 0; i < ls->ls_rsbtbl_size; i++) { |
729 | read_lock(&ls->ls_rsbtbl[i].lock); | 729 | spin_lock(&ls->ls_rsbtbl[i].lock); |
730 | list_for_each_entry(r, &ls->ls_rsbtbl[i].list, res_hashchain) { | 730 | list_for_each_entry(r, &ls->ls_rsbtbl[i].list, res_hashchain) { |
731 | list_add(&r->res_root_list, &ls->ls_root_list); | 731 | list_add(&r->res_root_list, &ls->ls_root_list); |
732 | dlm_hold_rsb(r); | 732 | dlm_hold_rsb(r); |
@@ -737,7 +737,7 @@ int dlm_create_root_list(struct dlm_ls *ls) | |||
737 | but no other recovery steps should do anything with them. */ | 737 | but no other recovery steps should do anything with them. */ |
738 | 738 | ||
739 | if (dlm_no_directory(ls)) { | 739 | if (dlm_no_directory(ls)) { |
740 | read_unlock(&ls->ls_rsbtbl[i].lock); | 740 | spin_unlock(&ls->ls_rsbtbl[i].lock); |
741 | continue; | 741 | continue; |
742 | } | 742 | } |
743 | 743 | ||
@@ -745,7 +745,7 @@ int dlm_create_root_list(struct dlm_ls *ls) | |||
745 | list_add(&r->res_root_list, &ls->ls_root_list); | 745 | list_add(&r->res_root_list, &ls->ls_root_list); |
746 | dlm_hold_rsb(r); | 746 | dlm_hold_rsb(r); |
747 | } | 747 | } |
748 | read_unlock(&ls->ls_rsbtbl[i].lock); | 748 | spin_unlock(&ls->ls_rsbtbl[i].lock); |
749 | } | 749 | } |
750 | out: | 750 | out: |
751 | up_write(&ls->ls_root_sem); | 751 | up_write(&ls->ls_root_sem); |
@@ -775,7 +775,7 @@ void dlm_clear_toss_list(struct dlm_ls *ls) | |||
775 | int i; | 775 | int i; |
776 | 776 | ||
777 | for (i = 0; i < ls->ls_rsbtbl_size; i++) { | 777 | for (i = 0; i < ls->ls_rsbtbl_size; i++) { |
778 | write_lock(&ls->ls_rsbtbl[i].lock); | 778 | spin_lock(&ls->ls_rsbtbl[i].lock); |
779 | list_for_each_entry_safe(r, safe, &ls->ls_rsbtbl[i].toss, | 779 | list_for_each_entry_safe(r, safe, &ls->ls_rsbtbl[i].toss, |
780 | res_hashchain) { | 780 | res_hashchain) { |
781 | if (dlm_no_directory(ls) || !is_master(r)) { | 781 | if (dlm_no_directory(ls) || !is_master(r)) { |
@@ -783,7 +783,7 @@ void dlm_clear_toss_list(struct dlm_ls *ls) | |||
783 | dlm_free_rsb(r); | 783 | dlm_free_rsb(r); |
784 | } | 784 | } |
785 | } | 785 | } |
786 | write_unlock(&ls->ls_rsbtbl[i].lock); | 786 | spin_unlock(&ls->ls_rsbtbl[i].lock); |
787 | } | 787 | } |
788 | } | 788 | } |
789 | 789 | ||
diff --git a/fs/dquot.c b/fs/dquot.c index 48c0571f831d..bca3cac4bee7 100644 --- a/fs/dquot.c +++ b/fs/dquot.c | |||
@@ -87,14 +87,17 @@ | |||
87 | #define __DQUOT_PARANOIA | 87 | #define __DQUOT_PARANOIA |
88 | 88 | ||
89 | /* | 89 | /* |
90 | * There are two quota SMP locks. dq_list_lock protects all lists with quotas | 90 | * There are three quota SMP locks. dq_list_lock protects all lists with quotas |
91 | * and quota formats and also dqstats structure containing statistics about the | 91 | * and quota formats, dqstats structure containing statistics about the lists |
92 | * lists. dq_data_lock protects data from dq_dqb and also mem_dqinfo structures | 92 | * dq_data_lock protects data from dq_dqb and also mem_dqinfo structures and |
93 | * and also guards consistency of dquot->dq_dqb with inode->i_blocks, i_bytes. | 93 | * also guards consistency of dquot->dq_dqb with inode->i_blocks, i_bytes. |
94 | * i_blocks and i_bytes updates itself are guarded by i_lock acquired directly | 94 | * i_blocks and i_bytes updates itself are guarded by i_lock acquired directly |
95 | * in inode_add_bytes() and inode_sub_bytes(). | 95 | * in inode_add_bytes() and inode_sub_bytes(). dq_state_lock protects |
96 | * modifications of quota state (on quotaon and quotaoff) and readers who care | ||
97 | * about latest values take it as well. | ||
96 | * | 98 | * |
97 | * The spinlock ordering is hence: dq_data_lock > dq_list_lock > i_lock | 99 | * The spinlock ordering is hence: dq_data_lock > dq_list_lock > i_lock, |
100 | * dq_list_lock > dq_state_lock | ||
98 | * | 101 | * |
99 | * Note that some things (eg. sb pointer, type, id) doesn't change during | 102 | * Note that some things (eg. sb pointer, type, id) doesn't change during |
100 | * the life of the dquot structure and so needn't to be protected by a lock | 103 | * the life of the dquot structure and so needn't to be protected by a lock |
@@ -103,12 +106,7 @@ | |||
103 | * operation is just reading pointers from inode (or not using them at all) the | 106 | * operation is just reading pointers from inode (or not using them at all) the |
104 | * read lock is enough. If pointers are altered function must hold write lock | 107 | * read lock is enough. If pointers are altered function must hold write lock |
105 | * (these locking rules also apply for S_NOQUOTA flag in the inode - note that | 108 | * (these locking rules also apply for S_NOQUOTA flag in the inode - note that |
106 | * for altering the flag i_mutex is also needed). If operation is holding | 109 | * for altering the flag i_mutex is also needed). |
107 | * reference to dquot in other way (e.g. quotactl ops) it must be guarded by | ||
108 | * dqonoff_mutex. | ||
109 | * This locking assures that: | ||
110 | * a) update/access to dquot pointers in inode is serialized | ||
111 | * b) everyone is guarded against invalidate_dquots() | ||
112 | * | 110 | * |
113 | * Each dquot has its dq_lock mutex. Locked dquots might not be referenced | 111 | * Each dquot has its dq_lock mutex. Locked dquots might not be referenced |
114 | * from inodes (dquot_alloc_space() and such don't check the dq_lock). | 112 | * from inodes (dquot_alloc_space() and such don't check the dq_lock). |
@@ -122,10 +120,17 @@ | |||
122 | * Lock ordering (including related VFS locks) is the following: | 120 | * Lock ordering (including related VFS locks) is the following: |
123 | * i_mutex > dqonoff_sem > journal_lock > dqptr_sem > dquot->dq_lock > | 121 | * i_mutex > dqonoff_sem > journal_lock > dqptr_sem > dquot->dq_lock > |
124 | * dqio_mutex | 122 | * dqio_mutex |
123 | * The lock ordering of dqptr_sem imposed by quota code is only dqonoff_sem > | ||
124 | * dqptr_sem. But filesystem has to count with the fact that functions such as | ||
125 | * dquot_alloc_space() acquire dqptr_sem and they usually have to be called | ||
126 | * from inside a transaction to keep filesystem consistency after a crash. Also | ||
127 | * filesystems usually want to do some IO on dquot from ->mark_dirty which is | ||
128 | * called with dqptr_sem held. | ||
125 | * i_mutex on quota files is special (it's below dqio_mutex) | 129 | * i_mutex on quota files is special (it's below dqio_mutex) |
126 | */ | 130 | */ |
127 | 131 | ||
128 | static DEFINE_SPINLOCK(dq_list_lock); | 132 | static DEFINE_SPINLOCK(dq_list_lock); |
133 | static DEFINE_SPINLOCK(dq_state_lock); | ||
129 | DEFINE_SPINLOCK(dq_data_lock); | 134 | DEFINE_SPINLOCK(dq_data_lock); |
130 | 135 | ||
131 | static char *quotatypes[] = INITQFNAMES; | 136 | static char *quotatypes[] = INITQFNAMES; |
@@ -428,7 +433,7 @@ static inline void do_destroy_dquot(struct dquot *dquot) | |||
428 | * quota is disabled and pointers from inodes removed so there cannot be new | 433 | * quota is disabled and pointers from inodes removed so there cannot be new |
429 | * quota users. There can still be some users of quotas due to inodes being | 434 | * quota users. There can still be some users of quotas due to inodes being |
430 | * just deleted or pruned by prune_icache() (those are not attached to any | 435 | * just deleted or pruned by prune_icache() (those are not attached to any |
431 | * list). We have to wait for such users. | 436 | * list) or parallel quotactl call. We have to wait for such users. |
432 | */ | 437 | */ |
433 | static void invalidate_dquots(struct super_block *sb, int type) | 438 | static void invalidate_dquots(struct super_block *sb, int type) |
434 | { | 439 | { |
@@ -600,7 +605,6 @@ static struct shrinker dqcache_shrinker = { | |||
600 | /* | 605 | /* |
601 | * Put reference to dquot | 606 | * Put reference to dquot |
602 | * NOTE: If you change this function please check whether dqput_blocks() works right... | 607 | * NOTE: If you change this function please check whether dqput_blocks() works right... |
603 | * MUST be called with either dqptr_sem or dqonoff_mutex held | ||
604 | */ | 608 | */ |
605 | void dqput(struct dquot *dquot) | 609 | void dqput(struct dquot *dquot) |
606 | { | 610 | { |
@@ -697,36 +701,30 @@ static struct dquot *get_empty_dquot(struct super_block *sb, int type) | |||
697 | } | 701 | } |
698 | 702 | ||
699 | /* | 703 | /* |
700 | * Check whether dquot is in memory. | ||
701 | * MUST be called with either dqptr_sem or dqonoff_mutex held | ||
702 | */ | ||
703 | int dquot_is_cached(struct super_block *sb, unsigned int id, int type) | ||
704 | { | ||
705 | unsigned int hashent = hashfn(sb, id, type); | ||
706 | int ret = 0; | ||
707 | |||
708 | if (!sb_has_quota_active(sb, type)) | ||
709 | return 0; | ||
710 | spin_lock(&dq_list_lock); | ||
711 | if (find_dquot(hashent, sb, id, type) != NODQUOT) | ||
712 | ret = 1; | ||
713 | spin_unlock(&dq_list_lock); | ||
714 | return ret; | ||
715 | } | ||
716 | |||
717 | /* | ||
718 | * Get reference to dquot | 704 | * Get reference to dquot |
719 | * MUST be called with either dqptr_sem or dqonoff_mutex held | 705 | * |
706 | * Locking is slightly tricky here. We are guarded from parallel quotaoff() | ||
707 | * destroying our dquot by: | ||
708 | * a) checking for quota flags under dq_list_lock and | ||
709 | * b) getting a reference to dquot before we release dq_list_lock | ||
720 | */ | 710 | */ |
721 | struct dquot *dqget(struct super_block *sb, unsigned int id, int type) | 711 | struct dquot *dqget(struct super_block *sb, unsigned int id, int type) |
722 | { | 712 | { |
723 | unsigned int hashent = hashfn(sb, id, type); | 713 | unsigned int hashent = hashfn(sb, id, type); |
724 | struct dquot *dquot, *empty = NODQUOT; | 714 | struct dquot *dquot = NODQUOT, *empty = NODQUOT; |
725 | 715 | ||
726 | if (!sb_has_quota_active(sb, type)) | 716 | if (!sb_has_quota_active(sb, type)) |
727 | return NODQUOT; | 717 | return NODQUOT; |
728 | we_slept: | 718 | we_slept: |
729 | spin_lock(&dq_list_lock); | 719 | spin_lock(&dq_list_lock); |
720 | spin_lock(&dq_state_lock); | ||
721 | if (!sb_has_quota_active(sb, type)) { | ||
722 | spin_unlock(&dq_state_lock); | ||
723 | spin_unlock(&dq_list_lock); | ||
724 | goto out; | ||
725 | } | ||
726 | spin_unlock(&dq_state_lock); | ||
727 | |||
730 | if ((dquot = find_dquot(hashent, sb, id, type)) == NODQUOT) { | 728 | if ((dquot = find_dquot(hashent, sb, id, type)) == NODQUOT) { |
731 | if (empty == NODQUOT) { | 729 | if (empty == NODQUOT) { |
732 | spin_unlock(&dq_list_lock); | 730 | spin_unlock(&dq_list_lock); |
@@ -735,6 +733,7 @@ we_slept: | |||
735 | goto we_slept; | 733 | goto we_slept; |
736 | } | 734 | } |
737 | dquot = empty; | 735 | dquot = empty; |
736 | empty = NODQUOT; | ||
738 | dquot->dq_id = id; | 737 | dquot->dq_id = id; |
739 | /* all dquots go on the inuse_list */ | 738 | /* all dquots go on the inuse_list */ |
740 | put_inuse(dquot); | 739 | put_inuse(dquot); |
@@ -749,8 +748,6 @@ we_slept: | |||
749 | dqstats.cache_hits++; | 748 | dqstats.cache_hits++; |
750 | dqstats.lookups++; | 749 | dqstats.lookups++; |
751 | spin_unlock(&dq_list_lock); | 750 | spin_unlock(&dq_list_lock); |
752 | if (empty) | ||
753 | do_destroy_dquot(empty); | ||
754 | } | 751 | } |
755 | /* Wait for dq_lock - after this we know that either dquot_release() is already | 752 | /* Wait for dq_lock - after this we know that either dquot_release() is already |
756 | * finished or it will be canceled due to dq_count > 1 test */ | 753 | * finished or it will be canceled due to dq_count > 1 test */ |
@@ -758,11 +755,15 @@ we_slept: | |||
758 | /* Read the dquot and instantiate it (everything done only if needed) */ | 755 | /* Read the dquot and instantiate it (everything done only if needed) */ |
759 | if (!test_bit(DQ_ACTIVE_B, &dquot->dq_flags) && sb->dq_op->acquire_dquot(dquot) < 0) { | 756 | if (!test_bit(DQ_ACTIVE_B, &dquot->dq_flags) && sb->dq_op->acquire_dquot(dquot) < 0) { |
760 | dqput(dquot); | 757 | dqput(dquot); |
761 | return NODQUOT; | 758 | dquot = NODQUOT; |
759 | goto out; | ||
762 | } | 760 | } |
763 | #ifdef __DQUOT_PARANOIA | 761 | #ifdef __DQUOT_PARANOIA |
764 | BUG_ON(!dquot->dq_sb); /* Has somebody invalidated entry under us? */ | 762 | BUG_ON(!dquot->dq_sb); /* Has somebody invalidated entry under us? */ |
765 | #endif | 763 | #endif |
764 | out: | ||
765 | if (empty) | ||
766 | do_destroy_dquot(empty); | ||
766 | 767 | ||
767 | return dquot; | 768 | return dquot; |
768 | } | 769 | } |
@@ -1198,63 +1199,76 @@ static int info_bdq_free(struct dquot *dquot, qsize_t space) | |||
1198 | } | 1199 | } |
1199 | /* | 1200 | /* |
1200 | * Initialize quota pointers in inode | 1201 | * Initialize quota pointers in inode |
1201 | * Transaction must be started at entry | 1202 | * We do things in a bit complicated way but by that we avoid calling |
1203 | * dqget() and thus filesystem callbacks under dqptr_sem. | ||
1202 | */ | 1204 | */ |
1203 | int dquot_initialize(struct inode *inode, int type) | 1205 | int dquot_initialize(struct inode *inode, int type) |
1204 | { | 1206 | { |
1205 | unsigned int id = 0; | 1207 | unsigned int id = 0; |
1206 | int cnt, ret = 0; | 1208 | int cnt, ret = 0; |
1209 | struct dquot *got[MAXQUOTAS] = { NODQUOT, NODQUOT }; | ||
1210 | struct super_block *sb = inode->i_sb; | ||
1207 | 1211 | ||
1208 | /* First test before acquiring mutex - solves deadlocks when we | 1212 | /* First test before acquiring mutex - solves deadlocks when we |
1209 | * re-enter the quota code and are already holding the mutex */ | 1213 | * re-enter the quota code and are already holding the mutex */ |
1210 | if (IS_NOQUOTA(inode)) | 1214 | if (IS_NOQUOTA(inode)) |
1211 | return 0; | 1215 | return 0; |
1212 | down_write(&sb_dqopt(inode->i_sb)->dqptr_sem); | 1216 | |
1217 | /* First get references to structures we might need. */ | ||
1218 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { | ||
1219 | if (type != -1 && cnt != type) | ||
1220 | continue; | ||
1221 | switch (cnt) { | ||
1222 | case USRQUOTA: | ||
1223 | id = inode->i_uid; | ||
1224 | break; | ||
1225 | case GRPQUOTA: | ||
1226 | id = inode->i_gid; | ||
1227 | break; | ||
1228 | } | ||
1229 | got[cnt] = dqget(sb, id, cnt); | ||
1230 | } | ||
1231 | |||
1232 | down_write(&sb_dqopt(sb)->dqptr_sem); | ||
1213 | /* Having dqptr_sem we know NOQUOTA flags can't be altered... */ | 1233 | /* Having dqptr_sem we know NOQUOTA flags can't be altered... */ |
1214 | if (IS_NOQUOTA(inode)) | 1234 | if (IS_NOQUOTA(inode)) |
1215 | goto out_err; | 1235 | goto out_err; |
1216 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { | 1236 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { |
1217 | if (type != -1 && cnt != type) | 1237 | if (type != -1 && cnt != type) |
1218 | continue; | 1238 | continue; |
1239 | /* Avoid races with quotaoff() */ | ||
1240 | if (!sb_has_quota_active(sb, cnt)) | ||
1241 | continue; | ||
1219 | if (inode->i_dquot[cnt] == NODQUOT) { | 1242 | if (inode->i_dquot[cnt] == NODQUOT) { |
1220 | switch (cnt) { | 1243 | inode->i_dquot[cnt] = got[cnt]; |
1221 | case USRQUOTA: | 1244 | got[cnt] = NODQUOT; |
1222 | id = inode->i_uid; | ||
1223 | break; | ||
1224 | case GRPQUOTA: | ||
1225 | id = inode->i_gid; | ||
1226 | break; | ||
1227 | } | ||
1228 | inode->i_dquot[cnt] = dqget(inode->i_sb, id, cnt); | ||
1229 | } | 1245 | } |
1230 | } | 1246 | } |
1231 | out_err: | 1247 | out_err: |
1232 | up_write(&sb_dqopt(inode->i_sb)->dqptr_sem); | 1248 | up_write(&sb_dqopt(sb)->dqptr_sem); |
1249 | /* Drop unused references */ | ||
1250 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) | ||
1251 | dqput(got[cnt]); | ||
1233 | return ret; | 1252 | return ret; |
1234 | } | 1253 | } |
1235 | 1254 | ||
1236 | /* | 1255 | /* |
1237 | * Release all quotas referenced by inode | 1256 | * Release all quotas referenced by inode |
1238 | * Transaction must be started at an entry | ||
1239 | */ | 1257 | */ |
1240 | int dquot_drop_locked(struct inode *inode) | 1258 | int dquot_drop(struct inode *inode) |
1241 | { | 1259 | { |
1242 | int cnt; | 1260 | int cnt; |
1261 | struct dquot *put[MAXQUOTAS]; | ||
1243 | 1262 | ||
1263 | down_write(&sb_dqopt(inode->i_sb)->dqptr_sem); | ||
1244 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { | 1264 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { |
1245 | if (inode->i_dquot[cnt] != NODQUOT) { | 1265 | put[cnt] = inode->i_dquot[cnt]; |
1246 | dqput(inode->i_dquot[cnt]); | 1266 | inode->i_dquot[cnt] = NODQUOT; |
1247 | inode->i_dquot[cnt] = NODQUOT; | ||
1248 | } | ||
1249 | } | 1267 | } |
1250 | return 0; | ||
1251 | } | ||
1252 | |||
1253 | int dquot_drop(struct inode *inode) | ||
1254 | { | ||
1255 | down_write(&sb_dqopt(inode->i_sb)->dqptr_sem); | ||
1256 | dquot_drop_locked(inode); | ||
1257 | up_write(&sb_dqopt(inode->i_sb)->dqptr_sem); | 1268 | up_write(&sb_dqopt(inode->i_sb)->dqptr_sem); |
1269 | |||
1270 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) | ||
1271 | dqput(put[cnt]); | ||
1258 | return 0; | 1272 | return 0; |
1259 | } | 1273 | } |
1260 | 1274 | ||
@@ -1470,8 +1484,9 @@ int dquot_transfer(struct inode *inode, struct iattr *iattr) | |||
1470 | qsize_t space; | 1484 | qsize_t space; |
1471 | struct dquot *transfer_from[MAXQUOTAS]; | 1485 | struct dquot *transfer_from[MAXQUOTAS]; |
1472 | struct dquot *transfer_to[MAXQUOTAS]; | 1486 | struct dquot *transfer_to[MAXQUOTAS]; |
1473 | int cnt, ret = NO_QUOTA, chuid = (iattr->ia_valid & ATTR_UID) && inode->i_uid != iattr->ia_uid, | 1487 | int cnt, ret = QUOTA_OK; |
1474 | chgid = (iattr->ia_valid & ATTR_GID) && inode->i_gid != iattr->ia_gid; | 1488 | int chuid = iattr->ia_valid & ATTR_UID && inode->i_uid != iattr->ia_uid, |
1489 | chgid = iattr->ia_valid & ATTR_GID && inode->i_gid != iattr->ia_gid; | ||
1475 | char warntype_to[MAXQUOTAS]; | 1490 | char warntype_to[MAXQUOTAS]; |
1476 | char warntype_from_inodes[MAXQUOTAS], warntype_from_space[MAXQUOTAS]; | 1491 | char warntype_from_inodes[MAXQUOTAS], warntype_from_space[MAXQUOTAS]; |
1477 | 1492 | ||
@@ -1479,21 +1494,11 @@ int dquot_transfer(struct inode *inode, struct iattr *iattr) | |||
1479 | * re-enter the quota code and are already holding the mutex */ | 1494 | * re-enter the quota code and are already holding the mutex */ |
1480 | if (IS_NOQUOTA(inode)) | 1495 | if (IS_NOQUOTA(inode)) |
1481 | return QUOTA_OK; | 1496 | return QUOTA_OK; |
1482 | /* Clear the arrays */ | 1497 | /* Initialize the arrays */ |
1483 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { | 1498 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { |
1484 | transfer_to[cnt] = transfer_from[cnt] = NODQUOT; | 1499 | transfer_from[cnt] = NODQUOT; |
1500 | transfer_to[cnt] = NODQUOT; | ||
1485 | warntype_to[cnt] = QUOTA_NL_NOWARN; | 1501 | warntype_to[cnt] = QUOTA_NL_NOWARN; |
1486 | } | ||
1487 | down_write(&sb_dqopt(inode->i_sb)->dqptr_sem); | ||
1488 | /* Now recheck reliably when holding dqptr_sem */ | ||
1489 | if (IS_NOQUOTA(inode)) { /* File without quota accounting? */ | ||
1490 | up_write(&sb_dqopt(inode->i_sb)->dqptr_sem); | ||
1491 | return QUOTA_OK; | ||
1492 | } | ||
1493 | /* First build the transfer_to list - here we can block on | ||
1494 | * reading/instantiating of dquots. We know that the transaction for | ||
1495 | * us was already started so we don't violate lock ranking here */ | ||
1496 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { | ||
1497 | switch (cnt) { | 1502 | switch (cnt) { |
1498 | case USRQUOTA: | 1503 | case USRQUOTA: |
1499 | if (!chuid) | 1504 | if (!chuid) |
@@ -1507,6 +1512,13 @@ int dquot_transfer(struct inode *inode, struct iattr *iattr) | |||
1507 | break; | 1512 | break; |
1508 | } | 1513 | } |
1509 | } | 1514 | } |
1515 | |||
1516 | down_write(&sb_dqopt(inode->i_sb)->dqptr_sem); | ||
1517 | /* Now recheck reliably when holding dqptr_sem */ | ||
1518 | if (IS_NOQUOTA(inode)) { /* File without quota accounting? */ | ||
1519 | up_write(&sb_dqopt(inode->i_sb)->dqptr_sem); | ||
1520 | goto put_all; | ||
1521 | } | ||
1510 | spin_lock(&dq_data_lock); | 1522 | spin_lock(&dq_data_lock); |
1511 | space = inode_get_bytes(inode); | 1523 | space = inode_get_bytes(inode); |
1512 | /* Build the transfer_from list and check the limits */ | 1524 | /* Build the transfer_from list and check the limits */ |
@@ -1517,7 +1529,7 @@ int dquot_transfer(struct inode *inode, struct iattr *iattr) | |||
1517 | if (check_idq(transfer_to[cnt], 1, warntype_to + cnt) == | 1529 | if (check_idq(transfer_to[cnt], 1, warntype_to + cnt) == |
1518 | NO_QUOTA || check_bdq(transfer_to[cnt], space, 0, | 1530 | NO_QUOTA || check_bdq(transfer_to[cnt], space, 0, |
1519 | warntype_to + cnt) == NO_QUOTA) | 1531 | warntype_to + cnt) == NO_QUOTA) |
1520 | goto warn_put_all; | 1532 | goto over_quota; |
1521 | } | 1533 | } |
1522 | 1534 | ||
1523 | /* | 1535 | /* |
@@ -1545,28 +1557,37 @@ int dquot_transfer(struct inode *inode, struct iattr *iattr) | |||
1545 | 1557 | ||
1546 | inode->i_dquot[cnt] = transfer_to[cnt]; | 1558 | inode->i_dquot[cnt] = transfer_to[cnt]; |
1547 | } | 1559 | } |
1548 | ret = QUOTA_OK; | ||
1549 | warn_put_all: | ||
1550 | spin_unlock(&dq_data_lock); | 1560 | spin_unlock(&dq_data_lock); |
1561 | up_write(&sb_dqopt(inode->i_sb)->dqptr_sem); | ||
1562 | |||
1551 | /* Dirtify all the dquots - this can block when journalling */ | 1563 | /* Dirtify all the dquots - this can block when journalling */ |
1552 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { | 1564 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { |
1553 | if (transfer_from[cnt]) | 1565 | if (transfer_from[cnt]) |
1554 | mark_dquot_dirty(transfer_from[cnt]); | 1566 | mark_dquot_dirty(transfer_from[cnt]); |
1555 | if (transfer_to[cnt]) | 1567 | if (transfer_to[cnt]) { |
1556 | mark_dquot_dirty(transfer_to[cnt]); | 1568 | mark_dquot_dirty(transfer_to[cnt]); |
1569 | /* The reference we got is transferred to the inode */ | ||
1570 | transfer_to[cnt] = NODQUOT; | ||
1571 | } | ||
1557 | } | 1572 | } |
1573 | warn_put_all: | ||
1558 | flush_warnings(transfer_to, warntype_to); | 1574 | flush_warnings(transfer_to, warntype_to); |
1559 | flush_warnings(transfer_from, warntype_from_inodes); | 1575 | flush_warnings(transfer_from, warntype_from_inodes); |
1560 | flush_warnings(transfer_from, warntype_from_space); | 1576 | flush_warnings(transfer_from, warntype_from_space); |
1561 | 1577 | put_all: | |
1562 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { | 1578 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { |
1563 | if (ret == QUOTA_OK && transfer_from[cnt] != NODQUOT) | 1579 | dqput(transfer_from[cnt]); |
1564 | dqput(transfer_from[cnt]); | 1580 | dqput(transfer_to[cnt]); |
1565 | if (ret == NO_QUOTA && transfer_to[cnt] != NODQUOT) | ||
1566 | dqput(transfer_to[cnt]); | ||
1567 | } | 1581 | } |
1568 | up_write(&sb_dqopt(inode->i_sb)->dqptr_sem); | ||
1569 | return ret; | 1582 | return ret; |
1583 | over_quota: | ||
1584 | spin_unlock(&dq_data_lock); | ||
1585 | up_write(&sb_dqopt(inode->i_sb)->dqptr_sem); | ||
1586 | /* Clear dquot pointers we don't want to dqput() */ | ||
1587 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) | ||
1588 | transfer_from[cnt] = NODQUOT; | ||
1589 | ret = NO_QUOTA; | ||
1590 | goto warn_put_all; | ||
1570 | } | 1591 | } |
1571 | 1592 | ||
1572 | /* Wrapper for transferring ownership of an inode */ | 1593 | /* Wrapper for transferring ownership of an inode */ |
@@ -1651,19 +1672,24 @@ int vfs_quota_disable(struct super_block *sb, int type, unsigned int flags) | |||
1651 | continue; | 1672 | continue; |
1652 | 1673 | ||
1653 | if (flags & DQUOT_SUSPENDED) { | 1674 | if (flags & DQUOT_SUSPENDED) { |
1675 | spin_lock(&dq_state_lock); | ||
1654 | dqopt->flags |= | 1676 | dqopt->flags |= |
1655 | dquot_state_flag(DQUOT_SUSPENDED, cnt); | 1677 | dquot_state_flag(DQUOT_SUSPENDED, cnt); |
1678 | spin_unlock(&dq_state_lock); | ||
1656 | } else { | 1679 | } else { |
1680 | spin_lock(&dq_state_lock); | ||
1657 | dqopt->flags &= ~dquot_state_flag(flags, cnt); | 1681 | dqopt->flags &= ~dquot_state_flag(flags, cnt); |
1658 | /* Turning off suspended quotas? */ | 1682 | /* Turning off suspended quotas? */ |
1659 | if (!sb_has_quota_loaded(sb, cnt) && | 1683 | if (!sb_has_quota_loaded(sb, cnt) && |
1660 | sb_has_quota_suspended(sb, cnt)) { | 1684 | sb_has_quota_suspended(sb, cnt)) { |
1661 | dqopt->flags &= ~dquot_state_flag( | 1685 | dqopt->flags &= ~dquot_state_flag( |
1662 | DQUOT_SUSPENDED, cnt); | 1686 | DQUOT_SUSPENDED, cnt); |
1687 | spin_unlock(&dq_state_lock); | ||
1663 | iput(dqopt->files[cnt]); | 1688 | iput(dqopt->files[cnt]); |
1664 | dqopt->files[cnt] = NULL; | 1689 | dqopt->files[cnt] = NULL; |
1665 | continue; | 1690 | continue; |
1666 | } | 1691 | } |
1692 | spin_unlock(&dq_state_lock); | ||
1667 | } | 1693 | } |
1668 | 1694 | ||
1669 | /* We still have to keep quota loaded? */ | 1695 | /* We still have to keep quota loaded? */ |
@@ -1830,7 +1856,9 @@ static int vfs_load_quota_inode(struct inode *inode, int type, int format_id, | |||
1830 | } | 1856 | } |
1831 | mutex_unlock(&dqopt->dqio_mutex); | 1857 | mutex_unlock(&dqopt->dqio_mutex); |
1832 | mutex_unlock(&inode->i_mutex); | 1858 | mutex_unlock(&inode->i_mutex); |
1859 | spin_lock(&dq_state_lock); | ||
1833 | dqopt->flags |= dquot_state_flag(flags, type); | 1860 | dqopt->flags |= dquot_state_flag(flags, type); |
1861 | spin_unlock(&dq_state_lock); | ||
1834 | 1862 | ||
1835 | add_dquot_ref(sb, type); | 1863 | add_dquot_ref(sb, type); |
1836 | mutex_unlock(&dqopt->dqonoff_mutex); | 1864 | mutex_unlock(&dqopt->dqonoff_mutex); |
@@ -1872,9 +1900,11 @@ static int vfs_quota_on_remount(struct super_block *sb, int type) | |||
1872 | } | 1900 | } |
1873 | inode = dqopt->files[type]; | 1901 | inode = dqopt->files[type]; |
1874 | dqopt->files[type] = NULL; | 1902 | dqopt->files[type] = NULL; |
1903 | spin_lock(&dq_state_lock); | ||
1875 | flags = dqopt->flags & dquot_state_flag(DQUOT_USAGE_ENABLED | | 1904 | flags = dqopt->flags & dquot_state_flag(DQUOT_USAGE_ENABLED | |
1876 | DQUOT_LIMITS_ENABLED, type); | 1905 | DQUOT_LIMITS_ENABLED, type); |
1877 | dqopt->flags &= ~dquot_state_flag(DQUOT_STATE_FLAGS, type); | 1906 | dqopt->flags &= ~dquot_state_flag(DQUOT_STATE_FLAGS, type); |
1907 | spin_unlock(&dq_state_lock); | ||
1878 | mutex_unlock(&dqopt->dqonoff_mutex); | 1908 | mutex_unlock(&dqopt->dqonoff_mutex); |
1879 | 1909 | ||
1880 | flags = dquot_generic_flag(flags, type); | 1910 | flags = dquot_generic_flag(flags, type); |
@@ -1952,7 +1982,9 @@ int vfs_quota_enable(struct inode *inode, int type, int format_id, | |||
1952 | ret = -EBUSY; | 1982 | ret = -EBUSY; |
1953 | goto out_lock; | 1983 | goto out_lock; |
1954 | } | 1984 | } |
1985 | spin_lock(&dq_state_lock); | ||
1955 | sb_dqopt(sb)->flags |= dquot_state_flag(flags, type); | 1986 | sb_dqopt(sb)->flags |= dquot_state_flag(flags, type); |
1987 | spin_unlock(&dq_state_lock); | ||
1956 | out_lock: | 1988 | out_lock: |
1957 | mutex_unlock(&dqopt->dqonoff_mutex); | 1989 | mutex_unlock(&dqopt->dqonoff_mutex); |
1958 | return ret; | 1990 | return ret; |
@@ -2039,14 +2071,12 @@ int vfs_get_dqblk(struct super_block *sb, int type, qid_t id, struct if_dqblk *d | |||
2039 | { | 2071 | { |
2040 | struct dquot *dquot; | 2072 | struct dquot *dquot; |
2041 | 2073 | ||
2042 | mutex_lock(&sb_dqopt(sb)->dqonoff_mutex); | 2074 | dquot = dqget(sb, id, type); |
2043 | if (!(dquot = dqget(sb, id, type))) { | 2075 | if (dquot == NODQUOT) |
2044 | mutex_unlock(&sb_dqopt(sb)->dqonoff_mutex); | ||
2045 | return -ESRCH; | 2076 | return -ESRCH; |
2046 | } | ||
2047 | do_get_dqblk(dquot, di); | 2077 | do_get_dqblk(dquot, di); |
2048 | dqput(dquot); | 2078 | dqput(dquot); |
2049 | mutex_unlock(&sb_dqopt(sb)->dqonoff_mutex); | 2079 | |
2050 | return 0; | 2080 | return 0; |
2051 | } | 2081 | } |
2052 | 2082 | ||
@@ -2130,7 +2160,6 @@ int vfs_set_dqblk(struct super_block *sb, int type, qid_t id, struct if_dqblk *d | |||
2130 | struct dquot *dquot; | 2160 | struct dquot *dquot; |
2131 | int rc; | 2161 | int rc; |
2132 | 2162 | ||
2133 | mutex_lock(&sb_dqopt(sb)->dqonoff_mutex); | ||
2134 | dquot = dqget(sb, id, type); | 2163 | dquot = dqget(sb, id, type); |
2135 | if (!dquot) { | 2164 | if (!dquot) { |
2136 | rc = -ESRCH; | 2165 | rc = -ESRCH; |
@@ -2139,7 +2168,6 @@ int vfs_set_dqblk(struct super_block *sb, int type, qid_t id, struct if_dqblk *d | |||
2139 | rc = do_set_dqblk(dquot, di); | 2168 | rc = do_set_dqblk(dquot, di); |
2140 | dqput(dquot); | 2169 | dqput(dquot); |
2141 | out: | 2170 | out: |
2142 | mutex_unlock(&sb_dqopt(sb)->dqonoff_mutex); | ||
2143 | return rc; | 2171 | return rc; |
2144 | } | 2172 | } |
2145 | 2173 | ||
@@ -2370,11 +2398,9 @@ EXPORT_SYMBOL(dquot_release); | |||
2370 | EXPORT_SYMBOL(dquot_mark_dquot_dirty); | 2398 | EXPORT_SYMBOL(dquot_mark_dquot_dirty); |
2371 | EXPORT_SYMBOL(dquot_initialize); | 2399 | EXPORT_SYMBOL(dquot_initialize); |
2372 | EXPORT_SYMBOL(dquot_drop); | 2400 | EXPORT_SYMBOL(dquot_drop); |
2373 | EXPORT_SYMBOL(dquot_drop_locked); | ||
2374 | EXPORT_SYMBOL(vfs_dq_drop); | 2401 | EXPORT_SYMBOL(vfs_dq_drop); |
2375 | EXPORT_SYMBOL(dqget); | 2402 | EXPORT_SYMBOL(dqget); |
2376 | EXPORT_SYMBOL(dqput); | 2403 | EXPORT_SYMBOL(dqput); |
2377 | EXPORT_SYMBOL(dquot_is_cached); | ||
2378 | EXPORT_SYMBOL(dquot_alloc_space); | 2404 | EXPORT_SYMBOL(dquot_alloc_space); |
2379 | EXPORT_SYMBOL(dquot_alloc_inode); | 2405 | EXPORT_SYMBOL(dquot_alloc_inode); |
2380 | EXPORT_SYMBOL(dquot_free_space); | 2406 | EXPORT_SYMBOL(dquot_free_space); |
diff --git a/fs/ecryptfs/Kconfig b/fs/ecryptfs/Kconfig new file mode 100644 index 000000000000..0c754e64232b --- /dev/null +++ b/fs/ecryptfs/Kconfig | |||
@@ -0,0 +1,11 @@ | |||
1 | config ECRYPT_FS | ||
2 | tristate "eCrypt filesystem layer support (EXPERIMENTAL)" | ||
3 | depends on EXPERIMENTAL && KEYS && CRYPTO && NET | ||
4 | help | ||
5 | Encrypted filesystem that operates on the VFS layer. See | ||
6 | <file:Documentation/filesystems/ecryptfs.txt> to learn more about | ||
7 | eCryptfs. Userspace components are required and can be | ||
8 | obtained from <http://ecryptfs.sf.net>. | ||
9 | |||
10 | To compile this file system support as a module, choose M here: the | ||
11 | module will be called ecryptfs. | ||
diff --git a/fs/ecryptfs/crypto.c b/fs/ecryptfs/crypto.c index c01e043670e2..f6caeb1d1106 100644 --- a/fs/ecryptfs/crypto.c +++ b/fs/ecryptfs/crypto.c | |||
@@ -1716,7 +1716,7 @@ static int ecryptfs_copy_filename(char **copied_name, size_t *copied_name_size, | |||
1716 | { | 1716 | { |
1717 | int rc = 0; | 1717 | int rc = 0; |
1718 | 1718 | ||
1719 | (*copied_name) = kmalloc((name_size + 2), GFP_KERNEL); | 1719 | (*copied_name) = kmalloc((name_size + 1), GFP_KERNEL); |
1720 | if (!(*copied_name)) { | 1720 | if (!(*copied_name)) { |
1721 | rc = -ENOMEM; | 1721 | rc = -ENOMEM; |
1722 | goto out; | 1722 | goto out; |
@@ -1726,7 +1726,7 @@ static int ecryptfs_copy_filename(char **copied_name, size_t *copied_name_size, | |||
1726 | * in printing out the | 1726 | * in printing out the |
1727 | * string in debug | 1727 | * string in debug |
1728 | * messages */ | 1728 | * messages */ |
1729 | (*copied_name_size) = (name_size + 1); | 1729 | (*copied_name_size) = name_size; |
1730 | out: | 1730 | out: |
1731 | return rc; | 1731 | return rc; |
1732 | } | 1732 | } |
diff --git a/fs/efs/Kconfig b/fs/efs/Kconfig new file mode 100644 index 000000000000..6ebfc1c207a8 --- /dev/null +++ b/fs/efs/Kconfig | |||
@@ -0,0 +1,14 @@ | |||
1 | config EFS_FS | ||
2 | tristate "EFS file system support (read only) (EXPERIMENTAL)" | ||
3 | depends on BLOCK && EXPERIMENTAL | ||
4 | help | ||
5 | EFS is an older file system used for non-ISO9660 CD-ROMs and hard | ||
6 | disk partitions by SGI's IRIX operating system (IRIX 6.0 and newer | ||
7 | uses the XFS file system for hard disk partitions however). | ||
8 | |||
9 | This implementation only offers read-only access. If you don't know | ||
10 | what all this is about, it's safe to say N. For more information | ||
11 | about EFS see its home page at <http://aeschi.ch.eu.org/efs/>. | ||
12 | |||
13 | To compile the EFS file system support as a module, choose M here: the | ||
14 | module will be called efs. | ||
diff --git a/fs/eventfd.c b/fs/eventfd.c index 08bf558d0408..5de2c2db3aa2 100644 --- a/fs/eventfd.c +++ b/fs/eventfd.c | |||
@@ -198,7 +198,7 @@ struct file *eventfd_fget(int fd) | |||
198 | return file; | 198 | return file; |
199 | } | 199 | } |
200 | 200 | ||
201 | asmlinkage long sys_eventfd2(unsigned int count, int flags) | 201 | SYSCALL_DEFINE2(eventfd2, unsigned int, count, int, flags) |
202 | { | 202 | { |
203 | int fd; | 203 | int fd; |
204 | struct eventfd_ctx *ctx; | 204 | struct eventfd_ctx *ctx; |
@@ -228,8 +228,7 @@ asmlinkage long sys_eventfd2(unsigned int count, int flags) | |||
228 | return fd; | 228 | return fd; |
229 | } | 229 | } |
230 | 230 | ||
231 | asmlinkage long sys_eventfd(unsigned int count) | 231 | SYSCALL_DEFINE1(eventfd, unsigned int, count) |
232 | { | 232 | { |
233 | return sys_eventfd2(count, 0); | 233 | return sys_eventfd2(count, 0); |
234 | } | 234 | } |
235 | |||
diff --git a/fs/eventpoll.c b/fs/eventpoll.c index 96355d505347..011b9b8c90c6 100644 --- a/fs/eventpoll.c +++ b/fs/eventpoll.c | |||
@@ -234,8 +234,6 @@ struct ep_pqueue { | |||
234 | /* | 234 | /* |
235 | * Configuration options available inside /proc/sys/fs/epoll/ | 235 | * Configuration options available inside /proc/sys/fs/epoll/ |
236 | */ | 236 | */ |
237 | /* Maximum number of epoll devices, per user */ | ||
238 | static int max_user_instances __read_mostly; | ||
239 | /* Maximum number of epoll watched descriptors, per user */ | 237 | /* Maximum number of epoll watched descriptors, per user */ |
240 | static int max_user_watches __read_mostly; | 238 | static int max_user_watches __read_mostly; |
241 | 239 | ||
@@ -261,14 +259,6 @@ static int zero; | |||
261 | 259 | ||
262 | ctl_table epoll_table[] = { | 260 | ctl_table epoll_table[] = { |
263 | { | 261 | { |
264 | .procname = "max_user_instances", | ||
265 | .data = &max_user_instances, | ||
266 | .maxlen = sizeof(int), | ||
267 | .mode = 0644, | ||
268 | .proc_handler = &proc_dointvec_minmax, | ||
269 | .extra1 = &zero, | ||
270 | }, | ||
271 | { | ||
272 | .procname = "max_user_watches", | 262 | .procname = "max_user_watches", |
273 | .data = &max_user_watches, | 263 | .data = &max_user_watches, |
274 | .maxlen = sizeof(int), | 264 | .maxlen = sizeof(int), |
@@ -491,7 +481,6 @@ static void ep_free(struct eventpoll *ep) | |||
491 | 481 | ||
492 | mutex_unlock(&epmutex); | 482 | mutex_unlock(&epmutex); |
493 | mutex_destroy(&ep->mtx); | 483 | mutex_destroy(&ep->mtx); |
494 | atomic_dec(&ep->user->epoll_devs); | ||
495 | free_uid(ep->user); | 484 | free_uid(ep->user); |
496 | kfree(ep); | 485 | kfree(ep); |
497 | } | 486 | } |
@@ -581,10 +570,6 @@ static int ep_alloc(struct eventpoll **pep) | |||
581 | struct eventpoll *ep; | 570 | struct eventpoll *ep; |
582 | 571 | ||
583 | user = get_current_user(); | 572 | user = get_current_user(); |
584 | error = -EMFILE; | ||
585 | if (unlikely(atomic_read(&user->epoll_devs) >= | ||
586 | max_user_instances)) | ||
587 | goto free_uid; | ||
588 | error = -ENOMEM; | 573 | error = -ENOMEM; |
589 | ep = kzalloc(sizeof(*ep), GFP_KERNEL); | 574 | ep = kzalloc(sizeof(*ep), GFP_KERNEL); |
590 | if (unlikely(!ep)) | 575 | if (unlikely(!ep)) |
@@ -1110,7 +1095,7 @@ retry: | |||
1110 | /* | 1095 | /* |
1111 | * Open an eventpoll file descriptor. | 1096 | * Open an eventpoll file descriptor. |
1112 | */ | 1097 | */ |
1113 | asmlinkage long sys_epoll_create1(int flags) | 1098 | SYSCALL_DEFINE1(epoll_create1, int, flags) |
1114 | { | 1099 | { |
1115 | int error, fd = -1; | 1100 | int error, fd = -1; |
1116 | struct eventpoll *ep; | 1101 | struct eventpoll *ep; |
@@ -1141,7 +1126,6 @@ asmlinkage long sys_epoll_create1(int flags) | |||
1141 | flags & O_CLOEXEC); | 1126 | flags & O_CLOEXEC); |
1142 | if (fd < 0) | 1127 | if (fd < 0) |
1143 | ep_free(ep); | 1128 | ep_free(ep); |
1144 | atomic_inc(&ep->user->epoll_devs); | ||
1145 | 1129 | ||
1146 | error_return: | 1130 | error_return: |
1147 | DNPRINTK(3, (KERN_INFO "[%p] eventpoll: sys_epoll_create(%d) = %d\n", | 1131 | DNPRINTK(3, (KERN_INFO "[%p] eventpoll: sys_epoll_create(%d) = %d\n", |
@@ -1150,7 +1134,7 @@ error_return: | |||
1150 | return fd; | 1134 | return fd; |
1151 | } | 1135 | } |
1152 | 1136 | ||
1153 | asmlinkage long sys_epoll_create(int size) | 1137 | SYSCALL_DEFINE1(epoll_create, int, size) |
1154 | { | 1138 | { |
1155 | if (size < 0) | 1139 | if (size < 0) |
1156 | return -EINVAL; | 1140 | return -EINVAL; |
@@ -1163,8 +1147,8 @@ asmlinkage long sys_epoll_create(int size) | |||
1163 | * the eventpoll file that enables the insertion/removal/change of | 1147 | * the eventpoll file that enables the insertion/removal/change of |
1164 | * file descriptors inside the interest set. | 1148 | * file descriptors inside the interest set. |
1165 | */ | 1149 | */ |
1166 | asmlinkage long sys_epoll_ctl(int epfd, int op, int fd, | 1150 | SYSCALL_DEFINE4(epoll_ctl, int, epfd, int, op, int, fd, |
1167 | struct epoll_event __user *event) | 1151 | struct epoll_event __user *, event) |
1168 | { | 1152 | { |
1169 | int error; | 1153 | int error; |
1170 | struct file *file, *tfile; | 1154 | struct file *file, *tfile; |
@@ -1261,8 +1245,8 @@ error_return: | |||
1261 | * Implement the event wait interface for the eventpoll file. It is the kernel | 1245 | * Implement the event wait interface for the eventpoll file. It is the kernel |
1262 | * part of the user space epoll_wait(2). | 1246 | * part of the user space epoll_wait(2). |
1263 | */ | 1247 | */ |
1264 | asmlinkage long sys_epoll_wait(int epfd, struct epoll_event __user *events, | 1248 | SYSCALL_DEFINE4(epoll_wait, int, epfd, struct epoll_event __user *, events, |
1265 | int maxevents, int timeout) | 1249 | int, maxevents, int, timeout) |
1266 | { | 1250 | { |
1267 | int error; | 1251 | int error; |
1268 | struct file *file; | 1252 | struct file *file; |
@@ -1319,9 +1303,9 @@ error_return: | |||
1319 | * Implement the event wait interface for the eventpoll file. It is the kernel | 1303 | * Implement the event wait interface for the eventpoll file. It is the kernel |
1320 | * part of the user space epoll_pwait(2). | 1304 | * part of the user space epoll_pwait(2). |
1321 | */ | 1305 | */ |
1322 | asmlinkage long sys_epoll_pwait(int epfd, struct epoll_event __user *events, | 1306 | SYSCALL_DEFINE6(epoll_pwait, int, epfd, struct epoll_event __user *, events, |
1323 | int maxevents, int timeout, const sigset_t __user *sigmask, | 1307 | int, maxevents, int, timeout, const sigset_t __user *, sigmask, |
1324 | size_t sigsetsize) | 1308 | size_t, sigsetsize) |
1325 | { | 1309 | { |
1326 | int error; | 1310 | int error; |
1327 | sigset_t ksigmask, sigsaved; | 1311 | sigset_t ksigmask, sigsaved; |
@@ -1366,8 +1350,10 @@ static int __init eventpoll_init(void) | |||
1366 | struct sysinfo si; | 1350 | struct sysinfo si; |
1367 | 1351 | ||
1368 | si_meminfo(&si); | 1352 | si_meminfo(&si); |
1369 | max_user_instances = 128; | 1353 | /* |
1370 | max_user_watches = (((si.totalram - si.totalhigh) / 32) << PAGE_SHIFT) / | 1354 | * Allows top 4% of lomem to be allocated for epoll watches (per user). |
1355 | */ | ||
1356 | max_user_watches = (((si.totalram - si.totalhigh) / 25) << PAGE_SHIFT) / | ||
1371 | EP_ITEM_COST; | 1357 | EP_ITEM_COST; |
1372 | 1358 | ||
1373 | /* Initialize the structure used to perform safe poll wait head wake ups */ | 1359 | /* Initialize the structure used to perform safe poll wait head wake ups */ |
@@ -99,7 +99,7 @@ static inline void put_binfmt(struct linux_binfmt * fmt) | |||
99 | * | 99 | * |
100 | * Also note that we take the address to load from from the file itself. | 100 | * Also note that we take the address to load from from the file itself. |
101 | */ | 101 | */ |
102 | asmlinkage long sys_uselib(const char __user * library) | 102 | SYSCALL_DEFINE1(uselib, const char __user *, library) |
103 | { | 103 | { |
104 | struct file *file; | 104 | struct file *file; |
105 | struct nameidata nd; | 105 | struct nameidata nd; |
@@ -1049,16 +1049,32 @@ EXPORT_SYMBOL(install_exec_creds); | |||
1049 | * - the caller must hold current->cred_exec_mutex to protect against | 1049 | * - the caller must hold current->cred_exec_mutex to protect against |
1050 | * PTRACE_ATTACH | 1050 | * PTRACE_ATTACH |
1051 | */ | 1051 | */ |
1052 | void check_unsafe_exec(struct linux_binprm *bprm) | 1052 | void check_unsafe_exec(struct linux_binprm *bprm, struct files_struct *files) |
1053 | { | 1053 | { |
1054 | struct task_struct *p = current; | 1054 | struct task_struct *p = current, *t; |
1055 | unsigned long flags; | ||
1056 | unsigned n_fs, n_files, n_sighand; | ||
1055 | 1057 | ||
1056 | bprm->unsafe = tracehook_unsafe_exec(p); | 1058 | bprm->unsafe = tracehook_unsafe_exec(p); |
1057 | 1059 | ||
1058 | if (atomic_read(&p->fs->count) > 1 || | 1060 | n_fs = 1; |
1059 | atomic_read(&p->files->count) > 1 || | 1061 | n_files = 1; |
1060 | atomic_read(&p->sighand->count) > 1) | 1062 | n_sighand = 1; |
1063 | lock_task_sighand(p, &flags); | ||
1064 | for (t = next_thread(p); t != p; t = next_thread(t)) { | ||
1065 | if (t->fs == p->fs) | ||
1066 | n_fs++; | ||
1067 | if (t->files == files) | ||
1068 | n_files++; | ||
1069 | n_sighand++; | ||
1070 | } | ||
1071 | |||
1072 | if (atomic_read(&p->fs->count) > n_fs || | ||
1073 | atomic_read(&p->files->count) > n_files || | ||
1074 | atomic_read(&p->sighand->count) > n_sighand) | ||
1061 | bprm->unsafe |= LSM_UNSAFE_SHARE; | 1075 | bprm->unsafe |= LSM_UNSAFE_SHARE; |
1076 | |||
1077 | unlock_task_sighand(p, &flags); | ||
1062 | } | 1078 | } |
1063 | 1079 | ||
1064 | /* | 1080 | /* |
@@ -1273,7 +1289,7 @@ int do_execve(char * filename, | |||
1273 | bprm->cred = prepare_exec_creds(); | 1289 | bprm->cred = prepare_exec_creds(); |
1274 | if (!bprm->cred) | 1290 | if (!bprm->cred) |
1275 | goto out_unlock; | 1291 | goto out_unlock; |
1276 | check_unsafe_exec(bprm); | 1292 | check_unsafe_exec(bprm, displaced); |
1277 | 1293 | ||
1278 | file = open_exec(filename); | 1294 | file = open_exec(filename); |
1279 | retval = PTR_ERR(file); | 1295 | retval = PTR_ERR(file); |
diff --git a/fs/ext2/dir.c b/fs/ext2/dir.c index 9a0fc400f91c..2999d72153b7 100644 --- a/fs/ext2/dir.c +++ b/fs/ext2/dir.c | |||
@@ -95,10 +95,13 @@ static int ext2_commit_chunk(struct page *page, loff_t pos, unsigned len) | |||
95 | mark_inode_dirty(dir); | 95 | mark_inode_dirty(dir); |
96 | } | 96 | } |
97 | 97 | ||
98 | if (IS_DIRSYNC(dir)) | 98 | if (IS_DIRSYNC(dir)) { |
99 | err = write_one_page(page, 1); | 99 | err = write_one_page(page, 1); |
100 | else | 100 | if (!err) |
101 | err = ext2_sync_inode(dir); | ||
102 | } else { | ||
101 | unlock_page(page); | 103 | unlock_page(page); |
104 | } | ||
102 | 105 | ||
103 | return err; | 106 | return err; |
104 | } | 107 | } |
diff --git a/fs/ext2/super.c b/fs/ext2/super.c index da8bdeaa2e6d..7c6e3606f0ec 100644 --- a/fs/ext2/super.c +++ b/fs/ext2/super.c | |||
@@ -1185,9 +1185,12 @@ static int ext2_remount (struct super_block * sb, int * flags, char * data) | |||
1185 | es = sbi->s_es; | 1185 | es = sbi->s_es; |
1186 | if (((sbi->s_mount_opt & EXT2_MOUNT_XIP) != | 1186 | if (((sbi->s_mount_opt & EXT2_MOUNT_XIP) != |
1187 | (old_mount_opt & EXT2_MOUNT_XIP)) && | 1187 | (old_mount_opt & EXT2_MOUNT_XIP)) && |
1188 | invalidate_inodes(sb)) | 1188 | invalidate_inodes(sb)) { |
1189 | ext2_warning(sb, __func__, "busy inodes while remounting "\ | 1189 | ext2_warning(sb, __func__, "refusing change of xip flag " |
1190 | "xip remain in cache (no functional problem)"); | 1190 | "with busy inodes while remounting"); |
1191 | sbi->s_mount_opt &= ~EXT2_MOUNT_XIP; | ||
1192 | sbi->s_mount_opt |= old_mount_opt & EXT2_MOUNT_XIP; | ||
1193 | } | ||
1191 | if ((*flags & MS_RDONLY) == (sb->s_flags & MS_RDONLY)) | 1194 | if ((*flags & MS_RDONLY) == (sb->s_flags & MS_RDONLY)) |
1192 | return 0; | 1195 | return 0; |
1193 | if (*flags & MS_RDONLY) { | 1196 | if (*flags & MS_RDONLY) { |
diff --git a/fs/ext3/namei.c b/fs/ext3/namei.c index 69a3d19ca9fd..4db4ffa1edad 100644 --- a/fs/ext3/namei.c +++ b/fs/ext3/namei.c | |||
@@ -1358,7 +1358,7 @@ static int make_indexed_dir(handle_t *handle, struct dentry *dentry, | |||
1358 | struct fake_dirent *fde; | 1358 | struct fake_dirent *fde; |
1359 | 1359 | ||
1360 | blocksize = dir->i_sb->s_blocksize; | 1360 | blocksize = dir->i_sb->s_blocksize; |
1361 | dxtrace(printk("Creating index\n")); | 1361 | dxtrace(printk(KERN_DEBUG "Creating index: inode %lu\n", dir->i_ino)); |
1362 | retval = ext3_journal_get_write_access(handle, bh); | 1362 | retval = ext3_journal_get_write_access(handle, bh); |
1363 | if (retval) { | 1363 | if (retval) { |
1364 | ext3_std_error(dir->i_sb, retval); | 1364 | ext3_std_error(dir->i_sb, retval); |
@@ -1367,6 +1367,19 @@ static int make_indexed_dir(handle_t *handle, struct dentry *dentry, | |||
1367 | } | 1367 | } |
1368 | root = (struct dx_root *) bh->b_data; | 1368 | root = (struct dx_root *) bh->b_data; |
1369 | 1369 | ||
1370 | /* The 0th block becomes the root, move the dirents out */ | ||
1371 | fde = &root->dotdot; | ||
1372 | de = (struct ext3_dir_entry_2 *)((char *)fde + | ||
1373 | ext3_rec_len_from_disk(fde->rec_len)); | ||
1374 | if ((char *) de >= (((char *) root) + blocksize)) { | ||
1375 | ext3_error(dir->i_sb, __func__, | ||
1376 | "invalid rec_len for '..' in inode %lu", | ||
1377 | dir->i_ino); | ||
1378 | brelse(bh); | ||
1379 | return -EIO; | ||
1380 | } | ||
1381 | len = ((char *) root) + blocksize - (char *) de; | ||
1382 | |||
1370 | bh2 = ext3_append (handle, dir, &block, &retval); | 1383 | bh2 = ext3_append (handle, dir, &block, &retval); |
1371 | if (!(bh2)) { | 1384 | if (!(bh2)) { |
1372 | brelse(bh); | 1385 | brelse(bh); |
@@ -1375,11 +1388,6 @@ static int make_indexed_dir(handle_t *handle, struct dentry *dentry, | |||
1375 | EXT3_I(dir)->i_flags |= EXT3_INDEX_FL; | 1388 | EXT3_I(dir)->i_flags |= EXT3_INDEX_FL; |
1376 | data1 = bh2->b_data; | 1389 | data1 = bh2->b_data; |
1377 | 1390 | ||
1378 | /* The 0th block becomes the root, move the dirents out */ | ||
1379 | fde = &root->dotdot; | ||
1380 | de = (struct ext3_dir_entry_2 *)((char *)fde + | ||
1381 | ext3_rec_len_from_disk(fde->rec_len)); | ||
1382 | len = ((char *) root) + blocksize - (char *) de; | ||
1383 | memcpy (data1, de, len); | 1391 | memcpy (data1, de, len); |
1384 | de = (struct ext3_dir_entry_2 *) data1; | 1392 | de = (struct ext3_dir_entry_2 *) data1; |
1385 | top = data1 + len; | 1393 | top = data1 + len; |
diff --git a/fs/ext3/super.c b/fs/ext3/super.c index b70d90e08a3c..4a970411a458 100644 --- a/fs/ext3/super.c +++ b/fs/ext3/super.c | |||
@@ -2428,12 +2428,13 @@ static void ext3_write_super (struct super_block * sb) | |||
2428 | 2428 | ||
2429 | static int ext3_sync_fs(struct super_block *sb, int wait) | 2429 | static int ext3_sync_fs(struct super_block *sb, int wait) |
2430 | { | 2430 | { |
2431 | sb->s_dirt = 0; | 2431 | tid_t target; |
2432 | if (wait) | ||
2433 | ext3_force_commit(sb); | ||
2434 | else | ||
2435 | journal_start_commit(EXT3_SB(sb)->s_journal, NULL); | ||
2436 | 2432 | ||
2433 | sb->s_dirt = 0; | ||
2434 | if (journal_start_commit(EXT3_SB(sb)->s_journal, &target)) { | ||
2435 | if (wait) | ||
2436 | log_wait_commit(EXT3_SB(sb)->s_journal, target); | ||
2437 | } | ||
2437 | return 0; | 2438 | return 0; |
2438 | } | 2439 | } |
2439 | 2440 | ||
diff --git a/fs/ext4/balloc.c b/fs/ext4/balloc.c index 6bba06b09dd1..9a50b8052dcf 100644 --- a/fs/ext4/balloc.c +++ b/fs/ext4/balloc.c | |||
@@ -684,15 +684,15 @@ ext4_fsblk_t ext4_count_free_blocks(struct super_block *sb) | |||
684 | gdp = ext4_get_group_desc(sb, i, NULL); | 684 | gdp = ext4_get_group_desc(sb, i, NULL); |
685 | if (!gdp) | 685 | if (!gdp) |
686 | continue; | 686 | continue; |
687 | desc_count += le16_to_cpu(gdp->bg_free_blocks_count); | 687 | desc_count += ext4_free_blks_count(sb, gdp); |
688 | brelse(bitmap_bh); | 688 | brelse(bitmap_bh); |
689 | bitmap_bh = ext4_read_block_bitmap(sb, i); | 689 | bitmap_bh = ext4_read_block_bitmap(sb, i); |
690 | if (bitmap_bh == NULL) | 690 | if (bitmap_bh == NULL) |
691 | continue; | 691 | continue; |
692 | 692 | ||
693 | x = ext4_count_free(bitmap_bh, sb->s_blocksize); | 693 | x = ext4_count_free(bitmap_bh, sb->s_blocksize); |
694 | printk(KERN_DEBUG "group %lu: stored = %d, counted = %u\n", | 694 | printk(KERN_DEBUG "group %u: stored = %d, counted = %u\n", |
695 | i, le16_to_cpu(gdp->bg_free_blocks_count), x); | 695 | i, ext4_free_blks_count(sb, gdp), x); |
696 | bitmap_count += x; | 696 | bitmap_count += x; |
697 | } | 697 | } |
698 | brelse(bitmap_bh); | 698 | brelse(bitmap_bh); |
diff --git a/fs/ext4/ext4.h b/fs/ext4/ext4.h index c668e4377d76..b0c87dce66a3 100644 --- a/fs/ext4/ext4.h +++ b/fs/ext4/ext4.h | |||
@@ -868,7 +868,7 @@ static inline unsigned ext4_rec_len_from_disk(__le16 dlen) | |||
868 | { | 868 | { |
869 | unsigned len = le16_to_cpu(dlen); | 869 | unsigned len = le16_to_cpu(dlen); |
870 | 870 | ||
871 | if (len == EXT4_MAX_REC_LEN) | 871 | if (len == EXT4_MAX_REC_LEN || len == 0) |
872 | return 1 << 16; | 872 | return 1 << 16; |
873 | return len; | 873 | return len; |
874 | } | 874 | } |
@@ -1206,8 +1206,11 @@ static inline void ext4_r_blocks_count_set(struct ext4_super_block *es, | |||
1206 | 1206 | ||
1207 | static inline loff_t ext4_isize(struct ext4_inode *raw_inode) | 1207 | static inline loff_t ext4_isize(struct ext4_inode *raw_inode) |
1208 | { | 1208 | { |
1209 | return ((loff_t)le32_to_cpu(raw_inode->i_size_high) << 32) | | 1209 | if (S_ISREG(le16_to_cpu(raw_inode->i_mode))) |
1210 | le32_to_cpu(raw_inode->i_size_lo); | 1210 | return ((loff_t)le32_to_cpu(raw_inode->i_size_high) << 32) | |
1211 | le32_to_cpu(raw_inode->i_size_lo); | ||
1212 | else | ||
1213 | return (loff_t) le32_to_cpu(raw_inode->i_size_lo); | ||
1211 | } | 1214 | } |
1212 | 1215 | ||
1213 | static inline void ext4_isize_set(struct ext4_inode *raw_inode, loff_t i_size) | 1216 | static inline void ext4_isize_set(struct ext4_inode *raw_inode, loff_t i_size) |
diff --git a/fs/ext4/extents.c b/fs/ext4/extents.c index 54bf0623a9ae..e2eab196875f 100644 --- a/fs/ext4/extents.c +++ b/fs/ext4/extents.c | |||
@@ -3048,7 +3048,7 @@ retry: | |||
3048 | WARN_ON(ret <= 0); | 3048 | WARN_ON(ret <= 0); |
3049 | printk(KERN_ERR "%s: ext4_ext_get_blocks " | 3049 | printk(KERN_ERR "%s: ext4_ext_get_blocks " |
3050 | "returned error inode#%lu, block=%u, " | 3050 | "returned error inode#%lu, block=%u, " |
3051 | "max_blocks=%lu", __func__, | 3051 | "max_blocks=%u", __func__, |
3052 | inode->i_ino, block, max_blocks); | 3052 | inode->i_ino, block, max_blocks); |
3053 | #endif | 3053 | #endif |
3054 | ext4_mark_inode_dirty(handle, inode); | 3054 | ext4_mark_inode_dirty(handle, inode); |
diff --git a/fs/ext4/ialloc.c b/fs/ext4/ialloc.c index 4fb86a0061d0..f18a919be70b 100644 --- a/fs/ext4/ialloc.c +++ b/fs/ext4/ialloc.c | |||
@@ -715,6 +715,13 @@ struct inode *ext4_new_inode(handle_t *handle, struct inode *dir, int mode) | |||
715 | 715 | ||
716 | if (sbi->s_log_groups_per_flex) { | 716 | if (sbi->s_log_groups_per_flex) { |
717 | ret2 = find_group_flex(sb, dir, &group); | 717 | ret2 = find_group_flex(sb, dir, &group); |
718 | if (ret2 == -1) { | ||
719 | ret2 = find_group_other(sb, dir, &group); | ||
720 | if (ret2 == 0 && printk_ratelimit()) | ||
721 | printk(KERN_NOTICE "ext4: find_group_flex " | ||
722 | "failed, fallback succeeded dir %lu\n", | ||
723 | dir->i_ino); | ||
724 | } | ||
718 | goto got_group; | 725 | goto got_group; |
719 | } | 726 | } |
720 | 727 | ||
diff --git a/fs/ext4/inode.c b/fs/ext4/inode.c index a6444cee0c7e..51cdd13e1c31 100644 --- a/fs/ext4/inode.c +++ b/fs/ext4/inode.c | |||
@@ -47,8 +47,10 @@ | |||
47 | static inline int ext4_begin_ordered_truncate(struct inode *inode, | 47 | static inline int ext4_begin_ordered_truncate(struct inode *inode, |
48 | loff_t new_size) | 48 | loff_t new_size) |
49 | { | 49 | { |
50 | return jbd2_journal_begin_ordered_truncate(&EXT4_I(inode)->jinode, | 50 | return jbd2_journal_begin_ordered_truncate( |
51 | new_size); | 51 | EXT4_SB(inode->i_sb)->s_journal, |
52 | &EXT4_I(inode)->jinode, | ||
53 | new_size); | ||
52 | } | 54 | } |
53 | 55 | ||
54 | static void ext4_invalidatepage(struct page *page, unsigned long offset); | 56 | static void ext4_invalidatepage(struct page *page, unsigned long offset); |
@@ -360,9 +362,9 @@ static int ext4_block_to_path(struct inode *inode, | |||
360 | final = ptrs; | 362 | final = ptrs; |
361 | } else { | 363 | } else { |
362 | ext4_warning(inode->i_sb, "ext4_block_to_path", | 364 | ext4_warning(inode->i_sb, "ext4_block_to_path", |
363 | "block %lu > max", | 365 | "block %lu > max in inode %lu", |
364 | i_block + direct_blocks + | 366 | i_block + direct_blocks + |
365 | indirect_blocks + double_blocks); | 367 | indirect_blocks + double_blocks, inode->i_ino); |
366 | } | 368 | } |
367 | if (boundary) | 369 | if (boundary) |
368 | *boundary = final - 1 - (i_block & (ptrs - 1)); | 370 | *boundary = final - 1 - (i_block & (ptrs - 1)); |
@@ -1366,6 +1368,10 @@ retry: | |||
1366 | goto out; | 1368 | goto out; |
1367 | } | 1369 | } |
1368 | 1370 | ||
1371 | /* We cannot recurse into the filesystem as the transaction is already | ||
1372 | * started */ | ||
1373 | flags |= AOP_FLAG_NOFS; | ||
1374 | |||
1369 | page = grab_cache_page_write_begin(mapping, index, flags); | 1375 | page = grab_cache_page_write_begin(mapping, index, flags); |
1370 | if (!page) { | 1376 | if (!page) { |
1371 | ext4_journal_stop(handle); | 1377 | ext4_journal_stop(handle); |
@@ -1375,7 +1381,7 @@ retry: | |||
1375 | *pagep = page; | 1381 | *pagep = page; |
1376 | 1382 | ||
1377 | ret = block_write_begin(file, mapping, pos, len, flags, pagep, fsdata, | 1383 | ret = block_write_begin(file, mapping, pos, len, flags, pagep, fsdata, |
1378 | ext4_get_block); | 1384 | ext4_get_block); |
1379 | 1385 | ||
1380 | if (!ret && ext4_should_journal_data(inode)) { | 1386 | if (!ret && ext4_should_journal_data(inode)) { |
1381 | ret = walk_page_buffers(handle, page_buffers(page), | 1387 | ret = walk_page_buffers(handle, page_buffers(page), |
@@ -2437,6 +2443,7 @@ static int ext4_da_writepages(struct address_space *mapping, | |||
2437 | int no_nrwrite_index_update; | 2443 | int no_nrwrite_index_update; |
2438 | int pages_written = 0; | 2444 | int pages_written = 0; |
2439 | long pages_skipped; | 2445 | long pages_skipped; |
2446 | int range_cyclic, cycled = 1, io_done = 0; | ||
2440 | int needed_blocks, ret = 0, nr_to_writebump = 0; | 2447 | int needed_blocks, ret = 0, nr_to_writebump = 0; |
2441 | struct ext4_sb_info *sbi = EXT4_SB(mapping->host->i_sb); | 2448 | struct ext4_sb_info *sbi = EXT4_SB(mapping->host->i_sb); |
2442 | 2449 | ||
@@ -2488,9 +2495,15 @@ static int ext4_da_writepages(struct address_space *mapping, | |||
2488 | if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) | 2495 | if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) |
2489 | range_whole = 1; | 2496 | range_whole = 1; |
2490 | 2497 | ||
2491 | if (wbc->range_cyclic) | 2498 | range_cyclic = wbc->range_cyclic; |
2499 | if (wbc->range_cyclic) { | ||
2492 | index = mapping->writeback_index; | 2500 | index = mapping->writeback_index; |
2493 | else | 2501 | if (index) |
2502 | cycled = 0; | ||
2503 | wbc->range_start = index << PAGE_CACHE_SHIFT; | ||
2504 | wbc->range_end = LLONG_MAX; | ||
2505 | wbc->range_cyclic = 0; | ||
2506 | } else | ||
2494 | index = wbc->range_start >> PAGE_CACHE_SHIFT; | 2507 | index = wbc->range_start >> PAGE_CACHE_SHIFT; |
2495 | 2508 | ||
2496 | mpd.wbc = wbc; | 2509 | mpd.wbc = wbc; |
@@ -2504,6 +2517,7 @@ static int ext4_da_writepages(struct address_space *mapping, | |||
2504 | wbc->no_nrwrite_index_update = 1; | 2517 | wbc->no_nrwrite_index_update = 1; |
2505 | pages_skipped = wbc->pages_skipped; | 2518 | pages_skipped = wbc->pages_skipped; |
2506 | 2519 | ||
2520 | retry: | ||
2507 | while (!ret && wbc->nr_to_write > 0) { | 2521 | while (!ret && wbc->nr_to_write > 0) { |
2508 | 2522 | ||
2509 | /* | 2523 | /* |
@@ -2546,6 +2560,7 @@ static int ext4_da_writepages(struct address_space *mapping, | |||
2546 | pages_written += mpd.pages_written; | 2560 | pages_written += mpd.pages_written; |
2547 | wbc->pages_skipped = pages_skipped; | 2561 | wbc->pages_skipped = pages_skipped; |
2548 | ret = 0; | 2562 | ret = 0; |
2563 | io_done = 1; | ||
2549 | } else if (wbc->nr_to_write) | 2564 | } else if (wbc->nr_to_write) |
2550 | /* | 2565 | /* |
2551 | * There is no more writeout needed | 2566 | * There is no more writeout needed |
@@ -2554,6 +2569,13 @@ static int ext4_da_writepages(struct address_space *mapping, | |||
2554 | */ | 2569 | */ |
2555 | break; | 2570 | break; |
2556 | } | 2571 | } |
2572 | if (!io_done && !cycled) { | ||
2573 | cycled = 1; | ||
2574 | index = 0; | ||
2575 | wbc->range_start = index << PAGE_CACHE_SHIFT; | ||
2576 | wbc->range_end = mapping->writeback_index - 1; | ||
2577 | goto retry; | ||
2578 | } | ||
2557 | if (pages_skipped != wbc->pages_skipped) | 2579 | if (pages_skipped != wbc->pages_skipped) |
2558 | printk(KERN_EMERG "This should not happen leaving %s " | 2580 | printk(KERN_EMERG "This should not happen leaving %s " |
2559 | "with nr_to_write = %ld ret = %d\n", | 2581 | "with nr_to_write = %ld ret = %d\n", |
@@ -2561,6 +2583,7 @@ static int ext4_da_writepages(struct address_space *mapping, | |||
2561 | 2583 | ||
2562 | /* Update index */ | 2584 | /* Update index */ |
2563 | index += pages_written; | 2585 | index += pages_written; |
2586 | wbc->range_cyclic = range_cyclic; | ||
2564 | if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) | 2587 | if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) |
2565 | /* | 2588 | /* |
2566 | * set the writeback_index so that range_cyclic | 2589 | * set the writeback_index so that range_cyclic |
@@ -2648,6 +2671,9 @@ retry: | |||
2648 | ret = PTR_ERR(handle); | 2671 | ret = PTR_ERR(handle); |
2649 | goto out; | 2672 | goto out; |
2650 | } | 2673 | } |
2674 | /* We cannot recurse into the filesystem as the transaction is already | ||
2675 | * started */ | ||
2676 | flags |= AOP_FLAG_NOFS; | ||
2651 | 2677 | ||
2652 | page = grab_cache_page_write_begin(mapping, index, flags); | 2678 | page = grab_cache_page_write_begin(mapping, index, flags); |
2653 | if (!page) { | 2679 | if (!page) { |
@@ -2821,9 +2847,6 @@ static sector_t ext4_bmap(struct address_space *mapping, sector_t block) | |||
2821 | filemap_write_and_wait(mapping); | 2847 | filemap_write_and_wait(mapping); |
2822 | } | 2848 | } |
2823 | 2849 | ||
2824 | BUG_ON(!EXT4_JOURNAL(inode) && | ||
2825 | EXT4_I(inode)->i_state & EXT4_STATE_JDATA); | ||
2826 | |||
2827 | if (EXT4_JOURNAL(inode) && EXT4_I(inode)->i_state & EXT4_STATE_JDATA) { | 2850 | if (EXT4_JOURNAL(inode) && EXT4_I(inode)->i_state & EXT4_STATE_JDATA) { |
2828 | /* | 2851 | /* |
2829 | * This is a REALLY heavyweight approach, but the use of | 2852 | * This is a REALLY heavyweight approach, but the use of |
@@ -3622,7 +3645,7 @@ static void ext4_free_data(handle_t *handle, struct inode *inode, | |||
3622 | * block pointed to itself, it would have been detached when | 3645 | * block pointed to itself, it would have been detached when |
3623 | * the block was cleared. Check for this instead of OOPSing. | 3646 | * the block was cleared. Check for this instead of OOPSing. |
3624 | */ | 3647 | */ |
3625 | if (bh2jh(this_bh)) | 3648 | if ((EXT4_JOURNAL(inode) == NULL) || bh2jh(this_bh)) |
3626 | ext4_handle_dirty_metadata(handle, inode, this_bh); | 3649 | ext4_handle_dirty_metadata(handle, inode, this_bh); |
3627 | else | 3650 | else |
3628 | ext4_error(inode->i_sb, __func__, | 3651 | ext4_error(inode->i_sb, __func__, |
diff --git a/fs/ext4/mballoc.c b/fs/ext4/mballoc.c index 918aec0c8a11..4415beeb0b62 100644 --- a/fs/ext4/mballoc.c +++ b/fs/ext4/mballoc.c | |||
@@ -3025,7 +3025,7 @@ ext4_mb_mark_diskspace_used(struct ext4_allocation_context *ac, | |||
3025 | goto out_err; | 3025 | goto out_err; |
3026 | 3026 | ||
3027 | ext4_debug("using block group %u(%d)\n", ac->ac_b_ex.fe_group, | 3027 | ext4_debug("using block group %u(%d)\n", ac->ac_b_ex.fe_group, |
3028 | gdp->bg_free_blocks_count); | 3028 | ext4_free_blks_count(sb, gdp)); |
3029 | 3029 | ||
3030 | err = ext4_journal_get_write_access(handle, gdp_bh); | 3030 | err = ext4_journal_get_write_access(handle, gdp_bh); |
3031 | if (err) | 3031 | if (err) |
@@ -3693,6 +3693,8 @@ ext4_mb_new_inode_pa(struct ext4_allocation_context *ac) | |||
3693 | pa->pa_free = pa->pa_len; | 3693 | pa->pa_free = pa->pa_len; |
3694 | atomic_set(&pa->pa_count, 1); | 3694 | atomic_set(&pa->pa_count, 1); |
3695 | spin_lock_init(&pa->pa_lock); | 3695 | spin_lock_init(&pa->pa_lock); |
3696 | INIT_LIST_HEAD(&pa->pa_inode_list); | ||
3697 | INIT_LIST_HEAD(&pa->pa_group_list); | ||
3696 | pa->pa_deleted = 0; | 3698 | pa->pa_deleted = 0; |
3697 | pa->pa_linear = 0; | 3699 | pa->pa_linear = 0; |
3698 | 3700 | ||
@@ -3755,6 +3757,7 @@ ext4_mb_new_group_pa(struct ext4_allocation_context *ac) | |||
3755 | atomic_set(&pa->pa_count, 1); | 3757 | atomic_set(&pa->pa_count, 1); |
3756 | spin_lock_init(&pa->pa_lock); | 3758 | spin_lock_init(&pa->pa_lock); |
3757 | INIT_LIST_HEAD(&pa->pa_inode_list); | 3759 | INIT_LIST_HEAD(&pa->pa_inode_list); |
3760 | INIT_LIST_HEAD(&pa->pa_group_list); | ||
3758 | pa->pa_deleted = 0; | 3761 | pa->pa_deleted = 0; |
3759 | pa->pa_linear = 1; | 3762 | pa->pa_linear = 1; |
3760 | 3763 | ||
@@ -4476,23 +4479,26 @@ static int ext4_mb_release_context(struct ext4_allocation_context *ac) | |||
4476 | pa->pa_free -= ac->ac_b_ex.fe_len; | 4479 | pa->pa_free -= ac->ac_b_ex.fe_len; |
4477 | pa->pa_len -= ac->ac_b_ex.fe_len; | 4480 | pa->pa_len -= ac->ac_b_ex.fe_len; |
4478 | spin_unlock(&pa->pa_lock); | 4481 | spin_unlock(&pa->pa_lock); |
4479 | /* | ||
4480 | * We want to add the pa to the right bucket. | ||
4481 | * Remove it from the list and while adding | ||
4482 | * make sure the list to which we are adding | ||
4483 | * doesn't grow big. | ||
4484 | */ | ||
4485 | if (likely(pa->pa_free)) { | ||
4486 | spin_lock(pa->pa_obj_lock); | ||
4487 | list_del_rcu(&pa->pa_inode_list); | ||
4488 | spin_unlock(pa->pa_obj_lock); | ||
4489 | ext4_mb_add_n_trim(ac); | ||
4490 | } | ||
4491 | } | 4482 | } |
4492 | ext4_mb_put_pa(ac, ac->ac_sb, pa); | ||
4493 | } | 4483 | } |
4494 | if (ac->alloc_semp) | 4484 | if (ac->alloc_semp) |
4495 | up_read(ac->alloc_semp); | 4485 | up_read(ac->alloc_semp); |
4486 | if (pa) { | ||
4487 | /* | ||
4488 | * We want to add the pa to the right bucket. | ||
4489 | * Remove it from the list and while adding | ||
4490 | * make sure the list to which we are adding | ||
4491 | * doesn't grow big. We need to release | ||
4492 | * alloc_semp before calling ext4_mb_add_n_trim() | ||
4493 | */ | ||
4494 | if (pa->pa_linear && likely(pa->pa_free)) { | ||
4495 | spin_lock(pa->pa_obj_lock); | ||
4496 | list_del_rcu(&pa->pa_inode_list); | ||
4497 | spin_unlock(pa->pa_obj_lock); | ||
4498 | ext4_mb_add_n_trim(ac); | ||
4499 | } | ||
4500 | ext4_mb_put_pa(ac, ac->ac_sb, pa); | ||
4501 | } | ||
4496 | if (ac->ac_bitmap_page) | 4502 | if (ac->ac_bitmap_page) |
4497 | page_cache_release(ac->ac_bitmap_page); | 4503 | page_cache_release(ac->ac_bitmap_page); |
4498 | if (ac->ac_buddy_page) | 4504 | if (ac->ac_buddy_page) |
diff --git a/fs/ext4/migrate.c b/fs/ext4/migrate.c index 734abca25e35..fe64d9f79852 100644 --- a/fs/ext4/migrate.c +++ b/fs/ext4/migrate.c | |||
@@ -481,7 +481,7 @@ int ext4_ext_migrate(struct inode *inode) | |||
481 | + 1); | 481 | + 1); |
482 | if (IS_ERR(handle)) { | 482 | if (IS_ERR(handle)) { |
483 | retval = PTR_ERR(handle); | 483 | retval = PTR_ERR(handle); |
484 | goto err_out; | 484 | return retval; |
485 | } | 485 | } |
486 | tmp_inode = ext4_new_inode(handle, | 486 | tmp_inode = ext4_new_inode(handle, |
487 | inode->i_sb->s_root->d_inode, | 487 | inode->i_sb->s_root->d_inode, |
@@ -489,8 +489,7 @@ int ext4_ext_migrate(struct inode *inode) | |||
489 | if (IS_ERR(tmp_inode)) { | 489 | if (IS_ERR(tmp_inode)) { |
490 | retval = -ENOMEM; | 490 | retval = -ENOMEM; |
491 | ext4_journal_stop(handle); | 491 | ext4_journal_stop(handle); |
492 | tmp_inode = NULL; | 492 | return retval; |
493 | goto err_out; | ||
494 | } | 493 | } |
495 | i_size_write(tmp_inode, i_size_read(inode)); | 494 | i_size_write(tmp_inode, i_size_read(inode)); |
496 | /* | 495 | /* |
@@ -618,8 +617,7 @@ err_out: | |||
618 | 617 | ||
619 | ext4_journal_stop(handle); | 618 | ext4_journal_stop(handle); |
620 | 619 | ||
621 | if (tmp_inode) | 620 | iput(tmp_inode); |
622 | iput(tmp_inode); | ||
623 | 621 | ||
624 | return retval; | 622 | return retval; |
625 | } | 623 | } |
diff --git a/fs/ext4/namei.c b/fs/ext4/namei.c index fec0b4c2f5f1..ba702bd7910d 100644 --- a/fs/ext4/namei.c +++ b/fs/ext4/namei.c | |||
@@ -1368,7 +1368,7 @@ static int make_indexed_dir(handle_t *handle, struct dentry *dentry, | |||
1368 | struct fake_dirent *fde; | 1368 | struct fake_dirent *fde; |
1369 | 1369 | ||
1370 | blocksize = dir->i_sb->s_blocksize; | 1370 | blocksize = dir->i_sb->s_blocksize; |
1371 | dxtrace(printk(KERN_DEBUG "Creating index\n")); | 1371 | dxtrace(printk(KERN_DEBUG "Creating index: inode %lu\n", dir->i_ino)); |
1372 | retval = ext4_journal_get_write_access(handle, bh); | 1372 | retval = ext4_journal_get_write_access(handle, bh); |
1373 | if (retval) { | 1373 | if (retval) { |
1374 | ext4_std_error(dir->i_sb, retval); | 1374 | ext4_std_error(dir->i_sb, retval); |
@@ -1377,6 +1377,20 @@ static int make_indexed_dir(handle_t *handle, struct dentry *dentry, | |||
1377 | } | 1377 | } |
1378 | root = (struct dx_root *) bh->b_data; | 1378 | root = (struct dx_root *) bh->b_data; |
1379 | 1379 | ||
1380 | /* The 0th block becomes the root, move the dirents out */ | ||
1381 | fde = &root->dotdot; | ||
1382 | de = (struct ext4_dir_entry_2 *)((char *)fde + | ||
1383 | ext4_rec_len_from_disk(fde->rec_len)); | ||
1384 | if ((char *) de >= (((char *) root) + blocksize)) { | ||
1385 | ext4_error(dir->i_sb, __func__, | ||
1386 | "invalid rec_len for '..' in inode %lu", | ||
1387 | dir->i_ino); | ||
1388 | brelse(bh); | ||
1389 | return -EIO; | ||
1390 | } | ||
1391 | len = ((char *) root) + blocksize - (char *) de; | ||
1392 | |||
1393 | /* Allocate new block for the 0th block's dirents */ | ||
1380 | bh2 = ext4_append(handle, dir, &block, &retval); | 1394 | bh2 = ext4_append(handle, dir, &block, &retval); |
1381 | if (!(bh2)) { | 1395 | if (!(bh2)) { |
1382 | brelse(bh); | 1396 | brelse(bh); |
@@ -1385,11 +1399,6 @@ static int make_indexed_dir(handle_t *handle, struct dentry *dentry, | |||
1385 | EXT4_I(dir)->i_flags |= EXT4_INDEX_FL; | 1399 | EXT4_I(dir)->i_flags |= EXT4_INDEX_FL; |
1386 | data1 = bh2->b_data; | 1400 | data1 = bh2->b_data; |
1387 | 1401 | ||
1388 | /* The 0th block becomes the root, move the dirents out */ | ||
1389 | fde = &root->dotdot; | ||
1390 | de = (struct ext4_dir_entry_2 *)((char *)fde + | ||
1391 | ext4_rec_len_from_disk(fde->rec_len)); | ||
1392 | len = ((char *) root) + blocksize - (char *) de; | ||
1393 | memcpy (data1, de, len); | 1402 | memcpy (data1, de, len); |
1394 | de = (struct ext4_dir_entry_2 *) data1; | 1403 | de = (struct ext4_dir_entry_2 *) data1; |
1395 | top = data1 + len; | 1404 | top = data1 + len; |
diff --git a/fs/ext4/resize.c b/fs/ext4/resize.c index c328be5d6885..c06886abd658 100644 --- a/fs/ext4/resize.c +++ b/fs/ext4/resize.c | |||
@@ -861,12 +861,13 @@ int ext4_group_add(struct super_block *sb, struct ext4_new_group_data *input) | |||
861 | gdp = (struct ext4_group_desc *)((char *)primary->b_data + | 861 | gdp = (struct ext4_group_desc *)((char *)primary->b_data + |
862 | gdb_off * EXT4_DESC_SIZE(sb)); | 862 | gdb_off * EXT4_DESC_SIZE(sb)); |
863 | 863 | ||
864 | memset(gdp, 0, EXT4_DESC_SIZE(sb)); | ||
864 | ext4_block_bitmap_set(sb, gdp, input->block_bitmap); /* LV FIXME */ | 865 | ext4_block_bitmap_set(sb, gdp, input->block_bitmap); /* LV FIXME */ |
865 | ext4_inode_bitmap_set(sb, gdp, input->inode_bitmap); /* LV FIXME */ | 866 | ext4_inode_bitmap_set(sb, gdp, input->inode_bitmap); /* LV FIXME */ |
866 | ext4_inode_table_set(sb, gdp, input->inode_table); /* LV FIXME */ | 867 | ext4_inode_table_set(sb, gdp, input->inode_table); /* LV FIXME */ |
867 | ext4_free_blks_set(sb, gdp, input->free_blocks_count); | 868 | ext4_free_blks_set(sb, gdp, input->free_blocks_count); |
868 | ext4_free_inodes_set(sb, gdp, EXT4_INODES_PER_GROUP(sb)); | 869 | ext4_free_inodes_set(sb, gdp, EXT4_INODES_PER_GROUP(sb)); |
869 | gdp->bg_flags |= cpu_to_le16(EXT4_BG_INODE_ZEROED); | 870 | gdp->bg_flags = cpu_to_le16(EXT4_BG_INODE_ZEROED); |
870 | gdp->bg_checksum = ext4_group_desc_csum(sbi, input->group, gdp); | 871 | gdp->bg_checksum = ext4_group_desc_csum(sbi, input->group, gdp); |
871 | 872 | ||
872 | /* | 873 | /* |
diff --git a/fs/ext4/super.c b/fs/ext4/super.c index e5f06a5f045e..a5732c58f676 100644 --- a/fs/ext4/super.c +++ b/fs/ext4/super.c | |||
@@ -3046,14 +3046,17 @@ static void ext4_write_super(struct super_block *sb) | |||
3046 | static int ext4_sync_fs(struct super_block *sb, int wait) | 3046 | static int ext4_sync_fs(struct super_block *sb, int wait) |
3047 | { | 3047 | { |
3048 | int ret = 0; | 3048 | int ret = 0; |
3049 | tid_t target; | ||
3049 | 3050 | ||
3050 | trace_mark(ext4_sync_fs, "dev %s wait %d", sb->s_id, wait); | 3051 | trace_mark(ext4_sync_fs, "dev %s wait %d", sb->s_id, wait); |
3051 | sb->s_dirt = 0; | 3052 | sb->s_dirt = 0; |
3052 | if (EXT4_SB(sb)->s_journal) { | 3053 | if (EXT4_SB(sb)->s_journal) { |
3053 | if (wait) | 3054 | if (jbd2_journal_start_commit(EXT4_SB(sb)->s_journal, |
3054 | ret = ext4_force_commit(sb); | 3055 | &target)) { |
3055 | else | 3056 | if (wait) |
3056 | jbd2_journal_start_commit(EXT4_SB(sb)->s_journal, NULL); | 3057 | jbd2_log_wait_commit(EXT4_SB(sb)->s_journal, |
3058 | target); | ||
3059 | } | ||
3057 | } else { | 3060 | } else { |
3058 | ext4_commit_super(sb, EXT4_SB(sb)->s_es, wait); | 3061 | ext4_commit_super(sb, EXT4_SB(sb)->s_es, wait); |
3059 | } | 3062 | } |
diff --git a/fs/fat/Kconfig b/fs/fat/Kconfig new file mode 100644 index 000000000000..d0a69ff25375 --- /dev/null +++ b/fs/fat/Kconfig | |||
@@ -0,0 +1,97 @@ | |||
1 | config FAT_FS | ||
2 | tristate | ||
3 | select NLS | ||
4 | help | ||
5 | If you want to use one of the FAT-based file systems (the MS-DOS and | ||
6 | VFAT (Windows 95) file systems), then you must say Y or M here | ||
7 | to include FAT support. You will then be able to mount partitions or | ||
8 | diskettes with FAT-based file systems and transparently access the | ||
9 | files on them, i.e. MSDOS files will look and behave just like all | ||
10 | other Unix files. | ||
11 | |||
12 | This FAT support is not a file system in itself, it only provides | ||
13 | the foundation for the other file systems. You will have to say Y or | ||
14 | M to at least one of "MSDOS fs support" or "VFAT fs support" in | ||
15 | order to make use of it. | ||
16 | |||
17 | Another way to read and write MSDOS floppies and hard drive | ||
18 | partitions from within Linux (but not transparently) is with the | ||
19 | mtools ("man mtools") program suite. You don't need to say Y here in | ||
20 | order to do that. | ||
21 | |||
22 | If you need to move large files on floppies between a DOS and a | ||
23 | Linux box, say Y here, mount the floppy under Linux with an MSDOS | ||
24 | file system and use GNU tar's M option. GNU tar is a program | ||
25 | available for Unix and DOS ("man tar" or "info tar"). | ||
26 | |||
27 | The FAT support will enlarge your kernel by about 37 KB. If unsure, | ||
28 | say Y. | ||
29 | |||
30 | To compile this as a module, choose M here: the module will be called | ||
31 | fat. Note that if you compile the FAT support as a module, you | ||
32 | cannot compile any of the FAT-based file systems into the kernel | ||
33 | -- they will have to be modules as well. | ||
34 | |||
35 | config MSDOS_FS | ||
36 | tristate "MSDOS fs support" | ||
37 | select FAT_FS | ||
38 | help | ||
39 | This allows you to mount MSDOS partitions of your hard drive (unless | ||
40 | they are compressed; to access compressed MSDOS partitions under | ||
41 | Linux, you can either use the DOS emulator DOSEMU, described in the | ||
42 | DOSEMU-HOWTO, available from | ||
43 | <http://www.tldp.org/docs.html#howto>, or try dmsdosfs in | ||
44 | <ftp://ibiblio.org/pub/Linux/system/filesystems/dosfs/>. If you | ||
45 | intend to use dosemu with a non-compressed MSDOS partition, say Y | ||
46 | here) and MSDOS floppies. This means that file access becomes | ||
47 | transparent, i.e. the MSDOS files look and behave just like all | ||
48 | other Unix files. | ||
49 | |||
50 | If you have Windows 95 or Windows NT installed on your MSDOS | ||
51 | partitions, you should use the VFAT file system (say Y to "VFAT fs | ||
52 | support" below), or you will not be able to see the long filenames | ||
53 | generated by Windows 95 / Windows NT. | ||
54 | |||
55 | This option will enlarge your kernel by about 7 KB. If unsure, | ||
56 | answer Y. This will only work if you said Y to "DOS FAT fs support" | ||
57 | as well. To compile this as a module, choose M here: the module will | ||
58 | be called msdos. | ||
59 | |||
60 | config VFAT_FS | ||
61 | tristate "VFAT (Windows-95) fs support" | ||
62 | select FAT_FS | ||
63 | help | ||
64 | This option provides support for normal Windows file systems with | ||
65 | long filenames. That includes non-compressed FAT-based file systems | ||
66 | used by Windows 95, Windows 98, Windows NT 4.0, and the Unix | ||
67 | programs from the mtools package. | ||
68 | |||
69 | The VFAT support enlarges your kernel by about 10 KB and it only | ||
70 | works if you said Y to the "DOS FAT fs support" above. Please read | ||
71 | the file <file:Documentation/filesystems/vfat.txt> for details. If | ||
72 | unsure, say Y. | ||
73 | |||
74 | To compile this as a module, choose M here: the module will be called | ||
75 | vfat. | ||
76 | |||
77 | config FAT_DEFAULT_CODEPAGE | ||
78 | int "Default codepage for FAT" | ||
79 | depends on MSDOS_FS || VFAT_FS | ||
80 | default 437 | ||
81 | help | ||
82 | This option should be set to the codepage of your FAT filesystems. | ||
83 | It can be overridden with the "codepage" mount option. | ||
84 | See <file:Documentation/filesystems/vfat.txt> for more information. | ||
85 | |||
86 | config FAT_DEFAULT_IOCHARSET | ||
87 | string "Default iocharset for FAT" | ||
88 | depends on VFAT_FS | ||
89 | default "iso8859-1" | ||
90 | help | ||
91 | Set this to the default input/output character set you'd | ||
92 | like FAT to use. It should probably match the character set | ||
93 | that most of your FAT filesystems use, and can be overridden | ||
94 | with the "iocharset" mount option for FAT filesystems. | ||
95 | Note that "utf8" is not recommended for FAT filesystems. | ||
96 | If unsure, you shouldn't set "utf8" here. | ||
97 | See <file:Documentation/filesystems/vfat.txt> for more information. | ||
diff --git a/fs/fcntl.c b/fs/fcntl.c index cdc141946724..bd215cc791da 100644 --- a/fs/fcntl.c +++ b/fs/fcntl.c | |||
@@ -50,7 +50,7 @@ static int get_close_on_exec(unsigned int fd) | |||
50 | return res; | 50 | return res; |
51 | } | 51 | } |
52 | 52 | ||
53 | asmlinkage long sys_dup3(unsigned int oldfd, unsigned int newfd, int flags) | 53 | SYSCALL_DEFINE3(dup3, unsigned int, oldfd, unsigned int, newfd, int, flags) |
54 | { | 54 | { |
55 | int err = -EBADF; | 55 | int err = -EBADF; |
56 | struct file * file, *tofree; | 56 | struct file * file, *tofree; |
@@ -113,7 +113,7 @@ out_unlock: | |||
113 | return err; | 113 | return err; |
114 | } | 114 | } |
115 | 115 | ||
116 | asmlinkage long sys_dup2(unsigned int oldfd, unsigned int newfd) | 116 | SYSCALL_DEFINE2(dup2, unsigned int, oldfd, unsigned int, newfd) |
117 | { | 117 | { |
118 | if (unlikely(newfd == oldfd)) { /* corner case */ | 118 | if (unlikely(newfd == oldfd)) { /* corner case */ |
119 | struct files_struct *files = current->files; | 119 | struct files_struct *files = current->files; |
@@ -126,7 +126,7 @@ asmlinkage long sys_dup2(unsigned int oldfd, unsigned int newfd) | |||
126 | return sys_dup3(oldfd, newfd, 0); | 126 | return sys_dup3(oldfd, newfd, 0); |
127 | } | 127 | } |
128 | 128 | ||
129 | asmlinkage long sys_dup(unsigned int fildes) | 129 | SYSCALL_DEFINE1(dup, unsigned int, fildes) |
130 | { | 130 | { |
131 | int ret = -EBADF; | 131 | int ret = -EBADF; |
132 | struct file *file = fget(fildes); | 132 | struct file *file = fget(fildes); |
@@ -335,7 +335,7 @@ static long do_fcntl(int fd, unsigned int cmd, unsigned long arg, | |||
335 | return err; | 335 | return err; |
336 | } | 336 | } |
337 | 337 | ||
338 | asmlinkage long sys_fcntl(unsigned int fd, unsigned int cmd, unsigned long arg) | 338 | SYSCALL_DEFINE3(fcntl, unsigned int, fd, unsigned int, cmd, unsigned long, arg) |
339 | { | 339 | { |
340 | struct file *filp; | 340 | struct file *filp; |
341 | long err = -EBADF; | 341 | long err = -EBADF; |
@@ -358,7 +358,8 @@ out: | |||
358 | } | 358 | } |
359 | 359 | ||
360 | #if BITS_PER_LONG == 32 | 360 | #if BITS_PER_LONG == 32 |
361 | asmlinkage long sys_fcntl64(unsigned int fd, unsigned int cmd, unsigned long arg) | 361 | SYSCALL_DEFINE3(fcntl64, unsigned int, fd, unsigned int, cmd, |
362 | unsigned long, arg) | ||
362 | { | 363 | { |
363 | struct file * filp; | 364 | struct file * filp; |
364 | long err; | 365 | long err; |
diff --git a/fs/filesystems.c b/fs/filesystems.c index d488dcd7f2bb..1aa70260e6d1 100644 --- a/fs/filesystems.c +++ b/fs/filesystems.c | |||
@@ -179,7 +179,7 @@ static int fs_maxindex(void) | |||
179 | /* | 179 | /* |
180 | * Whee.. Weird sysv syscall. | 180 | * Whee.. Weird sysv syscall. |
181 | */ | 181 | */ |
182 | asmlinkage long sys_sysfs(int option, unsigned long arg1, unsigned long arg2) | 182 | SYSCALL_DEFINE3(sysfs, int, option, unsigned long, arg1, unsigned long, arg2) |
183 | { | 183 | { |
184 | int retval = -EINVAL; | 184 | int retval = -EINVAL; |
185 | 185 | ||
diff --git a/fs/freevxfs/Kconfig b/fs/freevxfs/Kconfig new file mode 100644 index 000000000000..8dc1cd5c1efe --- /dev/null +++ b/fs/freevxfs/Kconfig | |||
@@ -0,0 +1,16 @@ | |||
1 | config VXFS_FS | ||
2 | tristate "FreeVxFS file system support (VERITAS VxFS(TM) compatible)" | ||
3 | depends on BLOCK | ||
4 | help | ||
5 | FreeVxFS is a file system driver that support the VERITAS VxFS(TM) | ||
6 | file system format. VERITAS VxFS(TM) is the standard file system | ||
7 | of SCO UnixWare (and possibly others) and optionally available | ||
8 | for Sunsoft Solaris, HP-UX and many other operating systems. | ||
9 | Currently only readonly access is supported. | ||
10 | |||
11 | NOTE: the file system type as used by mount(1), mount(2) and | ||
12 | fstab(5) is 'vxfs' as it describes the file system format, not | ||
13 | the actual driver. | ||
14 | |||
15 | To compile this as a module, choose M here: the module will be | ||
16 | called freevxfs. If unsure, say N. | ||
diff --git a/fs/fuse/Kconfig b/fs/fuse/Kconfig new file mode 100644 index 000000000000..0cf160a94eda --- /dev/null +++ b/fs/fuse/Kconfig | |||
@@ -0,0 +1,15 @@ | |||
1 | config FUSE_FS | ||
2 | tristate "FUSE (Filesystem in Userspace) support" | ||
3 | help | ||
4 | With FUSE it is possible to implement a fully functional filesystem | ||
5 | in a userspace program. | ||
6 | |||
7 | There's also companion library: libfuse. This library along with | ||
8 | utilities is available from the FUSE homepage: | ||
9 | <http://fuse.sourceforge.net/> | ||
10 | |||
11 | See <file:Documentation/filesystems/fuse.txt> for more information. | ||
12 | See <file:Documentation/Changes> for needed library/utility version. | ||
13 | |||
14 | If you want to develop a userspace FS, or if you want to use | ||
15 | a filesystem based on FUSE, answer Y or M. | ||
diff --git a/fs/fuse/dev.c b/fs/fuse/dev.c index e0c7ada08a1f..ba76b68c52ff 100644 --- a/fs/fuse/dev.c +++ b/fs/fuse/dev.c | |||
@@ -281,7 +281,8 @@ __releases(&fc->lock) | |||
281 | fc->blocked = 0; | 281 | fc->blocked = 0; |
282 | wake_up_all(&fc->blocked_waitq); | 282 | wake_up_all(&fc->blocked_waitq); |
283 | } | 283 | } |
284 | if (fc->num_background == FUSE_CONGESTION_THRESHOLD) { | 284 | if (fc->num_background == FUSE_CONGESTION_THRESHOLD && |
285 | fc->connected) { | ||
285 | clear_bdi_congested(&fc->bdi, READ); | 286 | clear_bdi_congested(&fc->bdi, READ); |
286 | clear_bdi_congested(&fc->bdi, WRITE); | 287 | clear_bdi_congested(&fc->bdi, WRITE); |
287 | } | 288 | } |
@@ -825,16 +826,21 @@ static int fuse_notify_poll(struct fuse_conn *fc, unsigned int size, | |||
825 | struct fuse_copy_state *cs) | 826 | struct fuse_copy_state *cs) |
826 | { | 827 | { |
827 | struct fuse_notify_poll_wakeup_out outarg; | 828 | struct fuse_notify_poll_wakeup_out outarg; |
828 | int err; | 829 | int err = -EINVAL; |
829 | 830 | ||
830 | if (size != sizeof(outarg)) | 831 | if (size != sizeof(outarg)) |
831 | return -EINVAL; | 832 | goto err; |
832 | 833 | ||
833 | err = fuse_copy_one(cs, &outarg, sizeof(outarg)); | 834 | err = fuse_copy_one(cs, &outarg, sizeof(outarg)); |
834 | if (err) | 835 | if (err) |
835 | return err; | 836 | goto err; |
836 | 837 | ||
838 | fuse_copy_finish(cs); | ||
837 | return fuse_notify_poll_wakeup(fc, &outarg); | 839 | return fuse_notify_poll_wakeup(fc, &outarg); |
840 | |||
841 | err: | ||
842 | fuse_copy_finish(cs); | ||
843 | return err; | ||
838 | } | 844 | } |
839 | 845 | ||
840 | static int fuse_notify(struct fuse_conn *fc, enum fuse_notify_code code, | 846 | static int fuse_notify(struct fuse_conn *fc, enum fuse_notify_code code, |
@@ -845,6 +851,7 @@ static int fuse_notify(struct fuse_conn *fc, enum fuse_notify_code code, | |||
845 | return fuse_notify_poll(fc, size, cs); | 851 | return fuse_notify_poll(fc, size, cs); |
846 | 852 | ||
847 | default: | 853 | default: |
854 | fuse_copy_finish(cs); | ||
848 | return -EINVAL; | 855 | return -EINVAL; |
849 | } | 856 | } |
850 | } | 857 | } |
@@ -923,7 +930,6 @@ static ssize_t fuse_dev_write(struct kiocb *iocb, const struct iovec *iov, | |||
923 | */ | 930 | */ |
924 | if (!oh.unique) { | 931 | if (!oh.unique) { |
925 | err = fuse_notify(fc, oh.error, nbytes - sizeof(oh), &cs); | 932 | err = fuse_notify(fc, oh.error, nbytes - sizeof(oh), &cs); |
926 | fuse_copy_finish(&cs); | ||
927 | return err ? err : nbytes; | 933 | return err ? err : nbytes; |
928 | } | 934 | } |
929 | 935 | ||
diff --git a/fs/fuse/file.c b/fs/fuse/file.c index e8162646a9b5..d9fdb7cec538 100644 --- a/fs/fuse/file.c +++ b/fs/fuse/file.c | |||
@@ -54,7 +54,7 @@ struct fuse_file *fuse_file_alloc(struct fuse_conn *fc) | |||
54 | ff->reserved_req = fuse_request_alloc(); | 54 | ff->reserved_req = fuse_request_alloc(); |
55 | if (!ff->reserved_req) { | 55 | if (!ff->reserved_req) { |
56 | kfree(ff); | 56 | kfree(ff); |
57 | ff = NULL; | 57 | return NULL; |
58 | } else { | 58 | } else { |
59 | INIT_LIST_HEAD(&ff->write_entry); | 59 | INIT_LIST_HEAD(&ff->write_entry); |
60 | atomic_set(&ff->count, 0); | 60 | atomic_set(&ff->count, 0); |
diff --git a/fs/fuse/inode.c b/fs/fuse/inode.c index 47c96fdca1ac..459b73dd45e1 100644 --- a/fs/fuse/inode.c +++ b/fs/fuse/inode.c | |||
@@ -292,6 +292,7 @@ static void fuse_put_super(struct super_block *sb) | |||
292 | list_del(&fc->entry); | 292 | list_del(&fc->entry); |
293 | fuse_ctl_remove_conn(fc); | 293 | fuse_ctl_remove_conn(fc); |
294 | mutex_unlock(&fuse_mutex); | 294 | mutex_unlock(&fuse_mutex); |
295 | bdi_destroy(&fc->bdi); | ||
295 | fuse_conn_put(fc); | 296 | fuse_conn_put(fc); |
296 | } | 297 | } |
297 | 298 | ||
@@ -532,7 +533,6 @@ void fuse_conn_put(struct fuse_conn *fc) | |||
532 | if (fc->destroy_req) | 533 | if (fc->destroy_req) |
533 | fuse_request_free(fc->destroy_req); | 534 | fuse_request_free(fc->destroy_req); |
534 | mutex_destroy(&fc->inst_mutex); | 535 | mutex_destroy(&fc->inst_mutex); |
535 | bdi_destroy(&fc->bdi); | ||
536 | fc->release(fc); | 536 | fc->release(fc); |
537 | } | 537 | } |
538 | } | 538 | } |
@@ -805,16 +805,18 @@ static int fuse_fill_super(struct super_block *sb, void *data, int silent) | |||
805 | int err; | 805 | int err; |
806 | int is_bdev = sb->s_bdev != NULL; | 806 | int is_bdev = sb->s_bdev != NULL; |
807 | 807 | ||
808 | err = -EINVAL; | ||
808 | if (sb->s_flags & MS_MANDLOCK) | 809 | if (sb->s_flags & MS_MANDLOCK) |
809 | return -EINVAL; | 810 | goto err; |
810 | 811 | ||
811 | if (!parse_fuse_opt((char *) data, &d, is_bdev)) | 812 | if (!parse_fuse_opt((char *) data, &d, is_bdev)) |
812 | return -EINVAL; | 813 | goto err; |
813 | 814 | ||
814 | if (is_bdev) { | 815 | if (is_bdev) { |
815 | #ifdef CONFIG_BLOCK | 816 | #ifdef CONFIG_BLOCK |
817 | err = -EINVAL; | ||
816 | if (!sb_set_blocksize(sb, d.blksize)) | 818 | if (!sb_set_blocksize(sb, d.blksize)) |
817 | return -EINVAL; | 819 | goto err; |
818 | #endif | 820 | #endif |
819 | } else { | 821 | } else { |
820 | sb->s_blocksize = PAGE_CACHE_SIZE; | 822 | sb->s_blocksize = PAGE_CACHE_SIZE; |
@@ -826,20 +828,22 @@ static int fuse_fill_super(struct super_block *sb, void *data, int silent) | |||
826 | sb->s_export_op = &fuse_export_operations; | 828 | sb->s_export_op = &fuse_export_operations; |
827 | 829 | ||
828 | file = fget(d.fd); | 830 | file = fget(d.fd); |
831 | err = -EINVAL; | ||
829 | if (!file) | 832 | if (!file) |
830 | return -EINVAL; | 833 | goto err; |
831 | 834 | ||
832 | if (file->f_op != &fuse_dev_operations) | 835 | if (file->f_op != &fuse_dev_operations) |
833 | return -EINVAL; | 836 | goto err_fput; |
834 | 837 | ||
835 | fc = kmalloc(sizeof(*fc), GFP_KERNEL); | 838 | fc = kmalloc(sizeof(*fc), GFP_KERNEL); |
839 | err = -ENOMEM; | ||
836 | if (!fc) | 840 | if (!fc) |
837 | return -ENOMEM; | 841 | goto err_fput; |
838 | 842 | ||
839 | err = fuse_conn_init(fc, sb); | 843 | err = fuse_conn_init(fc, sb); |
840 | if (err) { | 844 | if (err) { |
841 | kfree(fc); | 845 | kfree(fc); |
842 | return err; | 846 | goto err_fput; |
843 | } | 847 | } |
844 | 848 | ||
845 | fc->release = fuse_free_conn; | 849 | fc->release = fuse_free_conn; |
@@ -854,12 +858,12 @@ static int fuse_fill_super(struct super_block *sb, void *data, int silent) | |||
854 | err = -ENOMEM; | 858 | err = -ENOMEM; |
855 | root = fuse_get_root_inode(sb, d.rootmode); | 859 | root = fuse_get_root_inode(sb, d.rootmode); |
856 | if (!root) | 860 | if (!root) |
857 | goto err; | 861 | goto err_put_conn; |
858 | 862 | ||
859 | root_dentry = d_alloc_root(root); | 863 | root_dentry = d_alloc_root(root); |
860 | if (!root_dentry) { | 864 | if (!root_dentry) { |
861 | iput(root); | 865 | iput(root); |
862 | goto err; | 866 | goto err_put_conn; |
863 | } | 867 | } |
864 | 868 | ||
865 | init_req = fuse_request_alloc(); | 869 | init_req = fuse_request_alloc(); |
@@ -903,9 +907,11 @@ static int fuse_fill_super(struct super_block *sb, void *data, int silent) | |||
903 | fuse_request_free(init_req); | 907 | fuse_request_free(init_req); |
904 | err_put_root: | 908 | err_put_root: |
905 | dput(root_dentry); | 909 | dput(root_dentry); |
906 | err: | 910 | err_put_conn: |
907 | fput(file); | ||
908 | fuse_conn_put(fc); | 911 | fuse_conn_put(fc); |
912 | err_fput: | ||
913 | fput(file); | ||
914 | err: | ||
909 | return err; | 915 | return err; |
910 | } | 916 | } |
911 | 917 | ||
diff --git a/fs/hfs/Kconfig b/fs/hfs/Kconfig new file mode 100644 index 000000000000..b77c5bc20f8a --- /dev/null +++ b/fs/hfs/Kconfig | |||
@@ -0,0 +1,12 @@ | |||
1 | config HFS_FS | ||
2 | tristate "Apple Macintosh file system support (EXPERIMENTAL)" | ||
3 | depends on BLOCK && EXPERIMENTAL | ||
4 | select NLS | ||
5 | help | ||
6 | If you say Y here, you will be able to mount Macintosh-formatted | ||
7 | floppy disks and hard drive partitions with full read-write access. | ||
8 | Please read <file:Documentation/filesystems/hfs.txt> to learn about | ||
9 | the available mount options. | ||
10 | |||
11 | To compile this file system support as a module, choose M here: the | ||
12 | module will be called hfs. | ||
diff --git a/fs/hfsplus/Kconfig b/fs/hfsplus/Kconfig new file mode 100644 index 000000000000..a63371815aab --- /dev/null +++ b/fs/hfsplus/Kconfig | |||
@@ -0,0 +1,13 @@ | |||
1 | config HFSPLUS_FS | ||
2 | tristate "Apple Extended HFS file system support" | ||
3 | depends on BLOCK | ||
4 | select NLS | ||
5 | select NLS_UTF8 | ||
6 | help | ||
7 | If you say Y here, you will be able to mount extended format | ||
8 | Macintosh-formatted hard drive partitions with full read-write access. | ||
9 | |||
10 | This file system is often called HFS+ and was introduced with | ||
11 | MacOS 8. It includes all Mac specific filesystem data such as | ||
12 | data forks and creator codes, but it also has several UNIX | ||
13 | style features such as file ownership and permissions. | ||
diff --git a/fs/hpfs/Kconfig b/fs/hpfs/Kconfig new file mode 100644 index 000000000000..56bd15c5bf6c --- /dev/null +++ b/fs/hpfs/Kconfig | |||
@@ -0,0 +1,14 @@ | |||
1 | config HPFS_FS | ||
2 | tristate "OS/2 HPFS file system support" | ||
3 | depends on BLOCK | ||
4 | help | ||
5 | OS/2 is IBM's operating system for PC's, the same as Warp, and HPFS | ||
6 | is the file system used for organizing files on OS/2 hard disk | ||
7 | partitions. Say Y if you want to be able to read files from and | ||
8 | write files to an OS/2 HPFS partition on your hard drive. OS/2 | ||
9 | floppies however are in regular MSDOS format, so you don't need this | ||
10 | option in order to be able to read them. Read | ||
11 | <file:Documentation/filesystems/hpfs.txt>. | ||
12 | |||
13 | To compile this file system support as a module, choose M here: the | ||
14 | module will be called hpfs. If unsure, say N. | ||
diff --git a/fs/hugetlbfs/inode.c b/fs/hugetlbfs/inode.c index 6903d37af037..9b800d97a687 100644 --- a/fs/hugetlbfs/inode.c +++ b/fs/hugetlbfs/inode.c | |||
@@ -108,7 +108,8 @@ static int hugetlbfs_file_mmap(struct file *file, struct vm_area_struct *vma) | |||
108 | 108 | ||
109 | if (hugetlb_reserve_pages(inode, | 109 | if (hugetlb_reserve_pages(inode, |
110 | vma->vm_pgoff >> huge_page_order(h), | 110 | vma->vm_pgoff >> huge_page_order(h), |
111 | len >> huge_page_shift(h), vma)) | 111 | len >> huge_page_shift(h), vma, |
112 | vma->vm_flags)) | ||
112 | goto out; | 113 | goto out; |
113 | 114 | ||
114 | ret = 0; | 115 | ret = 0; |
@@ -947,7 +948,7 @@ static int can_do_hugetlb_shm(void) | |||
947 | can_do_mlock()); | 948 | can_do_mlock()); |
948 | } | 949 | } |
949 | 950 | ||
950 | struct file *hugetlb_file_setup(const char *name, size_t size) | 951 | struct file *hugetlb_file_setup(const char *name, size_t size, int acctflag) |
951 | { | 952 | { |
952 | int error = -ENOMEM; | 953 | int error = -ENOMEM; |
953 | struct file *file; | 954 | struct file *file; |
@@ -981,7 +982,8 @@ struct file *hugetlb_file_setup(const char *name, size_t size) | |||
981 | 982 | ||
982 | error = -ENOMEM; | 983 | error = -ENOMEM; |
983 | if (hugetlb_reserve_pages(inode, 0, | 984 | if (hugetlb_reserve_pages(inode, 0, |
984 | size >> huge_page_shift(hstate_inode(inode)), NULL)) | 985 | size >> huge_page_shift(hstate_inode(inode)), NULL, |
986 | acctflag)) | ||
985 | goto out_inode; | 987 | goto out_inode; |
986 | 988 | ||
987 | d_instantiate(dentry, inode); | 989 | d_instantiate(dentry, inode); |
diff --git a/fs/internal.h b/fs/internal.h index 53af885f1732..0d8ac497b3d5 100644 --- a/fs/internal.h +++ b/fs/internal.h | |||
@@ -43,7 +43,7 @@ extern void __init chrdev_init(void); | |||
43 | /* | 43 | /* |
44 | * exec.c | 44 | * exec.c |
45 | */ | 45 | */ |
46 | extern void check_unsafe_exec(struct linux_binprm *); | 46 | extern void check_unsafe_exec(struct linux_binprm *, struct files_struct *); |
47 | 47 | ||
48 | /* | 48 | /* |
49 | * namespace.c | 49 | * namespace.c |
diff --git a/fs/ioctl.c b/fs/ioctl.c index 20b0a8a24c6b..240ec63984cb 100644 --- a/fs/ioctl.c +++ b/fs/ioctl.c | |||
@@ -542,7 +542,7 @@ int do_vfs_ioctl(struct file *filp, unsigned int fd, unsigned int cmd, | |||
542 | return error; | 542 | return error; |
543 | } | 543 | } |
544 | 544 | ||
545 | asmlinkage long sys_ioctl(unsigned int fd, unsigned int cmd, unsigned long arg) | 545 | SYSCALL_DEFINE3(ioctl, unsigned int, fd, unsigned int, cmd, unsigned long, arg) |
546 | { | 546 | { |
547 | struct file *filp; | 547 | struct file *filp; |
548 | int error = -EBADF; | 548 | int error = -EBADF; |
diff --git a/fs/ioprio.c b/fs/ioprio.c index 1a39ac370942..c7c0b28d7d21 100644 --- a/fs/ioprio.c +++ b/fs/ioprio.c | |||
@@ -72,7 +72,7 @@ int set_task_ioprio(struct task_struct *task, int ioprio) | |||
72 | } | 72 | } |
73 | EXPORT_SYMBOL_GPL(set_task_ioprio); | 73 | EXPORT_SYMBOL_GPL(set_task_ioprio); |
74 | 74 | ||
75 | asmlinkage long sys_ioprio_set(int which, int who, int ioprio) | 75 | SYSCALL_DEFINE3(ioprio_set, int, which, int, who, int, ioprio) |
76 | { | 76 | { |
77 | int class = IOPRIO_PRIO_CLASS(ioprio); | 77 | int class = IOPRIO_PRIO_CLASS(ioprio); |
78 | int data = IOPRIO_PRIO_DATA(ioprio); | 78 | int data = IOPRIO_PRIO_DATA(ioprio); |
@@ -188,7 +188,7 @@ int ioprio_best(unsigned short aprio, unsigned short bprio) | |||
188 | return aprio; | 188 | return aprio; |
189 | } | 189 | } |
190 | 190 | ||
191 | asmlinkage long sys_ioprio_get(int which, int who) | 191 | SYSCALL_DEFINE2(ioprio_get, int, which, int, who) |
192 | { | 192 | { |
193 | struct task_struct *g, *p; | 193 | struct task_struct *g, *p; |
194 | struct user_struct *user; | 194 | struct user_struct *user; |
@@ -252,4 +252,3 @@ asmlinkage long sys_ioprio_get(int which, int who) | |||
252 | read_unlock(&tasklist_lock); | 252 | read_unlock(&tasklist_lock); |
253 | return ret; | 253 | return ret; |
254 | } | 254 | } |
255 | |||
diff --git a/fs/isofs/Kconfig b/fs/isofs/Kconfig new file mode 100644 index 000000000000..8ab9878e3671 --- /dev/null +++ b/fs/isofs/Kconfig | |||
@@ -0,0 +1,39 @@ | |||
1 | config ISO9660_FS | ||
2 | tristate "ISO 9660 CDROM file system support" | ||
3 | help | ||
4 | This is the standard file system used on CD-ROMs. It was previously | ||
5 | known as "High Sierra File System" and is called "hsfs" on other | ||
6 | Unix systems. The so-called Rock-Ridge extensions which allow for | ||
7 | long Unix filenames and symbolic links are also supported by this | ||
8 | driver. If you have a CD-ROM drive and want to do more with it than | ||
9 | just listen to audio CDs and watch its LEDs, say Y (and read | ||
10 | <file:Documentation/filesystems/isofs.txt> and the CD-ROM-HOWTO, | ||
11 | available from <http://www.tldp.org/docs.html#howto>), thereby | ||
12 | enlarging your kernel by about 27 KB; otherwise say N. | ||
13 | |||
14 | To compile this file system support as a module, choose M here: the | ||
15 | module will be called isofs. | ||
16 | |||
17 | config JOLIET | ||
18 | bool "Microsoft Joliet CDROM extensions" | ||
19 | depends on ISO9660_FS | ||
20 | select NLS | ||
21 | help | ||
22 | Joliet is a Microsoft extension for the ISO 9660 CD-ROM file system | ||
23 | which allows for long filenames in unicode format (unicode is the | ||
24 | new 16 bit character code, successor to ASCII, which encodes the | ||
25 | characters of almost all languages of the world; see | ||
26 | <http://www.unicode.org/> for more information). Say Y here if you | ||
27 | want to be able to read Joliet CD-ROMs under Linux. | ||
28 | |||
29 | config ZISOFS | ||
30 | bool "Transparent decompression extension" | ||
31 | depends on ISO9660_FS | ||
32 | select ZLIB_INFLATE | ||
33 | help | ||
34 | This is a Linux-specific extension to RockRidge which lets you store | ||
35 | data in compressed form on a CD-ROM and have it transparently | ||
36 | decompressed when the CD-ROM is accessed. See | ||
37 | <http://www.kernel.org/pub/linux/utils/fs/zisofs/> for the tools | ||
38 | necessary to create such a filesystem. Say Y here if you want to be | ||
39 | able to read such compressed CD-ROMs. | ||
diff --git a/fs/jbd/journal.c b/fs/jbd/journal.c index 9e4fa52d7dc8..e79c07812afa 100644 --- a/fs/jbd/journal.c +++ b/fs/jbd/journal.c | |||
@@ -427,7 +427,7 @@ int __log_space_left(journal_t *journal) | |||
427 | } | 427 | } |
428 | 428 | ||
429 | /* | 429 | /* |
430 | * Called under j_state_lock. Returns true if a transaction was started. | 430 | * Called under j_state_lock. Returns true if a transaction commit was started. |
431 | */ | 431 | */ |
432 | int __log_start_commit(journal_t *journal, tid_t target) | 432 | int __log_start_commit(journal_t *journal, tid_t target) |
433 | { | 433 | { |
@@ -495,7 +495,8 @@ int journal_force_commit_nested(journal_t *journal) | |||
495 | 495 | ||
496 | /* | 496 | /* |
497 | * Start a commit of the current running transaction (if any). Returns true | 497 | * Start a commit of the current running transaction (if any). Returns true |
498 | * if a transaction was started, and fills its tid in at *ptid | 498 | * if a transaction is going to be committed (or is currently already |
499 | * committing), and fills its tid in at *ptid | ||
499 | */ | 500 | */ |
500 | int journal_start_commit(journal_t *journal, tid_t *ptid) | 501 | int journal_start_commit(journal_t *journal, tid_t *ptid) |
501 | { | 502 | { |
@@ -505,15 +506,19 @@ int journal_start_commit(journal_t *journal, tid_t *ptid) | |||
505 | if (journal->j_running_transaction) { | 506 | if (journal->j_running_transaction) { |
506 | tid_t tid = journal->j_running_transaction->t_tid; | 507 | tid_t tid = journal->j_running_transaction->t_tid; |
507 | 508 | ||
508 | ret = __log_start_commit(journal, tid); | 509 | __log_start_commit(journal, tid); |
509 | if (ret && ptid) | 510 | /* There's a running transaction and we've just made sure |
511 | * it's commit has been scheduled. */ | ||
512 | if (ptid) | ||
510 | *ptid = tid; | 513 | *ptid = tid; |
511 | } else if (journal->j_committing_transaction && ptid) { | 514 | ret = 1; |
515 | } else if (journal->j_committing_transaction) { | ||
512 | /* | 516 | /* |
513 | * If ext3_write_super() recently started a commit, then we | 517 | * If ext3_write_super() recently started a commit, then we |
514 | * have to wait for completion of that transaction | 518 | * have to wait for completion of that transaction |
515 | */ | 519 | */ |
516 | *ptid = journal->j_committing_transaction->t_tid; | 520 | if (ptid) |
521 | *ptid = journal->j_committing_transaction->t_tid; | ||
517 | ret = 1; | 522 | ret = 1; |
518 | } | 523 | } |
519 | spin_unlock(&journal->j_state_lock); | 524 | spin_unlock(&journal->j_state_lock); |
diff --git a/fs/jbd2/journal.c b/fs/jbd2/journal.c index 56675306ed81..58144102bf25 100644 --- a/fs/jbd2/journal.c +++ b/fs/jbd2/journal.c | |||
@@ -37,10 +37,10 @@ | |||
37 | #include <linux/proc_fs.h> | 37 | #include <linux/proc_fs.h> |
38 | #include <linux/debugfs.h> | 38 | #include <linux/debugfs.h> |
39 | #include <linux/seq_file.h> | 39 | #include <linux/seq_file.h> |
40 | #include <linux/math64.h> | ||
40 | 41 | ||
41 | #include <asm/uaccess.h> | 42 | #include <asm/uaccess.h> |
42 | #include <asm/page.h> | 43 | #include <asm/page.h> |
43 | #include <asm/div64.h> | ||
44 | 44 | ||
45 | EXPORT_SYMBOL(jbd2_journal_start); | 45 | EXPORT_SYMBOL(jbd2_journal_start); |
46 | EXPORT_SYMBOL(jbd2_journal_restart); | 46 | EXPORT_SYMBOL(jbd2_journal_restart); |
@@ -450,7 +450,7 @@ int __jbd2_log_space_left(journal_t *journal) | |||
450 | } | 450 | } |
451 | 451 | ||
452 | /* | 452 | /* |
453 | * Called under j_state_lock. Returns true if a transaction was started. | 453 | * Called under j_state_lock. Returns true if a transaction commit was started. |
454 | */ | 454 | */ |
455 | int __jbd2_log_start_commit(journal_t *journal, tid_t target) | 455 | int __jbd2_log_start_commit(journal_t *journal, tid_t target) |
456 | { | 456 | { |
@@ -518,7 +518,8 @@ int jbd2_journal_force_commit_nested(journal_t *journal) | |||
518 | 518 | ||
519 | /* | 519 | /* |
520 | * Start a commit of the current running transaction (if any). Returns true | 520 | * Start a commit of the current running transaction (if any). Returns true |
521 | * if a transaction was started, and fills its tid in at *ptid | 521 | * if a transaction is going to be committed (or is currently already |
522 | * committing), and fills its tid in at *ptid | ||
522 | */ | 523 | */ |
523 | int jbd2_journal_start_commit(journal_t *journal, tid_t *ptid) | 524 | int jbd2_journal_start_commit(journal_t *journal, tid_t *ptid) |
524 | { | 525 | { |
@@ -528,15 +529,19 @@ int jbd2_journal_start_commit(journal_t *journal, tid_t *ptid) | |||
528 | if (journal->j_running_transaction) { | 529 | if (journal->j_running_transaction) { |
529 | tid_t tid = journal->j_running_transaction->t_tid; | 530 | tid_t tid = journal->j_running_transaction->t_tid; |
530 | 531 | ||
531 | ret = __jbd2_log_start_commit(journal, tid); | 532 | __jbd2_log_start_commit(journal, tid); |
532 | if (ret && ptid) | 533 | /* There's a running transaction and we've just made sure |
534 | * it's commit has been scheduled. */ | ||
535 | if (ptid) | ||
533 | *ptid = tid; | 536 | *ptid = tid; |
534 | } else if (journal->j_committing_transaction && ptid) { | 537 | ret = 1; |
538 | } else if (journal->j_committing_transaction) { | ||
535 | /* | 539 | /* |
536 | * If ext3_write_super() recently started a commit, then we | 540 | * If ext3_write_super() recently started a commit, then we |
537 | * have to wait for completion of that transaction | 541 | * have to wait for completion of that transaction |
538 | */ | 542 | */ |
539 | *ptid = journal->j_committing_transaction->t_tid; | 543 | if (ptid) |
544 | *ptid = journal->j_committing_transaction->t_tid; | ||
540 | ret = 1; | 545 | ret = 1; |
541 | } | 546 | } |
542 | spin_unlock(&journal->j_state_lock); | 547 | spin_unlock(&journal->j_state_lock); |
@@ -846,8 +851,8 @@ static int jbd2_seq_info_show(struct seq_file *seq, void *v) | |||
846 | jiffies_to_msecs(s->stats->u.run.rs_flushing / s->stats->ts_tid)); | 851 | jiffies_to_msecs(s->stats->u.run.rs_flushing / s->stats->ts_tid)); |
847 | seq_printf(seq, " %ums logging transaction\n", | 852 | seq_printf(seq, " %ums logging transaction\n", |
848 | jiffies_to_msecs(s->stats->u.run.rs_logging / s->stats->ts_tid)); | 853 | jiffies_to_msecs(s->stats->u.run.rs_logging / s->stats->ts_tid)); |
849 | seq_printf(seq, " %luus average transaction commit time\n", | 854 | seq_printf(seq, " %lluus average transaction commit time\n", |
850 | do_div(s->journal->j_average_commit_time, 1000)); | 855 | div_u64(s->journal->j_average_commit_time, 1000)); |
851 | seq_printf(seq, " %lu handles per transaction\n", | 856 | seq_printf(seq, " %lu handles per transaction\n", |
852 | s->stats->u.run.rs_handle_count / s->stats->ts_tid); | 857 | s->stats->u.run.rs_handle_count / s->stats->ts_tid); |
853 | seq_printf(seq, " %lu blocks per transaction\n", | 858 | seq_printf(seq, " %lu blocks per transaction\n", |
diff --git a/fs/jbd2/transaction.c b/fs/jbd2/transaction.c index 46b4e347ed7d..28ce21d8598e 100644 --- a/fs/jbd2/transaction.c +++ b/fs/jbd2/transaction.c | |||
@@ -2129,26 +2129,46 @@ done: | |||
2129 | } | 2129 | } |
2130 | 2130 | ||
2131 | /* | 2131 | /* |
2132 | * This function must be called when inode is journaled in ordered mode | 2132 | * File truncate and transaction commit interact with each other in a |
2133 | * before truncation happens. It starts writeout of truncated part in | 2133 | * non-trivial way. If a transaction writing data block A is |
2134 | * case it is in the committing transaction so that we stand to ordered | 2134 | * committing, we cannot discard the data by truncate until we have |
2135 | * mode consistency guarantees. | 2135 | * written them. Otherwise if we crashed after the transaction with |
2136 | * write has committed but before the transaction with truncate has | ||
2137 | * committed, we could see stale data in block A. This function is a | ||
2138 | * helper to solve this problem. It starts writeout of the truncated | ||
2139 | * part in case it is in the committing transaction. | ||
2140 | * | ||
2141 | * Filesystem code must call this function when inode is journaled in | ||
2142 | * ordered mode before truncation happens and after the inode has been | ||
2143 | * placed on orphan list with the new inode size. The second condition | ||
2144 | * avoids the race that someone writes new data and we start | ||
2145 | * committing the transaction after this function has been called but | ||
2146 | * before a transaction for truncate is started (and furthermore it | ||
2147 | * allows us to optimize the case where the addition to orphan list | ||
2148 | * happens in the same transaction as write --- we don't have to write | ||
2149 | * any data in such case). | ||
2136 | */ | 2150 | */ |
2137 | int jbd2_journal_begin_ordered_truncate(struct jbd2_inode *inode, | 2151 | int jbd2_journal_begin_ordered_truncate(journal_t *journal, |
2152 | struct jbd2_inode *jinode, | ||
2138 | loff_t new_size) | 2153 | loff_t new_size) |
2139 | { | 2154 | { |
2140 | journal_t *journal; | 2155 | transaction_t *inode_trans, *commit_trans; |
2141 | transaction_t *commit_trans; | ||
2142 | int ret = 0; | 2156 | int ret = 0; |
2143 | 2157 | ||
2144 | if (!inode->i_transaction && !inode->i_next_transaction) | 2158 | /* This is a quick check to avoid locking if not necessary */ |
2159 | if (!jinode->i_transaction) | ||
2145 | goto out; | 2160 | goto out; |
2146 | journal = inode->i_transaction->t_journal; | 2161 | /* Locks are here just to force reading of recent values, it is |
2162 | * enough that the transaction was not committing before we started | ||
2163 | * a transaction adding the inode to orphan list */ | ||
2147 | spin_lock(&journal->j_state_lock); | 2164 | spin_lock(&journal->j_state_lock); |
2148 | commit_trans = journal->j_committing_transaction; | 2165 | commit_trans = journal->j_committing_transaction; |
2149 | spin_unlock(&journal->j_state_lock); | 2166 | spin_unlock(&journal->j_state_lock); |
2150 | if (inode->i_transaction == commit_trans) { | 2167 | spin_lock(&journal->j_list_lock); |
2151 | ret = filemap_fdatawrite_range(inode->i_vfs_inode->i_mapping, | 2168 | inode_trans = jinode->i_transaction; |
2169 | spin_unlock(&journal->j_list_lock); | ||
2170 | if (inode_trans == commit_trans) { | ||
2171 | ret = filemap_fdatawrite_range(jinode->i_vfs_inode->i_mapping, | ||
2152 | new_size, LLONG_MAX); | 2172 | new_size, LLONG_MAX); |
2153 | if (ret) | 2173 | if (ret) |
2154 | jbd2_journal_abort(journal, ret); | 2174 | jbd2_journal_abort(journal, ret); |
diff --git a/fs/jfs/Kconfig b/fs/jfs/Kconfig new file mode 100644 index 000000000000..9ff619a6f9cc --- /dev/null +++ b/fs/jfs/Kconfig | |||
@@ -0,0 +1,49 @@ | |||
1 | config JFS_FS | ||
2 | tristate "JFS filesystem support" | ||
3 | select NLS | ||
4 | help | ||
5 | This is a port of IBM's Journaled Filesystem . More information is | ||
6 | available in the file <file:Documentation/filesystems/jfs.txt>. | ||
7 | |||
8 | If you do not intend to use the JFS filesystem, say N. | ||
9 | |||
10 | config JFS_POSIX_ACL | ||
11 | bool "JFS POSIX Access Control Lists" | ||
12 | depends on JFS_FS | ||
13 | select FS_POSIX_ACL | ||
14 | help | ||
15 | Posix Access Control Lists (ACLs) support permissions for users and | ||
16 | groups beyond the owner/group/world scheme. | ||
17 | |||
18 | To learn more about Access Control Lists, visit the Posix ACLs for | ||
19 | Linux website <http://acl.bestbits.at/>. | ||
20 | |||
21 | If you don't know what Access Control Lists are, say N | ||
22 | |||
23 | config JFS_SECURITY | ||
24 | bool "JFS Security Labels" | ||
25 | depends on JFS_FS | ||
26 | help | ||
27 | Security labels support alternative access control models | ||
28 | implemented by security modules like SELinux. This option | ||
29 | enables an extended attribute handler for file security | ||
30 | labels in the jfs filesystem. | ||
31 | |||
32 | If you are not using a security module that requires using | ||
33 | extended attributes for file security labels, say N. | ||
34 | |||
35 | config JFS_DEBUG | ||
36 | bool "JFS debugging" | ||
37 | depends on JFS_FS | ||
38 | help | ||
39 | If you are experiencing any problems with the JFS filesystem, say | ||
40 | Y here. This will result in additional debugging messages to be | ||
41 | written to the system log. Under normal circumstances, this | ||
42 | results in very little overhead. | ||
43 | |||
44 | config JFS_STATISTICS | ||
45 | bool "JFS statistics" | ||
46 | depends on JFS_FS | ||
47 | help | ||
48 | Enabling this option will cause statistics from the JFS file system | ||
49 | to be made available to the user in the /proc/fs/jfs/ directory. | ||
diff --git a/fs/lockd/svclock.c b/fs/lockd/svclock.c index 6063a8e4b9f3..763b78a6e9de 100644 --- a/fs/lockd/svclock.c +++ b/fs/lockd/svclock.c | |||
@@ -427,7 +427,7 @@ nlmsvc_lock(struct svc_rqst *rqstp, struct nlm_file *file, | |||
427 | goto out; | 427 | goto out; |
428 | case -EAGAIN: | 428 | case -EAGAIN: |
429 | ret = nlm_lck_denied; | 429 | ret = nlm_lck_denied; |
430 | goto out; | 430 | break; |
431 | case FILE_LOCK_DEFERRED: | 431 | case FILE_LOCK_DEFERRED: |
432 | if (wait) | 432 | if (wait) |
433 | break; | 433 | break; |
@@ -443,6 +443,10 @@ nlmsvc_lock(struct svc_rqst *rqstp, struct nlm_file *file, | |||
443 | goto out; | 443 | goto out; |
444 | } | 444 | } |
445 | 445 | ||
446 | ret = nlm_lck_denied; | ||
447 | if (!wait) | ||
448 | goto out; | ||
449 | |||
446 | ret = nlm_lck_blocked; | 450 | ret = nlm_lck_blocked; |
447 | 451 | ||
448 | /* Append to list of blocked */ | 452 | /* Append to list of blocked */ |
diff --git a/fs/locks.c b/fs/locks.c index 46a2e12f7d42..ec3deea29e37 100644 --- a/fs/locks.c +++ b/fs/locks.c | |||
@@ -1564,7 +1564,7 @@ EXPORT_SYMBOL(flock_lock_file_wait); | |||
1564 | * %LOCK_MAND can be combined with %LOCK_READ or %LOCK_WRITE to allow other | 1564 | * %LOCK_MAND can be combined with %LOCK_READ or %LOCK_WRITE to allow other |
1565 | * processes read and write access respectively. | 1565 | * processes read and write access respectively. |
1566 | */ | 1566 | */ |
1567 | asmlinkage long sys_flock(unsigned int fd, unsigned int cmd) | 1567 | SYSCALL_DEFINE2(flock, unsigned int, fd, unsigned int, cmd) |
1568 | { | 1568 | { |
1569 | struct file *filp; | 1569 | struct file *filp; |
1570 | struct file_lock *lock; | 1570 | struct file_lock *lock; |
diff --git a/fs/minix/Kconfig b/fs/minix/Kconfig new file mode 100644 index 000000000000..0fd7ca994264 --- /dev/null +++ b/fs/minix/Kconfig | |||
@@ -0,0 +1,17 @@ | |||
1 | config MINIX_FS | ||
2 | tristate "Minix file system support" | ||
3 | depends on BLOCK | ||
4 | help | ||
5 | Minix is a simple operating system used in many classes about OS's. | ||
6 | The minix file system (method to organize files on a hard disk | ||
7 | partition or a floppy disk) was the original file system for Linux, | ||
8 | but has been superseded by the second extended file system ext2fs. | ||
9 | You don't want to use the minix file system on your hard disk | ||
10 | because of certain built-in restrictions, but it is sometimes found | ||
11 | on older Linux floppy disks. This option will enlarge your kernel | ||
12 | by about 28 KB. If unsure, say N. | ||
13 | |||
14 | To compile this file system support as a module, choose M here: the | ||
15 | module will be called minix. Note that the file system of your root | ||
16 | partition (the one containing the directory /) cannot be compiled as | ||
17 | a module. | ||
diff --git a/fs/namei.c b/fs/namei.c index f05bed242422..bbc15c237558 100644 --- a/fs/namei.c +++ b/fs/namei.c | |||
@@ -1962,8 +1962,8 @@ static int may_mknod(mode_t mode) | |||
1962 | } | 1962 | } |
1963 | } | 1963 | } |
1964 | 1964 | ||
1965 | asmlinkage long sys_mknodat(int dfd, const char __user *filename, int mode, | 1965 | SYSCALL_DEFINE4(mknodat, int, dfd, const char __user *, filename, int, mode, |
1966 | unsigned dev) | 1966 | unsigned, dev) |
1967 | { | 1967 | { |
1968 | int error; | 1968 | int error; |
1969 | char *tmp; | 1969 | char *tmp; |
@@ -2017,7 +2017,7 @@ out_unlock: | |||
2017 | return error; | 2017 | return error; |
2018 | } | 2018 | } |
2019 | 2019 | ||
2020 | asmlinkage long sys_mknod(const char __user *filename, int mode, unsigned dev) | 2020 | SYSCALL_DEFINE3(mknod, const char __user *, filename, int, mode, unsigned, dev) |
2021 | { | 2021 | { |
2022 | return sys_mknodat(AT_FDCWD, filename, mode, dev); | 2022 | return sys_mknodat(AT_FDCWD, filename, mode, dev); |
2023 | } | 2023 | } |
@@ -2044,7 +2044,7 @@ int vfs_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
2044 | return error; | 2044 | return error; |
2045 | } | 2045 | } |
2046 | 2046 | ||
2047 | asmlinkage long sys_mkdirat(int dfd, const char __user *pathname, int mode) | 2047 | SYSCALL_DEFINE3(mkdirat, int, dfd, const char __user *, pathname, int, mode) |
2048 | { | 2048 | { |
2049 | int error = 0; | 2049 | int error = 0; |
2050 | char * tmp; | 2050 | char * tmp; |
@@ -2081,7 +2081,7 @@ out_err: | |||
2081 | return error; | 2081 | return error; |
2082 | } | 2082 | } |
2083 | 2083 | ||
2084 | asmlinkage long sys_mkdir(const char __user *pathname, int mode) | 2084 | SYSCALL_DEFINE2(mkdir, const char __user *, pathname, int, mode) |
2085 | { | 2085 | { |
2086 | return sys_mkdirat(AT_FDCWD, pathname, mode); | 2086 | return sys_mkdirat(AT_FDCWD, pathname, mode); |
2087 | } | 2087 | } |
@@ -2195,7 +2195,7 @@ exit1: | |||
2195 | return error; | 2195 | return error; |
2196 | } | 2196 | } |
2197 | 2197 | ||
2198 | asmlinkage long sys_rmdir(const char __user *pathname) | 2198 | SYSCALL_DEFINE1(rmdir, const char __user *, pathname) |
2199 | { | 2199 | { |
2200 | return do_rmdir(AT_FDCWD, pathname); | 2200 | return do_rmdir(AT_FDCWD, pathname); |
2201 | } | 2201 | } |
@@ -2291,7 +2291,7 @@ slashes: | |||
2291 | goto exit2; | 2291 | goto exit2; |
2292 | } | 2292 | } |
2293 | 2293 | ||
2294 | asmlinkage long sys_unlinkat(int dfd, const char __user *pathname, int flag) | 2294 | SYSCALL_DEFINE3(unlinkat, int, dfd, const char __user *, pathname, int, flag) |
2295 | { | 2295 | { |
2296 | if ((flag & ~AT_REMOVEDIR) != 0) | 2296 | if ((flag & ~AT_REMOVEDIR) != 0) |
2297 | return -EINVAL; | 2297 | return -EINVAL; |
@@ -2302,7 +2302,7 @@ asmlinkage long sys_unlinkat(int dfd, const char __user *pathname, int flag) | |||
2302 | return do_unlinkat(dfd, pathname); | 2302 | return do_unlinkat(dfd, pathname); |
2303 | } | 2303 | } |
2304 | 2304 | ||
2305 | asmlinkage long sys_unlink(const char __user *pathname) | 2305 | SYSCALL_DEFINE1(unlink, const char __user *, pathname) |
2306 | { | 2306 | { |
2307 | return do_unlinkat(AT_FDCWD, pathname); | 2307 | return do_unlinkat(AT_FDCWD, pathname); |
2308 | } | 2308 | } |
@@ -2328,8 +2328,8 @@ int vfs_symlink(struct inode *dir, struct dentry *dentry, const char *oldname) | |||
2328 | return error; | 2328 | return error; |
2329 | } | 2329 | } |
2330 | 2330 | ||
2331 | asmlinkage long sys_symlinkat(const char __user *oldname, | 2331 | SYSCALL_DEFINE3(symlinkat, const char __user *, oldname, |
2332 | int newdfd, const char __user *newname) | 2332 | int, newdfd, const char __user *, newname) |
2333 | { | 2333 | { |
2334 | int error; | 2334 | int error; |
2335 | char *from; | 2335 | char *from; |
@@ -2370,7 +2370,7 @@ out_putname: | |||
2370 | return error; | 2370 | return error; |
2371 | } | 2371 | } |
2372 | 2372 | ||
2373 | asmlinkage long sys_symlink(const char __user *oldname, const char __user *newname) | 2373 | SYSCALL_DEFINE2(symlink, const char __user *, oldname, const char __user *, newname) |
2374 | { | 2374 | { |
2375 | return sys_symlinkat(oldname, AT_FDCWD, newname); | 2375 | return sys_symlinkat(oldname, AT_FDCWD, newname); |
2376 | } | 2376 | } |
@@ -2422,9 +2422,8 @@ int vfs_link(struct dentry *old_dentry, struct inode *dir, struct dentry *new_de | |||
2422 | * with linux 2.0, and to avoid hard-linking to directories | 2422 | * with linux 2.0, and to avoid hard-linking to directories |
2423 | * and other special files. --ADM | 2423 | * and other special files. --ADM |
2424 | */ | 2424 | */ |
2425 | asmlinkage long sys_linkat(int olddfd, const char __user *oldname, | 2425 | SYSCALL_DEFINE5(linkat, int, olddfd, const char __user *, oldname, |
2426 | int newdfd, const char __user *newname, | 2426 | int, newdfd, const char __user *, newname, int, flags) |
2427 | int flags) | ||
2428 | { | 2427 | { |
2429 | struct dentry *new_dentry; | 2428 | struct dentry *new_dentry; |
2430 | struct nameidata nd; | 2429 | struct nameidata nd; |
@@ -2473,7 +2472,7 @@ out: | |||
2473 | return error; | 2472 | return error; |
2474 | } | 2473 | } |
2475 | 2474 | ||
2476 | asmlinkage long sys_link(const char __user *oldname, const char __user *newname) | 2475 | SYSCALL_DEFINE2(link, const char __user *, oldname, const char __user *, newname) |
2477 | { | 2476 | { |
2478 | return sys_linkat(AT_FDCWD, oldname, AT_FDCWD, newname, 0); | 2477 | return sys_linkat(AT_FDCWD, oldname, AT_FDCWD, newname, 0); |
2479 | } | 2478 | } |
@@ -2624,8 +2623,8 @@ int vfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
2624 | return error; | 2623 | return error; |
2625 | } | 2624 | } |
2626 | 2625 | ||
2627 | asmlinkage long sys_renameat(int olddfd, const char __user *oldname, | 2626 | SYSCALL_DEFINE4(renameat, int, olddfd, const char __user *, oldname, |
2628 | int newdfd, const char __user *newname) | 2627 | int, newdfd, const char __user *, newname) |
2629 | { | 2628 | { |
2630 | struct dentry *old_dir, *new_dir; | 2629 | struct dentry *old_dir, *new_dir; |
2631 | struct dentry *old_dentry, *new_dentry; | 2630 | struct dentry *old_dentry, *new_dentry; |
@@ -2718,7 +2717,7 @@ exit: | |||
2718 | return error; | 2717 | return error; |
2719 | } | 2718 | } |
2720 | 2719 | ||
2721 | asmlinkage long sys_rename(const char __user *oldname, const char __user *newname) | 2720 | SYSCALL_DEFINE2(rename, const char __user *, oldname, const char __user *, newname) |
2722 | { | 2721 | { |
2723 | return sys_renameat(AT_FDCWD, oldname, AT_FDCWD, newname); | 2722 | return sys_renameat(AT_FDCWD, oldname, AT_FDCWD, newname); |
2724 | } | 2723 | } |
diff --git a/fs/namespace.c b/fs/namespace.c index a40685d800a8..06f8e63f6cb1 100644 --- a/fs/namespace.c +++ b/fs/namespace.c | |||
@@ -614,9 +614,11 @@ static inline void __mntput(struct vfsmount *mnt) | |||
614 | */ | 614 | */ |
615 | for_each_possible_cpu(cpu) { | 615 | for_each_possible_cpu(cpu) { |
616 | struct mnt_writer *cpu_writer = &per_cpu(mnt_writers, cpu); | 616 | struct mnt_writer *cpu_writer = &per_cpu(mnt_writers, cpu); |
617 | if (cpu_writer->mnt != mnt) | ||
618 | continue; | ||
619 | spin_lock(&cpu_writer->lock); | 617 | spin_lock(&cpu_writer->lock); |
618 | if (cpu_writer->mnt != mnt) { | ||
619 | spin_unlock(&cpu_writer->lock); | ||
620 | continue; | ||
621 | } | ||
620 | atomic_add(cpu_writer->count, &mnt->__mnt_writers); | 622 | atomic_add(cpu_writer->count, &mnt->__mnt_writers); |
621 | cpu_writer->count = 0; | 623 | cpu_writer->count = 0; |
622 | /* | 624 | /* |
@@ -1128,7 +1130,7 @@ static int do_umount(struct vfsmount *mnt, int flags) | |||
1128 | * unixes. Our API is identical to OSF/1 to avoid making a mess of AMD | 1130 | * unixes. Our API is identical to OSF/1 to avoid making a mess of AMD |
1129 | */ | 1131 | */ |
1130 | 1132 | ||
1131 | asmlinkage long sys_umount(char __user * name, int flags) | 1133 | SYSCALL_DEFINE2(umount, char __user *, name, int, flags) |
1132 | { | 1134 | { |
1133 | struct path path; | 1135 | struct path path; |
1134 | int retval; | 1136 | int retval; |
@@ -1160,7 +1162,7 @@ out: | |||
1160 | /* | 1162 | /* |
1161 | * The 2.0 compatible umount. No flags. | 1163 | * The 2.0 compatible umount. No flags. |
1162 | */ | 1164 | */ |
1163 | asmlinkage long sys_oldumount(char __user * name) | 1165 | SYSCALL_DEFINE1(oldumount, char __user *, name) |
1164 | { | 1166 | { |
1165 | return sys_umount(name, 0); | 1167 | return sys_umount(name, 0); |
1166 | } | 1168 | } |
@@ -2045,9 +2047,8 @@ struct mnt_namespace *copy_mnt_ns(unsigned long flags, struct mnt_namespace *ns, | |||
2045 | return new_ns; | 2047 | return new_ns; |
2046 | } | 2048 | } |
2047 | 2049 | ||
2048 | asmlinkage long sys_mount(char __user * dev_name, char __user * dir_name, | 2050 | SYSCALL_DEFINE5(mount, char __user *, dev_name, char __user *, dir_name, |
2049 | char __user * type, unsigned long flags, | 2051 | char __user *, type, unsigned long, flags, void __user *, data) |
2050 | void __user * data) | ||
2051 | { | 2052 | { |
2052 | int retval; | 2053 | int retval; |
2053 | unsigned long data_page; | 2054 | unsigned long data_page; |
@@ -2172,8 +2173,8 @@ static void chroot_fs_refs(struct path *old_root, struct path *new_root) | |||
2172 | * though, so you may need to say mount --bind /nfs/my_root /nfs/my_root | 2173 | * though, so you may need to say mount --bind /nfs/my_root /nfs/my_root |
2173 | * first. | 2174 | * first. |
2174 | */ | 2175 | */ |
2175 | asmlinkage long sys_pivot_root(const char __user * new_root, | 2176 | SYSCALL_DEFINE2(pivot_root, const char __user *, new_root, |
2176 | const char __user * put_old) | 2177 | const char __user *, put_old) |
2177 | { | 2178 | { |
2178 | struct vfsmount *tmp; | 2179 | struct vfsmount *tmp; |
2179 | struct path new, old, parent_path, root_parent, root; | 2180 | struct path new, old, parent_path, root_parent, root; |
diff --git a/fs/ncpfs/Kconfig b/fs/ncpfs/Kconfig index 142808427b25..c931cf22a1f6 100644 --- a/fs/ncpfs/Kconfig +++ b/fs/ncpfs/Kconfig | |||
@@ -1,6 +1,27 @@ | |||
1 | # | 1 | # |
2 | # NCP Filesystem configuration | 2 | # NCP Filesystem configuration |
3 | # | 3 | # |
4 | config NCP_FS | ||
5 | tristate "NCP file system support (to mount NetWare volumes)" | ||
6 | depends on IPX!=n || INET | ||
7 | help | ||
8 | NCP (NetWare Core Protocol) is a protocol that runs over IPX and is | ||
9 | used by Novell NetWare clients to talk to file servers. It is to | ||
10 | IPX what NFS is to TCP/IP, if that helps. Saying Y here allows you | ||
11 | to mount NetWare file server volumes and to access them just like | ||
12 | any other Unix directory. For details, please read the file | ||
13 | <file:Documentation/filesystems/ncpfs.txt> in the kernel source and | ||
14 | the IPX-HOWTO from <http://www.tldp.org/docs.html#howto>. | ||
15 | |||
16 | You do not have to say Y here if you want your Linux box to act as a | ||
17 | file *server* for Novell NetWare clients. | ||
18 | |||
19 | General information about how to connect Linux, Windows machines and | ||
20 | Macs is on the WWW at <http://www.eats.com/linux_mac_win.html>. | ||
21 | |||
22 | To compile this as a module, choose M here: the module will be called | ||
23 | ncpfs. Say N unless you are connected to a Novell network. | ||
24 | |||
4 | config NCPFS_PACKET_SIGNING | 25 | config NCPFS_PACKET_SIGNING |
5 | bool "Packet signatures" | 26 | bool "Packet signatures" |
6 | depends on NCP_FS | 27 | depends on NCP_FS |
diff --git a/fs/nfs/Kconfig b/fs/nfs/Kconfig new file mode 100644 index 000000000000..36fe20d6eba2 --- /dev/null +++ b/fs/nfs/Kconfig | |||
@@ -0,0 +1,86 @@ | |||
1 | config NFS_FS | ||
2 | tristate "NFS client support" | ||
3 | depends on INET | ||
4 | select LOCKD | ||
5 | select SUNRPC | ||
6 | select NFS_ACL_SUPPORT if NFS_V3_ACL | ||
7 | help | ||
8 | Choose Y here if you want to access files residing on other | ||
9 | computers using Sun's Network File System protocol. To compile | ||
10 | this file system support as a module, choose M here: the module | ||
11 | will be called nfs. | ||
12 | |||
13 | To mount file systems exported by NFS servers, you also need to | ||
14 | install the user space mount.nfs command which can be found in | ||
15 | the Linux nfs-utils package, available from http://linux-nfs.org/. | ||
16 | Information about using the mount command is available in the | ||
17 | mount(8) man page. More detail about the Linux NFS client | ||
18 | implementation is available via the nfs(5) man page. | ||
19 | |||
20 | Below you can choose which versions of the NFS protocol are | ||
21 | available in the kernel to mount NFS servers. Support for NFS | ||
22 | version 2 (RFC 1094) is always available when NFS_FS is selected. | ||
23 | |||
24 | To configure a system which mounts its root file system via NFS | ||
25 | at boot time, say Y here, select "Kernel level IP | ||
26 | autoconfiguration" in the NETWORK menu, and select "Root file | ||
27 | system on NFS" below. You cannot compile this file system as a | ||
28 | module in this case. | ||
29 | |||
30 | If unsure, say N. | ||
31 | |||
32 | config NFS_V3 | ||
33 | bool "NFS client support for NFS version 3" | ||
34 | depends on NFS_FS | ||
35 | help | ||
36 | This option enables support for version 3 of the NFS protocol | ||
37 | (RFC 1813) in the kernel's NFS client. | ||
38 | |||
39 | If unsure, say Y. | ||
40 | |||
41 | config NFS_V3_ACL | ||
42 | bool "NFS client support for the NFSv3 ACL protocol extension" | ||
43 | depends on NFS_V3 | ||
44 | help | ||
45 | Some NFS servers support an auxiliary NFSv3 ACL protocol that | ||
46 | Sun added to Solaris but never became an official part of the | ||
47 | NFS version 3 protocol. This protocol extension allows | ||
48 | applications on NFS clients to manipulate POSIX Access Control | ||
49 | Lists on files residing on NFS servers. NFS servers enforce | ||
50 | ACLs on local files whether this protocol is available or not. | ||
51 | |||
52 | Choose Y here if your NFS server supports the Solaris NFSv3 ACL | ||
53 | protocol extension and you want your NFS client to allow | ||
54 | applications to access and modify ACLs on files on the server. | ||
55 | |||
56 | Most NFS servers don't support the Solaris NFSv3 ACL protocol | ||
57 | extension. You can choose N here or specify the "noacl" mount | ||
58 | option to prevent your NFS client from trying to use the NFSv3 | ||
59 | ACL protocol. | ||
60 | |||
61 | If unsure, say N. | ||
62 | |||
63 | config NFS_V4 | ||
64 | bool "NFS client support for NFS version 4 (EXPERIMENTAL)" | ||
65 | depends on NFS_FS && EXPERIMENTAL | ||
66 | select RPCSEC_GSS_KRB5 | ||
67 | help | ||
68 | This option enables support for version 4 of the NFS protocol | ||
69 | (RFC 3530) in the kernel's NFS client. | ||
70 | |||
71 | To mount NFS servers using NFSv4, you also need to install user | ||
72 | space programs which can be found in the Linux nfs-utils package, | ||
73 | available from http://linux-nfs.org/. | ||
74 | |||
75 | If unsure, say N. | ||
76 | |||
77 | config ROOT_NFS | ||
78 | bool "Root file system on NFS" | ||
79 | depends on NFS_FS=y && IP_PNP | ||
80 | help | ||
81 | If you want your system to mount its root file system via NFS, | ||
82 | choose Y here. This is common practice for managing systems | ||
83 | without local permanent storage. For details, read | ||
84 | <file:Documentation/filesystems/nfsroot.txt>. | ||
85 | |||
86 | Most people say N here. | ||
diff --git a/fs/nfsctl.c b/fs/nfsctl.c index b27451909dff..8f9a20556f79 100644 --- a/fs/nfsctl.c +++ b/fs/nfsctl.c | |||
@@ -86,8 +86,8 @@ static struct { | |||
86 | }, | 86 | }, |
87 | }; | 87 | }; |
88 | 88 | ||
89 | long | 89 | SYSCALL_DEFINE3(nfsservctl, int, cmd, struct nfsctl_arg __user *, arg, |
90 | asmlinkage sys_nfsservctl(int cmd, struct nfsctl_arg __user *arg, void __user *res) | 90 | void __user *, res) |
91 | { | 91 | { |
92 | struct file *file; | 92 | struct file *file; |
93 | void __user *p = &arg->u; | 93 | void __user *p = &arg->u; |
diff --git a/fs/nfsd/Kconfig b/fs/nfsd/Kconfig new file mode 100644 index 000000000000..44d7d04dab95 --- /dev/null +++ b/fs/nfsd/Kconfig | |||
@@ -0,0 +1,80 @@ | |||
1 | config NFSD | ||
2 | tristate "NFS server support" | ||
3 | depends on INET | ||
4 | select LOCKD | ||
5 | select SUNRPC | ||
6 | select EXPORTFS | ||
7 | select NFS_ACL_SUPPORT if NFSD_V2_ACL | ||
8 | help | ||
9 | Choose Y here if you want to allow other computers to access | ||
10 | files residing on this system using Sun's Network File System | ||
11 | protocol. To compile the NFS server support as a module, | ||
12 | choose M here: the module will be called nfsd. | ||
13 | |||
14 | You may choose to use a user-space NFS server instead, in which | ||
15 | case you can choose N here. | ||
16 | |||
17 | To export local file systems using NFS, you also need to install | ||
18 | user space programs which can be found in the Linux nfs-utils | ||
19 | package, available from http://linux-nfs.org/. More detail about | ||
20 | the Linux NFS server implementation is available via the | ||
21 | exports(5) man page. | ||
22 | |||
23 | Below you can choose which versions of the NFS protocol are | ||
24 | available to clients mounting the NFS server on this system. | ||
25 | Support for NFS version 2 (RFC 1094) is always available when | ||
26 | CONFIG_NFSD is selected. | ||
27 | |||
28 | If unsure, say N. | ||
29 | |||
30 | config NFSD_V2_ACL | ||
31 | bool | ||
32 | depends on NFSD | ||
33 | |||
34 | config NFSD_V3 | ||
35 | bool "NFS server support for NFS version 3" | ||
36 | depends on NFSD | ||
37 | help | ||
38 | This option enables support in your system's NFS server for | ||
39 | version 3 of the NFS protocol (RFC 1813). | ||
40 | |||
41 | If unsure, say Y. | ||
42 | |||
43 | config NFSD_V3_ACL | ||
44 | bool "NFS server support for the NFSv3 ACL protocol extension" | ||
45 | depends on NFSD_V3 | ||
46 | select NFSD_V2_ACL | ||
47 | help | ||
48 | Solaris NFS servers support an auxiliary NFSv3 ACL protocol that | ||
49 | never became an official part of the NFS version 3 protocol. | ||
50 | This protocol extension allows applications on NFS clients to | ||
51 | manipulate POSIX Access Control Lists on files residing on NFS | ||
52 | servers. NFS servers enforce POSIX ACLs on local files whether | ||
53 | this protocol is available or not. | ||
54 | |||
55 | This option enables support in your system's NFS server for the | ||
56 | NFSv3 ACL protocol extension allowing NFS clients to manipulate | ||
57 | POSIX ACLs on files exported by your system's NFS server. NFS | ||
58 | clients which support the Solaris NFSv3 ACL protocol can then | ||
59 | access and modify ACLs on your NFS server. | ||
60 | |||
61 | To store ACLs on your NFS server, you also need to enable ACL- | ||
62 | related CONFIG options for your local file systems of choice. | ||
63 | |||
64 | If unsure, say N. | ||
65 | |||
66 | config NFSD_V4 | ||
67 | bool "NFS server support for NFS version 4 (EXPERIMENTAL)" | ||
68 | depends on NFSD && PROC_FS && EXPERIMENTAL | ||
69 | select NFSD_V3 | ||
70 | select FS_POSIX_ACL | ||
71 | select RPCSEC_GSS_KRB5 | ||
72 | help | ||
73 | This option enables support in your system's NFS server for | ||
74 | version 4 of the NFS protocol (RFC 3530). | ||
75 | |||
76 | To export files using NFSv4, you need to install additional user | ||
77 | space programs which can be found in the Linux nfs-utils package, | ||
78 | available from http://linux-nfs.org/. | ||
79 | |||
80 | If unsure, say N. | ||
diff --git a/fs/nfsd/auth.c b/fs/nfsd/auth.c index c903e04aa217..5573508f707f 100644 --- a/fs/nfsd/auth.c +++ b/fs/nfsd/auth.c | |||
@@ -49,6 +49,8 @@ int nfsd_setuser(struct svc_rqst *rqstp, struct svc_export *exp) | |||
49 | new->fsuid = exp->ex_anon_uid; | 49 | new->fsuid = exp->ex_anon_uid; |
50 | new->fsgid = exp->ex_anon_gid; | 50 | new->fsgid = exp->ex_anon_gid; |
51 | gi = groups_alloc(0); | 51 | gi = groups_alloc(0); |
52 | if (!gi) | ||
53 | goto oom; | ||
52 | } else if (flags & NFSEXP_ROOTSQUASH) { | 54 | } else if (flags & NFSEXP_ROOTSQUASH) { |
53 | if (!new->fsuid) | 55 | if (!new->fsuid) |
54 | new->fsuid = exp->ex_anon_uid; | 56 | new->fsuid = exp->ex_anon_uid; |
@@ -85,6 +87,7 @@ int nfsd_setuser(struct svc_rqst *rqstp, struct svc_export *exp) | |||
85 | new->cap_effective = cap_raise_nfsd_set(new->cap_effective, | 87 | new->cap_effective = cap_raise_nfsd_set(new->cap_effective, |
86 | new->cap_permitted); | 88 | new->cap_permitted); |
87 | put_cred(override_creds(new)); | 89 | put_cred(override_creds(new)); |
90 | put_cred(new); | ||
88 | return 0; | 91 | return 0; |
89 | 92 | ||
90 | oom: | 93 | oom: |
diff --git a/fs/nfsd/nfs4state.c b/fs/nfsd/nfs4state.c index 88db7d3ec120..b6f60f48e94b 100644 --- a/fs/nfsd/nfs4state.c +++ b/fs/nfsd/nfs4state.c | |||
@@ -2871,7 +2871,6 @@ nfsd4_lockt(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate, | |||
2871 | file_lock.fl_owner = (fl_owner_t)lockt->lt_stateowner; | 2871 | file_lock.fl_owner = (fl_owner_t)lockt->lt_stateowner; |
2872 | file_lock.fl_pid = current->tgid; | 2872 | file_lock.fl_pid = current->tgid; |
2873 | file_lock.fl_flags = FL_POSIX; | 2873 | file_lock.fl_flags = FL_POSIX; |
2874 | file_lock.fl_lmops = &nfsd_posix_mng_ops; | ||
2875 | 2874 | ||
2876 | file_lock.fl_start = lockt->lt_offset; | 2875 | file_lock.fl_start = lockt->lt_offset; |
2877 | file_lock.fl_end = last_byte_offset(lockt->lt_offset, lockt->lt_length); | 2876 | file_lock.fl_end = last_byte_offset(lockt->lt_offset, lockt->lt_length); |
diff --git a/fs/notify/inotify/inotify.c b/fs/notify/inotify/inotify.c index dae3f28f30d4..331f2e88e284 100644 --- a/fs/notify/inotify/inotify.c +++ b/fs/notify/inotify/inotify.c | |||
@@ -156,7 +156,7 @@ static int inotify_handle_get_wd(struct inotify_handle *ih, | |||
156 | int ret; | 156 | int ret; |
157 | 157 | ||
158 | do { | 158 | do { |
159 | if (unlikely(!idr_pre_get(&ih->idr, GFP_KERNEL))) | 159 | if (unlikely(!idr_pre_get(&ih->idr, GFP_NOFS))) |
160 | return -ENOSPC; | 160 | return -ENOSPC; |
161 | ret = idr_get_new_above(&ih->idr, watch, ih->last_wd+1, &watch->wd); | 161 | ret = idr_get_new_above(&ih->idr, watch, ih->last_wd+1, &watch->wd); |
162 | } while (ret == -EAGAIN); | 162 | } while (ret == -EAGAIN); |
diff --git a/fs/notify/inotify/inotify_user.c b/fs/notify/inotify/inotify_user.c index 81b8644b0136..bed766e435b5 100644 --- a/fs/notify/inotify/inotify_user.c +++ b/fs/notify/inotify/inotify_user.c | |||
@@ -427,10 +427,61 @@ static unsigned int inotify_poll(struct file *file, poll_table *wait) | |||
427 | return ret; | 427 | return ret; |
428 | } | 428 | } |
429 | 429 | ||
430 | /* | ||
431 | * Get an inotify_kernel_event if one exists and is small | ||
432 | * enough to fit in "count". Return an error pointer if | ||
433 | * not large enough. | ||
434 | * | ||
435 | * Called with the device ev_mutex held. | ||
436 | */ | ||
437 | static struct inotify_kernel_event *get_one_event(struct inotify_device *dev, | ||
438 | size_t count) | ||
439 | { | ||
440 | size_t event_size = sizeof(struct inotify_event); | ||
441 | struct inotify_kernel_event *kevent; | ||
442 | |||
443 | if (list_empty(&dev->events)) | ||
444 | return NULL; | ||
445 | |||
446 | kevent = inotify_dev_get_event(dev); | ||
447 | if (kevent->name) | ||
448 | event_size += kevent->event.len; | ||
449 | |||
450 | if (event_size > count) | ||
451 | return ERR_PTR(-EINVAL); | ||
452 | |||
453 | remove_kevent(dev, kevent); | ||
454 | return kevent; | ||
455 | } | ||
456 | |||
457 | /* | ||
458 | * Copy an event to user space, returning how much we copied. | ||
459 | * | ||
460 | * We already checked that the event size is smaller than the | ||
461 | * buffer we had in "get_one_event()" above. | ||
462 | */ | ||
463 | static ssize_t copy_event_to_user(struct inotify_kernel_event *kevent, | ||
464 | char __user *buf) | ||
465 | { | ||
466 | size_t event_size = sizeof(struct inotify_event); | ||
467 | |||
468 | if (copy_to_user(buf, &kevent->event, event_size)) | ||
469 | return -EFAULT; | ||
470 | |||
471 | if (kevent->name) { | ||
472 | buf += event_size; | ||
473 | |||
474 | if (copy_to_user(buf, kevent->name, kevent->event.len)) | ||
475 | return -EFAULT; | ||
476 | |||
477 | event_size += kevent->event.len; | ||
478 | } | ||
479 | return event_size; | ||
480 | } | ||
481 | |||
430 | static ssize_t inotify_read(struct file *file, char __user *buf, | 482 | static ssize_t inotify_read(struct file *file, char __user *buf, |
431 | size_t count, loff_t *pos) | 483 | size_t count, loff_t *pos) |
432 | { | 484 | { |
433 | size_t event_size = sizeof (struct inotify_event); | ||
434 | struct inotify_device *dev; | 485 | struct inotify_device *dev; |
435 | char __user *start; | 486 | char __user *start; |
436 | int ret; | 487 | int ret; |
@@ -440,81 +491,43 @@ static ssize_t inotify_read(struct file *file, char __user *buf, | |||
440 | dev = file->private_data; | 491 | dev = file->private_data; |
441 | 492 | ||
442 | while (1) { | 493 | while (1) { |
494 | struct inotify_kernel_event *kevent; | ||
443 | 495 | ||
444 | prepare_to_wait(&dev->wq, &wait, TASK_INTERRUPTIBLE); | 496 | prepare_to_wait(&dev->wq, &wait, TASK_INTERRUPTIBLE); |
445 | 497 | ||
446 | mutex_lock(&dev->ev_mutex); | 498 | mutex_lock(&dev->ev_mutex); |
447 | if (!list_empty(&dev->events)) { | 499 | kevent = get_one_event(dev, count); |
448 | ret = 0; | ||
449 | break; | ||
450 | } | ||
451 | mutex_unlock(&dev->ev_mutex); | 500 | mutex_unlock(&dev->ev_mutex); |
452 | 501 | ||
453 | if (file->f_flags & O_NONBLOCK) { | 502 | if (kevent) { |
454 | ret = -EAGAIN; | 503 | ret = PTR_ERR(kevent); |
455 | break; | 504 | if (IS_ERR(kevent)) |
456 | } | 505 | break; |
457 | 506 | ret = copy_event_to_user(kevent, buf); | |
458 | if (signal_pending(current)) { | 507 | free_kevent(kevent); |
459 | ret = -EINTR; | 508 | if (ret < 0) |
460 | break; | 509 | break; |
510 | buf += ret; | ||
511 | count -= ret; | ||
512 | continue; | ||
461 | } | 513 | } |
462 | 514 | ||
463 | schedule(); | 515 | ret = -EAGAIN; |
464 | } | 516 | if (file->f_flags & O_NONBLOCK) |
465 | |||
466 | finish_wait(&dev->wq, &wait); | ||
467 | if (ret) | ||
468 | return ret; | ||
469 | |||
470 | while (1) { | ||
471 | struct inotify_kernel_event *kevent; | ||
472 | |||
473 | ret = buf - start; | ||
474 | if (list_empty(&dev->events)) | ||
475 | break; | 517 | break; |
476 | 518 | ret = -EINTR; | |
477 | kevent = inotify_dev_get_event(dev); | 519 | if (signal_pending(current)) |
478 | if (event_size + kevent->event.len > count) { | ||
479 | if (ret == 0 && count > 0) { | ||
480 | /* | ||
481 | * could not get a single event because we | ||
482 | * didn't have enough buffer space. | ||
483 | */ | ||
484 | ret = -EINVAL; | ||
485 | } | ||
486 | break; | 520 | break; |
487 | } | ||
488 | remove_kevent(dev, kevent); | ||
489 | 521 | ||
490 | /* | 522 | if (start != buf) |
491 | * Must perform the copy_to_user outside the mutex in order | ||
492 | * to avoid a lock order reversal with mmap_sem. | ||
493 | */ | ||
494 | mutex_unlock(&dev->ev_mutex); | ||
495 | |||
496 | if (copy_to_user(buf, &kevent->event, event_size)) { | ||
497 | ret = -EFAULT; | ||
498 | break; | 523 | break; |
499 | } | ||
500 | buf += event_size; | ||
501 | count -= event_size; | ||
502 | |||
503 | if (kevent->name) { | ||
504 | if (copy_to_user(buf, kevent->name, kevent->event.len)){ | ||
505 | ret = -EFAULT; | ||
506 | break; | ||
507 | } | ||
508 | buf += kevent->event.len; | ||
509 | count -= kevent->event.len; | ||
510 | } | ||
511 | |||
512 | free_kevent(kevent); | ||
513 | 524 | ||
514 | mutex_lock(&dev->ev_mutex); | 525 | schedule(); |
515 | } | 526 | } |
516 | mutex_unlock(&dev->ev_mutex); | ||
517 | 527 | ||
528 | finish_wait(&dev->wq, &wait); | ||
529 | if (start != buf && ret != -EFAULT) | ||
530 | ret = buf - start; | ||
518 | return ret; | 531 | return ret; |
519 | } | 532 | } |
520 | 533 | ||
@@ -576,7 +589,7 @@ static const struct inotify_operations inotify_user_ops = { | |||
576 | .destroy_watch = free_inotify_user_watch, | 589 | .destroy_watch = free_inotify_user_watch, |
577 | }; | 590 | }; |
578 | 591 | ||
579 | asmlinkage long sys_inotify_init1(int flags) | 592 | SYSCALL_DEFINE1(inotify_init1, int, flags) |
580 | { | 593 | { |
581 | struct inotify_device *dev; | 594 | struct inotify_device *dev; |
582 | struct inotify_handle *ih; | 595 | struct inotify_handle *ih; |
@@ -655,12 +668,13 @@ out_put_fd: | |||
655 | return ret; | 668 | return ret; |
656 | } | 669 | } |
657 | 670 | ||
658 | asmlinkage long sys_inotify_init(void) | 671 | SYSCALL_DEFINE0(inotify_init) |
659 | { | 672 | { |
660 | return sys_inotify_init1(0); | 673 | return sys_inotify_init1(0); |
661 | } | 674 | } |
662 | 675 | ||
663 | asmlinkage long sys_inotify_add_watch(int fd, const char __user *pathname, u32 mask) | 676 | SYSCALL_DEFINE3(inotify_add_watch, int, fd, const char __user *, pathname, |
677 | u32, mask) | ||
664 | { | 678 | { |
665 | struct inode *inode; | 679 | struct inode *inode; |
666 | struct inotify_device *dev; | 680 | struct inotify_device *dev; |
@@ -704,7 +718,7 @@ fput_and_out: | |||
704 | return ret; | 718 | return ret; |
705 | } | 719 | } |
706 | 720 | ||
707 | asmlinkage long sys_inotify_rm_watch(int fd, __s32 wd) | 721 | SYSCALL_DEFINE2(inotify_rm_watch, int, fd, __s32, wd) |
708 | { | 722 | { |
709 | struct file *filp; | 723 | struct file *filp; |
710 | struct inotify_device *dev; | 724 | struct inotify_device *dev; |
diff --git a/fs/ntfs/Kconfig b/fs/ntfs/Kconfig new file mode 100644 index 000000000000..f5a868cc9152 --- /dev/null +++ b/fs/ntfs/Kconfig | |||
@@ -0,0 +1,78 @@ | |||
1 | config NTFS_FS | ||
2 | tristate "NTFS file system support" | ||
3 | select NLS | ||
4 | help | ||
5 | NTFS is the file system of Microsoft Windows NT, 2000, XP and 2003. | ||
6 | |||
7 | Saying Y or M here enables read support. There is partial, but | ||
8 | safe, write support available. For write support you must also | ||
9 | say Y to "NTFS write support" below. | ||
10 | |||
11 | There are also a number of user-space tools available, called | ||
12 | ntfsprogs. These include ntfsundelete and ntfsresize, that work | ||
13 | without NTFS support enabled in the kernel. | ||
14 | |||
15 | This is a rewrite from scratch of Linux NTFS support and replaced | ||
16 | the old NTFS code starting with Linux 2.5.11. A backport to | ||
17 | the Linux 2.4 kernel series is separately available as a patch | ||
18 | from the project web site. | ||
19 | |||
20 | For more information see <file:Documentation/filesystems/ntfs.txt> | ||
21 | and <http://www.linux-ntfs.org/>. | ||
22 | |||
23 | To compile this file system support as a module, choose M here: the | ||
24 | module will be called ntfs. | ||
25 | |||
26 | If you are not using Windows NT, 2000, XP or 2003 in addition to | ||
27 | Linux on your computer it is safe to say N. | ||
28 | |||
29 | config NTFS_DEBUG | ||
30 | bool "NTFS debugging support" | ||
31 | depends on NTFS_FS | ||
32 | help | ||
33 | If you are experiencing any problems with the NTFS file system, say | ||
34 | Y here. This will result in additional consistency checks to be | ||
35 | performed by the driver as well as additional debugging messages to | ||
36 | be written to the system log. Note that debugging messages are | ||
37 | disabled by default. To enable them, supply the option debug_msgs=1 | ||
38 | at the kernel command line when booting the kernel or as an option | ||
39 | to insmod when loading the ntfs module. Once the driver is active, | ||
40 | you can enable debugging messages by doing (as root): | ||
41 | echo 1 > /proc/sys/fs/ntfs-debug | ||
42 | Replacing the "1" with "0" would disable debug messages. | ||
43 | |||
44 | If you leave debugging messages disabled, this results in little | ||
45 | overhead, but enabling debug messages results in very significant | ||
46 | slowdown of the system. | ||
47 | |||
48 | When reporting bugs, please try to have available a full dump of | ||
49 | debugging messages while the misbehaviour was occurring. | ||
50 | |||
51 | config NTFS_RW | ||
52 | bool "NTFS write support" | ||
53 | depends on NTFS_FS | ||
54 | help | ||
55 | This enables the partial, but safe, write support in the NTFS driver. | ||
56 | |||
57 | The only supported operation is overwriting existing files, without | ||
58 | changing the file length. No file or directory creation, deletion or | ||
59 | renaming is possible. Note only non-resident files can be written to | ||
60 | so you may find that some very small files (<500 bytes or so) cannot | ||
61 | be written to. | ||
62 | |||
63 | While we cannot guarantee that it will not damage any data, we have | ||
64 | so far not received a single report where the driver would have | ||
65 | damaged someones data so we assume it is perfectly safe to use. | ||
66 | |||
67 | Note: While write support is safe in this version (a rewrite from | ||
68 | scratch of the NTFS support), it should be noted that the old NTFS | ||
69 | write support, included in Linux 2.5.10 and before (since 1997), | ||
70 | is not safe. | ||
71 | |||
72 | This is currently useful with TopologiLinux. TopologiLinux is run | ||
73 | on top of any DOS/Microsoft Windows system without partitioning your | ||
74 | hard disk. Unlike other Linux distributions TopologiLinux does not | ||
75 | need its own partition. For more information see | ||
76 | <http://topologi-linux.sourceforge.net/> | ||
77 | |||
78 | It is perfectly safe to say N here. | ||
diff --git a/fs/ocfs2/Kconfig b/fs/ocfs2/Kconfig new file mode 100644 index 000000000000..701b7a3a872e --- /dev/null +++ b/fs/ocfs2/Kconfig | |||
@@ -0,0 +1,85 @@ | |||
1 | config OCFS2_FS | ||
2 | tristate "OCFS2 file system support" | ||
3 | depends on NET && SYSFS | ||
4 | select CONFIGFS_FS | ||
5 | select JBD2 | ||
6 | select CRC32 | ||
7 | select QUOTA | ||
8 | select QUOTA_TREE | ||
9 | help | ||
10 | OCFS2 is a general purpose extent based shared disk cluster file | ||
11 | system with many similarities to ext3. It supports 64 bit inode | ||
12 | numbers, and has automatically extending metadata groups which may | ||
13 | also make it attractive for non-clustered use. | ||
14 | |||
15 | You'll want to install the ocfs2-tools package in order to at least | ||
16 | get "mount.ocfs2". | ||
17 | |||
18 | Project web page: http://oss.oracle.com/projects/ocfs2 | ||
19 | Tools web page: http://oss.oracle.com/projects/ocfs2-tools | ||
20 | OCFS2 mailing lists: http://oss.oracle.com/projects/ocfs2/mailman/ | ||
21 | |||
22 | For more information on OCFS2, see the file | ||
23 | <file:Documentation/filesystems/ocfs2.txt>. | ||
24 | |||
25 | config OCFS2_FS_O2CB | ||
26 | tristate "O2CB Kernelspace Clustering" | ||
27 | depends on OCFS2_FS | ||
28 | default y | ||
29 | help | ||
30 | OCFS2 includes a simple kernelspace clustering package, the OCFS2 | ||
31 | Cluster Base. It only requires a very small userspace component | ||
32 | to configure it. This comes with the standard ocfs2-tools package. | ||
33 | O2CB is limited to maintaining a cluster for OCFS2 file systems. | ||
34 | It cannot manage any other cluster applications. | ||
35 | |||
36 | It is always safe to say Y here, as the clustering method is | ||
37 | run-time selectable. | ||
38 | |||
39 | config OCFS2_FS_USERSPACE_CLUSTER | ||
40 | tristate "OCFS2 Userspace Clustering" | ||
41 | depends on OCFS2_FS && DLM | ||
42 | default y | ||
43 | help | ||
44 | This option will allow OCFS2 to use userspace clustering services | ||
45 | in conjunction with the DLM in fs/dlm. If you are using a | ||
46 | userspace cluster manager, say Y here. | ||
47 | |||
48 | It is safe to say Y, as the clustering method is run-time | ||
49 | selectable. | ||
50 | |||
51 | config OCFS2_FS_STATS | ||
52 | bool "OCFS2 statistics" | ||
53 | depends on OCFS2_FS | ||
54 | default y | ||
55 | help | ||
56 | This option allows some fs statistics to be captured. Enabling | ||
57 | this option may increase the memory consumption. | ||
58 | |||
59 | config OCFS2_DEBUG_MASKLOG | ||
60 | bool "OCFS2 logging support" | ||
61 | depends on OCFS2_FS | ||
62 | default y | ||
63 | help | ||
64 | The ocfs2 filesystem has an extensive logging system. The system | ||
65 | allows selection of events to log via files in /sys/o2cb/logmask/. | ||
66 | This option will enlarge your kernel, but it allows debugging of | ||
67 | ocfs2 filesystem issues. | ||
68 | |||
69 | config OCFS2_DEBUG_FS | ||
70 | bool "OCFS2 expensive checks" | ||
71 | depends on OCFS2_FS | ||
72 | default n | ||
73 | help | ||
74 | This option will enable expensive consistency checks. Enable | ||
75 | this option for debugging only as it is likely to decrease | ||
76 | performance of the filesystem. | ||
77 | |||
78 | config OCFS2_FS_POSIX_ACL | ||
79 | bool "OCFS2 POSIX Access Control Lists" | ||
80 | depends on OCFS2_FS | ||
81 | select FS_POSIX_ACL | ||
82 | default n | ||
83 | help | ||
84 | Posix Access Control Lists (ACLs) support permissions for users and | ||
85 | groups beyond the owner/group/world scheme. | ||
diff --git a/fs/ocfs2/alloc.c b/fs/ocfs2/alloc.c index d861096c9d81..3a9e5deed74d 100644 --- a/fs/ocfs2/alloc.c +++ b/fs/ocfs2/alloc.c | |||
@@ -4796,6 +4796,29 @@ out: | |||
4796 | return ret; | 4796 | return ret; |
4797 | } | 4797 | } |
4798 | 4798 | ||
4799 | static int ocfs2_replace_extent_rec(struct inode *inode, | ||
4800 | handle_t *handle, | ||
4801 | struct ocfs2_path *path, | ||
4802 | struct ocfs2_extent_list *el, | ||
4803 | int split_index, | ||
4804 | struct ocfs2_extent_rec *split_rec) | ||
4805 | { | ||
4806 | int ret; | ||
4807 | |||
4808 | ret = ocfs2_path_bh_journal_access(handle, inode, path, | ||
4809 | path_num_items(path) - 1); | ||
4810 | if (ret) { | ||
4811 | mlog_errno(ret); | ||
4812 | goto out; | ||
4813 | } | ||
4814 | |||
4815 | el->l_recs[split_index] = *split_rec; | ||
4816 | |||
4817 | ocfs2_journal_dirty(handle, path_leaf_bh(path)); | ||
4818 | out: | ||
4819 | return ret; | ||
4820 | } | ||
4821 | |||
4799 | /* | 4822 | /* |
4800 | * Mark part or all of the extent record at split_index in the leaf | 4823 | * Mark part or all of the extent record at split_index in the leaf |
4801 | * pointed to by path as written. This removes the unwritten | 4824 | * pointed to by path as written. This removes the unwritten |
@@ -4885,7 +4908,9 @@ static int __ocfs2_mark_extent_written(struct inode *inode, | |||
4885 | 4908 | ||
4886 | if (ctxt.c_contig_type == CONTIG_NONE) { | 4909 | if (ctxt.c_contig_type == CONTIG_NONE) { |
4887 | if (ctxt.c_split_covers_rec) | 4910 | if (ctxt.c_split_covers_rec) |
4888 | el->l_recs[split_index] = *split_rec; | 4911 | ret = ocfs2_replace_extent_rec(inode, handle, |
4912 | path, el, | ||
4913 | split_index, split_rec); | ||
4889 | else | 4914 | else |
4890 | ret = ocfs2_split_and_insert(inode, handle, path, et, | 4915 | ret = ocfs2_split_and_insert(inode, handle, path, et, |
4891 | &last_eb_bh, split_index, | 4916 | &last_eb_bh, split_index, |
@@ -5390,6 +5415,9 @@ int ocfs2_remove_btree_range(struct inode *inode, | |||
5390 | goto out; | 5415 | goto out; |
5391 | } | 5416 | } |
5392 | 5417 | ||
5418 | vfs_dq_free_space_nodirty(inode, | ||
5419 | ocfs2_clusters_to_bytes(inode->i_sb, len)); | ||
5420 | |||
5393 | ret = ocfs2_remove_extent(inode, et, cpos, len, handle, meta_ac, | 5421 | ret = ocfs2_remove_extent(inode, et, cpos, len, handle, meta_ac, |
5394 | dealloc); | 5422 | dealloc); |
5395 | if (ret) { | 5423 | if (ret) { |
diff --git a/fs/ocfs2/dcache.c b/fs/ocfs2/dcache.c index b1cc7c381e88..e9d7c2038c0f 100644 --- a/fs/ocfs2/dcache.c +++ b/fs/ocfs2/dcache.c | |||
@@ -38,6 +38,7 @@ | |||
38 | #include "dlmglue.h" | 38 | #include "dlmglue.h" |
39 | #include "file.h" | 39 | #include "file.h" |
40 | #include "inode.h" | 40 | #include "inode.h" |
41 | #include "super.h" | ||
41 | 42 | ||
42 | 43 | ||
43 | static int ocfs2_dentry_revalidate(struct dentry *dentry, | 44 | static int ocfs2_dentry_revalidate(struct dentry *dentry, |
@@ -294,6 +295,34 @@ out_attach: | |||
294 | return ret; | 295 | return ret; |
295 | } | 296 | } |
296 | 297 | ||
298 | static DEFINE_SPINLOCK(dentry_list_lock); | ||
299 | |||
300 | /* We limit the number of dentry locks to drop in one go. We have | ||
301 | * this limit so that we don't starve other users of ocfs2_wq. */ | ||
302 | #define DL_INODE_DROP_COUNT 64 | ||
303 | |||
304 | /* Drop inode references from dentry locks */ | ||
305 | void ocfs2_drop_dl_inodes(struct work_struct *work) | ||
306 | { | ||
307 | struct ocfs2_super *osb = container_of(work, struct ocfs2_super, | ||
308 | dentry_lock_work); | ||
309 | struct ocfs2_dentry_lock *dl; | ||
310 | int drop_count = DL_INODE_DROP_COUNT; | ||
311 | |||
312 | spin_lock(&dentry_list_lock); | ||
313 | while (osb->dentry_lock_list && drop_count--) { | ||
314 | dl = osb->dentry_lock_list; | ||
315 | osb->dentry_lock_list = dl->dl_next; | ||
316 | spin_unlock(&dentry_list_lock); | ||
317 | iput(dl->dl_inode); | ||
318 | kfree(dl); | ||
319 | spin_lock(&dentry_list_lock); | ||
320 | } | ||
321 | if (osb->dentry_lock_list) | ||
322 | queue_work(ocfs2_wq, &osb->dentry_lock_work); | ||
323 | spin_unlock(&dentry_list_lock); | ||
324 | } | ||
325 | |||
297 | /* | 326 | /* |
298 | * ocfs2_dentry_iput() and friends. | 327 | * ocfs2_dentry_iput() and friends. |
299 | * | 328 | * |
@@ -318,16 +347,23 @@ out_attach: | |||
318 | static void ocfs2_drop_dentry_lock(struct ocfs2_super *osb, | 347 | static void ocfs2_drop_dentry_lock(struct ocfs2_super *osb, |
319 | struct ocfs2_dentry_lock *dl) | 348 | struct ocfs2_dentry_lock *dl) |
320 | { | 349 | { |
321 | iput(dl->dl_inode); | ||
322 | ocfs2_simple_drop_lockres(osb, &dl->dl_lockres); | 350 | ocfs2_simple_drop_lockres(osb, &dl->dl_lockres); |
323 | ocfs2_lock_res_free(&dl->dl_lockres); | 351 | ocfs2_lock_res_free(&dl->dl_lockres); |
324 | kfree(dl); | 352 | |
353 | /* We leave dropping of inode reference to ocfs2_wq as that can | ||
354 | * possibly lead to inode deletion which gets tricky */ | ||
355 | spin_lock(&dentry_list_lock); | ||
356 | if (!osb->dentry_lock_list) | ||
357 | queue_work(ocfs2_wq, &osb->dentry_lock_work); | ||
358 | dl->dl_next = osb->dentry_lock_list; | ||
359 | osb->dentry_lock_list = dl; | ||
360 | spin_unlock(&dentry_list_lock); | ||
325 | } | 361 | } |
326 | 362 | ||
327 | void ocfs2_dentry_lock_put(struct ocfs2_super *osb, | 363 | void ocfs2_dentry_lock_put(struct ocfs2_super *osb, |
328 | struct ocfs2_dentry_lock *dl) | 364 | struct ocfs2_dentry_lock *dl) |
329 | { | 365 | { |
330 | int unlock = 0; | 366 | int unlock; |
331 | 367 | ||
332 | BUG_ON(dl->dl_count == 0); | 368 | BUG_ON(dl->dl_count == 0); |
333 | 369 | ||
diff --git a/fs/ocfs2/dcache.h b/fs/ocfs2/dcache.h index c091c34d9883..d06e16c06640 100644 --- a/fs/ocfs2/dcache.h +++ b/fs/ocfs2/dcache.h | |||
@@ -29,8 +29,13 @@ | |||
29 | extern struct dentry_operations ocfs2_dentry_ops; | 29 | extern struct dentry_operations ocfs2_dentry_ops; |
30 | 30 | ||
31 | struct ocfs2_dentry_lock { | 31 | struct ocfs2_dentry_lock { |
32 | /* Use count of dentry lock */ | ||
32 | unsigned int dl_count; | 33 | unsigned int dl_count; |
33 | u64 dl_parent_blkno; | 34 | union { |
35 | /* Linked list of dentry locks to release */ | ||
36 | struct ocfs2_dentry_lock *dl_next; | ||
37 | u64 dl_parent_blkno; | ||
38 | }; | ||
34 | 39 | ||
35 | /* | 40 | /* |
36 | * The ocfs2_dentry_lock keeps an inode reference until | 41 | * The ocfs2_dentry_lock keeps an inode reference until |
@@ -47,6 +52,8 @@ int ocfs2_dentry_attach_lock(struct dentry *dentry, struct inode *inode, | |||
47 | void ocfs2_dentry_lock_put(struct ocfs2_super *osb, | 52 | void ocfs2_dentry_lock_put(struct ocfs2_super *osb, |
48 | struct ocfs2_dentry_lock *dl); | 53 | struct ocfs2_dentry_lock *dl); |
49 | 54 | ||
55 | void ocfs2_drop_dl_inodes(struct work_struct *work); | ||
56 | |||
50 | struct dentry *ocfs2_find_local_alias(struct inode *inode, u64 parent_blkno, | 57 | struct dentry *ocfs2_find_local_alias(struct inode *inode, u64 parent_blkno, |
51 | int skip_unhashed); | 58 | int skip_unhashed); |
52 | 59 | ||
diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c index 54e182a27caf..0a2813947853 100644 --- a/fs/ocfs2/dlm/dlmmaster.c +++ b/fs/ocfs2/dlm/dlmmaster.c | |||
@@ -1849,12 +1849,12 @@ int dlm_assert_master_handler(struct o2net_msg *msg, u32 len, void *data, | |||
1849 | if (!mle) { | 1849 | if (!mle) { |
1850 | if (res->owner != DLM_LOCK_RES_OWNER_UNKNOWN && | 1850 | if (res->owner != DLM_LOCK_RES_OWNER_UNKNOWN && |
1851 | res->owner != assert->node_idx) { | 1851 | res->owner != assert->node_idx) { |
1852 | mlog(ML_ERROR, "assert_master from " | 1852 | mlog(ML_ERROR, "DIE! Mastery assert from %u, " |
1853 | "%u, but current owner is " | 1853 | "but current owner is %u! (%.*s)\n", |
1854 | "%u! (%.*s)\n", | 1854 | assert->node_idx, res->owner, namelen, |
1855 | assert->node_idx, res->owner, | 1855 | name); |
1856 | namelen, name); | 1856 | __dlm_print_one_lock_resource(res); |
1857 | goto kill; | 1857 | BUG(); |
1858 | } | 1858 | } |
1859 | } else if (mle->type != DLM_MLE_MIGRATION) { | 1859 | } else if (mle->type != DLM_MLE_MIGRATION) { |
1860 | if (res->owner != DLM_LOCK_RES_OWNER_UNKNOWN) { | 1860 | if (res->owner != DLM_LOCK_RES_OWNER_UNKNOWN) { |
diff --git a/fs/ocfs2/dlm/dlmthread.c b/fs/ocfs2/dlm/dlmthread.c index d1295203029f..4060bb328bc8 100644 --- a/fs/ocfs2/dlm/dlmthread.c +++ b/fs/ocfs2/dlm/dlmthread.c | |||
@@ -181,8 +181,7 @@ static int dlm_purge_lockres(struct dlm_ctxt *dlm, | |||
181 | 181 | ||
182 | spin_lock(&res->spinlock); | 182 | spin_lock(&res->spinlock); |
183 | /* This ensures that clear refmap is sent after the set */ | 183 | /* This ensures that clear refmap is sent after the set */ |
184 | __dlm_wait_on_lockres_flags(res, (DLM_LOCK_RES_SETREF_INPROG | | 184 | __dlm_wait_on_lockres_flags(res, DLM_LOCK_RES_SETREF_INPROG); |
185 | DLM_LOCK_RES_MIGRATING)); | ||
186 | spin_unlock(&res->spinlock); | 185 | spin_unlock(&res->spinlock); |
187 | 186 | ||
188 | /* clear our bit from the master's refmap, ignore errors */ | 187 | /* clear our bit from the master's refmap, ignore errors */ |
diff --git a/fs/ocfs2/dlm/dlmunlock.c b/fs/ocfs2/dlm/dlmunlock.c index 86ca085ef324..fcf879ed6930 100644 --- a/fs/ocfs2/dlm/dlmunlock.c +++ b/fs/ocfs2/dlm/dlmunlock.c | |||
@@ -117,11 +117,11 @@ static enum dlm_status dlmunlock_common(struct dlm_ctxt *dlm, | |||
117 | else | 117 | else |
118 | BUG_ON(res->owner == dlm->node_num); | 118 | BUG_ON(res->owner == dlm->node_num); |
119 | 119 | ||
120 | spin_lock(&dlm->spinlock); | 120 | spin_lock(&dlm->ast_lock); |
121 | /* We want to be sure that we're not freeing a lock | 121 | /* We want to be sure that we're not freeing a lock |
122 | * that still has AST's pending... */ | 122 | * that still has AST's pending... */ |
123 | in_use = !list_empty(&lock->ast_list); | 123 | in_use = !list_empty(&lock->ast_list); |
124 | spin_unlock(&dlm->spinlock); | 124 | spin_unlock(&dlm->ast_lock); |
125 | if (in_use) { | 125 | if (in_use) { |
126 | mlog(ML_ERROR, "lockres %.*s: Someone is calling dlmunlock " | 126 | mlog(ML_ERROR, "lockres %.*s: Someone is calling dlmunlock " |
127 | "while waiting for an ast!", res->lockname.len, | 127 | "while waiting for an ast!", res->lockname.len, |
diff --git a/fs/ocfs2/dlmglue.c b/fs/ocfs2/dlmglue.c index b0c4cadd4c45..7219a86d34cc 100644 --- a/fs/ocfs2/dlmglue.c +++ b/fs/ocfs2/dlmglue.c | |||
@@ -320,9 +320,14 @@ static void ocfs2_schedule_blocked_lock(struct ocfs2_super *osb, | |||
320 | struct ocfs2_lock_res *lockres); | 320 | struct ocfs2_lock_res *lockres); |
321 | static inline void ocfs2_recover_from_dlm_error(struct ocfs2_lock_res *lockres, | 321 | static inline void ocfs2_recover_from_dlm_error(struct ocfs2_lock_res *lockres, |
322 | int convert); | 322 | int convert); |
323 | #define ocfs2_log_dlm_error(_func, _err, _lockres) do { \ | 323 | #define ocfs2_log_dlm_error(_func, _err, _lockres) do { \ |
324 | mlog(ML_ERROR, "DLM error %d while calling %s on resource %s\n", \ | 324 | if ((_lockres)->l_type != OCFS2_LOCK_TYPE_DENTRY) \ |
325 | _err, _func, _lockres->l_name); \ | 325 | mlog(ML_ERROR, "DLM error %d while calling %s on resource %s\n", \ |
326 | _err, _func, _lockres->l_name); \ | ||
327 | else \ | ||
328 | mlog(ML_ERROR, "DLM error %d while calling %s on resource %.*s%08x\n", \ | ||
329 | _err, _func, OCFS2_DENTRY_LOCK_INO_START - 1, (_lockres)->l_name, \ | ||
330 | (unsigned int)ocfs2_get_dentry_lock_ino(_lockres)); \ | ||
326 | } while (0) | 331 | } while (0) |
327 | static int ocfs2_downconvert_thread(void *arg); | 332 | static int ocfs2_downconvert_thread(void *arg); |
328 | static void ocfs2_downconvert_on_unlock(struct ocfs2_super *osb, | 333 | static void ocfs2_downconvert_on_unlock(struct ocfs2_super *osb, |
@@ -2860,6 +2865,10 @@ static void ocfs2_unlock_ast(void *opaque, int error) | |||
2860 | case OCFS2_UNLOCK_CANCEL_CONVERT: | 2865 | case OCFS2_UNLOCK_CANCEL_CONVERT: |
2861 | mlog(0, "Cancel convert success for %s\n", lockres->l_name); | 2866 | mlog(0, "Cancel convert success for %s\n", lockres->l_name); |
2862 | lockres->l_action = OCFS2_AST_INVALID; | 2867 | lockres->l_action = OCFS2_AST_INVALID; |
2868 | /* Downconvert thread may have requeued this lock, we | ||
2869 | * need to wake it. */ | ||
2870 | if (lockres->l_flags & OCFS2_LOCK_BLOCKED) | ||
2871 | ocfs2_wake_downconvert_thread(ocfs2_get_lockres_osb(lockres)); | ||
2863 | break; | 2872 | break; |
2864 | case OCFS2_UNLOCK_DROP_LOCK: | 2873 | case OCFS2_UNLOCK_DROP_LOCK: |
2865 | lockres->l_level = DLM_LOCK_IV; | 2874 | lockres->l_level = DLM_LOCK_IV; |
diff --git a/fs/ocfs2/journal.h b/fs/ocfs2/journal.h index 3c3532e1307c..172850a9a12a 100644 --- a/fs/ocfs2/journal.h +++ b/fs/ocfs2/journal.h | |||
@@ -513,8 +513,10 @@ static inline int ocfs2_jbd2_file_inode(handle_t *handle, struct inode *inode) | |||
513 | static inline int ocfs2_begin_ordered_truncate(struct inode *inode, | 513 | static inline int ocfs2_begin_ordered_truncate(struct inode *inode, |
514 | loff_t new_size) | 514 | loff_t new_size) |
515 | { | 515 | { |
516 | return jbd2_journal_begin_ordered_truncate(&OCFS2_I(inode)->ip_jinode, | 516 | return jbd2_journal_begin_ordered_truncate( |
517 | new_size); | 517 | OCFS2_SB(inode->i_sb)->journal->j_journal, |
518 | &OCFS2_I(inode)->ip_jinode, | ||
519 | new_size); | ||
518 | } | 520 | } |
519 | 521 | ||
520 | #endif /* OCFS2_JOURNAL_H */ | 522 | #endif /* OCFS2_JOURNAL_H */ |
diff --git a/fs/ocfs2/ocfs2.h b/fs/ocfs2/ocfs2.h index ad5c24a29edd..946d3c34b90b 100644 --- a/fs/ocfs2/ocfs2.h +++ b/fs/ocfs2/ocfs2.h | |||
@@ -210,6 +210,7 @@ struct ocfs2_journal; | |||
210 | struct ocfs2_slot_info; | 210 | struct ocfs2_slot_info; |
211 | struct ocfs2_recovery_map; | 211 | struct ocfs2_recovery_map; |
212 | struct ocfs2_quota_recovery; | 212 | struct ocfs2_quota_recovery; |
213 | struct ocfs2_dentry_lock; | ||
213 | struct ocfs2_super | 214 | struct ocfs2_super |
214 | { | 215 | { |
215 | struct task_struct *commit_task; | 216 | struct task_struct *commit_task; |
@@ -325,6 +326,11 @@ struct ocfs2_super | |||
325 | struct list_head blocked_lock_list; | 326 | struct list_head blocked_lock_list; |
326 | unsigned long blocked_lock_count; | 327 | unsigned long blocked_lock_count; |
327 | 328 | ||
329 | /* List of dentry locks to release. Anyone can add locks to | ||
330 | * the list, ocfs2_wq processes the list */ | ||
331 | struct ocfs2_dentry_lock *dentry_lock_list; | ||
332 | struct work_struct dentry_lock_work; | ||
333 | |||
328 | wait_queue_head_t osb_mount_event; | 334 | wait_queue_head_t osb_mount_event; |
329 | 335 | ||
330 | /* Truncate log info */ | 336 | /* Truncate log info */ |
@@ -335,6 +341,9 @@ struct ocfs2_super | |||
335 | struct ocfs2_node_map osb_recovering_orphan_dirs; | 341 | struct ocfs2_node_map osb_recovering_orphan_dirs; |
336 | unsigned int *osb_orphan_wipes; | 342 | unsigned int *osb_orphan_wipes; |
337 | wait_queue_head_t osb_wipe_event; | 343 | wait_queue_head_t osb_wipe_event; |
344 | |||
345 | /* used to protect metaecc calculation check of xattr. */ | ||
346 | spinlock_t osb_xattr_lock; | ||
338 | }; | 347 | }; |
339 | 348 | ||
340 | #define OCFS2_SB(sb) ((struct ocfs2_super *)(sb)->s_fs_info) | 349 | #define OCFS2_SB(sb) ((struct ocfs2_super *)(sb)->s_fs_info) |
diff --git a/fs/ocfs2/quota_global.c b/fs/ocfs2/quota_global.c index 6aff8f2d3e49..1ed0f7c86869 100644 --- a/fs/ocfs2/quota_global.c +++ b/fs/ocfs2/quota_global.c | |||
@@ -754,7 +754,9 @@ static int ocfs2_mark_dquot_dirty(struct dquot *dquot) | |||
754 | if (dquot->dq_flags & mask) | 754 | if (dquot->dq_flags & mask) |
755 | sync = 1; | 755 | sync = 1; |
756 | spin_unlock(&dq_data_lock); | 756 | spin_unlock(&dq_data_lock); |
757 | if (!sync) { | 757 | /* This is a slight hack but we can't afford getting global quota |
758 | * lock if we already have a transaction started. */ | ||
759 | if (!sync || journal_current_handle()) { | ||
758 | status = ocfs2_write_dquot(dquot); | 760 | status = ocfs2_write_dquot(dquot); |
759 | goto out; | 761 | goto out; |
760 | } | 762 | } |
@@ -810,171 +812,6 @@ out: | |||
810 | return status; | 812 | return status; |
811 | } | 813 | } |
812 | 814 | ||
813 | /* This is difficult. We have to lock quota inode and start transaction | ||
814 | * in this function but we don't want to take the penalty of exlusive | ||
815 | * quota file lock when we are just going to use cached structures. So | ||
816 | * we just take read lock check whether we have dquot cached and if so, | ||
817 | * we don't have to take the write lock... */ | ||
818 | static int ocfs2_dquot_initialize(struct inode *inode, int type) | ||
819 | { | ||
820 | handle_t *handle = NULL; | ||
821 | int status = 0; | ||
822 | struct super_block *sb = inode->i_sb; | ||
823 | struct ocfs2_mem_dqinfo *oinfo; | ||
824 | int exclusive = 0; | ||
825 | int cnt; | ||
826 | qid_t id; | ||
827 | |||
828 | mlog_entry_void(); | ||
829 | |||
830 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { | ||
831 | if (type != -1 && cnt != type) | ||
832 | continue; | ||
833 | if (!sb_has_quota_active(sb, cnt)) | ||
834 | continue; | ||
835 | oinfo = sb_dqinfo(sb, cnt)->dqi_priv; | ||
836 | status = ocfs2_lock_global_qf(oinfo, 0); | ||
837 | if (status < 0) | ||
838 | goto out; | ||
839 | /* This is just a performance optimization not a reliable test. | ||
840 | * Since we hold an inode lock, noone can actually release | ||
841 | * the structure until we are finished with initialization. */ | ||
842 | if (inode->i_dquot[cnt] != NODQUOT) { | ||
843 | ocfs2_unlock_global_qf(oinfo, 0); | ||
844 | continue; | ||
845 | } | ||
846 | /* When we have inode lock, we know that no dquot_release() can | ||
847 | * run and thus we can safely check whether we need to | ||
848 | * read+modify global file to get quota information or whether | ||
849 | * our node already has it. */ | ||
850 | if (cnt == USRQUOTA) | ||
851 | id = inode->i_uid; | ||
852 | else if (cnt == GRPQUOTA) | ||
853 | id = inode->i_gid; | ||
854 | else | ||
855 | BUG(); | ||
856 | /* Obtain exclusion from quota off... */ | ||
857 | down_write(&sb_dqopt(sb)->dqptr_sem); | ||
858 | exclusive = !dquot_is_cached(sb, id, cnt); | ||
859 | up_write(&sb_dqopt(sb)->dqptr_sem); | ||
860 | if (exclusive) { | ||
861 | status = ocfs2_lock_global_qf(oinfo, 1); | ||
862 | if (status < 0) { | ||
863 | exclusive = 0; | ||
864 | mlog_errno(status); | ||
865 | goto out_ilock; | ||
866 | } | ||
867 | handle = ocfs2_start_trans(OCFS2_SB(sb), | ||
868 | ocfs2_calc_qinit_credits(sb, cnt)); | ||
869 | if (IS_ERR(handle)) { | ||
870 | status = PTR_ERR(handle); | ||
871 | mlog_errno(status); | ||
872 | goto out_ilock; | ||
873 | } | ||
874 | } | ||
875 | dquot_initialize(inode, cnt); | ||
876 | if (exclusive) { | ||
877 | ocfs2_commit_trans(OCFS2_SB(sb), handle); | ||
878 | ocfs2_unlock_global_qf(oinfo, 1); | ||
879 | } | ||
880 | ocfs2_unlock_global_qf(oinfo, 0); | ||
881 | } | ||
882 | mlog_exit(0); | ||
883 | return 0; | ||
884 | out_ilock: | ||
885 | if (exclusive) | ||
886 | ocfs2_unlock_global_qf(oinfo, 1); | ||
887 | ocfs2_unlock_global_qf(oinfo, 0); | ||
888 | out: | ||
889 | mlog_exit(status); | ||
890 | return status; | ||
891 | } | ||
892 | |||
893 | static int ocfs2_dquot_drop_slow(struct inode *inode) | ||
894 | { | ||
895 | int status = 0; | ||
896 | int cnt; | ||
897 | int got_lock[MAXQUOTAS] = {0, 0}; | ||
898 | handle_t *handle; | ||
899 | struct super_block *sb = inode->i_sb; | ||
900 | struct ocfs2_mem_dqinfo *oinfo; | ||
901 | |||
902 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { | ||
903 | if (!sb_has_quota_active(sb, cnt)) | ||
904 | continue; | ||
905 | oinfo = sb_dqinfo(sb, cnt)->dqi_priv; | ||
906 | status = ocfs2_lock_global_qf(oinfo, 1); | ||
907 | if (status < 0) | ||
908 | goto out; | ||
909 | got_lock[cnt] = 1; | ||
910 | } | ||
911 | handle = ocfs2_start_trans(OCFS2_SB(sb), | ||
912 | ocfs2_calc_qinit_credits(sb, USRQUOTA) + | ||
913 | ocfs2_calc_qinit_credits(sb, GRPQUOTA)); | ||
914 | if (IS_ERR(handle)) { | ||
915 | status = PTR_ERR(handle); | ||
916 | mlog_errno(status); | ||
917 | goto out; | ||
918 | } | ||
919 | dquot_drop(inode); | ||
920 | ocfs2_commit_trans(OCFS2_SB(sb), handle); | ||
921 | out: | ||
922 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) | ||
923 | if (got_lock[cnt]) { | ||
924 | oinfo = sb_dqinfo(sb, cnt)->dqi_priv; | ||
925 | ocfs2_unlock_global_qf(oinfo, 1); | ||
926 | } | ||
927 | return status; | ||
928 | } | ||
929 | |||
930 | /* See the comment before ocfs2_dquot_initialize. */ | ||
931 | static int ocfs2_dquot_drop(struct inode *inode) | ||
932 | { | ||
933 | int status = 0; | ||
934 | struct super_block *sb = inode->i_sb; | ||
935 | struct ocfs2_mem_dqinfo *oinfo; | ||
936 | int exclusive = 0; | ||
937 | int cnt; | ||
938 | int got_lock[MAXQUOTAS] = {0, 0}; | ||
939 | |||
940 | mlog_entry_void(); | ||
941 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { | ||
942 | if (!sb_has_quota_active(sb, cnt)) | ||
943 | continue; | ||
944 | oinfo = sb_dqinfo(sb, cnt)->dqi_priv; | ||
945 | status = ocfs2_lock_global_qf(oinfo, 0); | ||
946 | if (status < 0) | ||
947 | goto out; | ||
948 | got_lock[cnt] = 1; | ||
949 | } | ||
950 | /* Lock against anyone releasing references so that when when we check | ||
951 | * we know we are not going to be last ones to release dquot */ | ||
952 | down_write(&sb_dqopt(sb)->dqptr_sem); | ||
953 | /* Urgh, this is a terrible hack :( */ | ||
954 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) { | ||
955 | if (inode->i_dquot[cnt] != NODQUOT && | ||
956 | atomic_read(&inode->i_dquot[cnt]->dq_count) > 1) { | ||
957 | exclusive = 1; | ||
958 | break; | ||
959 | } | ||
960 | } | ||
961 | if (!exclusive) | ||
962 | dquot_drop_locked(inode); | ||
963 | up_write(&sb_dqopt(sb)->dqptr_sem); | ||
964 | out: | ||
965 | for (cnt = 0; cnt < MAXQUOTAS; cnt++) | ||
966 | if (got_lock[cnt]) { | ||
967 | oinfo = sb_dqinfo(sb, cnt)->dqi_priv; | ||
968 | ocfs2_unlock_global_qf(oinfo, 0); | ||
969 | } | ||
970 | /* In case we bailed out because we had to do expensive locking | ||
971 | * do it now... */ | ||
972 | if (exclusive) | ||
973 | status = ocfs2_dquot_drop_slow(inode); | ||
974 | mlog_exit(status); | ||
975 | return status; | ||
976 | } | ||
977 | |||
978 | static struct dquot *ocfs2_alloc_dquot(struct super_block *sb, int type) | 815 | static struct dquot *ocfs2_alloc_dquot(struct super_block *sb, int type) |
979 | { | 816 | { |
980 | struct ocfs2_dquot *dquot = | 817 | struct ocfs2_dquot *dquot = |
@@ -991,8 +828,8 @@ static void ocfs2_destroy_dquot(struct dquot *dquot) | |||
991 | } | 828 | } |
992 | 829 | ||
993 | struct dquot_operations ocfs2_quota_operations = { | 830 | struct dquot_operations ocfs2_quota_operations = { |
994 | .initialize = ocfs2_dquot_initialize, | 831 | .initialize = dquot_initialize, |
995 | .drop = ocfs2_dquot_drop, | 832 | .drop = dquot_drop, |
996 | .alloc_space = dquot_alloc_space, | 833 | .alloc_space = dquot_alloc_space, |
997 | .alloc_inode = dquot_alloc_inode, | 834 | .alloc_inode = dquot_alloc_inode, |
998 | .free_space = dquot_free_space, | 835 | .free_space = dquot_free_space, |
diff --git a/fs/ocfs2/super.c b/fs/ocfs2/super.c index 43ed11345b59..7ac83a81ee55 100644 --- a/fs/ocfs2/super.c +++ b/fs/ocfs2/super.c | |||
@@ -1537,6 +1537,13 @@ static int ocfs2_get_sector(struct super_block *sb, | |||
1537 | unlock_buffer(*bh); | 1537 | unlock_buffer(*bh); |
1538 | ll_rw_block(READ, 1, bh); | 1538 | ll_rw_block(READ, 1, bh); |
1539 | wait_on_buffer(*bh); | 1539 | wait_on_buffer(*bh); |
1540 | if (!buffer_uptodate(*bh)) { | ||
1541 | mlog_errno(-EIO); | ||
1542 | brelse(*bh); | ||
1543 | *bh = NULL; | ||
1544 | return -EIO; | ||
1545 | } | ||
1546 | |||
1540 | return 0; | 1547 | return 0; |
1541 | } | 1548 | } |
1542 | 1549 | ||
@@ -1747,6 +1754,7 @@ static int ocfs2_initialize_super(struct super_block *sb, | |||
1747 | INIT_LIST_HEAD(&osb->blocked_lock_list); | 1754 | INIT_LIST_HEAD(&osb->blocked_lock_list); |
1748 | osb->blocked_lock_count = 0; | 1755 | osb->blocked_lock_count = 0; |
1749 | spin_lock_init(&osb->osb_lock); | 1756 | spin_lock_init(&osb->osb_lock); |
1757 | spin_lock_init(&osb->osb_xattr_lock); | ||
1750 | ocfs2_init_inode_steal_slot(osb); | 1758 | ocfs2_init_inode_steal_slot(osb); |
1751 | 1759 | ||
1752 | atomic_set(&osb->alloc_stats.moves, 0); | 1760 | atomic_set(&osb->alloc_stats.moves, 0); |
@@ -1887,6 +1895,9 @@ static int ocfs2_initialize_super(struct super_block *sb, | |||
1887 | INIT_WORK(&journal->j_recovery_work, ocfs2_complete_recovery); | 1895 | INIT_WORK(&journal->j_recovery_work, ocfs2_complete_recovery); |
1888 | journal->j_state = OCFS2_JOURNAL_FREE; | 1896 | journal->j_state = OCFS2_JOURNAL_FREE; |
1889 | 1897 | ||
1898 | INIT_WORK(&osb->dentry_lock_work, ocfs2_drop_dl_inodes); | ||
1899 | osb->dentry_lock_list = NULL; | ||
1900 | |||
1890 | /* get some pseudo constants for clustersize bits */ | 1901 | /* get some pseudo constants for clustersize bits */ |
1891 | osb->s_clustersize_bits = | 1902 | osb->s_clustersize_bits = |
1892 | le32_to_cpu(di->id2.i_super.s_clustersize_bits); | 1903 | le32_to_cpu(di->id2.i_super.s_clustersize_bits); |
diff --git a/fs/ocfs2/xattr.c b/fs/ocfs2/xattr.c index e1d638af6ac3..4ddd788add67 100644 --- a/fs/ocfs2/xattr.c +++ b/fs/ocfs2/xattr.c | |||
@@ -82,13 +82,14 @@ struct ocfs2_xattr_set_ctxt { | |||
82 | 82 | ||
83 | #define OCFS2_XATTR_ROOT_SIZE (sizeof(struct ocfs2_xattr_def_value_root)) | 83 | #define OCFS2_XATTR_ROOT_SIZE (sizeof(struct ocfs2_xattr_def_value_root)) |
84 | #define OCFS2_XATTR_INLINE_SIZE 80 | 84 | #define OCFS2_XATTR_INLINE_SIZE 80 |
85 | #define OCFS2_XATTR_HEADER_GAP 4 | ||
85 | #define OCFS2_XATTR_FREE_IN_IBODY (OCFS2_MIN_XATTR_INLINE_SIZE \ | 86 | #define OCFS2_XATTR_FREE_IN_IBODY (OCFS2_MIN_XATTR_INLINE_SIZE \ |
86 | - sizeof(struct ocfs2_xattr_header) \ | 87 | - sizeof(struct ocfs2_xattr_header) \ |
87 | - sizeof(__u32)) | 88 | - OCFS2_XATTR_HEADER_GAP) |
88 | #define OCFS2_XATTR_FREE_IN_BLOCK(ptr) ((ptr)->i_sb->s_blocksize \ | 89 | #define OCFS2_XATTR_FREE_IN_BLOCK(ptr) ((ptr)->i_sb->s_blocksize \ |
89 | - sizeof(struct ocfs2_xattr_block) \ | 90 | - sizeof(struct ocfs2_xattr_block) \ |
90 | - sizeof(struct ocfs2_xattr_header) \ | 91 | - sizeof(struct ocfs2_xattr_header) \ |
91 | - sizeof(__u32)) | 92 | - OCFS2_XATTR_HEADER_GAP) |
92 | 93 | ||
93 | static struct ocfs2_xattr_def_value_root def_xv = { | 94 | static struct ocfs2_xattr_def_value_root def_xv = { |
94 | .xv.xr_list.l_count = cpu_to_le16(1), | 95 | .xv.xr_list.l_count = cpu_to_le16(1), |
@@ -274,10 +275,12 @@ static int ocfs2_read_xattr_bucket(struct ocfs2_xattr_bucket *bucket, | |||
274 | bucket->bu_blocks, bucket->bu_bhs, 0, | 275 | bucket->bu_blocks, bucket->bu_bhs, 0, |
275 | NULL); | 276 | NULL); |
276 | if (!rc) { | 277 | if (!rc) { |
278 | spin_lock(&OCFS2_SB(bucket->bu_inode->i_sb)->osb_xattr_lock); | ||
277 | rc = ocfs2_validate_meta_ecc_bhs(bucket->bu_inode->i_sb, | 279 | rc = ocfs2_validate_meta_ecc_bhs(bucket->bu_inode->i_sb, |
278 | bucket->bu_bhs, | 280 | bucket->bu_bhs, |
279 | bucket->bu_blocks, | 281 | bucket->bu_blocks, |
280 | &bucket_xh(bucket)->xh_check); | 282 | &bucket_xh(bucket)->xh_check); |
283 | spin_unlock(&OCFS2_SB(bucket->bu_inode->i_sb)->osb_xattr_lock); | ||
281 | if (rc) | 284 | if (rc) |
282 | mlog_errno(rc); | 285 | mlog_errno(rc); |
283 | } | 286 | } |
@@ -310,9 +313,11 @@ static void ocfs2_xattr_bucket_journal_dirty(handle_t *handle, | |||
310 | { | 313 | { |
311 | int i; | 314 | int i; |
312 | 315 | ||
316 | spin_lock(&OCFS2_SB(bucket->bu_inode->i_sb)->osb_xattr_lock); | ||
313 | ocfs2_compute_meta_ecc_bhs(bucket->bu_inode->i_sb, | 317 | ocfs2_compute_meta_ecc_bhs(bucket->bu_inode->i_sb, |
314 | bucket->bu_bhs, bucket->bu_blocks, | 318 | bucket->bu_bhs, bucket->bu_blocks, |
315 | &bucket_xh(bucket)->xh_check); | 319 | &bucket_xh(bucket)->xh_check); |
320 | spin_unlock(&OCFS2_SB(bucket->bu_inode->i_sb)->osb_xattr_lock); | ||
316 | 321 | ||
317 | for (i = 0; i < bucket->bu_blocks; i++) | 322 | for (i = 0; i < bucket->bu_blocks; i++) |
318 | ocfs2_journal_dirty(handle, bucket->bu_bhs[i]); | 323 | ocfs2_journal_dirty(handle, bucket->bu_bhs[i]); |
@@ -1507,7 +1512,7 @@ static int ocfs2_xattr_set_entry(struct inode *inode, | |||
1507 | last += 1; | 1512 | last += 1; |
1508 | } | 1513 | } |
1509 | 1514 | ||
1510 | free = min_offs - ((void *)last - xs->base) - sizeof(__u32); | 1515 | free = min_offs - ((void *)last - xs->base) - OCFS2_XATTR_HEADER_GAP; |
1511 | if (free < 0) | 1516 | if (free < 0) |
1512 | return -EIO; | 1517 | return -EIO; |
1513 | 1518 | ||
@@ -2190,7 +2195,7 @@ static int ocfs2_xattr_can_be_in_inode(struct inode *inode, | |||
2190 | last += 1; | 2195 | last += 1; |
2191 | } | 2196 | } |
2192 | 2197 | ||
2193 | free = min_offs - ((void *)last - xs->base) - sizeof(__u32); | 2198 | free = min_offs - ((void *)last - xs->base) - OCFS2_XATTR_HEADER_GAP; |
2194 | if (free < 0) | 2199 | if (free < 0) |
2195 | return 0; | 2200 | return 0; |
2196 | 2201 | ||
@@ -2592,8 +2597,9 @@ static int __ocfs2_xattr_set_handle(struct inode *inode, | |||
2592 | 2597 | ||
2593 | if (!ret) { | 2598 | if (!ret) { |
2594 | /* Update inode ctime. */ | 2599 | /* Update inode ctime. */ |
2595 | ret = ocfs2_journal_access(ctxt->handle, inode, xis->inode_bh, | 2600 | ret = ocfs2_journal_access_di(ctxt->handle, inode, |
2596 | OCFS2_JOURNAL_ACCESS_WRITE); | 2601 | xis->inode_bh, |
2602 | OCFS2_JOURNAL_ACCESS_WRITE); | ||
2597 | if (ret) { | 2603 | if (ret) { |
2598 | mlog_errno(ret); | 2604 | mlog_errno(ret); |
2599 | goto out; | 2605 | goto out; |
@@ -4729,13 +4735,6 @@ static int ocfs2_xattr_bucket_value_truncate(struct inode *inode, | |||
4729 | vb.vb_xv = (struct ocfs2_xattr_value_root *) | 4735 | vb.vb_xv = (struct ocfs2_xattr_value_root *) |
4730 | (vb.vb_bh->b_data + offset % blocksize); | 4736 | (vb.vb_bh->b_data + offset % blocksize); |
4731 | 4737 | ||
4732 | ret = ocfs2_xattr_bucket_journal_access(ctxt->handle, bucket, | ||
4733 | OCFS2_JOURNAL_ACCESS_WRITE); | ||
4734 | if (ret) { | ||
4735 | mlog_errno(ret); | ||
4736 | goto out; | ||
4737 | } | ||
4738 | |||
4739 | /* | 4738 | /* |
4740 | * From here on out we have to dirty the bucket. The generic | 4739 | * From here on out we have to dirty the bucket. The generic |
4741 | * value calls only modify one of the bucket's bhs, but we need | 4740 | * value calls only modify one of the bucket's bhs, but we need |
@@ -4748,12 +4747,18 @@ static int ocfs2_xattr_bucket_value_truncate(struct inode *inode, | |||
4748 | ret = ocfs2_xattr_value_truncate(inode, &vb, len, ctxt); | 4747 | ret = ocfs2_xattr_value_truncate(inode, &vb, len, ctxt); |
4749 | if (ret) { | 4748 | if (ret) { |
4750 | mlog_errno(ret); | 4749 | mlog_errno(ret); |
4751 | goto out_dirty; | 4750 | goto out; |
4751 | } | ||
4752 | |||
4753 | ret = ocfs2_xattr_bucket_journal_access(ctxt->handle, bucket, | ||
4754 | OCFS2_JOURNAL_ACCESS_WRITE); | ||
4755 | if (ret) { | ||
4756 | mlog_errno(ret); | ||
4757 | goto out; | ||
4752 | } | 4758 | } |
4753 | 4759 | ||
4754 | xe->xe_value_size = cpu_to_le64(len); | 4760 | xe->xe_value_size = cpu_to_le64(len); |
4755 | 4761 | ||
4756 | out_dirty: | ||
4757 | ocfs2_xattr_bucket_journal_dirty(ctxt->handle, bucket); | 4762 | ocfs2_xattr_bucket_journal_dirty(ctxt->handle, bucket); |
4758 | 4763 | ||
4759 | out: | 4764 | out: |
@@ -5061,8 +5066,8 @@ try_again: | |||
5061 | xh_free_start = le16_to_cpu(xh->xh_free_start); | 5066 | xh_free_start = le16_to_cpu(xh->xh_free_start); |
5062 | header_size = sizeof(struct ocfs2_xattr_header) + | 5067 | header_size = sizeof(struct ocfs2_xattr_header) + |
5063 | count * sizeof(struct ocfs2_xattr_entry); | 5068 | count * sizeof(struct ocfs2_xattr_entry); |
5064 | max_free = OCFS2_XATTR_BUCKET_SIZE - | 5069 | max_free = OCFS2_XATTR_BUCKET_SIZE - header_size - |
5065 | le16_to_cpu(xh->xh_name_value_len) - header_size; | 5070 | le16_to_cpu(xh->xh_name_value_len) - OCFS2_XATTR_HEADER_GAP; |
5066 | 5071 | ||
5067 | mlog_bug_on_msg(header_size > blocksize, "bucket %llu has header size " | 5072 | mlog_bug_on_msg(header_size > blocksize, "bucket %llu has header size " |
5068 | "of %u which exceed block size\n", | 5073 | "of %u which exceed block size\n", |
@@ -5095,7 +5100,7 @@ try_again: | |||
5095 | need = 0; | 5100 | need = 0; |
5096 | } | 5101 | } |
5097 | 5102 | ||
5098 | free = xh_free_start - header_size; | 5103 | free = xh_free_start - header_size - OCFS2_XATTR_HEADER_GAP; |
5099 | /* | 5104 | /* |
5100 | * We need to make sure the new name/value pair | 5105 | * We need to make sure the new name/value pair |
5101 | * can exist in the same block. | 5106 | * can exist in the same block. |
@@ -5128,7 +5133,8 @@ try_again: | |||
5128 | } | 5133 | } |
5129 | 5134 | ||
5130 | xh_free_start = le16_to_cpu(xh->xh_free_start); | 5135 | xh_free_start = le16_to_cpu(xh->xh_free_start); |
5131 | free = xh_free_start - header_size; | 5136 | free = xh_free_start - header_size |
5137 | - OCFS2_XATTR_HEADER_GAP; | ||
5132 | if (xh_free_start % blocksize < need) | 5138 | if (xh_free_start % blocksize < need) |
5133 | free -= xh_free_start % blocksize; | 5139 | free -= xh_free_start % blocksize; |
5134 | 5140 | ||
diff --git a/fs/omfs/Kconfig b/fs/omfs/Kconfig new file mode 100644 index 000000000000..b1b9a0aba6fd --- /dev/null +++ b/fs/omfs/Kconfig | |||
@@ -0,0 +1,13 @@ | |||
1 | config OMFS_FS | ||
2 | tristate "SonicBlue Optimized MPEG File System support" | ||
3 | depends on BLOCK | ||
4 | select CRC_ITU_T | ||
5 | help | ||
6 | This is the proprietary file system used by the Rio Karma music | ||
7 | player and ReplayTV DVR. Despite the name, this filesystem is not | ||
8 | more efficient than a standard FS for MPEG files, in fact likely | ||
9 | the opposite is true. Say Y if you have either of these devices | ||
10 | and wish to mount its disk. | ||
11 | |||
12 | To compile this file system support as a module, choose M here: the | ||
13 | module will be called omfs. If unsure, say N. | ||
@@ -122,7 +122,7 @@ static int vfs_statfs64(struct dentry *dentry, struct statfs64 *buf) | |||
122 | return 0; | 122 | return 0; |
123 | } | 123 | } |
124 | 124 | ||
125 | asmlinkage long sys_statfs(const char __user *pathname, struct statfs __user * buf) | 125 | SYSCALL_DEFINE2(statfs, const char __user *, pathname, struct statfs __user *, buf) |
126 | { | 126 | { |
127 | struct path path; | 127 | struct path path; |
128 | int error; | 128 | int error; |
@@ -138,8 +138,7 @@ asmlinkage long sys_statfs(const char __user *pathname, struct statfs __user * b | |||
138 | return error; | 138 | return error; |
139 | } | 139 | } |
140 | 140 | ||
141 | 141 | SYSCALL_DEFINE3(statfs64, const char __user *, pathname, size_t, sz, struct statfs64 __user *, buf) | |
142 | asmlinkage long sys_statfs64(const char __user *pathname, size_t sz, struct statfs64 __user *buf) | ||
143 | { | 142 | { |
144 | struct path path; | 143 | struct path path; |
145 | long error; | 144 | long error; |
@@ -157,8 +156,7 @@ asmlinkage long sys_statfs64(const char __user *pathname, size_t sz, struct stat | |||
157 | return error; | 156 | return error; |
158 | } | 157 | } |
159 | 158 | ||
160 | 159 | SYSCALL_DEFINE2(fstatfs, unsigned int, fd, struct statfs __user *, buf) | |
161 | asmlinkage long sys_fstatfs(unsigned int fd, struct statfs __user * buf) | ||
162 | { | 160 | { |
163 | struct file * file; | 161 | struct file * file; |
164 | struct statfs tmp; | 162 | struct statfs tmp; |
@@ -176,7 +174,7 @@ out: | |||
176 | return error; | 174 | return error; |
177 | } | 175 | } |
178 | 176 | ||
179 | asmlinkage long sys_fstatfs64(unsigned int fd, size_t sz, struct statfs64 __user *buf) | 177 | SYSCALL_DEFINE3(fstatfs64, unsigned int, fd, size_t, sz, struct statfs64 __user *, buf) |
180 | { | 178 | { |
181 | struct file * file; | 179 | struct file * file; |
182 | struct statfs64 tmp; | 180 | struct statfs64 tmp; |
@@ -289,7 +287,7 @@ out: | |||
289 | return error; | 287 | return error; |
290 | } | 288 | } |
291 | 289 | ||
292 | asmlinkage long sys_truncate(const char __user * path, unsigned long length) | 290 | SYSCALL_DEFINE2(truncate, const char __user *, path, unsigned long, length) |
293 | { | 291 | { |
294 | /* on 32-bit boxen it will cut the range 2^31--2^32-1 off */ | 292 | /* on 32-bit boxen it will cut the range 2^31--2^32-1 off */ |
295 | return do_sys_truncate(path, (long)length); | 293 | return do_sys_truncate(path, (long)length); |
@@ -341,7 +339,7 @@ out: | |||
341 | return error; | 339 | return error; |
342 | } | 340 | } |
343 | 341 | ||
344 | asmlinkage long sys_ftruncate(unsigned int fd, unsigned long length) | 342 | SYSCALL_DEFINE2(ftruncate, unsigned int, fd, unsigned long, length) |
345 | { | 343 | { |
346 | long ret = do_sys_ftruncate(fd, length, 1); | 344 | long ret = do_sys_ftruncate(fd, length, 1); |
347 | /* avoid REGPARM breakage on x86: */ | 345 | /* avoid REGPARM breakage on x86: */ |
@@ -351,21 +349,35 @@ asmlinkage long sys_ftruncate(unsigned int fd, unsigned long length) | |||
351 | 349 | ||
352 | /* LFS versions of truncate are only needed on 32 bit machines */ | 350 | /* LFS versions of truncate are only needed on 32 bit machines */ |
353 | #if BITS_PER_LONG == 32 | 351 | #if BITS_PER_LONG == 32 |
354 | asmlinkage long sys_truncate64(const char __user * path, loff_t length) | 352 | SYSCALL_DEFINE(truncate64)(const char __user * path, loff_t length) |
355 | { | 353 | { |
356 | return do_sys_truncate(path, length); | 354 | return do_sys_truncate(path, length); |
357 | } | 355 | } |
356 | #ifdef CONFIG_HAVE_SYSCALL_WRAPPERS | ||
357 | asmlinkage long SyS_truncate64(long path, loff_t length) | ||
358 | { | ||
359 | return SYSC_truncate64((const char __user *) path, length); | ||
360 | } | ||
361 | SYSCALL_ALIAS(sys_truncate64, SyS_truncate64); | ||
362 | #endif | ||
358 | 363 | ||
359 | asmlinkage long sys_ftruncate64(unsigned int fd, loff_t length) | 364 | SYSCALL_DEFINE(ftruncate64)(unsigned int fd, loff_t length) |
360 | { | 365 | { |
361 | long ret = do_sys_ftruncate(fd, length, 0); | 366 | long ret = do_sys_ftruncate(fd, length, 0); |
362 | /* avoid REGPARM breakage on x86: */ | 367 | /* avoid REGPARM breakage on x86: */ |
363 | asmlinkage_protect(2, ret, fd, length); | 368 | asmlinkage_protect(2, ret, fd, length); |
364 | return ret; | 369 | return ret; |
365 | } | 370 | } |
371 | #ifdef CONFIG_HAVE_SYSCALL_WRAPPERS | ||
372 | asmlinkage long SyS_ftruncate64(long fd, loff_t length) | ||
373 | { | ||
374 | return SYSC_ftruncate64((unsigned int) fd, length); | ||
375 | } | ||
376 | SYSCALL_ALIAS(sys_ftruncate64, SyS_ftruncate64); | ||
366 | #endif | 377 | #endif |
378 | #endif /* BITS_PER_LONG == 32 */ | ||
367 | 379 | ||
368 | asmlinkage long sys_fallocate(int fd, int mode, loff_t offset, loff_t len) | 380 | SYSCALL_DEFINE(fallocate)(int fd, int mode, loff_t offset, loff_t len) |
369 | { | 381 | { |
370 | struct file *file; | 382 | struct file *file; |
371 | struct inode *inode; | 383 | struct inode *inode; |
@@ -422,13 +434,20 @@ out_fput: | |||
422 | out: | 434 | out: |
423 | return ret; | 435 | return ret; |
424 | } | 436 | } |
437 | #ifdef CONFIG_HAVE_SYSCALL_WRAPPERS | ||
438 | asmlinkage long SyS_fallocate(long fd, long mode, loff_t offset, loff_t len) | ||
439 | { | ||
440 | return SYSC_fallocate((int)fd, (int)mode, offset, len); | ||
441 | } | ||
442 | SYSCALL_ALIAS(sys_fallocate, SyS_fallocate); | ||
443 | #endif | ||
425 | 444 | ||
426 | /* | 445 | /* |
427 | * access() needs to use the real uid/gid, not the effective uid/gid. | 446 | * access() needs to use the real uid/gid, not the effective uid/gid. |
428 | * We do this by temporarily clearing all FS-related capabilities and | 447 | * We do this by temporarily clearing all FS-related capabilities and |
429 | * switching the fsuid/fsgid around to the real ones. | 448 | * switching the fsuid/fsgid around to the real ones. |
430 | */ | 449 | */ |
431 | asmlinkage long sys_faccessat(int dfd, const char __user *filename, int mode) | 450 | SYSCALL_DEFINE3(faccessat, int, dfd, const char __user *, filename, int, mode) |
432 | { | 451 | { |
433 | const struct cred *old_cred; | 452 | const struct cred *old_cred; |
434 | struct cred *override_cred; | 453 | struct cred *override_cred; |
@@ -498,12 +517,12 @@ out: | |||
498 | return res; | 517 | return res; |
499 | } | 518 | } |
500 | 519 | ||
501 | asmlinkage long sys_access(const char __user *filename, int mode) | 520 | SYSCALL_DEFINE2(access, const char __user *, filename, int, mode) |
502 | { | 521 | { |
503 | return sys_faccessat(AT_FDCWD, filename, mode); | 522 | return sys_faccessat(AT_FDCWD, filename, mode); |
504 | } | 523 | } |
505 | 524 | ||
506 | asmlinkage long sys_chdir(const char __user * filename) | 525 | SYSCALL_DEFINE1(chdir, const char __user *, filename) |
507 | { | 526 | { |
508 | struct path path; | 527 | struct path path; |
509 | int error; | 528 | int error; |
@@ -524,7 +543,7 @@ out: | |||
524 | return error; | 543 | return error; |
525 | } | 544 | } |
526 | 545 | ||
527 | asmlinkage long sys_fchdir(unsigned int fd) | 546 | SYSCALL_DEFINE1(fchdir, unsigned int, fd) |
528 | { | 547 | { |
529 | struct file *file; | 548 | struct file *file; |
530 | struct inode *inode; | 549 | struct inode *inode; |
@@ -550,7 +569,7 @@ out: | |||
550 | return error; | 569 | return error; |
551 | } | 570 | } |
552 | 571 | ||
553 | asmlinkage long sys_chroot(const char __user * filename) | 572 | SYSCALL_DEFINE1(chroot, const char __user *, filename) |
554 | { | 573 | { |
555 | struct path path; | 574 | struct path path; |
556 | int error; | 575 | int error; |
@@ -575,7 +594,7 @@ out: | |||
575 | return error; | 594 | return error; |
576 | } | 595 | } |
577 | 596 | ||
578 | asmlinkage long sys_fchmod(unsigned int fd, mode_t mode) | 597 | SYSCALL_DEFINE2(fchmod, unsigned int, fd, mode_t, mode) |
579 | { | 598 | { |
580 | struct inode * inode; | 599 | struct inode * inode; |
581 | struct dentry * dentry; | 600 | struct dentry * dentry; |
@@ -609,8 +628,7 @@ out: | |||
609 | return err; | 628 | return err; |
610 | } | 629 | } |
611 | 630 | ||
612 | asmlinkage long sys_fchmodat(int dfd, const char __user *filename, | 631 | SYSCALL_DEFINE3(fchmodat, int, dfd, const char __user *, filename, mode_t, mode) |
613 | mode_t mode) | ||
614 | { | 632 | { |
615 | struct path path; | 633 | struct path path; |
616 | struct inode *inode; | 634 | struct inode *inode; |
@@ -639,7 +657,7 @@ out: | |||
639 | return error; | 657 | return error; |
640 | } | 658 | } |
641 | 659 | ||
642 | asmlinkage long sys_chmod(const char __user *filename, mode_t mode) | 660 | SYSCALL_DEFINE2(chmod, const char __user *, filename, mode_t, mode) |
643 | { | 661 | { |
644 | return sys_fchmodat(AT_FDCWD, filename, mode); | 662 | return sys_fchmodat(AT_FDCWD, filename, mode); |
645 | } | 663 | } |
@@ -669,7 +687,7 @@ static int chown_common(struct dentry * dentry, uid_t user, gid_t group) | |||
669 | return error; | 687 | return error; |
670 | } | 688 | } |
671 | 689 | ||
672 | asmlinkage long sys_chown(const char __user * filename, uid_t user, gid_t group) | 690 | SYSCALL_DEFINE3(chown, const char __user *, filename, uid_t, user, gid_t, group) |
673 | { | 691 | { |
674 | struct path path; | 692 | struct path path; |
675 | int error; | 693 | int error; |
@@ -688,8 +706,8 @@ out: | |||
688 | return error; | 706 | return error; |
689 | } | 707 | } |
690 | 708 | ||
691 | asmlinkage long sys_fchownat(int dfd, const char __user *filename, uid_t user, | 709 | SYSCALL_DEFINE5(fchownat, int, dfd, const char __user *, filename, uid_t, user, |
692 | gid_t group, int flag) | 710 | gid_t, group, int, flag) |
693 | { | 711 | { |
694 | struct path path; | 712 | struct path path; |
695 | int error = -EINVAL; | 713 | int error = -EINVAL; |
@@ -713,7 +731,7 @@ out: | |||
713 | return error; | 731 | return error; |
714 | } | 732 | } |
715 | 733 | ||
716 | asmlinkage long sys_lchown(const char __user * filename, uid_t user, gid_t group) | 734 | SYSCALL_DEFINE3(lchown, const char __user *, filename, uid_t, user, gid_t, group) |
717 | { | 735 | { |
718 | struct path path; | 736 | struct path path; |
719 | int error; | 737 | int error; |
@@ -732,8 +750,7 @@ out: | |||
732 | return error; | 750 | return error; |
733 | } | 751 | } |
734 | 752 | ||
735 | 753 | SYSCALL_DEFINE3(fchown, unsigned int, fd, uid_t, user, gid_t, group) | |
736 | asmlinkage long sys_fchown(unsigned int fd, uid_t user, gid_t group) | ||
737 | { | 754 | { |
738 | struct file * file; | 755 | struct file * file; |
739 | int error = -EBADF; | 756 | int error = -EBADF; |
@@ -1029,7 +1046,7 @@ long do_sys_open(int dfd, const char __user *filename, int flags, int mode) | |||
1029 | return fd; | 1046 | return fd; |
1030 | } | 1047 | } |
1031 | 1048 | ||
1032 | asmlinkage long sys_open(const char __user *filename, int flags, int mode) | 1049 | SYSCALL_DEFINE3(open, const char __user *, filename, int, flags, int, mode) |
1033 | { | 1050 | { |
1034 | long ret; | 1051 | long ret; |
1035 | 1052 | ||
@@ -1042,8 +1059,8 @@ asmlinkage long sys_open(const char __user *filename, int flags, int mode) | |||
1042 | return ret; | 1059 | return ret; |
1043 | } | 1060 | } |
1044 | 1061 | ||
1045 | asmlinkage long sys_openat(int dfd, const char __user *filename, int flags, | 1062 | SYSCALL_DEFINE4(openat, int, dfd, const char __user *, filename, int, flags, |
1046 | int mode) | 1063 | int, mode) |
1047 | { | 1064 | { |
1048 | long ret; | 1065 | long ret; |
1049 | 1066 | ||
@@ -1062,7 +1079,7 @@ asmlinkage long sys_openat(int dfd, const char __user *filename, int flags, | |||
1062 | * For backward compatibility? Maybe this should be moved | 1079 | * For backward compatibility? Maybe this should be moved |
1063 | * into arch/i386 instead? | 1080 | * into arch/i386 instead? |
1064 | */ | 1081 | */ |
1065 | asmlinkage long sys_creat(const char __user * pathname, int mode) | 1082 | SYSCALL_DEFINE2(creat, const char __user *, pathname, int, mode) |
1066 | { | 1083 | { |
1067 | return sys_open(pathname, O_CREAT | O_WRONLY | O_TRUNC, mode); | 1084 | return sys_open(pathname, O_CREAT | O_WRONLY | O_TRUNC, mode); |
1068 | } | 1085 | } |
@@ -1098,7 +1115,7 @@ EXPORT_SYMBOL(filp_close); | |||
1098 | * releasing the fd. This ensures that one clone task can't release | 1115 | * releasing the fd. This ensures that one clone task can't release |
1099 | * an fd while another clone is opening it. | 1116 | * an fd while another clone is opening it. |
1100 | */ | 1117 | */ |
1101 | asmlinkage long sys_close(unsigned int fd) | 1118 | SYSCALL_DEFINE1(close, unsigned int, fd) |
1102 | { | 1119 | { |
1103 | struct file * filp; | 1120 | struct file * filp; |
1104 | struct files_struct *files = current->files; | 1121 | struct files_struct *files = current->files; |
@@ -1131,14 +1148,13 @@ out_unlock: | |||
1131 | spin_unlock(&files->file_lock); | 1148 | spin_unlock(&files->file_lock); |
1132 | return -EBADF; | 1149 | return -EBADF; |
1133 | } | 1150 | } |
1134 | |||
1135 | EXPORT_SYMBOL(sys_close); | 1151 | EXPORT_SYMBOL(sys_close); |
1136 | 1152 | ||
1137 | /* | 1153 | /* |
1138 | * This routine simulates a hangup on the tty, to arrange that users | 1154 | * This routine simulates a hangup on the tty, to arrange that users |
1139 | * are given clean terminals at login time. | 1155 | * are given clean terminals at login time. |
1140 | */ | 1156 | */ |
1141 | asmlinkage long sys_vhangup(void) | 1157 | SYSCALL_DEFINE0(vhangup) |
1142 | { | 1158 | { |
1143 | if (capable(CAP_SYS_TTY_CONFIG)) { | 1159 | if (capable(CAP_SYS_TTY_CONFIG)) { |
1144 | tty_vhangup_self(); | 1160 | tty_vhangup_self(); |
@@ -1043,7 +1043,7 @@ int do_pipe(int *fd) | |||
1043 | * sys_pipe() is the normal C calling standard for creating | 1043 | * sys_pipe() is the normal C calling standard for creating |
1044 | * a pipe. It's not the way Unix traditionally does this, though. | 1044 | * a pipe. It's not the way Unix traditionally does this, though. |
1045 | */ | 1045 | */ |
1046 | asmlinkage long __weak sys_pipe2(int __user *fildes, int flags) | 1046 | SYSCALL_DEFINE2(pipe2, int __user *, fildes, int, flags) |
1047 | { | 1047 | { |
1048 | int fd[2]; | 1048 | int fd[2]; |
1049 | int error; | 1049 | int error; |
@@ -1059,7 +1059,7 @@ asmlinkage long __weak sys_pipe2(int __user *fildes, int flags) | |||
1059 | return error; | 1059 | return error; |
1060 | } | 1060 | } |
1061 | 1061 | ||
1062 | asmlinkage long __weak sys_pipe(int __user *fildes) | 1062 | SYSCALL_DEFINE1(pipe, int __user *, fildes) |
1063 | { | 1063 | { |
1064 | return sys_pipe2(fildes, 0); | 1064 | return sys_pipe2(fildes, 0); |
1065 | } | 1065 | } |
diff --git a/fs/proc/inode.c b/fs/proc/inode.c index 3e76bb9b3ad6..d8bb5c671f42 100644 --- a/fs/proc/inode.c +++ b/fs/proc/inode.c | |||
@@ -485,8 +485,10 @@ struct inode *proc_get_inode(struct super_block *sb, unsigned int ino, | |||
485 | } | 485 | } |
486 | } | 486 | } |
487 | unlock_new_inode(inode); | 487 | unlock_new_inode(inode); |
488 | } else | 488 | } else { |
489 | module_put(de->owner); | 489 | module_put(de->owner); |
490 | de_put(de); | ||
491 | } | ||
490 | return inode; | 492 | return inode; |
491 | 493 | ||
492 | out_ino: | 494 | out_ino: |
diff --git a/fs/proc/page.c b/fs/proc/page.c index 767d95a6d1b1..2d1345112a42 100644 --- a/fs/proc/page.c +++ b/fs/proc/page.c | |||
@@ -107,7 +107,7 @@ static ssize_t kpageflags_read(struct file *file, char __user *buf, | |||
107 | else | 107 | else |
108 | kflags = ppage->flags; | 108 | kflags = ppage->flags; |
109 | 109 | ||
110 | uflags = kpf_copy_bit(KPF_LOCKED, PG_locked, kflags) | | 110 | uflags = kpf_copy_bit(kflags, KPF_LOCKED, PG_locked) | |
111 | kpf_copy_bit(kflags, KPF_ERROR, PG_error) | | 111 | kpf_copy_bit(kflags, KPF_ERROR, PG_error) | |
112 | kpf_copy_bit(kflags, KPF_REFERENCED, PG_referenced) | | 112 | kpf_copy_bit(kflags, KPF_REFERENCED, PG_referenced) | |
113 | kpf_copy_bit(kflags, KPF_UPTODATE, PG_uptodate) | | 113 | kpf_copy_bit(kflags, KPF_UPTODATE, PG_uptodate) | |
diff --git a/fs/qnx4/Kconfig b/fs/qnx4/Kconfig new file mode 100644 index 000000000000..be8e0e1445b6 --- /dev/null +++ b/fs/qnx4/Kconfig | |||
@@ -0,0 +1,25 @@ | |||
1 | config QNX4FS_FS | ||
2 | tristate "QNX4 file system support (read only)" | ||
3 | depends on BLOCK | ||
4 | help | ||
5 | This is the file system used by the real-time operating systems | ||
6 | QNX 4 and QNX 6 (the latter is also called QNX RTP). | ||
7 | Further information is available at <http://www.qnx.com/>. | ||
8 | Say Y if you intend to mount QNX hard disks or floppies. | ||
9 | Unless you say Y to "QNX4FS read-write support" below, you will | ||
10 | only be able to read these file systems. | ||
11 | |||
12 | To compile this file system support as a module, choose M here: the | ||
13 | module will be called qnx4. | ||
14 | |||
15 | If you don't know whether you need it, then you don't need it: | ||
16 | answer N. | ||
17 | |||
18 | config QNX4FS_RW | ||
19 | bool "QNX4FS write support (DANGEROUS)" | ||
20 | depends on QNX4FS_FS && EXPERIMENTAL && BROKEN | ||
21 | help | ||
22 | Say Y if you want to test write support for QNX4 file systems. | ||
23 | |||
24 | It's currently broken, so for now: | ||
25 | answer N. | ||
diff --git a/fs/quota.c b/fs/quota.c index 4a8c94f05f76..d76ada914f98 100644 --- a/fs/quota.c +++ b/fs/quota.c | |||
@@ -371,7 +371,8 @@ static inline struct super_block *quotactl_block(const char __user *special) | |||
371 | * calls. Maybe we need to add the process quotas etc. in the future, | 371 | * calls. Maybe we need to add the process quotas etc. in the future, |
372 | * but we probably should use rlimits for that. | 372 | * but we probably should use rlimits for that. |
373 | */ | 373 | */ |
374 | asmlinkage long sys_quotactl(unsigned int cmd, const char __user *special, qid_t id, void __user *addr) | 374 | SYSCALL_DEFINE4(quotactl, unsigned int, cmd, const char __user *, special, |
375 | qid_t, id, void __user *, addr) | ||
375 | { | 376 | { |
376 | uint cmds, type; | 377 | uint cmds, type; |
377 | struct super_block *sb = NULL; | 378 | struct super_block *sb = NULL; |
diff --git a/fs/read_write.c b/fs/read_write.c index 5cc6924eb158..400fe81c973e 100644 --- a/fs/read_write.c +++ b/fs/read_write.c | |||
@@ -147,7 +147,7 @@ loff_t vfs_llseek(struct file *file, loff_t offset, int origin) | |||
147 | } | 147 | } |
148 | EXPORT_SYMBOL(vfs_llseek); | 148 | EXPORT_SYMBOL(vfs_llseek); |
149 | 149 | ||
150 | asmlinkage off_t sys_lseek(unsigned int fd, off_t offset, unsigned int origin) | 150 | SYSCALL_DEFINE3(lseek, unsigned int, fd, off_t, offset, unsigned int, origin) |
151 | { | 151 | { |
152 | off_t retval; | 152 | off_t retval; |
153 | struct file * file; | 153 | struct file * file; |
@@ -171,9 +171,9 @@ bad: | |||
171 | } | 171 | } |
172 | 172 | ||
173 | #ifdef __ARCH_WANT_SYS_LLSEEK | 173 | #ifdef __ARCH_WANT_SYS_LLSEEK |
174 | asmlinkage long sys_llseek(unsigned int fd, unsigned long offset_high, | 174 | SYSCALL_DEFINE5(llseek, unsigned int, fd, unsigned long, offset_high, |
175 | unsigned long offset_low, loff_t __user * result, | 175 | unsigned long, offset_low, loff_t __user *, result, |
176 | unsigned int origin) | 176 | unsigned int, origin) |
177 | { | 177 | { |
178 | int retval; | 178 | int retval; |
179 | struct file * file; | 179 | struct file * file; |
@@ -369,7 +369,7 @@ static inline void file_pos_write(struct file *file, loff_t pos) | |||
369 | file->f_pos = pos; | 369 | file->f_pos = pos; |
370 | } | 370 | } |
371 | 371 | ||
372 | asmlinkage ssize_t sys_read(unsigned int fd, char __user * buf, size_t count) | 372 | SYSCALL_DEFINE3(read, unsigned int, fd, char __user *, buf, size_t, count) |
373 | { | 373 | { |
374 | struct file *file; | 374 | struct file *file; |
375 | ssize_t ret = -EBADF; | 375 | ssize_t ret = -EBADF; |
@@ -386,7 +386,8 @@ asmlinkage ssize_t sys_read(unsigned int fd, char __user * buf, size_t count) | |||
386 | return ret; | 386 | return ret; |
387 | } | 387 | } |
388 | 388 | ||
389 | asmlinkage ssize_t sys_write(unsigned int fd, const char __user * buf, size_t count) | 389 | SYSCALL_DEFINE3(write, unsigned int, fd, const char __user *, buf, |
390 | size_t, count) | ||
390 | { | 391 | { |
391 | struct file *file; | 392 | struct file *file; |
392 | ssize_t ret = -EBADF; | 393 | ssize_t ret = -EBADF; |
@@ -403,8 +404,8 @@ asmlinkage ssize_t sys_write(unsigned int fd, const char __user * buf, size_t co | |||
403 | return ret; | 404 | return ret; |
404 | } | 405 | } |
405 | 406 | ||
406 | asmlinkage ssize_t sys_pread64(unsigned int fd, char __user *buf, | 407 | SYSCALL_DEFINE(pread64)(unsigned int fd, char __user *buf, |
407 | size_t count, loff_t pos) | 408 | size_t count, loff_t pos) |
408 | { | 409 | { |
409 | struct file *file; | 410 | struct file *file; |
410 | ssize_t ret = -EBADF; | 411 | ssize_t ret = -EBADF; |
@@ -423,9 +424,17 @@ asmlinkage ssize_t sys_pread64(unsigned int fd, char __user *buf, | |||
423 | 424 | ||
424 | return ret; | 425 | return ret; |
425 | } | 426 | } |
427 | #ifdef CONFIG_HAVE_SYSCALL_WRAPPERS | ||
428 | asmlinkage long SyS_pread64(long fd, long buf, long count, loff_t pos) | ||
429 | { | ||
430 | return SYSC_pread64((unsigned int) fd, (char __user *) buf, | ||
431 | (size_t) count, pos); | ||
432 | } | ||
433 | SYSCALL_ALIAS(sys_pread64, SyS_pread64); | ||
434 | #endif | ||
426 | 435 | ||
427 | asmlinkage ssize_t sys_pwrite64(unsigned int fd, const char __user *buf, | 436 | SYSCALL_DEFINE(pwrite64)(unsigned int fd, const char __user *buf, |
428 | size_t count, loff_t pos) | 437 | size_t count, loff_t pos) |
429 | { | 438 | { |
430 | struct file *file; | 439 | struct file *file; |
431 | ssize_t ret = -EBADF; | 440 | ssize_t ret = -EBADF; |
@@ -444,6 +453,14 @@ asmlinkage ssize_t sys_pwrite64(unsigned int fd, const char __user *buf, | |||
444 | 453 | ||
445 | return ret; | 454 | return ret; |
446 | } | 455 | } |
456 | #ifdef CONFIG_HAVE_SYSCALL_WRAPPERS | ||
457 | asmlinkage long SyS_pwrite64(long fd, long buf, long count, loff_t pos) | ||
458 | { | ||
459 | return SYSC_pwrite64((unsigned int) fd, (const char __user *) buf, | ||
460 | (size_t) count, pos); | ||
461 | } | ||
462 | SYSCALL_ALIAS(sys_pwrite64, SyS_pwrite64); | ||
463 | #endif | ||
447 | 464 | ||
448 | /* | 465 | /* |
449 | * Reduce an iovec's length in-place. Return the resulting number of segments | 466 | * Reduce an iovec's length in-place. Return the resulting number of segments |
@@ -672,8 +689,8 @@ ssize_t vfs_writev(struct file *file, const struct iovec __user *vec, | |||
672 | 689 | ||
673 | EXPORT_SYMBOL(vfs_writev); | 690 | EXPORT_SYMBOL(vfs_writev); |
674 | 691 | ||
675 | asmlinkage ssize_t | 692 | SYSCALL_DEFINE3(readv, unsigned long, fd, const struct iovec __user *, vec, |
676 | sys_readv(unsigned long fd, const struct iovec __user *vec, unsigned long vlen) | 693 | unsigned long, vlen) |
677 | { | 694 | { |
678 | struct file *file; | 695 | struct file *file; |
679 | ssize_t ret = -EBADF; | 696 | ssize_t ret = -EBADF; |
@@ -693,8 +710,8 @@ sys_readv(unsigned long fd, const struct iovec __user *vec, unsigned long vlen) | |||
693 | return ret; | 710 | return ret; |
694 | } | 711 | } |
695 | 712 | ||
696 | asmlinkage ssize_t | 713 | SYSCALL_DEFINE3(writev, unsigned long, fd, const struct iovec __user *, vec, |
697 | sys_writev(unsigned long fd, const struct iovec __user *vec, unsigned long vlen) | 714 | unsigned long, vlen) |
698 | { | 715 | { |
699 | struct file *file; | 716 | struct file *file; |
700 | ssize_t ret = -EBADF; | 717 | ssize_t ret = -EBADF; |
@@ -812,7 +829,7 @@ out: | |||
812 | return retval; | 829 | return retval; |
813 | } | 830 | } |
814 | 831 | ||
815 | asmlinkage ssize_t sys_sendfile(int out_fd, int in_fd, off_t __user *offset, size_t count) | 832 | SYSCALL_DEFINE4(sendfile, int, out_fd, int, in_fd, off_t __user *, offset, size_t, count) |
816 | { | 833 | { |
817 | loff_t pos; | 834 | loff_t pos; |
818 | off_t off; | 835 | off_t off; |
@@ -831,7 +848,7 @@ asmlinkage ssize_t sys_sendfile(int out_fd, int in_fd, off_t __user *offset, siz | |||
831 | return do_sendfile(out_fd, in_fd, NULL, count, 0); | 848 | return do_sendfile(out_fd, in_fd, NULL, count, 0); |
832 | } | 849 | } |
833 | 850 | ||
834 | asmlinkage ssize_t sys_sendfile64(int out_fd, int in_fd, loff_t __user *offset, size_t count) | 851 | SYSCALL_DEFINE4(sendfile64, int, out_fd, int, in_fd, loff_t __user *, offset, size_t, count) |
835 | { | 852 | { |
836 | loff_t pos; | 853 | loff_t pos; |
837 | ssize_t ret; | 854 | ssize_t ret; |
diff --git a/fs/readdir.c b/fs/readdir.c index b318d9b5af2e..7723401f8d8b 100644 --- a/fs/readdir.c +++ b/fs/readdir.c | |||
@@ -102,7 +102,8 @@ efault: | |||
102 | return -EFAULT; | 102 | return -EFAULT; |
103 | } | 103 | } |
104 | 104 | ||
105 | asmlinkage long old_readdir(unsigned int fd, struct old_linux_dirent __user * dirent, unsigned int count) | 105 | SYSCALL_DEFINE3(old_readdir, unsigned int, fd, |
106 | struct old_linux_dirent __user *, dirent, unsigned int, count) | ||
106 | { | 107 | { |
107 | int error; | 108 | int error; |
108 | struct file * file; | 109 | struct file * file; |
@@ -187,7 +188,8 @@ efault: | |||
187 | return -EFAULT; | 188 | return -EFAULT; |
188 | } | 189 | } |
189 | 190 | ||
190 | asmlinkage long sys_getdents(unsigned int fd, struct linux_dirent __user * dirent, unsigned int count) | 191 | SYSCALL_DEFINE3(getdents, unsigned int, fd, |
192 | struct linux_dirent __user *, dirent, unsigned int, count) | ||
191 | { | 193 | { |
192 | struct file * file; | 194 | struct file * file; |
193 | struct linux_dirent __user * lastdirent; | 195 | struct linux_dirent __user * lastdirent; |
@@ -268,7 +270,8 @@ efault: | |||
268 | return -EFAULT; | 270 | return -EFAULT; |
269 | } | 271 | } |
270 | 272 | ||
271 | asmlinkage long sys_getdents64(unsigned int fd, struct linux_dirent64 __user * dirent, unsigned int count) | 273 | SYSCALL_DEFINE3(getdents64, unsigned int, fd, |
274 | struct linux_dirent64 __user *, dirent, unsigned int, count) | ||
272 | { | 275 | { |
273 | struct file * file; | 276 | struct file * file; |
274 | struct linux_dirent64 __user * lastdirent; | 277 | struct linux_dirent64 __user * lastdirent; |
diff --git a/fs/reiserfs/Kconfig b/fs/reiserfs/Kconfig new file mode 100644 index 000000000000..949b8c6addc8 --- /dev/null +++ b/fs/reiserfs/Kconfig | |||
@@ -0,0 +1,85 @@ | |||
1 | config REISERFS_FS | ||
2 | tristate "Reiserfs support" | ||
3 | help | ||
4 | Stores not just filenames but the files themselves in a balanced | ||
5 | tree. Uses journalling. | ||
6 | |||
7 | Balanced trees are more efficient than traditional file system | ||
8 | architectural foundations. | ||
9 | |||
10 | In general, ReiserFS is as fast as ext2, but is very efficient with | ||
11 | large directories and small files. Additional patches are needed | ||
12 | for NFS and quotas, please see <http://www.namesys.com/> for links. | ||
13 | |||
14 | It is more easily extended to have features currently found in | ||
15 | database and keyword search systems than block allocation based file | ||
16 | systems are. The next version will be so extended, and will support | ||
17 | plugins consistent with our motto ``It takes more than a license to | ||
18 | make source code open.'' | ||
19 | |||
20 | Read <http://www.namesys.com/> to learn more about reiserfs. | ||
21 | |||
22 | Sponsored by Threshold Networks, Emusic.com, and Bigstorage.com. | ||
23 | |||
24 | If you like it, you can pay us to add new features to it that you | ||
25 | need, buy a support contract, or pay us to port it to another OS. | ||
26 | |||
27 | config REISERFS_CHECK | ||
28 | bool "Enable reiserfs debug mode" | ||
29 | depends on REISERFS_FS | ||
30 | help | ||
31 | If you set this to Y, then ReiserFS will perform every check it can | ||
32 | possibly imagine of its internal consistency throughout its | ||
33 | operation. It will also go substantially slower. More than once we | ||
34 | have forgotten that this was on, and then gone despondent over the | ||
35 | latest benchmarks.:-) Use of this option allows our team to go all | ||
36 | out in checking for consistency when debugging without fear of its | ||
37 | effect on end users. If you are on the verge of sending in a bug | ||
38 | report, say Y and you might get a useful error message. Almost | ||
39 | everyone should say N. | ||
40 | |||
41 | config REISERFS_PROC_INFO | ||
42 | bool "Stats in /proc/fs/reiserfs" | ||
43 | depends on REISERFS_FS && PROC_FS | ||
44 | help | ||
45 | Create under /proc/fs/reiserfs a hierarchy of files, displaying | ||
46 | various ReiserFS statistics and internal data at the expense of | ||
47 | making your kernel or module slightly larger (+8 KB). This also | ||
48 | increases the amount of kernel memory required for each mount. | ||
49 | Almost everyone but ReiserFS developers and people fine-tuning | ||
50 | reiserfs or tracing problems should say N. | ||
51 | |||
52 | config REISERFS_FS_XATTR | ||
53 | bool "ReiserFS extended attributes" | ||
54 | depends on REISERFS_FS | ||
55 | help | ||
56 | Extended attributes are name:value pairs associated with inodes by | ||
57 | the kernel or by users (see the attr(5) manual page, or visit | ||
58 | <http://acl.bestbits.at/> for details). | ||
59 | |||
60 | If unsure, say N. | ||
61 | |||
62 | config REISERFS_FS_POSIX_ACL | ||
63 | bool "ReiserFS POSIX Access Control Lists" | ||
64 | depends on REISERFS_FS_XATTR | ||
65 | select FS_POSIX_ACL | ||
66 | help | ||
67 | Posix Access Control Lists (ACLs) support permissions for users and | ||
68 | groups beyond the owner/group/world scheme. | ||
69 | |||
70 | To learn more about Access Control Lists, visit the Posix ACLs for | ||
71 | Linux website <http://acl.bestbits.at/>. | ||
72 | |||
73 | If you don't know what Access Control Lists are, say N | ||
74 | |||
75 | config REISERFS_FS_SECURITY | ||
76 | bool "ReiserFS Security Labels" | ||
77 | depends on REISERFS_FS_XATTR | ||
78 | help | ||
79 | Security labels support alternative access control models | ||
80 | implemented by security modules like SELinux. This option | ||
81 | enables an extended attribute handler for file security | ||
82 | labels in the ReiserFS filesystem. | ||
83 | |||
84 | If you are not using a security module that requires using | ||
85 | extended attributes for file security labels, say N. | ||
diff --git a/fs/romfs/Kconfig b/fs/romfs/Kconfig new file mode 100644 index 000000000000..1a17020f9faf --- /dev/null +++ b/fs/romfs/Kconfig | |||
@@ -0,0 +1,16 @@ | |||
1 | config ROMFS_FS | ||
2 | tristate "ROM file system support" | ||
3 | depends on BLOCK | ||
4 | ---help--- | ||
5 | This is a very small read-only file system mainly intended for | ||
6 | initial ram disks of installation disks, but it could be used for | ||
7 | other read-only media as well. Read | ||
8 | <file:Documentation/filesystems/romfs.txt> for details. | ||
9 | |||
10 | To compile this file system support as a module, choose M here: the | ||
11 | module will be called romfs. Note that the file system of your | ||
12 | root partition (the one containing the directory /) cannot be a | ||
13 | module. | ||
14 | |||
15 | If you don't know whether you need it, then you don't need it: | ||
16 | answer N. | ||
diff --git a/fs/select.c b/fs/select.c index 08b91beed806..0fe0e1469df3 100644 --- a/fs/select.c +++ b/fs/select.c | |||
@@ -557,8 +557,8 @@ out_nofds: | |||
557 | return ret; | 557 | return ret; |
558 | } | 558 | } |
559 | 559 | ||
560 | asmlinkage long sys_select(int n, fd_set __user *inp, fd_set __user *outp, | 560 | SYSCALL_DEFINE5(select, int, n, fd_set __user *, inp, fd_set __user *, outp, |
561 | fd_set __user *exp, struct timeval __user *tvp) | 561 | fd_set __user *, exp, struct timeval __user *, tvp) |
562 | { | 562 | { |
563 | struct timespec end_time, *to = NULL; | 563 | struct timespec end_time, *to = NULL; |
564 | struct timeval tv; | 564 | struct timeval tv; |
@@ -582,9 +582,9 @@ asmlinkage long sys_select(int n, fd_set __user *inp, fd_set __user *outp, | |||
582 | } | 582 | } |
583 | 583 | ||
584 | #ifdef HAVE_SET_RESTORE_SIGMASK | 584 | #ifdef HAVE_SET_RESTORE_SIGMASK |
585 | asmlinkage long sys_pselect7(int n, fd_set __user *inp, fd_set __user *outp, | 585 | static long do_pselect(int n, fd_set __user *inp, fd_set __user *outp, |
586 | fd_set __user *exp, struct timespec __user *tsp, | 586 | fd_set __user *exp, struct timespec __user *tsp, |
587 | const sigset_t __user *sigmask, size_t sigsetsize) | 587 | const sigset_t __user *sigmask, size_t sigsetsize) |
588 | { | 588 | { |
589 | sigset_t ksigmask, sigsaved; | 589 | sigset_t ksigmask, sigsaved; |
590 | struct timespec ts, end_time, *to = NULL; | 590 | struct timespec ts, end_time, *to = NULL; |
@@ -610,7 +610,7 @@ asmlinkage long sys_pselect7(int n, fd_set __user *inp, fd_set __user *outp, | |||
610 | sigprocmask(SIG_SETMASK, &ksigmask, &sigsaved); | 610 | sigprocmask(SIG_SETMASK, &ksigmask, &sigsaved); |
611 | } | 611 | } |
612 | 612 | ||
613 | ret = core_sys_select(n, inp, outp, exp, &end_time); | 613 | ret = core_sys_select(n, inp, outp, exp, to); |
614 | ret = poll_select_copy_remaining(&end_time, tsp, 0, ret); | 614 | ret = poll_select_copy_remaining(&end_time, tsp, 0, ret); |
615 | 615 | ||
616 | if (ret == -ERESTARTNOHAND) { | 616 | if (ret == -ERESTARTNOHAND) { |
@@ -636,8 +636,9 @@ asmlinkage long sys_pselect7(int n, fd_set __user *inp, fd_set __user *outp, | |||
636 | * which has a pointer to the sigset_t itself followed by a size_t containing | 636 | * which has a pointer to the sigset_t itself followed by a size_t containing |
637 | * the sigset size. | 637 | * the sigset size. |
638 | */ | 638 | */ |
639 | asmlinkage long sys_pselect6(int n, fd_set __user *inp, fd_set __user *outp, | 639 | SYSCALL_DEFINE6(pselect6, int, n, fd_set __user *, inp, fd_set __user *, outp, |
640 | fd_set __user *exp, struct timespec __user *tsp, void __user *sig) | 640 | fd_set __user *, exp, struct timespec __user *, tsp, |
641 | void __user *, sig) | ||
641 | { | 642 | { |
642 | size_t sigsetsize = 0; | 643 | size_t sigsetsize = 0; |
643 | sigset_t __user *up = NULL; | 644 | sigset_t __user *up = NULL; |
@@ -650,7 +651,7 @@ asmlinkage long sys_pselect6(int n, fd_set __user *inp, fd_set __user *outp, | |||
650 | return -EFAULT; | 651 | return -EFAULT; |
651 | } | 652 | } |
652 | 653 | ||
653 | return sys_pselect7(n, inp, outp, exp, tsp, up, sigsetsize); | 654 | return do_pselect(n, inp, outp, exp, tsp, up, sigsetsize); |
654 | } | 655 | } |
655 | #endif /* HAVE_SET_RESTORE_SIGMASK */ | 656 | #endif /* HAVE_SET_RESTORE_SIGMASK */ |
656 | 657 | ||
@@ -854,8 +855,8 @@ static long do_restart_poll(struct restart_block *restart_block) | |||
854 | return ret; | 855 | return ret; |
855 | } | 856 | } |
856 | 857 | ||
857 | asmlinkage long sys_poll(struct pollfd __user *ufds, unsigned int nfds, | 858 | SYSCALL_DEFINE3(poll, struct pollfd __user *, ufds, unsigned int, nfds, |
858 | long timeout_msecs) | 859 | long, timeout_msecs) |
859 | { | 860 | { |
860 | struct timespec end_time, *to = NULL; | 861 | struct timespec end_time, *to = NULL; |
861 | int ret; | 862 | int ret; |
@@ -889,9 +890,9 @@ asmlinkage long sys_poll(struct pollfd __user *ufds, unsigned int nfds, | |||
889 | } | 890 | } |
890 | 891 | ||
891 | #ifdef HAVE_SET_RESTORE_SIGMASK | 892 | #ifdef HAVE_SET_RESTORE_SIGMASK |
892 | asmlinkage long sys_ppoll(struct pollfd __user *ufds, unsigned int nfds, | 893 | SYSCALL_DEFINE5(ppoll, struct pollfd __user *, ufds, unsigned int, nfds, |
893 | struct timespec __user *tsp, const sigset_t __user *sigmask, | 894 | struct timespec __user *, tsp, const sigset_t __user *, sigmask, |
894 | size_t sigsetsize) | 895 | size_t, sigsetsize) |
895 | { | 896 | { |
896 | sigset_t ksigmask, sigsaved; | 897 | sigset_t ksigmask, sigsaved; |
897 | struct timespec ts, end_time, *to = NULL; | 898 | struct timespec ts, end_time, *to = NULL; |
diff --git a/fs/seq_file.c b/fs/seq_file.c index b569ff1c4dc8..a1a4cfe19210 100644 --- a/fs/seq_file.c +++ b/fs/seq_file.c | |||
@@ -48,12 +48,78 @@ int seq_open(struct file *file, const struct seq_operations *op) | |||
48 | */ | 48 | */ |
49 | file->f_version = 0; | 49 | file->f_version = 0; |
50 | 50 | ||
51 | /* SEQ files support lseek, but not pread/pwrite */ | 51 | /* |
52 | file->f_mode &= ~(FMODE_PREAD | FMODE_PWRITE); | 52 | * seq_files support lseek() and pread(). They do not implement |
53 | * write() at all, but we clear FMODE_PWRITE here for historical | ||
54 | * reasons. | ||
55 | * | ||
56 | * If a client of seq_files a) implements file.write() and b) wishes to | ||
57 | * support pwrite() then that client will need to implement its own | ||
58 | * file.open() which calls seq_open() and then sets FMODE_PWRITE. | ||
59 | */ | ||
60 | file->f_mode &= ~FMODE_PWRITE; | ||
53 | return 0; | 61 | return 0; |
54 | } | 62 | } |
55 | EXPORT_SYMBOL(seq_open); | 63 | EXPORT_SYMBOL(seq_open); |
56 | 64 | ||
65 | static int traverse(struct seq_file *m, loff_t offset) | ||
66 | { | ||
67 | loff_t pos = 0, index; | ||
68 | int error = 0; | ||
69 | void *p; | ||
70 | |||
71 | m->version = 0; | ||
72 | index = 0; | ||
73 | m->count = m->from = 0; | ||
74 | if (!offset) { | ||
75 | m->index = index; | ||
76 | return 0; | ||
77 | } | ||
78 | if (!m->buf) { | ||
79 | m->buf = kmalloc(m->size = PAGE_SIZE, GFP_KERNEL); | ||
80 | if (!m->buf) | ||
81 | return -ENOMEM; | ||
82 | } | ||
83 | p = m->op->start(m, &index); | ||
84 | while (p) { | ||
85 | error = PTR_ERR(p); | ||
86 | if (IS_ERR(p)) | ||
87 | break; | ||
88 | error = m->op->show(m, p); | ||
89 | if (error < 0) | ||
90 | break; | ||
91 | if (unlikely(error)) { | ||
92 | error = 0; | ||
93 | m->count = 0; | ||
94 | } | ||
95 | if (m->count == m->size) | ||
96 | goto Eoverflow; | ||
97 | if (pos + m->count > offset) { | ||
98 | m->from = offset - pos; | ||
99 | m->count -= m->from; | ||
100 | m->index = index; | ||
101 | break; | ||
102 | } | ||
103 | pos += m->count; | ||
104 | m->count = 0; | ||
105 | if (pos == offset) { | ||
106 | index++; | ||
107 | m->index = index; | ||
108 | break; | ||
109 | } | ||
110 | p = m->op->next(m, p, &index); | ||
111 | } | ||
112 | m->op->stop(m, p); | ||
113 | m->index = index; | ||
114 | return error; | ||
115 | |||
116 | Eoverflow: | ||
117 | m->op->stop(m, p); | ||
118 | kfree(m->buf); | ||
119 | m->buf = kmalloc(m->size <<= 1, GFP_KERNEL); | ||
120 | return !m->buf ? -ENOMEM : -EAGAIN; | ||
121 | } | ||
122 | |||
57 | /** | 123 | /** |
58 | * seq_read - ->read() method for sequential files. | 124 | * seq_read - ->read() method for sequential files. |
59 | * @file: the file to read from | 125 | * @file: the file to read from |
@@ -73,6 +139,22 @@ ssize_t seq_read(struct file *file, char __user *buf, size_t size, loff_t *ppos) | |||
73 | int err = 0; | 139 | int err = 0; |
74 | 140 | ||
75 | mutex_lock(&m->lock); | 141 | mutex_lock(&m->lock); |
142 | |||
143 | /* Don't assume *ppos is where we left it */ | ||
144 | if (unlikely(*ppos != m->read_pos)) { | ||
145 | m->read_pos = *ppos; | ||
146 | while ((err = traverse(m, *ppos)) == -EAGAIN) | ||
147 | ; | ||
148 | if (err) { | ||
149 | /* With prejudice... */ | ||
150 | m->read_pos = 0; | ||
151 | m->version = 0; | ||
152 | m->index = 0; | ||
153 | m->count = 0; | ||
154 | goto Done; | ||
155 | } | ||
156 | } | ||
157 | |||
76 | /* | 158 | /* |
77 | * seq_file->op->..m_start/m_stop/m_next may do special actions | 159 | * seq_file->op->..m_start/m_stop/m_next may do special actions |
78 | * or optimisations based on the file->f_version, so we want to | 160 | * or optimisations based on the file->f_version, so we want to |
@@ -172,8 +254,10 @@ Fill: | |||
172 | Done: | 254 | Done: |
173 | if (!copied) | 255 | if (!copied) |
174 | copied = err; | 256 | copied = err; |
175 | else | 257 | else { |
176 | *ppos += copied; | 258 | *ppos += copied; |
259 | m->read_pos += copied; | ||
260 | } | ||
177 | file->f_version = m->version; | 261 | file->f_version = m->version; |
178 | mutex_unlock(&m->lock); | 262 | mutex_unlock(&m->lock); |
179 | return copied; | 263 | return copied; |
@@ -186,63 +270,6 @@ Efault: | |||
186 | } | 270 | } |
187 | EXPORT_SYMBOL(seq_read); | 271 | EXPORT_SYMBOL(seq_read); |
188 | 272 | ||
189 | static int traverse(struct seq_file *m, loff_t offset) | ||
190 | { | ||
191 | loff_t pos = 0, index; | ||
192 | int error = 0; | ||
193 | void *p; | ||
194 | |||
195 | m->version = 0; | ||
196 | index = 0; | ||
197 | m->count = m->from = 0; | ||
198 | if (!offset) { | ||
199 | m->index = index; | ||
200 | return 0; | ||
201 | } | ||
202 | if (!m->buf) { | ||
203 | m->buf = kmalloc(m->size = PAGE_SIZE, GFP_KERNEL); | ||
204 | if (!m->buf) | ||
205 | return -ENOMEM; | ||
206 | } | ||
207 | p = m->op->start(m, &index); | ||
208 | while (p) { | ||
209 | error = PTR_ERR(p); | ||
210 | if (IS_ERR(p)) | ||
211 | break; | ||
212 | error = m->op->show(m, p); | ||
213 | if (error < 0) | ||
214 | break; | ||
215 | if (unlikely(error)) { | ||
216 | error = 0; | ||
217 | m->count = 0; | ||
218 | } | ||
219 | if (m->count == m->size) | ||
220 | goto Eoverflow; | ||
221 | if (pos + m->count > offset) { | ||
222 | m->from = offset - pos; | ||
223 | m->count -= m->from; | ||
224 | m->index = index; | ||
225 | break; | ||
226 | } | ||
227 | pos += m->count; | ||
228 | m->count = 0; | ||
229 | if (pos == offset) { | ||
230 | index++; | ||
231 | m->index = index; | ||
232 | break; | ||
233 | } | ||
234 | p = m->op->next(m, p, &index); | ||
235 | } | ||
236 | m->op->stop(m, p); | ||
237 | return error; | ||
238 | |||
239 | Eoverflow: | ||
240 | m->op->stop(m, p); | ||
241 | kfree(m->buf); | ||
242 | m->buf = kmalloc(m->size <<= 1, GFP_KERNEL); | ||
243 | return !m->buf ? -ENOMEM : -EAGAIN; | ||
244 | } | ||
245 | |||
246 | /** | 273 | /** |
247 | * seq_lseek - ->llseek() method for sequential files. | 274 | * seq_lseek - ->llseek() method for sequential files. |
248 | * @file: the file in question | 275 | * @file: the file in question |
@@ -265,16 +292,18 @@ loff_t seq_lseek(struct file *file, loff_t offset, int origin) | |||
265 | if (offset < 0) | 292 | if (offset < 0) |
266 | break; | 293 | break; |
267 | retval = offset; | 294 | retval = offset; |
268 | if (offset != file->f_pos) { | 295 | if (offset != m->read_pos) { |
269 | while ((retval=traverse(m, offset)) == -EAGAIN) | 296 | while ((retval=traverse(m, offset)) == -EAGAIN) |
270 | ; | 297 | ; |
271 | if (retval) { | 298 | if (retval) { |
272 | /* with extreme prejudice... */ | 299 | /* with extreme prejudice... */ |
273 | file->f_pos = 0; | 300 | file->f_pos = 0; |
301 | m->read_pos = 0; | ||
274 | m->version = 0; | 302 | m->version = 0; |
275 | m->index = 0; | 303 | m->index = 0; |
276 | m->count = 0; | 304 | m->count = 0; |
277 | } else { | 305 | } else { |
306 | m->read_pos = offset; | ||
278 | retval = file->f_pos = offset; | 307 | retval = file->f_pos = offset; |
279 | } | 308 | } |
280 | } | 309 | } |
diff --git a/fs/signalfd.c b/fs/signalfd.c index 9c39bc7f8431..b07565c94386 100644 --- a/fs/signalfd.c +++ b/fs/signalfd.c | |||
@@ -205,8 +205,8 @@ static const struct file_operations signalfd_fops = { | |||
205 | .read = signalfd_read, | 205 | .read = signalfd_read, |
206 | }; | 206 | }; |
207 | 207 | ||
208 | asmlinkage long sys_signalfd4(int ufd, sigset_t __user *user_mask, | 208 | SYSCALL_DEFINE4(signalfd4, int, ufd, sigset_t __user *, user_mask, |
209 | size_t sizemask, int flags) | 209 | size_t, sizemask, int, flags) |
210 | { | 210 | { |
211 | sigset_t sigmask; | 211 | sigset_t sigmask; |
212 | struct signalfd_ctx *ctx; | 212 | struct signalfd_ctx *ctx; |
@@ -259,8 +259,8 @@ asmlinkage long sys_signalfd4(int ufd, sigset_t __user *user_mask, | |||
259 | return ufd; | 259 | return ufd; |
260 | } | 260 | } |
261 | 261 | ||
262 | asmlinkage long sys_signalfd(int ufd, sigset_t __user *user_mask, | 262 | SYSCALL_DEFINE3(signalfd, int, ufd, sigset_t __user *, user_mask, |
263 | size_t sizemask) | 263 | size_t, sizemask) |
264 | { | 264 | { |
265 | return sys_signalfd4(ufd, user_mask, sizemask, 0); | 265 | return sys_signalfd4(ufd, user_mask, sizemask, 0); |
266 | } | 266 | } |
diff --git a/fs/smbfs/Kconfig b/fs/smbfs/Kconfig new file mode 100644 index 000000000000..e668127c8b2e --- /dev/null +++ b/fs/smbfs/Kconfig | |||
@@ -0,0 +1,55 @@ | |||
1 | config SMB_FS | ||
2 | tristate "SMB file system support (OBSOLETE, please use CIFS)" | ||
3 | depends on INET | ||
4 | select NLS | ||
5 | help | ||
6 | SMB (Server Message Block) is the protocol Windows for Workgroups | ||
7 | (WfW), Windows 95/98, Windows NT and OS/2 Lan Manager use to share | ||
8 | files and printers over local networks. Saying Y here allows you to | ||
9 | mount their file systems (often called "shares" in this context) and | ||
10 | access them just like any other Unix directory. Currently, this | ||
11 | works only if the Windows machines use TCP/IP as the underlying | ||
12 | transport protocol, and not NetBEUI. For details, read | ||
13 | <file:Documentation/filesystems/smbfs.txt> and the SMB-HOWTO, | ||
14 | available from <http://www.tldp.org/docs.html#howto>. | ||
15 | |||
16 | Note: if you just want your box to act as an SMB *server* and make | ||
17 | files and printing services available to Windows clients (which need | ||
18 | to have a TCP/IP stack), you don't need to say Y here; you can use | ||
19 | the program SAMBA (available from <ftp://ftp.samba.org/pub/samba/>) | ||
20 | for that. | ||
21 | |||
22 | General information about how to connect Linux, Windows machines and | ||
23 | Macs is on the WWW at <http://www.eats.com/linux_mac_win.html>. | ||
24 | |||
25 | To compile the SMB support as a module, choose M here: | ||
26 | the module will be called smbfs. Most people say N, however. | ||
27 | |||
28 | config SMB_NLS_DEFAULT | ||
29 | bool "Use a default NLS" | ||
30 | depends on SMB_FS | ||
31 | help | ||
32 | Enabling this will make smbfs use nls translations by default. You | ||
33 | need to specify the local charset (CONFIG_NLS_DEFAULT) in the nls | ||
34 | settings and you need to give the default nls for the SMB server as | ||
35 | CONFIG_SMB_NLS_REMOTE. | ||
36 | |||
37 | The nls settings can be changed at mount time, if your smbmount | ||
38 | supports that, using the codepage and iocharset parameters. | ||
39 | |||
40 | smbmount from samba 2.2.0 or later supports this. | ||
41 | |||
42 | config SMB_NLS_REMOTE | ||
43 | string "Default Remote NLS Option" | ||
44 | depends on SMB_NLS_DEFAULT | ||
45 | default "cp437" | ||
46 | help | ||
47 | This setting allows you to specify a default value for which | ||
48 | codepage the server uses. If this field is left blank no | ||
49 | translations will be done by default. The local codepage/charset | ||
50 | default to CONFIG_NLS_DEFAULT. | ||
51 | |||
52 | The nls settings can be changed at mount time, if your smbmount | ||
53 | supports that, using the codepage and iocharset parameters. | ||
54 | |||
55 | smbmount from samba 2.2.0 or later supports this. | ||
diff --git a/fs/splice.c b/fs/splice.c index a54b3e3f10a7..4ed0ba44a966 100644 --- a/fs/splice.c +++ b/fs/splice.c | |||
@@ -1435,8 +1435,8 @@ static long vmsplice_to_pipe(struct file *file, const struct iovec __user *iov, | |||
1435 | * Currently we punt and implement it as a normal copy, see pipe_to_user(). | 1435 | * Currently we punt and implement it as a normal copy, see pipe_to_user(). |
1436 | * | 1436 | * |
1437 | */ | 1437 | */ |
1438 | asmlinkage long sys_vmsplice(int fd, const struct iovec __user *iov, | 1438 | SYSCALL_DEFINE4(vmsplice, int, fd, const struct iovec __user *, iov, |
1439 | unsigned long nr_segs, unsigned int flags) | 1439 | unsigned long, nr_segs, unsigned int, flags) |
1440 | { | 1440 | { |
1441 | struct file *file; | 1441 | struct file *file; |
1442 | long error; | 1442 | long error; |
@@ -1461,9 +1461,9 @@ asmlinkage long sys_vmsplice(int fd, const struct iovec __user *iov, | |||
1461 | return error; | 1461 | return error; |
1462 | } | 1462 | } |
1463 | 1463 | ||
1464 | asmlinkage long sys_splice(int fd_in, loff_t __user *off_in, | 1464 | SYSCALL_DEFINE6(splice, int, fd_in, loff_t __user *, off_in, |
1465 | int fd_out, loff_t __user *off_out, | 1465 | int, fd_out, loff_t __user *, off_out, |
1466 | size_t len, unsigned int flags) | 1466 | size_t, len, unsigned int, flags) |
1467 | { | 1467 | { |
1468 | long error; | 1468 | long error; |
1469 | struct file *in, *out; | 1469 | struct file *in, *out; |
@@ -1685,7 +1685,7 @@ static long do_tee(struct file *in, struct file *out, size_t len, | |||
1685 | return ret; | 1685 | return ret; |
1686 | } | 1686 | } |
1687 | 1687 | ||
1688 | asmlinkage long sys_tee(int fdin, int fdout, size_t len, unsigned int flags) | 1688 | SYSCALL_DEFINE4(tee, int, fdin, int, fdout, size_t, len, unsigned int, flags) |
1689 | { | 1689 | { |
1690 | struct file *in; | 1690 | struct file *in; |
1691 | int error, fput_in; | 1691 | int error, fput_in; |
diff --git a/fs/squashfs/Kconfig b/fs/squashfs/Kconfig new file mode 100644 index 000000000000..25a00d19d686 --- /dev/null +++ b/fs/squashfs/Kconfig | |||
@@ -0,0 +1,51 @@ | |||
1 | config SQUASHFS | ||
2 | tristate "SquashFS 4.0 - Squashed file system support" | ||
3 | depends on BLOCK | ||
4 | select ZLIB_INFLATE | ||
5 | help | ||
6 | Saying Y here includes support for SquashFS 4.0 (a Compressed | ||
7 | Read-Only File System). Squashfs is a highly compressed read-only | ||
8 | filesystem for Linux. It uses zlib compression to compress both | ||
9 | files, inodes and directories. Inodes in the system are very small | ||
10 | and all blocks are packed to minimise data overhead. Block sizes | ||
11 | greater than 4K are supported up to a maximum of 1 Mbytes (default | ||
12 | block size 128K). SquashFS 4.0 supports 64 bit filesystems and files | ||
13 | (larger than 4GB), full uid/gid information, hard links and | ||
14 | timestamps. | ||
15 | |||
16 | Squashfs is intended for general read-only filesystem use, for | ||
17 | archival use (i.e. in cases where a .tar.gz file may be used), and in | ||
18 | embedded systems where low overhead is needed. Further information | ||
19 | and tools are available from http://squashfs.sourceforge.net. | ||
20 | |||
21 | If you want to compile this as a module ( = code which can be | ||
22 | inserted in and removed from the running kernel whenever you want), | ||
23 | say M here and read <file:Documentation/modules.txt>. The module | ||
24 | will be called squashfs. Note that the root file system (the one | ||
25 | containing the directory /) cannot be compiled as a module. | ||
26 | |||
27 | If unsure, say N. | ||
28 | |||
29 | config SQUASHFS_EMBEDDED | ||
30 | |||
31 | bool "Additional option for memory-constrained systems" | ||
32 | depends on SQUASHFS | ||
33 | default n | ||
34 | help | ||
35 | Saying Y here allows you to specify cache size. | ||
36 | |||
37 | If unsure, say N. | ||
38 | |||
39 | config SQUASHFS_FRAGMENT_CACHE_SIZE | ||
40 | int "Number of fragments cached" if SQUASHFS_EMBEDDED | ||
41 | depends on SQUASHFS | ||
42 | default "3" | ||
43 | help | ||
44 | By default SquashFS caches the last 3 fragments read from | ||
45 | the filesystem. Increasing this amount may mean SquashFS | ||
46 | has to re-read fragments less often from disk, at the expense | ||
47 | of extra system memory. Decreasing this amount will mean | ||
48 | SquashFS uses less memory at the expense of extra reads from disk. | ||
49 | |||
50 | Note there must be at least one cached fragment. Anything | ||
51 | much more than three will probably not make much difference. | ||
diff --git a/fs/squashfs/squashfs_fs.h b/fs/squashfs/squashfs_fs.h index 6840da1bf21e..283daafc568e 100644 --- a/fs/squashfs/squashfs_fs.h +++ b/fs/squashfs/squashfs_fs.h | |||
@@ -26,7 +26,6 @@ | |||
26 | #define SQUASHFS_CACHED_FRAGMENTS CONFIG_SQUASHFS_FRAGMENT_CACHE_SIZE | 26 | #define SQUASHFS_CACHED_FRAGMENTS CONFIG_SQUASHFS_FRAGMENT_CACHE_SIZE |
27 | #define SQUASHFS_MAJOR 4 | 27 | #define SQUASHFS_MAJOR 4 |
28 | #define SQUASHFS_MINOR 0 | 28 | #define SQUASHFS_MINOR 0 |
29 | #define SQUASHFS_MAGIC 0x73717368 | ||
30 | #define SQUASHFS_START 0 | 29 | #define SQUASHFS_START 0 |
31 | 30 | ||
32 | /* size of metadata (inode and directory) blocks */ | 31 | /* size of metadata (inode and directory) blocks */ |
diff --git a/fs/squashfs/super.c b/fs/squashfs/super.c index a0466d7467b2..071df5b5b491 100644 --- a/fs/squashfs/super.c +++ b/fs/squashfs/super.c | |||
@@ -35,6 +35,7 @@ | |||
35 | #include <linux/init.h> | 35 | #include <linux/init.h> |
36 | #include <linux/module.h> | 36 | #include <linux/module.h> |
37 | #include <linux/zlib.h> | 37 | #include <linux/zlib.h> |
38 | #include <linux/magic.h> | ||
38 | 39 | ||
39 | #include "squashfs_fs.h" | 40 | #include "squashfs_fs.h" |
40 | #include "squashfs_fs_sb.h" | 41 | #include "squashfs_fs_sb.h" |
@@ -152,7 +152,7 @@ static int cp_old_stat(struct kstat *stat, struct __old_kernel_stat __user * sta | |||
152 | return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; | 152 | return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; |
153 | } | 153 | } |
154 | 154 | ||
155 | asmlinkage long sys_stat(char __user * filename, struct __old_kernel_stat __user * statbuf) | 155 | SYSCALL_DEFINE2(stat, char __user *, filename, struct __old_kernel_stat __user *, statbuf) |
156 | { | 156 | { |
157 | struct kstat stat; | 157 | struct kstat stat; |
158 | int error = vfs_stat_fd(AT_FDCWD, filename, &stat); | 158 | int error = vfs_stat_fd(AT_FDCWD, filename, &stat); |
@@ -162,7 +162,8 @@ asmlinkage long sys_stat(char __user * filename, struct __old_kernel_stat __user | |||
162 | 162 | ||
163 | return error; | 163 | return error; |
164 | } | 164 | } |
165 | asmlinkage long sys_lstat(char __user * filename, struct __old_kernel_stat __user * statbuf) | 165 | |
166 | SYSCALL_DEFINE2(lstat, char __user *, filename, struct __old_kernel_stat __user *, statbuf) | ||
166 | { | 167 | { |
167 | struct kstat stat; | 168 | struct kstat stat; |
168 | int error = vfs_lstat_fd(AT_FDCWD, filename, &stat); | 169 | int error = vfs_lstat_fd(AT_FDCWD, filename, &stat); |
@@ -172,7 +173,8 @@ asmlinkage long sys_lstat(char __user * filename, struct __old_kernel_stat __use | |||
172 | 173 | ||
173 | return error; | 174 | return error; |
174 | } | 175 | } |
175 | asmlinkage long sys_fstat(unsigned int fd, struct __old_kernel_stat __user * statbuf) | 176 | |
177 | SYSCALL_DEFINE2(fstat, unsigned int, fd, struct __old_kernel_stat __user *, statbuf) | ||
176 | { | 178 | { |
177 | struct kstat stat; | 179 | struct kstat stat; |
178 | int error = vfs_fstat(fd, &stat); | 180 | int error = vfs_fstat(fd, &stat); |
@@ -235,7 +237,7 @@ static int cp_new_stat(struct kstat *stat, struct stat __user *statbuf) | |||
235 | return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; | 237 | return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; |
236 | } | 238 | } |
237 | 239 | ||
238 | asmlinkage long sys_newstat(char __user *filename, struct stat __user *statbuf) | 240 | SYSCALL_DEFINE2(newstat, char __user *, filename, struct stat __user *, statbuf) |
239 | { | 241 | { |
240 | struct kstat stat; | 242 | struct kstat stat; |
241 | int error = vfs_stat_fd(AT_FDCWD, filename, &stat); | 243 | int error = vfs_stat_fd(AT_FDCWD, filename, &stat); |
@@ -246,7 +248,7 @@ asmlinkage long sys_newstat(char __user *filename, struct stat __user *statbuf) | |||
246 | return error; | 248 | return error; |
247 | } | 249 | } |
248 | 250 | ||
249 | asmlinkage long sys_newlstat(char __user *filename, struct stat __user *statbuf) | 251 | SYSCALL_DEFINE2(newlstat, char __user *, filename, struct stat __user *, statbuf) |
250 | { | 252 | { |
251 | struct kstat stat; | 253 | struct kstat stat; |
252 | int error = vfs_lstat_fd(AT_FDCWD, filename, &stat); | 254 | int error = vfs_lstat_fd(AT_FDCWD, filename, &stat); |
@@ -258,8 +260,8 @@ asmlinkage long sys_newlstat(char __user *filename, struct stat __user *statbuf) | |||
258 | } | 260 | } |
259 | 261 | ||
260 | #if !defined(__ARCH_WANT_STAT64) || defined(__ARCH_WANT_SYS_NEWFSTATAT) | 262 | #if !defined(__ARCH_WANT_STAT64) || defined(__ARCH_WANT_SYS_NEWFSTATAT) |
261 | asmlinkage long sys_newfstatat(int dfd, char __user *filename, | 263 | SYSCALL_DEFINE4(newfstatat, int, dfd, char __user *, filename, |
262 | struct stat __user *statbuf, int flag) | 264 | struct stat __user *, statbuf, int, flag) |
263 | { | 265 | { |
264 | struct kstat stat; | 266 | struct kstat stat; |
265 | int error = -EINVAL; | 267 | int error = -EINVAL; |
@@ -280,7 +282,7 @@ out: | |||
280 | } | 282 | } |
281 | #endif | 283 | #endif |
282 | 284 | ||
283 | asmlinkage long sys_newfstat(unsigned int fd, struct stat __user *statbuf) | 285 | SYSCALL_DEFINE2(newfstat, unsigned int, fd, struct stat __user *, statbuf) |
284 | { | 286 | { |
285 | struct kstat stat; | 287 | struct kstat stat; |
286 | int error = vfs_fstat(fd, &stat); | 288 | int error = vfs_fstat(fd, &stat); |
@@ -291,8 +293,8 @@ asmlinkage long sys_newfstat(unsigned int fd, struct stat __user *statbuf) | |||
291 | return error; | 293 | return error; |
292 | } | 294 | } |
293 | 295 | ||
294 | asmlinkage long sys_readlinkat(int dfd, const char __user *pathname, | 296 | SYSCALL_DEFINE4(readlinkat, int, dfd, const char __user *, pathname, |
295 | char __user *buf, int bufsiz) | 297 | char __user *, buf, int, bufsiz) |
296 | { | 298 | { |
297 | struct path path; | 299 | struct path path; |
298 | int error; | 300 | int error; |
@@ -318,8 +320,8 @@ asmlinkage long sys_readlinkat(int dfd, const char __user *pathname, | |||
318 | return error; | 320 | return error; |
319 | } | 321 | } |
320 | 322 | ||
321 | asmlinkage long sys_readlink(const char __user *path, char __user *buf, | 323 | SYSCALL_DEFINE3(readlink, const char __user *, path, char __user *, buf, |
322 | int bufsiz) | 324 | int, bufsiz) |
323 | { | 325 | { |
324 | return sys_readlinkat(AT_FDCWD, path, buf, bufsiz); | 326 | return sys_readlinkat(AT_FDCWD, path, buf, bufsiz); |
325 | } | 327 | } |
@@ -365,7 +367,7 @@ static long cp_new_stat64(struct kstat *stat, struct stat64 __user *statbuf) | |||
365 | return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; | 367 | return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; |
366 | } | 368 | } |
367 | 369 | ||
368 | asmlinkage long sys_stat64(char __user * filename, struct stat64 __user * statbuf) | 370 | SYSCALL_DEFINE2(stat64, char __user *, filename, struct stat64 __user *, statbuf) |
369 | { | 371 | { |
370 | struct kstat stat; | 372 | struct kstat stat; |
371 | int error = vfs_stat(filename, &stat); | 373 | int error = vfs_stat(filename, &stat); |
@@ -375,7 +377,8 @@ asmlinkage long sys_stat64(char __user * filename, struct stat64 __user * statbu | |||
375 | 377 | ||
376 | return error; | 378 | return error; |
377 | } | 379 | } |
378 | asmlinkage long sys_lstat64(char __user * filename, struct stat64 __user * statbuf) | 380 | |
381 | SYSCALL_DEFINE2(lstat64, char __user *, filename, struct stat64 __user *, statbuf) | ||
379 | { | 382 | { |
380 | struct kstat stat; | 383 | struct kstat stat; |
381 | int error = vfs_lstat(filename, &stat); | 384 | int error = vfs_lstat(filename, &stat); |
@@ -385,7 +388,8 @@ asmlinkage long sys_lstat64(char __user * filename, struct stat64 __user * statb | |||
385 | 388 | ||
386 | return error; | 389 | return error; |
387 | } | 390 | } |
388 | asmlinkage long sys_fstat64(unsigned long fd, struct stat64 __user * statbuf) | 391 | |
392 | SYSCALL_DEFINE2(fstat64, unsigned long, fd, struct stat64 __user *, statbuf) | ||
389 | { | 393 | { |
390 | struct kstat stat; | 394 | struct kstat stat; |
391 | int error = vfs_fstat(fd, &stat); | 395 | int error = vfs_fstat(fd, &stat); |
@@ -396,8 +400,8 @@ asmlinkage long sys_fstat64(unsigned long fd, struct stat64 __user * statbuf) | |||
396 | return error; | 400 | return error; |
397 | } | 401 | } |
398 | 402 | ||
399 | asmlinkage long sys_fstatat64(int dfd, char __user *filename, | 403 | SYSCALL_DEFINE4(fstatat64, int, dfd, char __user *, filename, |
400 | struct stat64 __user *statbuf, int flag) | 404 | struct stat64 __user *, statbuf, int, flag) |
401 | { | 405 | { |
402 | struct kstat stat; | 406 | struct kstat stat; |
403 | int error = -EINVAL; | 407 | int error = -EINVAL; |
diff --git a/fs/super.c b/fs/super.c index ed080c417167..8349ed6b1412 100644 --- a/fs/super.c +++ b/fs/super.c | |||
@@ -82,7 +82,22 @@ static struct super_block *alloc_super(struct file_system_type *type) | |||
82 | * lock ordering than usbfs: | 82 | * lock ordering than usbfs: |
83 | */ | 83 | */ |
84 | lockdep_set_class(&s->s_lock, &type->s_lock_key); | 84 | lockdep_set_class(&s->s_lock, &type->s_lock_key); |
85 | down_write(&s->s_umount); | 85 | /* |
86 | * sget() can have s_umount recursion. | ||
87 | * | ||
88 | * When it cannot find a suitable sb, it allocates a new | ||
89 | * one (this one), and tries again to find a suitable old | ||
90 | * one. | ||
91 | * | ||
92 | * In case that succeeds, it will acquire the s_umount | ||
93 | * lock of the old one. Since these are clearly distrinct | ||
94 | * locks, and this object isn't exposed yet, there's no | ||
95 | * risk of deadlocks. | ||
96 | * | ||
97 | * Annotate this by putting this lock in a different | ||
98 | * subclass. | ||
99 | */ | ||
100 | down_write_nested(&s->s_umount, SINGLE_DEPTH_NESTING); | ||
86 | s->s_count = S_BIAS; | 101 | s->s_count = S_BIAS; |
87 | atomic_set(&s->s_active, 1); | 102 | atomic_set(&s->s_active, 1); |
88 | mutex_init(&s->s_vfs_rename_mutex); | 103 | mutex_init(&s->s_vfs_rename_mutex); |
@@ -301,7 +316,7 @@ void generic_shutdown_super(struct super_block *sb) | |||
301 | /* | 316 | /* |
302 | * wait for asynchronous fs operations to finish before going further | 317 | * wait for asynchronous fs operations to finish before going further |
303 | */ | 318 | */ |
304 | async_synchronize_full_special(&sb->s_async_list); | 319 | async_synchronize_full_domain(&sb->s_async_list); |
305 | 320 | ||
306 | /* bad name - it should be evict_inodes() */ | 321 | /* bad name - it should be evict_inodes() */ |
307 | invalidate_inodes(sb); | 322 | invalidate_inodes(sb); |
@@ -470,7 +485,7 @@ restart: | |||
470 | sb->s_count++; | 485 | sb->s_count++; |
471 | spin_unlock(&sb_lock); | 486 | spin_unlock(&sb_lock); |
472 | down_read(&sb->s_umount); | 487 | down_read(&sb->s_umount); |
473 | async_synchronize_full_special(&sb->s_async_list); | 488 | async_synchronize_full_domain(&sb->s_async_list); |
474 | if (sb->s_root && (wait || sb->s_dirt)) | 489 | if (sb->s_root && (wait || sb->s_dirt)) |
475 | sb->s_op->sync_fs(sb, wait); | 490 | sb->s_op->sync_fs(sb, wait); |
476 | up_read(&sb->s_umount); | 491 | up_read(&sb->s_umount); |
@@ -544,7 +559,7 @@ rescan: | |||
544 | return NULL; | 559 | return NULL; |
545 | } | 560 | } |
546 | 561 | ||
547 | asmlinkage long sys_ustat(unsigned dev, struct ustat __user * ubuf) | 562 | SYSCALL_DEFINE2(ustat, unsigned, dev, struct ustat __user *, ubuf) |
548 | { | 563 | { |
549 | struct super_block *s; | 564 | struct super_block *s; |
550 | struct ustat tmp; | 565 | struct ustat tmp; |
@@ -36,7 +36,7 @@ static void do_sync(unsigned long wait) | |||
36 | laptop_sync_completion(); | 36 | laptop_sync_completion(); |
37 | } | 37 | } |
38 | 38 | ||
39 | asmlinkage long sys_sync(void) | 39 | SYSCALL_DEFINE0(sync) |
40 | { | 40 | { |
41 | do_sync(1); | 41 | do_sync(1); |
42 | return 0; | 42 | return 0; |
@@ -144,12 +144,12 @@ static int do_fsync(unsigned int fd, int datasync) | |||
144 | return ret; | 144 | return ret; |
145 | } | 145 | } |
146 | 146 | ||
147 | asmlinkage long sys_fsync(unsigned int fd) | 147 | SYSCALL_DEFINE1(fsync, unsigned int, fd) |
148 | { | 148 | { |
149 | return do_fsync(fd, 0); | 149 | return do_fsync(fd, 0); |
150 | } | 150 | } |
151 | 151 | ||
152 | asmlinkage long sys_fdatasync(unsigned int fd) | 152 | SYSCALL_DEFINE1(fdatasync, unsigned int, fd) |
153 | { | 153 | { |
154 | return do_fsync(fd, 1); | 154 | return do_fsync(fd, 1); |
155 | } | 155 | } |
@@ -201,8 +201,8 @@ asmlinkage long sys_fdatasync(unsigned int fd) | |||
201 | * already-instantiated disk blocks, there are no guarantees here that the data | 201 | * already-instantiated disk blocks, there are no guarantees here that the data |
202 | * will be available after a crash. | 202 | * will be available after a crash. |
203 | */ | 203 | */ |
204 | asmlinkage long sys_sync_file_range(int fd, loff_t offset, loff_t nbytes, | 204 | SYSCALL_DEFINE(sync_file_range)(int fd, loff_t offset, loff_t nbytes, |
205 | unsigned int flags) | 205 | unsigned int flags) |
206 | { | 206 | { |
207 | int ret; | 207 | int ret; |
208 | struct file *file; | 208 | struct file *file; |
@@ -262,14 +262,32 @@ out_put: | |||
262 | out: | 262 | out: |
263 | return ret; | 263 | return ret; |
264 | } | 264 | } |
265 | #ifdef CONFIG_HAVE_SYSCALL_WRAPPERS | ||
266 | asmlinkage long SyS_sync_file_range(long fd, loff_t offset, loff_t nbytes, | ||
267 | long flags) | ||
268 | { | ||
269 | return SYSC_sync_file_range((int) fd, offset, nbytes, | ||
270 | (unsigned int) flags); | ||
271 | } | ||
272 | SYSCALL_ALIAS(sys_sync_file_range, SyS_sync_file_range); | ||
273 | #endif | ||
265 | 274 | ||
266 | /* It would be nice if people remember that not all the world's an i386 | 275 | /* It would be nice if people remember that not all the world's an i386 |
267 | when they introduce new system calls */ | 276 | when they introduce new system calls */ |
268 | asmlinkage long sys_sync_file_range2(int fd, unsigned int flags, | 277 | SYSCALL_DEFINE(sync_file_range2)(int fd, unsigned int flags, |
269 | loff_t offset, loff_t nbytes) | 278 | loff_t offset, loff_t nbytes) |
270 | { | 279 | { |
271 | return sys_sync_file_range(fd, offset, nbytes, flags); | 280 | return sys_sync_file_range(fd, offset, nbytes, flags); |
272 | } | 281 | } |
282 | #ifdef CONFIG_HAVE_SYSCALL_WRAPPERS | ||
283 | asmlinkage long SyS_sync_file_range2(long fd, long flags, | ||
284 | loff_t offset, loff_t nbytes) | ||
285 | { | ||
286 | return SYSC_sync_file_range2((int) fd, (unsigned int) flags, | ||
287 | offset, nbytes); | ||
288 | } | ||
289 | SYSCALL_ALIAS(sys_sync_file_range2, SyS_sync_file_range2); | ||
290 | #endif | ||
273 | 291 | ||
274 | /* | 292 | /* |
275 | * `endbyte' is inclusive | 293 | * `endbyte' is inclusive |
diff --git a/fs/sysfs/Kconfig b/fs/sysfs/Kconfig new file mode 100644 index 000000000000..f4b67588b9d6 --- /dev/null +++ b/fs/sysfs/Kconfig | |||
@@ -0,0 +1,23 @@ | |||
1 | config SYSFS | ||
2 | bool "sysfs file system support" if EMBEDDED | ||
3 | default y | ||
4 | help | ||
5 | The sysfs filesystem is a virtual filesystem that the kernel uses to | ||
6 | export internal kernel objects, their attributes, and their | ||
7 | relationships to one another. | ||
8 | |||
9 | Users can use sysfs to ascertain useful information about the running | ||
10 | kernel, such as the devices the kernel has discovered on each bus and | ||
11 | which driver each is bound to. sysfs can also be used to tune devices | ||
12 | and other kernel subsystems. | ||
13 | |||
14 | Some system agents rely on the information in sysfs to operate. | ||
15 | /sbin/hotplug uses device and object attributes in sysfs to assist in | ||
16 | delegating policy decisions, like persistently naming devices. | ||
17 | |||
18 | sysfs is currently used by the block subsystem to mount the root | ||
19 | partition. If sysfs is disabled you must specify the boot device on | ||
20 | the kernel boot command line via its major and minor numbers. For | ||
21 | example, "root=03:01" for /dev/hda1. | ||
22 | |||
23 | Designers of embedded systems may wish to say N here to conserve space. | ||
diff --git a/fs/sysfs/bin.c b/fs/sysfs/bin.c index 66f6e58a7e4b..f2c478c3424e 100644 --- a/fs/sysfs/bin.c +++ b/fs/sysfs/bin.c | |||
@@ -63,6 +63,9 @@ read(struct file *file, char __user *userbuf, size_t bytes, loff_t *off) | |||
63 | int count = min_t(size_t, bytes, PAGE_SIZE); | 63 | int count = min_t(size_t, bytes, PAGE_SIZE); |
64 | char *temp; | 64 | char *temp; |
65 | 65 | ||
66 | if (!bytes) | ||
67 | return 0; | ||
68 | |||
66 | if (size) { | 69 | if (size) { |
67 | if (offs > size) | 70 | if (offs > size) |
68 | return 0; | 71 | return 0; |
@@ -131,6 +134,9 @@ static ssize_t write(struct file *file, const char __user *userbuf, | |||
131 | int count = min_t(size_t, bytes, PAGE_SIZE); | 134 | int count = min_t(size_t, bytes, PAGE_SIZE); |
132 | char *temp; | 135 | char *temp; |
133 | 136 | ||
137 | if (!bytes) | ||
138 | return 0; | ||
139 | |||
134 | if (size) { | 140 | if (size) { |
135 | if (offs > size) | 141 | if (offs > size) |
136 | return 0; | 142 | return 0; |
diff --git a/fs/sysv/Kconfig b/fs/sysv/Kconfig new file mode 100644 index 000000000000..33aeb4b75db1 --- /dev/null +++ b/fs/sysv/Kconfig | |||
@@ -0,0 +1,36 @@ | |||
1 | config SYSV_FS | ||
2 | tristate "System V/Xenix/V7/Coherent file system support" | ||
3 | depends on BLOCK | ||
4 | help | ||
5 | SCO, Xenix and Coherent are commercial Unix systems for Intel | ||
6 | machines, and Version 7 was used on the DEC PDP-11. Saying Y | ||
7 | here would allow you to read from their floppies and hard disk | ||
8 | partitions. | ||
9 | |||
10 | If you have floppies or hard disk partitions like that, it is likely | ||
11 | that they contain binaries from those other Unix systems; in order | ||
12 | to run these binaries, you will want to install linux-abi which is | ||
13 | a set of kernel modules that lets you run SCO, Xenix, Wyse, | ||
14 | UnixWare, Dell Unix and System V programs under Linux. It is | ||
15 | available via FTP (user: ftp) from | ||
16 | <ftp://ftp.openlinux.org/pub/people/hch/linux-abi/>). | ||
17 | NOTE: that will work only for binaries from Intel-based systems; | ||
18 | PDP ones will have to wait until somebody ports Linux to -11 ;-) | ||
19 | |||
20 | If you only intend to mount files from some other Unix over the | ||
21 | network using NFS, you don't need the System V file system support | ||
22 | (but you need NFS file system support obviously). | ||
23 | |||
24 | Note that this option is generally not needed for floppies, since a | ||
25 | good portable way to transport files and directories between unixes | ||
26 | (and even other operating systems) is given by the tar program ("man | ||
27 | tar" or preferably "info tar"). Note also that this option has | ||
28 | nothing whatsoever to do with the option "System V IPC". Read about | ||
29 | the System V file system in | ||
30 | <file:Documentation/filesystems/sysv-fs.txt>. | ||
31 | Saying Y here will enlarge your kernel by about 27 KB. | ||
32 | |||
33 | To compile this as a module, choose M here: the module will be called | ||
34 | sysv. | ||
35 | |||
36 | If you haven't heard about all of this before, it's safe to say N. | ||
diff --git a/fs/timerfd.c b/fs/timerfd.c index 0862f0e49d0c..b042bd7034b1 100644 --- a/fs/timerfd.c +++ b/fs/timerfd.c | |||
@@ -177,7 +177,7 @@ static struct file *timerfd_fget(int fd) | |||
177 | return file; | 177 | return file; |
178 | } | 178 | } |
179 | 179 | ||
180 | asmlinkage long sys_timerfd_create(int clockid, int flags) | 180 | SYSCALL_DEFINE2(timerfd_create, int, clockid, int, flags) |
181 | { | 181 | { |
182 | int ufd; | 182 | int ufd; |
183 | struct timerfd_ctx *ctx; | 183 | struct timerfd_ctx *ctx; |
@@ -186,10 +186,9 @@ asmlinkage long sys_timerfd_create(int clockid, int flags) | |||
186 | BUILD_BUG_ON(TFD_CLOEXEC != O_CLOEXEC); | 186 | BUILD_BUG_ON(TFD_CLOEXEC != O_CLOEXEC); |
187 | BUILD_BUG_ON(TFD_NONBLOCK != O_NONBLOCK); | 187 | BUILD_BUG_ON(TFD_NONBLOCK != O_NONBLOCK); |
188 | 188 | ||
189 | if (flags & ~(TFD_CLOEXEC | TFD_NONBLOCK)) | 189 | if ((flags & ~TFD_CREATE_FLAGS) || |
190 | return -EINVAL; | 190 | (clockid != CLOCK_MONOTONIC && |
191 | if (clockid != CLOCK_MONOTONIC && | 191 | clockid != CLOCK_REALTIME)) |
192 | clockid != CLOCK_REALTIME) | ||
193 | return -EINVAL; | 192 | return -EINVAL; |
194 | 193 | ||
195 | ctx = kzalloc(sizeof(*ctx), GFP_KERNEL); | 194 | ctx = kzalloc(sizeof(*ctx), GFP_KERNEL); |
@@ -201,16 +200,16 @@ asmlinkage long sys_timerfd_create(int clockid, int flags) | |||
201 | hrtimer_init(&ctx->tmr, clockid, HRTIMER_MODE_ABS); | 200 | hrtimer_init(&ctx->tmr, clockid, HRTIMER_MODE_ABS); |
202 | 201 | ||
203 | ufd = anon_inode_getfd("[timerfd]", &timerfd_fops, ctx, | 202 | ufd = anon_inode_getfd("[timerfd]", &timerfd_fops, ctx, |
204 | flags & (O_CLOEXEC | O_NONBLOCK)); | 203 | flags & TFD_SHARED_FCNTL_FLAGS); |
205 | if (ufd < 0) | 204 | if (ufd < 0) |
206 | kfree(ctx); | 205 | kfree(ctx); |
207 | 206 | ||
208 | return ufd; | 207 | return ufd; |
209 | } | 208 | } |
210 | 209 | ||
211 | asmlinkage long sys_timerfd_settime(int ufd, int flags, | 210 | SYSCALL_DEFINE4(timerfd_settime, int, ufd, int, flags, |
212 | const struct itimerspec __user *utmr, | 211 | const struct itimerspec __user *, utmr, |
213 | struct itimerspec __user *otmr) | 212 | struct itimerspec __user *, otmr) |
214 | { | 213 | { |
215 | struct file *file; | 214 | struct file *file; |
216 | struct timerfd_ctx *ctx; | 215 | struct timerfd_ctx *ctx; |
@@ -219,7 +218,8 @@ asmlinkage long sys_timerfd_settime(int ufd, int flags, | |||
219 | if (copy_from_user(&ktmr, utmr, sizeof(ktmr))) | 218 | if (copy_from_user(&ktmr, utmr, sizeof(ktmr))) |
220 | return -EFAULT; | 219 | return -EFAULT; |
221 | 220 | ||
222 | if (!timespec_valid(&ktmr.it_value) || | 221 | if ((flags & ~TFD_SETTIME_FLAGS) || |
222 | !timespec_valid(&ktmr.it_value) || | ||
223 | !timespec_valid(&ktmr.it_interval)) | 223 | !timespec_valid(&ktmr.it_interval)) |
224 | return -EINVAL; | 224 | return -EINVAL; |
225 | 225 | ||
@@ -265,7 +265,7 @@ asmlinkage long sys_timerfd_settime(int ufd, int flags, | |||
265 | return 0; | 265 | return 0; |
266 | } | 266 | } |
267 | 267 | ||
268 | asmlinkage long sys_timerfd_gettime(int ufd, struct itimerspec __user *otmr) | 268 | SYSCALL_DEFINE2(timerfd_gettime, int, ufd, struct itimerspec __user *, otmr) |
269 | { | 269 | { |
270 | struct file *file; | 270 | struct file *file; |
271 | struct timerfd_ctx *ctx; | 271 | struct timerfd_ctx *ctx; |
diff --git a/fs/ubifs/budget.c b/fs/ubifs/budget.c index 175f9c590b77..f393620890ee 100644 --- a/fs/ubifs/budget.c +++ b/fs/ubifs/budget.c | |||
@@ -689,7 +689,7 @@ long long ubifs_reported_space(const struct ubifs_info *c, long long free) | |||
689 | } | 689 | } |
690 | 690 | ||
691 | /** | 691 | /** |
692 | * ubifs_get_free_space - return amount of free space. | 692 | * ubifs_get_free_space_nolock - return amount of free space. |
693 | * @c: UBIFS file-system description object | 693 | * @c: UBIFS file-system description object |
694 | * | 694 | * |
695 | * This function calculates amount of free space to report to user-space. | 695 | * This function calculates amount of free space to report to user-space. |
@@ -704,16 +704,14 @@ long long ubifs_reported_space(const struct ubifs_info *c, long long free) | |||
704 | * traditional file-systems, because they have way less overhead than UBIFS. | 704 | * traditional file-systems, because they have way less overhead than UBIFS. |
705 | * So, to keep users happy, UBIFS tries to take the overhead into account. | 705 | * So, to keep users happy, UBIFS tries to take the overhead into account. |
706 | */ | 706 | */ |
707 | long long ubifs_get_free_space(struct ubifs_info *c) | 707 | long long ubifs_get_free_space_nolock(struct ubifs_info *c) |
708 | { | 708 | { |
709 | int min_idx_lebs, rsvd_idx_lebs, lebs; | 709 | int rsvd_idx_lebs, lebs; |
710 | long long available, outstanding, free; | 710 | long long available, outstanding, free; |
711 | 711 | ||
712 | spin_lock(&c->space_lock); | 712 | ubifs_assert(c->min_idx_lebs == ubifs_calc_min_idx_lebs(c)); |
713 | min_idx_lebs = c->min_idx_lebs; | ||
714 | ubifs_assert(min_idx_lebs == ubifs_calc_min_idx_lebs(c)); | ||
715 | outstanding = c->budg_data_growth + c->budg_dd_growth; | 713 | outstanding = c->budg_data_growth + c->budg_dd_growth; |
716 | available = ubifs_calc_available(c, min_idx_lebs); | 714 | available = ubifs_calc_available(c, c->min_idx_lebs); |
717 | 715 | ||
718 | /* | 716 | /* |
719 | * When reporting free space to user-space, UBIFS guarantees that it is | 717 | * When reporting free space to user-space, UBIFS guarantees that it is |
@@ -726,15 +724,14 @@ long long ubifs_get_free_space(struct ubifs_info *c) | |||
726 | * Note, the calculations below are similar to what we have in | 724 | * Note, the calculations below are similar to what we have in |
727 | * 'do_budget_space()', so refer there for comments. | 725 | * 'do_budget_space()', so refer there for comments. |
728 | */ | 726 | */ |
729 | if (min_idx_lebs > c->lst.idx_lebs) | 727 | if (c->min_idx_lebs > c->lst.idx_lebs) |
730 | rsvd_idx_lebs = min_idx_lebs - c->lst.idx_lebs; | 728 | rsvd_idx_lebs = c->min_idx_lebs - c->lst.idx_lebs; |
731 | else | 729 | else |
732 | rsvd_idx_lebs = 0; | 730 | rsvd_idx_lebs = 0; |
733 | lebs = c->lst.empty_lebs + c->freeable_cnt + c->idx_gc_cnt - | 731 | lebs = c->lst.empty_lebs + c->freeable_cnt + c->idx_gc_cnt - |
734 | c->lst.taken_empty_lebs; | 732 | c->lst.taken_empty_lebs; |
735 | lebs -= rsvd_idx_lebs; | 733 | lebs -= rsvd_idx_lebs; |
736 | available += lebs * (c->dark_wm - c->leb_overhead); | 734 | available += lebs * (c->dark_wm - c->leb_overhead); |
737 | spin_unlock(&c->space_lock); | ||
738 | 735 | ||
739 | if (available > outstanding) | 736 | if (available > outstanding) |
740 | free = ubifs_reported_space(c, available - outstanding); | 737 | free = ubifs_reported_space(c, available - outstanding); |
@@ -742,3 +739,21 @@ long long ubifs_get_free_space(struct ubifs_info *c) | |||
742 | free = 0; | 739 | free = 0; |
743 | return free; | 740 | return free; |
744 | } | 741 | } |
742 | |||
743 | /** | ||
744 | * ubifs_get_free_space - return amount of free space. | ||
745 | * @c: UBIFS file-system description object | ||
746 | * | ||
747 | * This function calculates and retuns amount of free space to report to | ||
748 | * user-space. | ||
749 | */ | ||
750 | long long ubifs_get_free_space(struct ubifs_info *c) | ||
751 | { | ||
752 | long long free; | ||
753 | |||
754 | spin_lock(&c->space_lock); | ||
755 | free = ubifs_get_free_space_nolock(c); | ||
756 | spin_unlock(&c->space_lock); | ||
757 | |||
758 | return free; | ||
759 | } | ||
diff --git a/fs/ubifs/debug.c b/fs/ubifs/debug.c index 792c5a16c182..e975bd82f38b 100644 --- a/fs/ubifs/debug.c +++ b/fs/ubifs/debug.c | |||
@@ -620,9 +620,11 @@ void dbg_dump_budg(struct ubifs_info *c) | |||
620 | c->dark_wm, c->dead_wm, c->max_idx_node_sz); | 620 | c->dark_wm, c->dead_wm, c->max_idx_node_sz); |
621 | printk(KERN_DEBUG "\tgc_lnum %d, ihead_lnum %d\n", | 621 | printk(KERN_DEBUG "\tgc_lnum %d, ihead_lnum %d\n", |
622 | c->gc_lnum, c->ihead_lnum); | 622 | c->gc_lnum, c->ihead_lnum); |
623 | for (i = 0; i < c->jhead_cnt; i++) | 623 | /* If we are in R/O mode, journal heads do not exist */ |
624 | printk(KERN_DEBUG "\tjhead %d\t LEB %d\n", | 624 | if (c->jheads) |
625 | c->jheads[i].wbuf.jhead, c->jheads[i].wbuf.lnum); | 625 | for (i = 0; i < c->jhead_cnt; i++) |
626 | printk(KERN_DEBUG "\tjhead %d\t LEB %d\n", | ||
627 | c->jheads[i].wbuf.jhead, c->jheads[i].wbuf.lnum); | ||
626 | for (rb = rb_first(&c->buds); rb; rb = rb_next(rb)) { | 628 | for (rb = rb_first(&c->buds); rb; rb = rb_next(rb)) { |
627 | bud = rb_entry(rb, struct ubifs_bud, rb); | 629 | bud = rb_entry(rb, struct ubifs_bud, rb); |
628 | printk(KERN_DEBUG "\tbud LEB %d\n", bud->lnum); | 630 | printk(KERN_DEBUG "\tbud LEB %d\n", bud->lnum); |
@@ -637,10 +639,7 @@ void dbg_dump_budg(struct ubifs_info *c) | |||
637 | /* Print budgeting predictions */ | 639 | /* Print budgeting predictions */ |
638 | available = ubifs_calc_available(c, c->min_idx_lebs); | 640 | available = ubifs_calc_available(c, c->min_idx_lebs); |
639 | outstanding = c->budg_data_growth + c->budg_dd_growth; | 641 | outstanding = c->budg_data_growth + c->budg_dd_growth; |
640 | if (available > outstanding) | 642 | free = ubifs_get_free_space_nolock(c); |
641 | free = ubifs_reported_space(c, available - outstanding); | ||
642 | else | ||
643 | free = 0; | ||
644 | printk(KERN_DEBUG "Budgeting predictions:\n"); | 643 | printk(KERN_DEBUG "Budgeting predictions:\n"); |
645 | printk(KERN_DEBUG "\tavailable: %lld, outstanding %lld, free %lld\n", | 644 | printk(KERN_DEBUG "\tavailable: %lld, outstanding %lld, free %lld\n", |
646 | available, outstanding, free); | 645 | available, outstanding, free); |
@@ -861,6 +860,65 @@ void dbg_dump_index(struct ubifs_info *c) | |||
861 | } | 860 | } |
862 | 861 | ||
863 | /** | 862 | /** |
863 | * dbg_save_space_info - save information about flash space. | ||
864 | * @c: UBIFS file-system description object | ||
865 | * | ||
866 | * This function saves information about UBIFS free space, dirty space, etc, in | ||
867 | * order to check it later. | ||
868 | */ | ||
869 | void dbg_save_space_info(struct ubifs_info *c) | ||
870 | { | ||
871 | struct ubifs_debug_info *d = c->dbg; | ||
872 | |||
873 | ubifs_get_lp_stats(c, &d->saved_lst); | ||
874 | |||
875 | spin_lock(&c->space_lock); | ||
876 | d->saved_free = ubifs_get_free_space_nolock(c); | ||
877 | spin_unlock(&c->space_lock); | ||
878 | } | ||
879 | |||
880 | /** | ||
881 | * dbg_check_space_info - check flash space information. | ||
882 | * @c: UBIFS file-system description object | ||
883 | * | ||
884 | * This function compares current flash space information with the information | ||
885 | * which was saved when the 'dbg_save_space_info()' function was called. | ||
886 | * Returns zero if the information has not changed, and %-EINVAL it it has | ||
887 | * changed. | ||
888 | */ | ||
889 | int dbg_check_space_info(struct ubifs_info *c) | ||
890 | { | ||
891 | struct ubifs_debug_info *d = c->dbg; | ||
892 | struct ubifs_lp_stats lst; | ||
893 | long long avail, free; | ||
894 | |||
895 | spin_lock(&c->space_lock); | ||
896 | avail = ubifs_calc_available(c, c->min_idx_lebs); | ||
897 | spin_unlock(&c->space_lock); | ||
898 | free = ubifs_get_free_space(c); | ||
899 | |||
900 | if (free != d->saved_free) { | ||
901 | ubifs_err("free space changed from %lld to %lld", | ||
902 | d->saved_free, free); | ||
903 | goto out; | ||
904 | } | ||
905 | |||
906 | return 0; | ||
907 | |||
908 | out: | ||
909 | ubifs_msg("saved lprops statistics dump"); | ||
910 | dbg_dump_lstats(&d->saved_lst); | ||
911 | ubifs_get_lp_stats(c, &lst); | ||
912 | ubifs_msg("current lprops statistics dump"); | ||
913 | dbg_dump_lstats(&d->saved_lst); | ||
914 | spin_lock(&c->space_lock); | ||
915 | dbg_dump_budg(c); | ||
916 | spin_unlock(&c->space_lock); | ||
917 | dump_stack(); | ||
918 | return -EINVAL; | ||
919 | } | ||
920 | |||
921 | /** | ||
864 | * dbg_check_synced_i_size - check synchronized inode size. | 922 | * dbg_check_synced_i_size - check synchronized inode size. |
865 | * @inode: inode to check | 923 | * @inode: inode to check |
866 | * | 924 | * |
@@ -1349,7 +1407,7 @@ int dbg_check_tnc(struct ubifs_info *c, int extra) | |||
1349 | * @c: UBIFS file-system description object | 1407 | * @c: UBIFS file-system description object |
1350 | * @leaf_cb: called for each leaf node | 1408 | * @leaf_cb: called for each leaf node |
1351 | * @znode_cb: called for each indexing node | 1409 | * @znode_cb: called for each indexing node |
1352 | * @priv: private date which is passed to callbacks | 1410 | * @priv: private data which is passed to callbacks |
1353 | * | 1411 | * |
1354 | * This function walks the UBIFS index and calls the @leaf_cb for each leaf | 1412 | * This function walks the UBIFS index and calls the @leaf_cb for each leaf |
1355 | * node and @znode_cb for each indexing node. Returns zero in case of success | 1413 | * node and @znode_cb for each indexing node. Returns zero in case of success |
@@ -2409,7 +2467,7 @@ void ubifs_debugging_exit(struct ubifs_info *c) | |||
2409 | * Root directory for UBIFS stuff in debugfs. Contains sub-directories which | 2467 | * Root directory for UBIFS stuff in debugfs. Contains sub-directories which |
2410 | * contain the stuff specific to particular file-system mounts. | 2468 | * contain the stuff specific to particular file-system mounts. |
2411 | */ | 2469 | */ |
2412 | static struct dentry *debugfs_rootdir; | 2470 | static struct dentry *dfs_rootdir; |
2413 | 2471 | ||
2414 | /** | 2472 | /** |
2415 | * dbg_debugfs_init - initialize debugfs file-system. | 2473 | * dbg_debugfs_init - initialize debugfs file-system. |
@@ -2421,9 +2479,9 @@ static struct dentry *debugfs_rootdir; | |||
2421 | */ | 2479 | */ |
2422 | int dbg_debugfs_init(void) | 2480 | int dbg_debugfs_init(void) |
2423 | { | 2481 | { |
2424 | debugfs_rootdir = debugfs_create_dir("ubifs", NULL); | 2482 | dfs_rootdir = debugfs_create_dir("ubifs", NULL); |
2425 | if (IS_ERR(debugfs_rootdir)) { | 2483 | if (IS_ERR(dfs_rootdir)) { |
2426 | int err = PTR_ERR(debugfs_rootdir); | 2484 | int err = PTR_ERR(dfs_rootdir); |
2427 | ubifs_err("cannot create \"ubifs\" debugfs directory, " | 2485 | ubifs_err("cannot create \"ubifs\" debugfs directory, " |
2428 | "error %d\n", err); | 2486 | "error %d\n", err); |
2429 | return err; | 2487 | return err; |
@@ -2437,7 +2495,7 @@ int dbg_debugfs_init(void) | |||
2437 | */ | 2495 | */ |
2438 | void dbg_debugfs_exit(void) | 2496 | void dbg_debugfs_exit(void) |
2439 | { | 2497 | { |
2440 | debugfs_remove(debugfs_rootdir); | 2498 | debugfs_remove(dfs_rootdir); |
2441 | } | 2499 | } |
2442 | 2500 | ||
2443 | static int open_debugfs_file(struct inode *inode, struct file *file) | 2501 | static int open_debugfs_file(struct inode *inode, struct file *file) |
@@ -2452,13 +2510,13 @@ static ssize_t write_debugfs_file(struct file *file, const char __user *buf, | |||
2452 | struct ubifs_info *c = file->private_data; | 2510 | struct ubifs_info *c = file->private_data; |
2453 | struct ubifs_debug_info *d = c->dbg; | 2511 | struct ubifs_debug_info *d = c->dbg; |
2454 | 2512 | ||
2455 | if (file->f_path.dentry == d->dump_lprops) | 2513 | if (file->f_path.dentry == d->dfs_dump_lprops) |
2456 | dbg_dump_lprops(c); | 2514 | dbg_dump_lprops(c); |
2457 | else if (file->f_path.dentry == d->dump_budg) { | 2515 | else if (file->f_path.dentry == d->dfs_dump_budg) { |
2458 | spin_lock(&c->space_lock); | 2516 | spin_lock(&c->space_lock); |
2459 | dbg_dump_budg(c); | 2517 | dbg_dump_budg(c); |
2460 | spin_unlock(&c->space_lock); | 2518 | spin_unlock(&c->space_lock); |
2461 | } else if (file->f_path.dentry == d->dump_tnc) { | 2519 | } else if (file->f_path.dentry == d->dfs_dump_tnc) { |
2462 | mutex_lock(&c->tnc_mutex); | 2520 | mutex_lock(&c->tnc_mutex); |
2463 | dbg_dump_tnc(c); | 2521 | dbg_dump_tnc(c); |
2464 | mutex_unlock(&c->tnc_mutex); | 2522 | mutex_unlock(&c->tnc_mutex); |
@@ -2469,7 +2527,7 @@ static ssize_t write_debugfs_file(struct file *file, const char __user *buf, | |||
2469 | return count; | 2527 | return count; |
2470 | } | 2528 | } |
2471 | 2529 | ||
2472 | static const struct file_operations debugfs_fops = { | 2530 | static const struct file_operations dfs_fops = { |
2473 | .open = open_debugfs_file, | 2531 | .open = open_debugfs_file, |
2474 | .write = write_debugfs_file, | 2532 | .write = write_debugfs_file, |
2475 | .owner = THIS_MODULE, | 2533 | .owner = THIS_MODULE, |
@@ -2494,36 +2552,32 @@ int dbg_debugfs_init_fs(struct ubifs_info *c) | |||
2494 | struct dentry *dent; | 2552 | struct dentry *dent; |
2495 | struct ubifs_debug_info *d = c->dbg; | 2553 | struct ubifs_debug_info *d = c->dbg; |
2496 | 2554 | ||
2497 | sprintf(d->debugfs_dir_name, "ubi%d_%d", c->vi.ubi_num, c->vi.vol_id); | 2555 | sprintf(d->dfs_dir_name, "ubi%d_%d", c->vi.ubi_num, c->vi.vol_id); |
2498 | d->debugfs_dir = debugfs_create_dir(d->debugfs_dir_name, | 2556 | d->dfs_dir = debugfs_create_dir(d->dfs_dir_name, dfs_rootdir); |
2499 | debugfs_rootdir); | 2557 | if (IS_ERR(d->dfs_dir)) { |
2500 | if (IS_ERR(d->debugfs_dir)) { | 2558 | err = PTR_ERR(d->dfs_dir); |
2501 | err = PTR_ERR(d->debugfs_dir); | ||
2502 | ubifs_err("cannot create \"%s\" debugfs directory, error %d\n", | 2559 | ubifs_err("cannot create \"%s\" debugfs directory, error %d\n", |
2503 | d->debugfs_dir_name, err); | 2560 | d->dfs_dir_name, err); |
2504 | goto out; | 2561 | goto out; |
2505 | } | 2562 | } |
2506 | 2563 | ||
2507 | fname = "dump_lprops"; | 2564 | fname = "dump_lprops"; |
2508 | dent = debugfs_create_file(fname, S_IWUGO, d->debugfs_dir, c, | 2565 | dent = debugfs_create_file(fname, S_IWUGO, d->dfs_dir, c, &dfs_fops); |
2509 | &debugfs_fops); | ||
2510 | if (IS_ERR(dent)) | 2566 | if (IS_ERR(dent)) |
2511 | goto out_remove; | 2567 | goto out_remove; |
2512 | d->dump_lprops = dent; | 2568 | d->dfs_dump_lprops = dent; |
2513 | 2569 | ||
2514 | fname = "dump_budg"; | 2570 | fname = "dump_budg"; |
2515 | dent = debugfs_create_file(fname, S_IWUGO, d->debugfs_dir, c, | 2571 | dent = debugfs_create_file(fname, S_IWUGO, d->dfs_dir, c, &dfs_fops); |
2516 | &debugfs_fops); | ||
2517 | if (IS_ERR(dent)) | 2572 | if (IS_ERR(dent)) |
2518 | goto out_remove; | 2573 | goto out_remove; |
2519 | d->dump_budg = dent; | 2574 | d->dfs_dump_budg = dent; |
2520 | 2575 | ||
2521 | fname = "dump_tnc"; | 2576 | fname = "dump_tnc"; |
2522 | dent = debugfs_create_file(fname, S_IWUGO, d->debugfs_dir, c, | 2577 | dent = debugfs_create_file(fname, S_IWUGO, d->dfs_dir, c, &dfs_fops); |
2523 | &debugfs_fops); | ||
2524 | if (IS_ERR(dent)) | 2578 | if (IS_ERR(dent)) |
2525 | goto out_remove; | 2579 | goto out_remove; |
2526 | d->dump_tnc = dent; | 2580 | d->dfs_dump_tnc = dent; |
2527 | 2581 | ||
2528 | return 0; | 2582 | return 0; |
2529 | 2583 | ||
@@ -2531,7 +2585,7 @@ out_remove: | |||
2531 | err = PTR_ERR(dent); | 2585 | err = PTR_ERR(dent); |
2532 | ubifs_err("cannot create \"%s\" debugfs directory, error %d\n", | 2586 | ubifs_err("cannot create \"%s\" debugfs directory, error %d\n", |
2533 | fname, err); | 2587 | fname, err); |
2534 | debugfs_remove_recursive(d->debugfs_dir); | 2588 | debugfs_remove_recursive(d->dfs_dir); |
2535 | out: | 2589 | out: |
2536 | return err; | 2590 | return err; |
2537 | } | 2591 | } |
@@ -2542,7 +2596,7 @@ out: | |||
2542 | */ | 2596 | */ |
2543 | void dbg_debugfs_exit_fs(struct ubifs_info *c) | 2597 | void dbg_debugfs_exit_fs(struct ubifs_info *c) |
2544 | { | 2598 | { |
2545 | debugfs_remove_recursive(c->dbg->debugfs_dir); | 2599 | debugfs_remove_recursive(c->dbg->dfs_dir); |
2546 | } | 2600 | } |
2547 | 2601 | ||
2548 | #endif /* CONFIG_UBIFS_FS_DEBUG */ | 2602 | #endif /* CONFIG_UBIFS_FS_DEBUG */ |
diff --git a/fs/ubifs/debug.h b/fs/ubifs/debug.h index 9820d6999f7e..c1cd73b2e06e 100644 --- a/fs/ubifs/debug.h +++ b/fs/ubifs/debug.h | |||
@@ -41,15 +41,17 @@ | |||
41 | * @chk_lpt_wastage: used by LPT tree size checker | 41 | * @chk_lpt_wastage: used by LPT tree size checker |
42 | * @chk_lpt_lebs: used by LPT tree size checker | 42 | * @chk_lpt_lebs: used by LPT tree size checker |
43 | * @new_nhead_offs: used by LPT tree size checker | 43 | * @new_nhead_offs: used by LPT tree size checker |
44 | * @new_ihead_lnum: used by debugging to check ihead_lnum | 44 | * @new_ihead_lnum: used by debugging to check @c->ihead_lnum |
45 | * @new_ihead_offs: used by debugging to check ihead_offs | 45 | * @new_ihead_offs: used by debugging to check @c->ihead_offs |
46 | * | 46 | * |
47 | * debugfs_dir_name: name of debugfs directory containing this file-system's | 47 | * @saved_lst: saved lprops statistics (used by 'dbg_save_space_info()') |
48 | * files | 48 | * @saved_free: saved free space (used by 'dbg_save_space_info()') |
49 | * debugfs_dir: direntry object of the file-system debugfs directory | 49 | * |
50 | * dump_lprops: "dump lprops" debugfs knob | 50 | * dfs_dir_name: name of debugfs directory containing this file-system's files |
51 | * dump_budg: "dump budgeting information" debugfs knob | 51 | * dfs_dir: direntry object of the file-system debugfs directory |
52 | * dump_tnc: "dump TNC" debugfs knob | 52 | * dfs_dump_lprops: "dump lprops" debugfs knob |
53 | * dfs_dump_budg: "dump budgeting information" debugfs knob | ||
54 | * dfs_dump_tnc: "dump TNC" debugfs knob | ||
53 | */ | 55 | */ |
54 | struct ubifs_debug_info { | 56 | struct ubifs_debug_info { |
55 | void *buf; | 57 | void *buf; |
@@ -69,11 +71,14 @@ struct ubifs_debug_info { | |||
69 | int new_ihead_lnum; | 71 | int new_ihead_lnum; |
70 | int new_ihead_offs; | 72 | int new_ihead_offs; |
71 | 73 | ||
72 | char debugfs_dir_name[100]; | 74 | struct ubifs_lp_stats saved_lst; |
73 | struct dentry *debugfs_dir; | 75 | long long saved_free; |
74 | struct dentry *dump_lprops; | 76 | |
75 | struct dentry *dump_budg; | 77 | char dfs_dir_name[100]; |
76 | struct dentry *dump_tnc; | 78 | struct dentry *dfs_dir; |
79 | struct dentry *dfs_dump_lprops; | ||
80 | struct dentry *dfs_dump_budg; | ||
81 | struct dentry *dfs_dump_tnc; | ||
77 | }; | 82 | }; |
78 | 83 | ||
79 | #define ubifs_assert(expr) do { \ | 84 | #define ubifs_assert(expr) do { \ |
@@ -297,7 +302,8 @@ int dbg_walk_index(struct ubifs_info *c, dbg_leaf_callback leaf_cb, | |||
297 | dbg_znode_callback znode_cb, void *priv); | 302 | dbg_znode_callback znode_cb, void *priv); |
298 | 303 | ||
299 | /* Checking functions */ | 304 | /* Checking functions */ |
300 | 305 | void dbg_save_space_info(struct ubifs_info *c); | |
306 | int dbg_check_space_info(struct ubifs_info *c); | ||
301 | int dbg_check_lprops(struct ubifs_info *c); | 307 | int dbg_check_lprops(struct ubifs_info *c); |
302 | int dbg_old_index_check_init(struct ubifs_info *c, struct ubifs_zbranch *zroot); | 308 | int dbg_old_index_check_init(struct ubifs_info *c, struct ubifs_zbranch *zroot); |
303 | int dbg_check_old_index(struct ubifs_info *c, struct ubifs_zbranch *zroot); | 309 | int dbg_check_old_index(struct ubifs_info *c, struct ubifs_zbranch *zroot); |
@@ -439,6 +445,8 @@ void dbg_debugfs_exit_fs(struct ubifs_info *c); | |||
439 | 445 | ||
440 | #define dbg_walk_index(c, leaf_cb, znode_cb, priv) 0 | 446 | #define dbg_walk_index(c, leaf_cb, znode_cb, priv) 0 |
441 | #define dbg_old_index_check_init(c, zroot) 0 | 447 | #define dbg_old_index_check_init(c, zroot) 0 |
448 | #define dbg_save_space_info(c) ({}) | ||
449 | #define dbg_check_space_info(c) 0 | ||
442 | #define dbg_check_old_index(c, zroot) 0 | 450 | #define dbg_check_old_index(c, zroot) 0 |
443 | #define dbg_check_cats(c) 0 | 451 | #define dbg_check_cats(c) 0 |
444 | #define dbg_check_ltab(c) 0 | 452 | #define dbg_check_ltab(c) 0 |
diff --git a/fs/ubifs/dir.c b/fs/ubifs/dir.c index f448ab1f9c38..f55d523c52bb 100644 --- a/fs/ubifs/dir.c +++ b/fs/ubifs/dir.c | |||
@@ -482,30 +482,29 @@ static int ubifs_dir_release(struct inode *dir, struct file *file) | |||
482 | } | 482 | } |
483 | 483 | ||
484 | /** | 484 | /** |
485 | * lock_2_inodes - lock two UBIFS inodes. | 485 | * lock_2_inodes - a wrapper for locking two UBIFS inodes. |
486 | * @inode1: first inode | 486 | * @inode1: first inode |
487 | * @inode2: second inode | 487 | * @inode2: second inode |
488 | * | ||
489 | * We do not implement any tricks to guarantee strict lock ordering, because | ||
490 | * VFS has already done it for us on the @i_mutex. So this is just a simple | ||
491 | * wrapper function. | ||
488 | */ | 492 | */ |
489 | static void lock_2_inodes(struct inode *inode1, struct inode *inode2) | 493 | static void lock_2_inodes(struct inode *inode1, struct inode *inode2) |
490 | { | 494 | { |
491 | if (inode1->i_ino < inode2->i_ino) { | 495 | mutex_lock_nested(&ubifs_inode(inode1)->ui_mutex, WB_MUTEX_1); |
492 | mutex_lock_nested(&ubifs_inode(inode1)->ui_mutex, WB_MUTEX_2); | 496 | mutex_lock_nested(&ubifs_inode(inode2)->ui_mutex, WB_MUTEX_2); |
493 | mutex_lock_nested(&ubifs_inode(inode2)->ui_mutex, WB_MUTEX_3); | ||
494 | } else { | ||
495 | mutex_lock_nested(&ubifs_inode(inode2)->ui_mutex, WB_MUTEX_2); | ||
496 | mutex_lock_nested(&ubifs_inode(inode1)->ui_mutex, WB_MUTEX_3); | ||
497 | } | ||
498 | } | 497 | } |
499 | 498 | ||
500 | /** | 499 | /** |
501 | * unlock_2_inodes - unlock two UBIFS inodes inodes. | 500 | * unlock_2_inodes - a wrapper for unlocking two UBIFS inodes. |
502 | * @inode1: first inode | 501 | * @inode1: first inode |
503 | * @inode2: second inode | 502 | * @inode2: second inode |
504 | */ | 503 | */ |
505 | static void unlock_2_inodes(struct inode *inode1, struct inode *inode2) | 504 | static void unlock_2_inodes(struct inode *inode1, struct inode *inode2) |
506 | { | 505 | { |
507 | mutex_unlock(&ubifs_inode(inode1)->ui_mutex); | ||
508 | mutex_unlock(&ubifs_inode(inode2)->ui_mutex); | 506 | mutex_unlock(&ubifs_inode(inode2)->ui_mutex); |
507 | mutex_unlock(&ubifs_inode(inode1)->ui_mutex); | ||
509 | } | 508 | } |
510 | 509 | ||
511 | static int ubifs_link(struct dentry *old_dentry, struct inode *dir, | 510 | static int ubifs_link(struct dentry *old_dentry, struct inode *dir, |
@@ -527,6 +526,8 @@ static int ubifs_link(struct dentry *old_dentry, struct inode *dir, | |||
527 | dbg_gen("dent '%.*s' to ino %lu (nlink %d) in dir ino %lu", | 526 | dbg_gen("dent '%.*s' to ino %lu (nlink %d) in dir ino %lu", |
528 | dentry->d_name.len, dentry->d_name.name, inode->i_ino, | 527 | dentry->d_name.len, dentry->d_name.name, inode->i_ino, |
529 | inode->i_nlink, dir->i_ino); | 528 | inode->i_nlink, dir->i_ino); |
529 | ubifs_assert(mutex_is_locked(&dir->i_mutex)); | ||
530 | ubifs_assert(mutex_is_locked(&inode->i_mutex)); | ||
530 | err = dbg_check_synced_i_size(inode); | 531 | err = dbg_check_synced_i_size(inode); |
531 | if (err) | 532 | if (err) |
532 | return err; | 533 | return err; |
@@ -580,6 +581,8 @@ static int ubifs_unlink(struct inode *dir, struct dentry *dentry) | |||
580 | dbg_gen("dent '%.*s' from ino %lu (nlink %d) in dir ino %lu", | 581 | dbg_gen("dent '%.*s' from ino %lu (nlink %d) in dir ino %lu", |
581 | dentry->d_name.len, dentry->d_name.name, inode->i_ino, | 582 | dentry->d_name.len, dentry->d_name.name, inode->i_ino, |
582 | inode->i_nlink, dir->i_ino); | 583 | inode->i_nlink, dir->i_ino); |
584 | ubifs_assert(mutex_is_locked(&dir->i_mutex)); | ||
585 | ubifs_assert(mutex_is_locked(&inode->i_mutex)); | ||
583 | err = dbg_check_synced_i_size(inode); | 586 | err = dbg_check_synced_i_size(inode); |
584 | if (err) | 587 | if (err) |
585 | return err; | 588 | return err; |
@@ -667,7 +670,8 @@ static int ubifs_rmdir(struct inode *dir, struct dentry *dentry) | |||
667 | 670 | ||
668 | dbg_gen("directory '%.*s', ino %lu in dir ino %lu", dentry->d_name.len, | 671 | dbg_gen("directory '%.*s', ino %lu in dir ino %lu", dentry->d_name.len, |
669 | dentry->d_name.name, inode->i_ino, dir->i_ino); | 672 | dentry->d_name.name, inode->i_ino, dir->i_ino); |
670 | 673 | ubifs_assert(mutex_is_locked(&dir->i_mutex)); | |
674 | ubifs_assert(mutex_is_locked(&inode->i_mutex)); | ||
671 | err = check_dir_empty(c, dentry->d_inode); | 675 | err = check_dir_empty(c, dentry->d_inode); |
672 | if (err) | 676 | if (err) |
673 | return err; | 677 | return err; |
@@ -922,59 +926,30 @@ out_budg: | |||
922 | } | 926 | } |
923 | 927 | ||
924 | /** | 928 | /** |
925 | * lock_3_inodes - lock three UBIFS inodes for rename. | 929 | * lock_3_inodes - a wrapper for locking three UBIFS inodes. |
926 | * @inode1: first inode | 930 | * @inode1: first inode |
927 | * @inode2: second inode | 931 | * @inode2: second inode |
928 | * @inode3: third inode | 932 | * @inode3: third inode |
929 | * | 933 | * |
930 | * For 'ubifs_rename()', @inode1 may be the same as @inode2 whereas @inode3 may | 934 | * This function is used for 'ubifs_rename()' and @inode1 may be the same as |
931 | * be null. | 935 | * @inode2 whereas @inode3 may be %NULL. |
936 | * | ||
937 | * We do not implement any tricks to guarantee strict lock ordering, because | ||
938 | * VFS has already done it for us on the @i_mutex. So this is just a simple | ||
939 | * wrapper function. | ||
932 | */ | 940 | */ |
933 | static void lock_3_inodes(struct inode *inode1, struct inode *inode2, | 941 | static void lock_3_inodes(struct inode *inode1, struct inode *inode2, |
934 | struct inode *inode3) | 942 | struct inode *inode3) |
935 | { | 943 | { |
936 | struct inode *i1, *i2, *i3; | 944 | mutex_lock_nested(&ubifs_inode(inode1)->ui_mutex, WB_MUTEX_1); |
937 | 945 | if (inode2 != inode1) | |
938 | if (!inode3) { | 946 | mutex_lock_nested(&ubifs_inode(inode2)->ui_mutex, WB_MUTEX_2); |
939 | if (inode1 != inode2) { | 947 | if (inode3) |
940 | lock_2_inodes(inode1, inode2); | 948 | mutex_lock_nested(&ubifs_inode(inode3)->ui_mutex, WB_MUTEX_3); |
941 | return; | ||
942 | } | ||
943 | mutex_lock_nested(&ubifs_inode(inode1)->ui_mutex, WB_MUTEX_1); | ||
944 | return; | ||
945 | } | ||
946 | |||
947 | if (inode1 == inode2) { | ||
948 | lock_2_inodes(inode1, inode3); | ||
949 | return; | ||
950 | } | ||
951 | |||
952 | /* 3 different inodes */ | ||
953 | if (inode1 < inode2) { | ||
954 | i3 = inode2; | ||
955 | if (inode1 < inode3) { | ||
956 | i1 = inode1; | ||
957 | i2 = inode3; | ||
958 | } else { | ||
959 | i1 = inode3; | ||
960 | i2 = inode1; | ||
961 | } | ||
962 | } else { | ||
963 | i3 = inode1; | ||
964 | if (inode2 < inode3) { | ||
965 | i1 = inode2; | ||
966 | i2 = inode3; | ||
967 | } else { | ||
968 | i1 = inode3; | ||
969 | i2 = inode2; | ||
970 | } | ||
971 | } | ||
972 | mutex_lock_nested(&ubifs_inode(i1)->ui_mutex, WB_MUTEX_1); | ||
973 | lock_2_inodes(i2, i3); | ||
974 | } | 949 | } |
975 | 950 | ||
976 | /** | 951 | /** |
977 | * unlock_3_inodes - unlock three UBIFS inodes for rename. | 952 | * unlock_3_inodes - a wrapper for unlocking three UBIFS inodes for rename. |
978 | * @inode1: first inode | 953 | * @inode1: first inode |
979 | * @inode2: second inode | 954 | * @inode2: second inode |
980 | * @inode3: third inode | 955 | * @inode3: third inode |
@@ -982,11 +957,11 @@ static void lock_3_inodes(struct inode *inode1, struct inode *inode2, | |||
982 | static void unlock_3_inodes(struct inode *inode1, struct inode *inode2, | 957 | static void unlock_3_inodes(struct inode *inode1, struct inode *inode2, |
983 | struct inode *inode3) | 958 | struct inode *inode3) |
984 | { | 959 | { |
985 | mutex_unlock(&ubifs_inode(inode1)->ui_mutex); | ||
986 | if (inode1 != inode2) | ||
987 | mutex_unlock(&ubifs_inode(inode2)->ui_mutex); | ||
988 | if (inode3) | 960 | if (inode3) |
989 | mutex_unlock(&ubifs_inode(inode3)->ui_mutex); | 961 | mutex_unlock(&ubifs_inode(inode3)->ui_mutex); |
962 | if (inode1 != inode2) | ||
963 | mutex_unlock(&ubifs_inode(inode2)->ui_mutex); | ||
964 | mutex_unlock(&ubifs_inode(inode1)->ui_mutex); | ||
990 | } | 965 | } |
991 | 966 | ||
992 | static int ubifs_rename(struct inode *old_dir, struct dentry *old_dentry, | 967 | static int ubifs_rename(struct inode *old_dir, struct dentry *old_dentry, |
@@ -1020,6 +995,11 @@ static int ubifs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
1020 | "dir ino %lu", old_dentry->d_name.len, old_dentry->d_name.name, | 995 | "dir ino %lu", old_dentry->d_name.len, old_dentry->d_name.name, |
1021 | old_inode->i_ino, old_dir->i_ino, new_dentry->d_name.len, | 996 | old_inode->i_ino, old_dir->i_ino, new_dentry->d_name.len, |
1022 | new_dentry->d_name.name, new_dir->i_ino); | 997 | new_dentry->d_name.name, new_dir->i_ino); |
998 | ubifs_assert(mutex_is_locked(&old_dir->i_mutex)); | ||
999 | ubifs_assert(mutex_is_locked(&new_dir->i_mutex)); | ||
1000 | if (unlink) | ||
1001 | ubifs_assert(mutex_is_locked(&new_inode->i_mutex)); | ||
1002 | |||
1023 | 1003 | ||
1024 | if (unlink && is_dir) { | 1004 | if (unlink && is_dir) { |
1025 | err = check_dir_empty(c, new_inode); | 1005 | err = check_dir_empty(c, new_inode); |
@@ -1199,7 +1179,7 @@ int ubifs_getattr(struct vfsmount *mnt, struct dentry *dentry, | |||
1199 | return 0; | 1179 | return 0; |
1200 | } | 1180 | } |
1201 | 1181 | ||
1202 | struct inode_operations ubifs_dir_inode_operations = { | 1182 | const struct inode_operations ubifs_dir_inode_operations = { |
1203 | .lookup = ubifs_lookup, | 1183 | .lookup = ubifs_lookup, |
1204 | .create = ubifs_create, | 1184 | .create = ubifs_create, |
1205 | .link = ubifs_link, | 1185 | .link = ubifs_link, |
@@ -1219,7 +1199,7 @@ struct inode_operations ubifs_dir_inode_operations = { | |||
1219 | #endif | 1199 | #endif |
1220 | }; | 1200 | }; |
1221 | 1201 | ||
1222 | struct file_operations ubifs_dir_operations = { | 1202 | const struct file_operations ubifs_dir_operations = { |
1223 | .llseek = ubifs_dir_llseek, | 1203 | .llseek = ubifs_dir_llseek, |
1224 | .release = ubifs_dir_release, | 1204 | .release = ubifs_dir_release, |
1225 | .read = generic_read_dir, | 1205 | .read = generic_read_dir, |
diff --git a/fs/ubifs/file.c b/fs/ubifs/file.c index bf37374567fa..93b6de51f261 100644 --- a/fs/ubifs/file.c +++ b/fs/ubifs/file.c | |||
@@ -432,7 +432,6 @@ static int ubifs_write_begin(struct file *file, struct address_space *mapping, | |||
432 | int uninitialized_var(err), appending = !!(pos + len > inode->i_size); | 432 | int uninitialized_var(err), appending = !!(pos + len > inode->i_size); |
433 | struct page *page; | 433 | struct page *page; |
434 | 434 | ||
435 | |||
436 | ubifs_assert(ubifs_inode(inode)->ui_size == inode->i_size); | 435 | ubifs_assert(ubifs_inode(inode)->ui_size == inode->i_size); |
437 | 436 | ||
438 | if (unlikely(c->ro_media)) | 437 | if (unlikely(c->ro_media)) |
@@ -1541,7 +1540,7 @@ static int ubifs_file_mmap(struct file *file, struct vm_area_struct *vma) | |||
1541 | return 0; | 1540 | return 0; |
1542 | } | 1541 | } |
1543 | 1542 | ||
1544 | struct address_space_operations ubifs_file_address_operations = { | 1543 | const struct address_space_operations ubifs_file_address_operations = { |
1545 | .readpage = ubifs_readpage, | 1544 | .readpage = ubifs_readpage, |
1546 | .writepage = ubifs_writepage, | 1545 | .writepage = ubifs_writepage, |
1547 | .write_begin = ubifs_write_begin, | 1546 | .write_begin = ubifs_write_begin, |
@@ -1551,7 +1550,7 @@ struct address_space_operations ubifs_file_address_operations = { | |||
1551 | .releasepage = ubifs_releasepage, | 1550 | .releasepage = ubifs_releasepage, |
1552 | }; | 1551 | }; |
1553 | 1552 | ||
1554 | struct inode_operations ubifs_file_inode_operations = { | 1553 | const struct inode_operations ubifs_file_inode_operations = { |
1555 | .setattr = ubifs_setattr, | 1554 | .setattr = ubifs_setattr, |
1556 | .getattr = ubifs_getattr, | 1555 | .getattr = ubifs_getattr, |
1557 | #ifdef CONFIG_UBIFS_FS_XATTR | 1556 | #ifdef CONFIG_UBIFS_FS_XATTR |
@@ -1562,14 +1561,14 @@ struct inode_operations ubifs_file_inode_operations = { | |||
1562 | #endif | 1561 | #endif |
1563 | }; | 1562 | }; |
1564 | 1563 | ||
1565 | struct inode_operations ubifs_symlink_inode_operations = { | 1564 | const struct inode_operations ubifs_symlink_inode_operations = { |
1566 | .readlink = generic_readlink, | 1565 | .readlink = generic_readlink, |
1567 | .follow_link = ubifs_follow_link, | 1566 | .follow_link = ubifs_follow_link, |
1568 | .setattr = ubifs_setattr, | 1567 | .setattr = ubifs_setattr, |
1569 | .getattr = ubifs_getattr, | 1568 | .getattr = ubifs_getattr, |
1570 | }; | 1569 | }; |
1571 | 1570 | ||
1572 | struct file_operations ubifs_file_operations = { | 1571 | const struct file_operations ubifs_file_operations = { |
1573 | .llseek = generic_file_llseek, | 1572 | .llseek = generic_file_llseek, |
1574 | .read = do_sync_read, | 1573 | .read = do_sync_read, |
1575 | .write = do_sync_write, | 1574 | .write = do_sync_write, |
diff --git a/fs/ubifs/gc.c b/fs/ubifs/gc.c index 9832f9abe28e..a711d33b3d3e 100644 --- a/fs/ubifs/gc.c +++ b/fs/ubifs/gc.c | |||
@@ -31,6 +31,26 @@ | |||
31 | * to be reused. Garbage collection will cause the number of dirty index nodes | 31 | * to be reused. Garbage collection will cause the number of dirty index nodes |
32 | * to grow, however sufficient space is reserved for the index to ensure the | 32 | * to grow, however sufficient space is reserved for the index to ensure the |
33 | * commit will never run out of space. | 33 | * commit will never run out of space. |
34 | * | ||
35 | * Notes about dead watermark. At current UBIFS implementation we assume that | ||
36 | * LEBs which have less than @c->dead_wm bytes of free + dirty space are full | ||
37 | * and not worth garbage-collecting. The dead watermark is one min. I/O unit | ||
38 | * size, or min. UBIFS node size, depending on what is greater. Indeed, UBIFS | ||
39 | * Garbage Collector has to synchronize the GC head's write buffer before | ||
40 | * returning, so this is about wasting one min. I/O unit. However, UBIFS GC can | ||
41 | * actually reclaim even very small pieces of dirty space by garbage collecting | ||
42 | * enough dirty LEBs, but we do not bother doing this at this implementation. | ||
43 | * | ||
44 | * Notes about dark watermark. The results of GC work depends on how big are | ||
45 | * the UBIFS nodes GC deals with. Large nodes make GC waste more space. Indeed, | ||
46 | * if GC move data from LEB A to LEB B and nodes in LEB A are large, GC would | ||
47 | * have to waste large pieces of free space at the end of LEB B, because nodes | ||
48 | * from LEB A would not fit. And the worst situation is when all nodes are of | ||
49 | * maximum size. So dark watermark is the amount of free + dirty space in LEB | ||
50 | * which are guaranteed to be reclaimable. If LEB has less space, the GC migh | ||
51 | * be unable to reclaim it. So, LEBs with free + dirty greater than dark | ||
52 | * watermark are "good" LEBs from GC's point of few. The other LEBs are not so | ||
53 | * good, and GC takes extra care when moving them. | ||
34 | */ | 54 | */ |
35 | 55 | ||
36 | #include <linux/pagemap.h> | 56 | #include <linux/pagemap.h> |
@@ -381,7 +401,7 @@ int ubifs_garbage_collect_leb(struct ubifs_info *c, struct ubifs_lprops *lp) | |||
381 | 401 | ||
382 | /* | 402 | /* |
383 | * Don't release the LEB until after the next commit, because | 403 | * Don't release the LEB until after the next commit, because |
384 | * it may contain date which is needed for recovery. So | 404 | * it may contain data which is needed for recovery. So |
385 | * although we freed this LEB, it will become usable only after | 405 | * although we freed this LEB, it will become usable only after |
386 | * the commit. | 406 | * the commit. |
387 | */ | 407 | */ |
@@ -810,8 +830,9 @@ out: | |||
810 | * ubifs_destroy_idx_gc - destroy idx_gc list. | 830 | * ubifs_destroy_idx_gc - destroy idx_gc list. |
811 | * @c: UBIFS file-system description object | 831 | * @c: UBIFS file-system description object |
812 | * | 832 | * |
813 | * This function destroys the idx_gc list. It is called when unmounting or | 833 | * This function destroys the @c->idx_gc list. It is called when unmounting |
814 | * remounting read-only so locks are not needed. | 834 | * so locks are not needed. Returns zero in case of success and a negative |
835 | * error code in case of failure. | ||
815 | */ | 836 | */ |
816 | void ubifs_destroy_idx_gc(struct ubifs_info *c) | 837 | void ubifs_destroy_idx_gc(struct ubifs_info *c) |
817 | { | 838 | { |
@@ -824,7 +845,6 @@ void ubifs_destroy_idx_gc(struct ubifs_info *c) | |||
824 | list_del(&idx_gc->list); | 845 | list_del(&idx_gc->list); |
825 | kfree(idx_gc); | 846 | kfree(idx_gc); |
826 | } | 847 | } |
827 | |||
828 | } | 848 | } |
829 | 849 | ||
830 | /** | 850 | /** |
diff --git a/fs/ubifs/io.c b/fs/ubifs/io.c index 01682713af69..e8e632a1dcdf 100644 --- a/fs/ubifs/io.c +++ b/fs/ubifs/io.c | |||
@@ -29,7 +29,7 @@ | |||
29 | * would have been wasted for padding to the nearest minimal I/O unit boundary. | 29 | * would have been wasted for padding to the nearest minimal I/O unit boundary. |
30 | * Instead, data first goes to the write-buffer and is flushed when the | 30 | * Instead, data first goes to the write-buffer and is flushed when the |
31 | * buffer is full or when it is not used for some time (by timer). This is | 31 | * buffer is full or when it is not used for some time (by timer). This is |
32 | * similarto the mechanism is used by JFFS2. | 32 | * similar to the mechanism is used by JFFS2. |
33 | * | 33 | * |
34 | * Write-buffers are defined by 'struct ubifs_wbuf' objects and protected by | 34 | * Write-buffers are defined by 'struct ubifs_wbuf' objects and protected by |
35 | * mutexes defined inside these objects. Since sometimes upper-level code | 35 | * mutexes defined inside these objects. Since sometimes upper-level code |
@@ -75,7 +75,7 @@ void ubifs_ro_mode(struct ubifs_info *c, int err) | |||
75 | * @lnum: logical eraseblock number | 75 | * @lnum: logical eraseblock number |
76 | * @offs: offset within the logical eraseblock | 76 | * @offs: offset within the logical eraseblock |
77 | * @quiet: print no messages | 77 | * @quiet: print no messages |
78 | * @chk_crc: indicates whether to always check the CRC | 78 | * @must_chk_crc: indicates whether to always check the CRC |
79 | * | 79 | * |
80 | * This function checks node magic number and CRC checksum. This function also | 80 | * This function checks node magic number and CRC checksum. This function also |
81 | * validates node length to prevent UBIFS from becoming crazy when an attacker | 81 | * validates node length to prevent UBIFS from becoming crazy when an attacker |
@@ -83,11 +83,17 @@ void ubifs_ro_mode(struct ubifs_info *c, int err) | |||
83 | * node length in the common header could cause UBIFS to read memory outside of | 83 | * node length in the common header could cause UBIFS to read memory outside of |
84 | * allocated buffer when checking the CRC checksum. | 84 | * allocated buffer when checking the CRC checksum. |
85 | * | 85 | * |
86 | * This function returns zero in case of success %-EUCLEAN in case of bad CRC | 86 | * This function may skip data nodes CRC checking if @c->no_chk_data_crc is |
87 | * or magic. | 87 | * true, which is controlled by corresponding UBIFS mount option. However, if |
88 | * @must_chk_crc is true, then @c->no_chk_data_crc is ignored and CRC is | ||
89 | * checked. Similarly, if @c->always_chk_crc is true, @c->no_chk_data_crc is | ||
90 | * ignored and CRC is checked. | ||
91 | * | ||
92 | * This function returns zero in case of success and %-EUCLEAN in case of bad | ||
93 | * CRC or magic. | ||
88 | */ | 94 | */ |
89 | int ubifs_check_node(const struct ubifs_info *c, const void *buf, int lnum, | 95 | int ubifs_check_node(const struct ubifs_info *c, const void *buf, int lnum, |
90 | int offs, int quiet, int chk_crc) | 96 | int offs, int quiet, int must_chk_crc) |
91 | { | 97 | { |
92 | int err = -EINVAL, type, node_len; | 98 | int err = -EINVAL, type, node_len; |
93 | uint32_t crc, node_crc, magic; | 99 | uint32_t crc, node_crc, magic; |
@@ -123,9 +129,9 @@ int ubifs_check_node(const struct ubifs_info *c, const void *buf, int lnum, | |||
123 | node_len > c->ranges[type].max_len) | 129 | node_len > c->ranges[type].max_len) |
124 | goto out_len; | 130 | goto out_len; |
125 | 131 | ||
126 | if (!chk_crc && type == UBIFS_DATA_NODE && !c->always_chk_crc) | 132 | if (!must_chk_crc && type == UBIFS_DATA_NODE && !c->always_chk_crc && |
127 | if (c->no_chk_data_crc) | 133 | c->no_chk_data_crc) |
128 | return 0; | 134 | return 0; |
129 | 135 | ||
130 | crc = crc32(UBIFS_CRC32_INIT, buf + 8, node_len - 8); | 136 | crc = crc32(UBIFS_CRC32_INIT, buf + 8, node_len - 8); |
131 | node_crc = le32_to_cpu(ch->crc); | 137 | node_crc = le32_to_cpu(ch->crc); |
diff --git a/fs/ubifs/journal.c b/fs/ubifs/journal.c index 9b7c54e0cd2a..a11ca0958a23 100644 --- a/fs/ubifs/journal.c +++ b/fs/ubifs/journal.c | |||
@@ -208,7 +208,7 @@ again: | |||
208 | offs = 0; | 208 | offs = 0; |
209 | 209 | ||
210 | out: | 210 | out: |
211 | err = ubifs_wbuf_seek_nolock(wbuf, lnum, offs, UBI_SHORTTERM); | 211 | err = ubifs_wbuf_seek_nolock(wbuf, lnum, offs, wbuf->dtype); |
212 | if (err) | 212 | if (err) |
213 | goto out_unlock; | 213 | goto out_unlock; |
214 | 214 | ||
diff --git a/fs/ubifs/lprops.c b/fs/ubifs/lprops.c index dfd2bcece27a..4cdd284dea56 100644 --- a/fs/ubifs/lprops.c +++ b/fs/ubifs/lprops.c | |||
@@ -635,10 +635,10 @@ const struct ubifs_lprops *ubifs_change_lp(struct ubifs_info *c, | |||
635 | * @c: UBIFS file-system description object | 635 | * @c: UBIFS file-system description object |
636 | * @st: return statistics | 636 | * @st: return statistics |
637 | */ | 637 | */ |
638 | void ubifs_get_lp_stats(struct ubifs_info *c, struct ubifs_lp_stats *st) | 638 | void ubifs_get_lp_stats(struct ubifs_info *c, struct ubifs_lp_stats *lst) |
639 | { | 639 | { |
640 | spin_lock(&c->space_lock); | 640 | spin_lock(&c->space_lock); |
641 | memcpy(st, &c->lst, sizeof(struct ubifs_lp_stats)); | 641 | memcpy(lst, &c->lst, sizeof(struct ubifs_lp_stats)); |
642 | spin_unlock(&c->space_lock); | 642 | spin_unlock(&c->space_lock); |
643 | } | 643 | } |
644 | 644 | ||
@@ -678,6 +678,9 @@ int ubifs_change_one_lp(struct ubifs_info *c, int lnum, int free, int dirty, | |||
678 | 678 | ||
679 | out: | 679 | out: |
680 | ubifs_release_lprops(c); | 680 | ubifs_release_lprops(c); |
681 | if (err) | ||
682 | ubifs_err("cannot change properties of LEB %d, error %d", | ||
683 | lnum, err); | ||
681 | return err; | 684 | return err; |
682 | } | 685 | } |
683 | 686 | ||
@@ -714,6 +717,9 @@ int ubifs_update_one_lp(struct ubifs_info *c, int lnum, int free, int dirty, | |||
714 | 717 | ||
715 | out: | 718 | out: |
716 | ubifs_release_lprops(c); | 719 | ubifs_release_lprops(c); |
720 | if (err) | ||
721 | ubifs_err("cannot update properties of LEB %d, error %d", | ||
722 | lnum, err); | ||
717 | return err; | 723 | return err; |
718 | } | 724 | } |
719 | 725 | ||
@@ -737,6 +743,8 @@ int ubifs_read_one_lp(struct ubifs_info *c, int lnum, struct ubifs_lprops *lp) | |||
737 | lpp = ubifs_lpt_lookup(c, lnum); | 743 | lpp = ubifs_lpt_lookup(c, lnum); |
738 | if (IS_ERR(lpp)) { | 744 | if (IS_ERR(lpp)) { |
739 | err = PTR_ERR(lpp); | 745 | err = PTR_ERR(lpp); |
746 | ubifs_err("cannot read properties of LEB %d, error %d", | ||
747 | lnum, err); | ||
740 | goto out; | 748 | goto out; |
741 | } | 749 | } |
742 | 750 | ||
diff --git a/fs/ubifs/lpt_commit.c b/fs/ubifs/lpt_commit.c index 96ca95707175..3216a1f277f8 100644 --- a/fs/ubifs/lpt_commit.c +++ b/fs/ubifs/lpt_commit.c | |||
@@ -556,23 +556,23 @@ no_space: | |||
556 | } | 556 | } |
557 | 557 | ||
558 | /** | 558 | /** |
559 | * next_pnode - find next pnode. | 559 | * next_pnode_to_dirty - find next pnode to dirty. |
560 | * @c: UBIFS file-system description object | 560 | * @c: UBIFS file-system description object |
561 | * @pnode: pnode | 561 | * @pnode: pnode |
562 | * | 562 | * |
563 | * This function returns the next pnode or %NULL if there are no more pnodes. | 563 | * This function returns the next pnode to dirty or %NULL if there are no more |
564 | * pnodes. Note that pnodes that have never been written (lnum == 0) are | ||
565 | * skipped. | ||
564 | */ | 566 | */ |
565 | static struct ubifs_pnode *next_pnode(struct ubifs_info *c, | 567 | static struct ubifs_pnode *next_pnode_to_dirty(struct ubifs_info *c, |
566 | struct ubifs_pnode *pnode) | 568 | struct ubifs_pnode *pnode) |
567 | { | 569 | { |
568 | struct ubifs_nnode *nnode; | 570 | struct ubifs_nnode *nnode; |
569 | int iip; | 571 | int iip; |
570 | 572 | ||
571 | /* Try to go right */ | 573 | /* Try to go right */ |
572 | nnode = pnode->parent; | 574 | nnode = pnode->parent; |
573 | iip = pnode->iip + 1; | 575 | for (iip = pnode->iip + 1; iip < UBIFS_LPT_FANOUT; iip++) { |
574 | if (iip < UBIFS_LPT_FANOUT) { | ||
575 | /* We assume here that LEB zero is never an LPT LEB */ | ||
576 | if (nnode->nbranch[iip].lnum) | 576 | if (nnode->nbranch[iip].lnum) |
577 | return ubifs_get_pnode(c, nnode, iip); | 577 | return ubifs_get_pnode(c, nnode, iip); |
578 | } | 578 | } |
@@ -583,8 +583,11 @@ static struct ubifs_pnode *next_pnode(struct ubifs_info *c, | |||
583 | nnode = nnode->parent; | 583 | nnode = nnode->parent; |
584 | if (!nnode) | 584 | if (!nnode) |
585 | return NULL; | 585 | return NULL; |
586 | /* We assume here that LEB zero is never an LPT LEB */ | 586 | for (; iip < UBIFS_LPT_FANOUT; iip++) { |
587 | } while (iip >= UBIFS_LPT_FANOUT || !nnode->nbranch[iip].lnum); | 587 | if (nnode->nbranch[iip].lnum) |
588 | break; | ||
589 | } | ||
590 | } while (iip >= UBIFS_LPT_FANOUT); | ||
588 | 591 | ||
589 | /* Go right */ | 592 | /* Go right */ |
590 | nnode = ubifs_get_nnode(c, nnode, iip); | 593 | nnode = ubifs_get_nnode(c, nnode, iip); |
@@ -593,12 +596,29 @@ static struct ubifs_pnode *next_pnode(struct ubifs_info *c, | |||
593 | 596 | ||
594 | /* Go down to level 1 */ | 597 | /* Go down to level 1 */ |
595 | while (nnode->level > 1) { | 598 | while (nnode->level > 1) { |
596 | nnode = ubifs_get_nnode(c, nnode, 0); | 599 | for (iip = 0; iip < UBIFS_LPT_FANOUT; iip++) { |
600 | if (nnode->nbranch[iip].lnum) | ||
601 | break; | ||
602 | } | ||
603 | if (iip >= UBIFS_LPT_FANOUT) { | ||
604 | /* | ||
605 | * Should not happen, but we need to keep going | ||
606 | * if it does. | ||
607 | */ | ||
608 | iip = 0; | ||
609 | } | ||
610 | nnode = ubifs_get_nnode(c, nnode, iip); | ||
597 | if (IS_ERR(nnode)) | 611 | if (IS_ERR(nnode)) |
598 | return (void *)nnode; | 612 | return (void *)nnode; |
599 | } | 613 | } |
600 | 614 | ||
601 | return ubifs_get_pnode(c, nnode, 0); | 615 | for (iip = 0; iip < UBIFS_LPT_FANOUT; iip++) |
616 | if (nnode->nbranch[iip].lnum) | ||
617 | break; | ||
618 | if (iip >= UBIFS_LPT_FANOUT) | ||
619 | /* Should not happen, but we need to keep going if it does */ | ||
620 | iip = 0; | ||
621 | return ubifs_get_pnode(c, nnode, iip); | ||
602 | } | 622 | } |
603 | 623 | ||
604 | /** | 624 | /** |
@@ -688,7 +708,7 @@ static int make_tree_dirty(struct ubifs_info *c) | |||
688 | pnode = pnode_lookup(c, 0); | 708 | pnode = pnode_lookup(c, 0); |
689 | while (pnode) { | 709 | while (pnode) { |
690 | do_make_pnode_dirty(c, pnode); | 710 | do_make_pnode_dirty(c, pnode); |
691 | pnode = next_pnode(c, pnode); | 711 | pnode = next_pnode_to_dirty(c, pnode); |
692 | if (IS_ERR(pnode)) | 712 | if (IS_ERR(pnode)) |
693 | return PTR_ERR(pnode); | 713 | return PTR_ERR(pnode); |
694 | } | 714 | } |
diff --git a/fs/ubifs/master.c b/fs/ubifs/master.c index 71d5493bf565..a88f33801b98 100644 --- a/fs/ubifs/master.c +++ b/fs/ubifs/master.c | |||
@@ -354,7 +354,7 @@ int ubifs_write_master(struct ubifs_info *c) | |||
354 | int err, lnum, offs, len; | 354 | int err, lnum, offs, len; |
355 | 355 | ||
356 | if (c->ro_media) | 356 | if (c->ro_media) |
357 | return -EINVAL; | 357 | return -EROFS; |
358 | 358 | ||
359 | lnum = UBIFS_MST_LNUM; | 359 | lnum = UBIFS_MST_LNUM; |
360 | offs = c->mst_offs + c->mst_node_alsz; | 360 | offs = c->mst_offs + c->mst_node_alsz; |
diff --git a/fs/ubifs/orphan.c b/fs/ubifs/orphan.c index 9e6f403f170e..152a7b34a141 100644 --- a/fs/ubifs/orphan.c +++ b/fs/ubifs/orphan.c | |||
@@ -46,7 +46,7 @@ | |||
46 | * Orphans are accumulated in a rb-tree. When an inode's link count drops to | 46 | * Orphans are accumulated in a rb-tree. When an inode's link count drops to |
47 | * zero, the inode number is added to the rb-tree. It is removed from the tree | 47 | * zero, the inode number is added to the rb-tree. It is removed from the tree |
48 | * when the inode is deleted. Any new orphans that are in the orphan tree when | 48 | * when the inode is deleted. Any new orphans that are in the orphan tree when |
49 | * the commit is run, are written to the orphan area in 1 or more orph nodes. | 49 | * the commit is run, are written to the orphan area in 1 or more orphan nodes. |
50 | * If the orphan area is full, it is consolidated to make space. There is | 50 | * If the orphan area is full, it is consolidated to make space. There is |
51 | * always enough space because validation prevents the user from creating more | 51 | * always enough space because validation prevents the user from creating more |
52 | * than the maximum number of orphans allowed. | 52 | * than the maximum number of orphans allowed. |
@@ -231,7 +231,7 @@ static int tot_avail_orphs(struct ubifs_info *c) | |||
231 | } | 231 | } |
232 | 232 | ||
233 | /** | 233 | /** |
234 | * do_write_orph_node - write a node | 234 | * do_write_orph_node - write a node to the orphan head. |
235 | * @c: UBIFS file-system description object | 235 | * @c: UBIFS file-system description object |
236 | * @len: length of node | 236 | * @len: length of node |
237 | * @atomic: write atomically | 237 | * @atomic: write atomically |
@@ -264,11 +264,11 @@ static int do_write_orph_node(struct ubifs_info *c, int len, int atomic) | |||
264 | } | 264 | } |
265 | 265 | ||
266 | /** | 266 | /** |
267 | * write_orph_node - write an orph node | 267 | * write_orph_node - write an orphan node. |
268 | * @c: UBIFS file-system description object | 268 | * @c: UBIFS file-system description object |
269 | * @atomic: write atomically | 269 | * @atomic: write atomically |
270 | * | 270 | * |
271 | * This function builds an orph node from the cnext list and writes it to the | 271 | * This function builds an orphan node from the cnext list and writes it to the |
272 | * orphan head. On success, %0 is returned, otherwise a negative error code | 272 | * orphan head. On success, %0 is returned, otherwise a negative error code |
273 | * is returned. | 273 | * is returned. |
274 | */ | 274 | */ |
@@ -326,11 +326,11 @@ static int write_orph_node(struct ubifs_info *c, int atomic) | |||
326 | } | 326 | } |
327 | 327 | ||
328 | /** | 328 | /** |
329 | * write_orph_nodes - write orph nodes until there are no more to commit | 329 | * write_orph_nodes - write orphan nodes until there are no more to commit. |
330 | * @c: UBIFS file-system description object | 330 | * @c: UBIFS file-system description object |
331 | * @atomic: write atomically | 331 | * @atomic: write atomically |
332 | * | 332 | * |
333 | * This function writes orph nodes for all the orphans to commit. On success, | 333 | * This function writes orphan nodes for all the orphans to commit. On success, |
334 | * %0 is returned, otherwise a negative error code is returned. | 334 | * %0 is returned, otherwise a negative error code is returned. |
335 | */ | 335 | */ |
336 | static int write_orph_nodes(struct ubifs_info *c, int atomic) | 336 | static int write_orph_nodes(struct ubifs_info *c, int atomic) |
@@ -478,14 +478,14 @@ int ubifs_orphan_end_commit(struct ubifs_info *c) | |||
478 | } | 478 | } |
479 | 479 | ||
480 | /** | 480 | /** |
481 | * clear_orphans - erase all LEBs used for orphans. | 481 | * ubifs_clear_orphans - erase all LEBs used for orphans. |
482 | * @c: UBIFS file-system description object | 482 | * @c: UBIFS file-system description object |
483 | * | 483 | * |
484 | * If recovery is not required, then the orphans from the previous session | 484 | * If recovery is not required, then the orphans from the previous session |
485 | * are not needed. This function locates the LEBs used to record | 485 | * are not needed. This function locates the LEBs used to record |
486 | * orphans, and un-maps them. | 486 | * orphans, and un-maps them. |
487 | */ | 487 | */ |
488 | static int clear_orphans(struct ubifs_info *c) | 488 | int ubifs_clear_orphans(struct ubifs_info *c) |
489 | { | 489 | { |
490 | int lnum, err; | 490 | int lnum, err; |
491 | 491 | ||
@@ -547,9 +547,9 @@ static int insert_dead_orphan(struct ubifs_info *c, ino_t inum) | |||
547 | * do_kill_orphans - remove orphan inodes from the index. | 547 | * do_kill_orphans - remove orphan inodes from the index. |
548 | * @c: UBIFS file-system description object | 548 | * @c: UBIFS file-system description object |
549 | * @sleb: scanned LEB | 549 | * @sleb: scanned LEB |
550 | * @last_cmt_no: cmt_no of last orph node read is passed and returned here | 550 | * @last_cmt_no: cmt_no of last orphan node read is passed and returned here |
551 | * @outofdate: whether the LEB is out of date is returned here | 551 | * @outofdate: whether the LEB is out of date is returned here |
552 | * @last_flagged: whether the end orph node is encountered | 552 | * @last_flagged: whether the end orphan node is encountered |
553 | * | 553 | * |
554 | * This function is a helper to the 'kill_orphans()' function. It goes through | 554 | * This function is a helper to the 'kill_orphans()' function. It goes through |
555 | * every orphan node in a LEB and for every inode number recorded, removes | 555 | * every orphan node in a LEB and for every inode number recorded, removes |
@@ -580,8 +580,8 @@ static int do_kill_orphans(struct ubifs_info *c, struct ubifs_scan_leb *sleb, | |||
580 | /* | 580 | /* |
581 | * The commit number on the master node may be less, because | 581 | * The commit number on the master node may be less, because |
582 | * of a failed commit. If there are several failed commits in a | 582 | * of a failed commit. If there are several failed commits in a |
583 | * row, the commit number written on orph nodes will continue to | 583 | * row, the commit number written on orphan nodes will continue |
584 | * increase (because the commit number is adjusted here) even | 584 | * to increase (because the commit number is adjusted here) even |
585 | * though the commit number on the master node stays the same | 585 | * though the commit number on the master node stays the same |
586 | * because the master node has not been re-written. | 586 | * because the master node has not been re-written. |
587 | */ | 587 | */ |
@@ -589,9 +589,9 @@ static int do_kill_orphans(struct ubifs_info *c, struct ubifs_scan_leb *sleb, | |||
589 | c->cmt_no = cmt_no; | 589 | c->cmt_no = cmt_no; |
590 | if (cmt_no < *last_cmt_no && *last_flagged) { | 590 | if (cmt_no < *last_cmt_no && *last_flagged) { |
591 | /* | 591 | /* |
592 | * The last orph node had a higher commit number and was | 592 | * The last orphan node had a higher commit number and |
593 | * flagged as the last written for that commit number. | 593 | * was flagged as the last written for that commit |
594 | * That makes this orph node, out of date. | 594 | * number. That makes this orphan node, out of date. |
595 | */ | 595 | */ |
596 | if (!first) { | 596 | if (!first) { |
597 | ubifs_err("out of order commit number %llu in " | 597 | ubifs_err("out of order commit number %llu in " |
@@ -658,10 +658,10 @@ static int kill_orphans(struct ubifs_info *c) | |||
658 | /* | 658 | /* |
659 | * Orph nodes always start at c->orph_first and are written to each | 659 | * Orph nodes always start at c->orph_first and are written to each |
660 | * successive LEB in turn. Generally unused LEBs will have been unmapped | 660 | * successive LEB in turn. Generally unused LEBs will have been unmapped |
661 | * but may contain out of date orph nodes if the unmap didn't go | 661 | * but may contain out of date orphan nodes if the unmap didn't go |
662 | * through. In addition, the last orph node written for each commit is | 662 | * through. In addition, the last orphan node written for each commit is |
663 | * marked (top bit of orph->cmt_no is set to 1). It is possible that | 663 | * marked (top bit of orph->cmt_no is set to 1). It is possible that |
664 | * there are orph nodes from the next commit (i.e. the commit did not | 664 | * there are orphan nodes from the next commit (i.e. the commit did not |
665 | * complete successfully). In that case, no orphans will have been lost | 665 | * complete successfully). In that case, no orphans will have been lost |
666 | * due to the way that orphans are written, and any orphans added will | 666 | * due to the way that orphans are written, and any orphans added will |
667 | * be valid orphans anyway and so can be deleted. | 667 | * be valid orphans anyway and so can be deleted. |
@@ -718,7 +718,7 @@ int ubifs_mount_orphans(struct ubifs_info *c, int unclean, int read_only) | |||
718 | if (unclean) | 718 | if (unclean) |
719 | err = kill_orphans(c); | 719 | err = kill_orphans(c); |
720 | else if (!read_only) | 720 | else if (!read_only) |
721 | err = clear_orphans(c); | 721 | err = ubifs_clear_orphans(c); |
722 | 722 | ||
723 | return err; | 723 | return err; |
724 | } | 724 | } |
diff --git a/fs/ubifs/super.c b/fs/ubifs/super.c index 89556ee72518..1182b66a5491 100644 --- a/fs/ubifs/super.c +++ b/fs/ubifs/super.c | |||
@@ -397,6 +397,7 @@ static int ubifs_statfs(struct dentry *dentry, struct kstatfs *buf) | |||
397 | buf->f_namelen = UBIFS_MAX_NLEN; | 397 | buf->f_namelen = UBIFS_MAX_NLEN; |
398 | buf->f_fsid.val[0] = le32_to_cpu(uuid[0]) ^ le32_to_cpu(uuid[2]); | 398 | buf->f_fsid.val[0] = le32_to_cpu(uuid[0]) ^ le32_to_cpu(uuid[2]); |
399 | buf->f_fsid.val[1] = le32_to_cpu(uuid[1]) ^ le32_to_cpu(uuid[3]); | 399 | buf->f_fsid.val[1] = le32_to_cpu(uuid[1]) ^ le32_to_cpu(uuid[3]); |
400 | ubifs_assert(buf->f_bfree <= c->block_cnt); | ||
400 | return 0; | 401 | return 0; |
401 | } | 402 | } |
402 | 403 | ||
@@ -432,33 +433,24 @@ static int ubifs_sync_fs(struct super_block *sb, int wait) | |||
432 | int i, err; | 433 | int i, err; |
433 | struct ubifs_info *c = sb->s_fs_info; | 434 | struct ubifs_info *c = sb->s_fs_info; |
434 | struct writeback_control wbc = { | 435 | struct writeback_control wbc = { |
435 | .sync_mode = wait ? WB_SYNC_ALL : WB_SYNC_NONE, | 436 | .sync_mode = WB_SYNC_ALL, |
436 | .range_start = 0, | 437 | .range_start = 0, |
437 | .range_end = LLONG_MAX, | 438 | .range_end = LLONG_MAX, |
438 | .nr_to_write = LONG_MAX, | 439 | .nr_to_write = LONG_MAX, |
439 | }; | 440 | }; |
440 | 441 | ||
441 | /* | 442 | /* |
442 | * Note by akpm about WB_SYNC_NONE used above: zero @wait is just an | 443 | * Zero @wait is just an advisory thing to help the file system shove |
443 | * advisory thing to help the file system shove lots of data into the | 444 | * lots of data into the queues, and there will be the second |
444 | * queues. If some gets missed then it'll be picked up on the second | ||
445 | * '->sync_fs()' call, with non-zero @wait. | 445 | * '->sync_fs()' call, with non-zero @wait. |
446 | */ | 446 | */ |
447 | if (!wait) | ||
448 | return 0; | ||
447 | 449 | ||
448 | if (sb->s_flags & MS_RDONLY) | 450 | if (sb->s_flags & MS_RDONLY) |
449 | return 0; | 451 | return 0; |
450 | 452 | ||
451 | /* | 453 | /* |
452 | * Synchronize write buffers, because 'ubifs_run_commit()' does not | ||
453 | * do this if it waits for an already running commit. | ||
454 | */ | ||
455 | for (i = 0; i < c->jhead_cnt; i++) { | ||
456 | err = ubifs_wbuf_sync(&c->jheads[i].wbuf); | ||
457 | if (err) | ||
458 | return err; | ||
459 | } | ||
460 | |||
461 | /* | ||
462 | * VFS calls '->sync_fs()' before synchronizing all dirty inodes and | 454 | * VFS calls '->sync_fs()' before synchronizing all dirty inodes and |
463 | * pages, so synchronize them first, then commit the journal. Strictly | 455 | * pages, so synchronize them first, then commit the journal. Strictly |
464 | * speaking, it is not necessary to commit the journal here, | 456 | * speaking, it is not necessary to commit the journal here, |
@@ -469,6 +461,16 @@ static int ubifs_sync_fs(struct super_block *sb, int wait) | |||
469 | */ | 461 | */ |
470 | generic_sync_sb_inodes(sb, &wbc); | 462 | generic_sync_sb_inodes(sb, &wbc); |
471 | 463 | ||
464 | /* | ||
465 | * Synchronize write buffers, because 'ubifs_run_commit()' does not | ||
466 | * do this if it waits for an already running commit. | ||
467 | */ | ||
468 | for (i = 0; i < c->jhead_cnt; i++) { | ||
469 | err = ubifs_wbuf_sync(&c->jheads[i].wbuf); | ||
470 | if (err) | ||
471 | return err; | ||
472 | } | ||
473 | |||
472 | err = ubifs_run_commit(c); | 474 | err = ubifs_run_commit(c); |
473 | if (err) | 475 | if (err) |
474 | return err; | 476 | return err; |
@@ -572,15 +574,8 @@ static int init_constants_early(struct ubifs_info *c) | |||
572 | c->ranges[UBIFS_IDX_NODE].max_len = INT_MAX; | 574 | c->ranges[UBIFS_IDX_NODE].max_len = INT_MAX; |
573 | 575 | ||
574 | /* | 576 | /* |
575 | * Initialize dead and dark LEB space watermarks. | 577 | * Initialize dead and dark LEB space watermarks. See gc.c for comments |
576 | * | 578 | * about these values. |
577 | * Dead space is the space which cannot be used. Its watermark is | ||
578 | * equivalent to min. I/O unit or minimum node size if it is greater | ||
579 | * then min. I/O unit. | ||
580 | * | ||
581 | * Dark space is the space which might be used, or might not, depending | ||
582 | * on which node should be written to the LEB. Its watermark is | ||
583 | * equivalent to maximum UBIFS node size. | ||
584 | */ | 579 | */ |
585 | c->dead_wm = ALIGN(MIN_WRITE_SZ, c->min_io_size); | 580 | c->dead_wm = ALIGN(MIN_WRITE_SZ, c->min_io_size); |
586 | c->dark_wm = ALIGN(UBIFS_MAX_NODE_SZ, c->min_io_size); | 581 | c->dark_wm = ALIGN(UBIFS_MAX_NODE_SZ, c->min_io_size); |
@@ -741,12 +736,12 @@ static void init_constants_master(struct ubifs_info *c) | |||
741 | * take_gc_lnum - reserve GC LEB. | 736 | * take_gc_lnum - reserve GC LEB. |
742 | * @c: UBIFS file-system description object | 737 | * @c: UBIFS file-system description object |
743 | * | 738 | * |
744 | * This function ensures that the LEB reserved for garbage collection is | 739 | * This function ensures that the LEB reserved for garbage collection is marked |
745 | * unmapped and is marked as "taken" in lprops. We also have to set free space | 740 | * as "taken" in lprops. We also have to set free space to LEB size and dirty |
746 | * to LEB size and dirty space to zero, because lprops may contain out-of-date | 741 | * space to zero, because lprops may contain out-of-date information if the |
747 | * information if the file-system was un-mounted before it has been committed. | 742 | * file-system was un-mounted before it has been committed. This function |
748 | * This function returns zero in case of success and a negative error code in | 743 | * returns zero in case of success and a negative error code in case of |
749 | * case of failure. | 744 | * failure. |
750 | */ | 745 | */ |
751 | static int take_gc_lnum(struct ubifs_info *c) | 746 | static int take_gc_lnum(struct ubifs_info *c) |
752 | { | 747 | { |
@@ -757,10 +752,6 @@ static int take_gc_lnum(struct ubifs_info *c) | |||
757 | return -EINVAL; | 752 | return -EINVAL; |
758 | } | 753 | } |
759 | 754 | ||
760 | err = ubifs_leb_unmap(c, c->gc_lnum); | ||
761 | if (err) | ||
762 | return err; | ||
763 | |||
764 | /* And we have to tell lprops that this LEB is taken */ | 755 | /* And we have to tell lprops that this LEB is taken */ |
765 | err = ubifs_change_one_lp(c, c->gc_lnum, c->leb_size, 0, | 756 | err = ubifs_change_one_lp(c, c->gc_lnum, c->leb_size, 0, |
766 | LPROPS_TAKEN, 0, 0); | 757 | LPROPS_TAKEN, 0, 0); |
@@ -966,13 +957,16 @@ static int ubifs_parse_options(struct ubifs_info *c, char *options, | |||
966 | 957 | ||
967 | token = match_token(p, tokens, args); | 958 | token = match_token(p, tokens, args); |
968 | switch (token) { | 959 | switch (token) { |
960 | /* | ||
961 | * %Opt_fast_unmount and %Opt_norm_unmount options are ignored. | ||
962 | * We accepte them in order to be backware-compatible. But this | ||
963 | * should be removed at some point. | ||
964 | */ | ||
969 | case Opt_fast_unmount: | 965 | case Opt_fast_unmount: |
970 | c->mount_opts.unmount_mode = 2; | 966 | c->mount_opts.unmount_mode = 2; |
971 | c->fast_unmount = 1; | ||
972 | break; | 967 | break; |
973 | case Opt_norm_unmount: | 968 | case Opt_norm_unmount: |
974 | c->mount_opts.unmount_mode = 1; | 969 | c->mount_opts.unmount_mode = 1; |
975 | c->fast_unmount = 0; | ||
976 | break; | 970 | break; |
977 | case Opt_bulk_read: | 971 | case Opt_bulk_read: |
978 | c->mount_opts.bulk_read = 2; | 972 | c->mount_opts.bulk_read = 2; |
@@ -1094,12 +1088,7 @@ static int check_free_space(struct ubifs_info *c) | |||
1094 | ubifs_err("insufficient free space to mount in read/write mode"); | 1088 | ubifs_err("insufficient free space to mount in read/write mode"); |
1095 | dbg_dump_budg(c); | 1089 | dbg_dump_budg(c); |
1096 | dbg_dump_lprops(c); | 1090 | dbg_dump_lprops(c); |
1097 | /* | 1091 | return -ENOSPC; |
1098 | * We return %-EINVAL instead of %-ENOSPC because it seems to | ||
1099 | * be the closest error code mentioned in the mount function | ||
1100 | * documentation. | ||
1101 | */ | ||
1102 | return -EINVAL; | ||
1103 | } | 1092 | } |
1104 | return 0; | 1093 | return 0; |
1105 | } | 1094 | } |
@@ -1286,10 +1275,19 @@ static int mount_ubifs(struct ubifs_info *c) | |||
1286 | if (err) | 1275 | if (err) |
1287 | goto out_orphans; | 1276 | goto out_orphans; |
1288 | err = ubifs_rcvry_gc_commit(c); | 1277 | err = ubifs_rcvry_gc_commit(c); |
1289 | } else | 1278 | } else { |
1290 | err = take_gc_lnum(c); | 1279 | err = take_gc_lnum(c); |
1291 | if (err) | 1280 | if (err) |
1292 | goto out_orphans; | 1281 | goto out_orphans; |
1282 | |||
1283 | /* | ||
1284 | * GC LEB may contain garbage if there was an unclean | ||
1285 | * reboot, and it should be un-mapped. | ||
1286 | */ | ||
1287 | err = ubifs_leb_unmap(c, c->gc_lnum); | ||
1288 | if (err) | ||
1289 | return err; | ||
1290 | } | ||
1293 | 1291 | ||
1294 | err = dbg_check_lprops(c); | 1292 | err = dbg_check_lprops(c); |
1295 | if (err) | 1293 | if (err) |
@@ -1298,6 +1296,16 @@ static int mount_ubifs(struct ubifs_info *c) | |||
1298 | err = ubifs_recover_size(c); | 1296 | err = ubifs_recover_size(c); |
1299 | if (err) | 1297 | if (err) |
1300 | goto out_orphans; | 1298 | goto out_orphans; |
1299 | } else { | ||
1300 | /* | ||
1301 | * Even if we mount read-only, we have to set space in GC LEB | ||
1302 | * to proper value because this affects UBIFS free space | ||
1303 | * reporting. We do not want to have a situation when | ||
1304 | * re-mounting from R/O to R/W changes amount of free space. | ||
1305 | */ | ||
1306 | err = take_gc_lnum(c); | ||
1307 | if (err) | ||
1308 | goto out_orphans; | ||
1301 | } | 1309 | } |
1302 | 1310 | ||
1303 | spin_lock(&ubifs_infos_lock); | 1311 | spin_lock(&ubifs_infos_lock); |
@@ -1310,14 +1318,17 @@ static int mount_ubifs(struct ubifs_info *c) | |||
1310 | else { | 1318 | else { |
1311 | c->need_recovery = 0; | 1319 | c->need_recovery = 0; |
1312 | ubifs_msg("recovery completed"); | 1320 | ubifs_msg("recovery completed"); |
1321 | /* GC LEB has to be empty and taken at this point */ | ||
1322 | ubifs_assert(c->lst.taken_empty_lebs == 1); | ||
1313 | } | 1323 | } |
1314 | } | 1324 | } else |
1325 | ubifs_assert(c->lst.taken_empty_lebs == 1); | ||
1315 | 1326 | ||
1316 | err = dbg_debugfs_init_fs(c); | 1327 | err = dbg_check_filesystem(c); |
1317 | if (err) | 1328 | if (err) |
1318 | goto out_infos; | 1329 | goto out_infos; |
1319 | 1330 | ||
1320 | err = dbg_check_filesystem(c); | 1331 | err = dbg_debugfs_init_fs(c); |
1321 | if (err) | 1332 | if (err) |
1322 | goto out_infos; | 1333 | goto out_infos; |
1323 | 1334 | ||
@@ -1351,7 +1362,6 @@ static int mount_ubifs(struct ubifs_info *c) | |||
1351 | c->uuid[4], c->uuid[5], c->uuid[6], c->uuid[7], | 1362 | c->uuid[4], c->uuid[5], c->uuid[6], c->uuid[7], |
1352 | c->uuid[8], c->uuid[9], c->uuid[10], c->uuid[11], | 1363 | c->uuid[8], c->uuid[9], c->uuid[10], c->uuid[11], |
1353 | c->uuid[12], c->uuid[13], c->uuid[14], c->uuid[15]); | 1364 | c->uuid[12], c->uuid[13], c->uuid[14], c->uuid[15]); |
1354 | dbg_msg("fast unmount: %d", c->fast_unmount); | ||
1355 | dbg_msg("big_lpt %d", c->big_lpt); | 1365 | dbg_msg("big_lpt %d", c->big_lpt); |
1356 | dbg_msg("log LEBs: %d (%d - %d)", | 1366 | dbg_msg("log LEBs: %d (%d - %d)", |
1357 | c->log_lebs, UBIFS_LOG_LNUM, c->log_last); | 1367 | c->log_lebs, UBIFS_LOG_LNUM, c->log_last); |
@@ -1475,10 +1485,8 @@ static int ubifs_remount_rw(struct ubifs_info *c) | |||
1475 | { | 1485 | { |
1476 | int err, lnum; | 1486 | int err, lnum; |
1477 | 1487 | ||
1478 | if (c->ro_media) | ||
1479 | return -EINVAL; | ||
1480 | |||
1481 | mutex_lock(&c->umount_mutex); | 1488 | mutex_lock(&c->umount_mutex); |
1489 | dbg_save_space_info(c); | ||
1482 | c->remounting_rw = 1; | 1490 | c->remounting_rw = 1; |
1483 | c->always_chk_crc = 1; | 1491 | c->always_chk_crc = 1; |
1484 | 1492 | ||
@@ -1514,6 +1522,12 @@ static int ubifs_remount_rw(struct ubifs_info *c) | |||
1514 | err = ubifs_recover_inl_heads(c, c->sbuf); | 1522 | err = ubifs_recover_inl_heads(c, c->sbuf); |
1515 | if (err) | 1523 | if (err) |
1516 | goto out; | 1524 | goto out; |
1525 | } else { | ||
1526 | /* A readonly mount is not allowed to have orphans */ | ||
1527 | ubifs_assert(c->tot_orphans == 0); | ||
1528 | err = ubifs_clear_orphans(c); | ||
1529 | if (err) | ||
1530 | goto out; | ||
1517 | } | 1531 | } |
1518 | 1532 | ||
1519 | if (!(c->mst_node->flags & cpu_to_le32(UBIFS_MST_DIRTY))) { | 1533 | if (!(c->mst_node->flags & cpu_to_le32(UBIFS_MST_DIRTY))) { |
@@ -1569,7 +1583,7 @@ static int ubifs_remount_rw(struct ubifs_info *c) | |||
1569 | if (c->need_recovery) | 1583 | if (c->need_recovery) |
1570 | err = ubifs_rcvry_gc_commit(c); | 1584 | err = ubifs_rcvry_gc_commit(c); |
1571 | else | 1585 | else |
1572 | err = take_gc_lnum(c); | 1586 | err = ubifs_leb_unmap(c, c->gc_lnum); |
1573 | if (err) | 1587 | if (err) |
1574 | goto out; | 1588 | goto out; |
1575 | 1589 | ||
@@ -1582,8 +1596,9 @@ static int ubifs_remount_rw(struct ubifs_info *c) | |||
1582 | c->vfs_sb->s_flags &= ~MS_RDONLY; | 1596 | c->vfs_sb->s_flags &= ~MS_RDONLY; |
1583 | c->remounting_rw = 0; | 1597 | c->remounting_rw = 0; |
1584 | c->always_chk_crc = 0; | 1598 | c->always_chk_crc = 0; |
1599 | err = dbg_check_space_info(c); | ||
1585 | mutex_unlock(&c->umount_mutex); | 1600 | mutex_unlock(&c->umount_mutex); |
1586 | return 0; | 1601 | return err; |
1587 | 1602 | ||
1588 | out: | 1603 | out: |
1589 | vfree(c->orph_buf); | 1604 | vfree(c->orph_buf); |
@@ -1603,43 +1618,18 @@ out: | |||
1603 | } | 1618 | } |
1604 | 1619 | ||
1605 | /** | 1620 | /** |
1606 | * commit_on_unmount - commit the journal when un-mounting. | ||
1607 | * @c: UBIFS file-system description object | ||
1608 | * | ||
1609 | * This function is called during un-mounting and re-mounting, and it commits | ||
1610 | * the journal unless the "fast unmount" mode is enabled. | ||
1611 | */ | ||
1612 | static void commit_on_unmount(struct ubifs_info *c) | ||
1613 | { | ||
1614 | struct super_block *sb = c->vfs_sb; | ||
1615 | long long bud_bytes; | ||
1616 | |||
1617 | /* | ||
1618 | * This function is called before the background thread is stopped, so | ||
1619 | * we may race with ongoing commit, which means we have to take | ||
1620 | * @c->bud_lock to access @c->bud_bytes. | ||
1621 | */ | ||
1622 | spin_lock(&c->buds_lock); | ||
1623 | bud_bytes = c->bud_bytes; | ||
1624 | spin_unlock(&c->buds_lock); | ||
1625 | |||
1626 | if (!c->fast_unmount && !(sb->s_flags & MS_RDONLY) && bud_bytes) | ||
1627 | ubifs_run_commit(c); | ||
1628 | } | ||
1629 | |||
1630 | /** | ||
1631 | * ubifs_remount_ro - re-mount in read-only mode. | 1621 | * ubifs_remount_ro - re-mount in read-only mode. |
1632 | * @c: UBIFS file-system description object | 1622 | * @c: UBIFS file-system description object |
1633 | * | 1623 | * |
1634 | * We rely on VFS to have stopped writing. Possibly the background thread could | 1624 | * We assume VFS has stopped writing. Possibly the background thread could be |
1635 | * be running a commit, however kthread_stop will wait in that case. | 1625 | * running a commit, however kthread_stop will wait in that case. |
1636 | */ | 1626 | */ |
1637 | static void ubifs_remount_ro(struct ubifs_info *c) | 1627 | static void ubifs_remount_ro(struct ubifs_info *c) |
1638 | { | 1628 | { |
1639 | int i, err; | 1629 | int i, err; |
1640 | 1630 | ||
1641 | ubifs_assert(!c->need_recovery); | 1631 | ubifs_assert(!c->need_recovery); |
1642 | commit_on_unmount(c); | 1632 | ubifs_assert(!(c->vfs_sb->s_flags & MS_RDONLY)); |
1643 | 1633 | ||
1644 | mutex_lock(&c->umount_mutex); | 1634 | mutex_lock(&c->umount_mutex); |
1645 | if (c->bgt) { | 1635 | if (c->bgt) { |
@@ -1647,27 +1637,29 @@ static void ubifs_remount_ro(struct ubifs_info *c) | |||
1647 | c->bgt = NULL; | 1637 | c->bgt = NULL; |
1648 | } | 1638 | } |
1649 | 1639 | ||
1640 | dbg_save_space_info(c); | ||
1641 | |||
1650 | for (i = 0; i < c->jhead_cnt; i++) { | 1642 | for (i = 0; i < c->jhead_cnt; i++) { |
1651 | ubifs_wbuf_sync(&c->jheads[i].wbuf); | 1643 | ubifs_wbuf_sync(&c->jheads[i].wbuf); |
1652 | del_timer_sync(&c->jheads[i].wbuf.timer); | 1644 | del_timer_sync(&c->jheads[i].wbuf.timer); |
1653 | } | 1645 | } |
1654 | 1646 | ||
1655 | if (!c->ro_media) { | 1647 | c->mst_node->flags &= ~cpu_to_le32(UBIFS_MST_DIRTY); |
1656 | c->mst_node->flags &= ~cpu_to_le32(UBIFS_MST_DIRTY); | 1648 | c->mst_node->flags |= cpu_to_le32(UBIFS_MST_NO_ORPHS); |
1657 | c->mst_node->flags |= cpu_to_le32(UBIFS_MST_NO_ORPHS); | 1649 | c->mst_node->gc_lnum = cpu_to_le32(c->gc_lnum); |
1658 | c->mst_node->gc_lnum = cpu_to_le32(c->gc_lnum); | 1650 | err = ubifs_write_master(c); |
1659 | err = ubifs_write_master(c); | 1651 | if (err) |
1660 | if (err) | 1652 | ubifs_ro_mode(c, err); |
1661 | ubifs_ro_mode(c, err); | ||
1662 | } | ||
1663 | 1653 | ||
1664 | ubifs_destroy_idx_gc(c); | ||
1665 | free_wbufs(c); | 1654 | free_wbufs(c); |
1666 | vfree(c->orph_buf); | 1655 | vfree(c->orph_buf); |
1667 | c->orph_buf = NULL; | 1656 | c->orph_buf = NULL; |
1668 | vfree(c->ileb_buf); | 1657 | vfree(c->ileb_buf); |
1669 | c->ileb_buf = NULL; | 1658 | c->ileb_buf = NULL; |
1670 | ubifs_lpt_free(c, 1); | 1659 | ubifs_lpt_free(c, 1); |
1660 | err = dbg_check_space_info(c); | ||
1661 | if (err) | ||
1662 | ubifs_ro_mode(c, err); | ||
1671 | mutex_unlock(&c->umount_mutex); | 1663 | mutex_unlock(&c->umount_mutex); |
1672 | } | 1664 | } |
1673 | 1665 | ||
@@ -1760,11 +1752,20 @@ static int ubifs_remount_fs(struct super_block *sb, int *flags, char *data) | |||
1760 | } | 1752 | } |
1761 | 1753 | ||
1762 | if ((sb->s_flags & MS_RDONLY) && !(*flags & MS_RDONLY)) { | 1754 | if ((sb->s_flags & MS_RDONLY) && !(*flags & MS_RDONLY)) { |
1755 | if (c->ro_media) { | ||
1756 | ubifs_msg("cannot re-mount due to prior errors"); | ||
1757 | return -EROFS; | ||
1758 | } | ||
1763 | err = ubifs_remount_rw(c); | 1759 | err = ubifs_remount_rw(c); |
1764 | if (err) | 1760 | if (err) |
1765 | return err; | 1761 | return err; |
1766 | } else if (!(sb->s_flags & MS_RDONLY) && (*flags & MS_RDONLY)) | 1762 | } else if (!(sb->s_flags & MS_RDONLY) && (*flags & MS_RDONLY)) { |
1763 | if (c->ro_media) { | ||
1764 | ubifs_msg("cannot re-mount due to prior errors"); | ||
1765 | return -EROFS; | ||
1766 | } | ||
1767 | ubifs_remount_ro(c); | 1767 | ubifs_remount_ro(c); |
1768 | } | ||
1768 | 1769 | ||
1769 | if (c->bulk_read == 1) | 1770 | if (c->bulk_read == 1) |
1770 | bu_init(c); | 1771 | bu_init(c); |
@@ -1774,10 +1775,11 @@ static int ubifs_remount_fs(struct super_block *sb, int *flags, char *data) | |||
1774 | c->bu.buf = NULL; | 1775 | c->bu.buf = NULL; |
1775 | } | 1776 | } |
1776 | 1777 | ||
1778 | ubifs_assert(c->lst.taken_empty_lebs == 1); | ||
1777 | return 0; | 1779 | return 0; |
1778 | } | 1780 | } |
1779 | 1781 | ||
1780 | struct super_operations ubifs_super_operations = { | 1782 | const struct super_operations ubifs_super_operations = { |
1781 | .alloc_inode = ubifs_alloc_inode, | 1783 | .alloc_inode = ubifs_alloc_inode, |
1782 | .destroy_inode = ubifs_destroy_inode, | 1784 | .destroy_inode = ubifs_destroy_inode, |
1783 | .put_super = ubifs_put_super, | 1785 | .put_super = ubifs_put_super, |
@@ -2044,15 +2046,6 @@ out_close: | |||
2044 | 2046 | ||
2045 | static void ubifs_kill_sb(struct super_block *sb) | 2047 | static void ubifs_kill_sb(struct super_block *sb) |
2046 | { | 2048 | { |
2047 | struct ubifs_info *c = sb->s_fs_info; | ||
2048 | |||
2049 | /* | ||
2050 | * We do 'commit_on_unmount()' here instead of 'ubifs_put_super()' | ||
2051 | * in order to be outside BKL. | ||
2052 | */ | ||
2053 | if (sb->s_root) | ||
2054 | commit_on_unmount(c); | ||
2055 | /* The un-mount routine is actually done in put_super() */ | ||
2056 | generic_shutdown_super(sb); | 2049 | generic_shutdown_super(sb); |
2057 | } | 2050 | } |
2058 | 2051 | ||
diff --git a/fs/ubifs/tnc.c b/fs/ubifs/tnc.c index f7e36f545527..fa28a84c6a1b 100644 --- a/fs/ubifs/tnc.c +++ b/fs/ubifs/tnc.c | |||
@@ -443,6 +443,11 @@ static int tnc_read_node_nm(struct ubifs_info *c, struct ubifs_zbranch *zbr, | |||
443 | * This function performs that same function as ubifs_read_node except that | 443 | * This function performs that same function as ubifs_read_node except that |
444 | * it does not require that there is actually a node present and instead | 444 | * it does not require that there is actually a node present and instead |
445 | * the return code indicates if a node was read. | 445 | * the return code indicates if a node was read. |
446 | * | ||
447 | * Note, this function does not check CRC of data nodes if @c->no_chk_data_crc | ||
448 | * is true (it is controlled by corresponding mount option). However, if | ||
449 | * @c->always_chk_crc is true, @c->no_chk_data_crc is ignored and CRC is always | ||
450 | * checked. | ||
446 | */ | 451 | */ |
447 | static int try_read_node(const struct ubifs_info *c, void *buf, int type, | 452 | static int try_read_node(const struct ubifs_info *c, void *buf, int type, |
448 | int len, int lnum, int offs) | 453 | int len, int lnum, int offs) |
@@ -470,9 +475,8 @@ static int try_read_node(const struct ubifs_info *c, void *buf, int type, | |||
470 | if (node_len != len) | 475 | if (node_len != len) |
471 | return 0; | 476 | return 0; |
472 | 477 | ||
473 | if (type == UBIFS_DATA_NODE && !c->always_chk_crc) | 478 | if (type == UBIFS_DATA_NODE && !c->always_chk_crc && c->no_chk_data_crc) |
474 | if (c->no_chk_data_crc) | 479 | return 1; |
475 | return 0; | ||
476 | 480 | ||
477 | crc = crc32(UBIFS_CRC32_INIT, buf + 8, node_len - 8); | 481 | crc = crc32(UBIFS_CRC32_INIT, buf + 8, node_len - 8); |
478 | node_crc = le32_to_cpu(ch->crc); | 482 | node_crc = le32_to_cpu(ch->crc); |
@@ -1506,7 +1510,7 @@ out: | |||
1506 | * | 1510 | * |
1507 | * Note, if the bulk-read buffer length (@bu->buf_len) is known, this function | 1511 | * Note, if the bulk-read buffer length (@bu->buf_len) is known, this function |
1508 | * makes sure bulk-read nodes fit the buffer. Otherwise, this function prepares | 1512 | * makes sure bulk-read nodes fit the buffer. Otherwise, this function prepares |
1509 | * maxumum possible amount of nodes for bulk-read. | 1513 | * maximum possible amount of nodes for bulk-read. |
1510 | */ | 1514 | */ |
1511 | int ubifs_tnc_get_bu_keys(struct ubifs_info *c, struct bu_info *bu) | 1515 | int ubifs_tnc_get_bu_keys(struct ubifs_info *c, struct bu_info *bu) |
1512 | { | 1516 | { |
diff --git a/fs/ubifs/ubifs.h b/fs/ubifs/ubifs.h index fc2a4cc66d03..039a68bee29a 100644 --- a/fs/ubifs/ubifs.h +++ b/fs/ubifs/ubifs.h | |||
@@ -426,9 +426,9 @@ struct ubifs_unclean_leb { | |||
426 | * LEB properties flags. | 426 | * LEB properties flags. |
427 | * | 427 | * |
428 | * LPROPS_UNCAT: not categorized | 428 | * LPROPS_UNCAT: not categorized |
429 | * LPROPS_DIRTY: dirty > 0, not index | 429 | * LPROPS_DIRTY: dirty > free, dirty >= @c->dead_wm, not index |
430 | * LPROPS_DIRTY_IDX: dirty + free > @c->min_idx_node_sze and index | 430 | * LPROPS_DIRTY_IDX: dirty + free > @c->min_idx_node_sze and index |
431 | * LPROPS_FREE: free > 0, not empty, not index | 431 | * LPROPS_FREE: free > 0, dirty < @c->dead_wm, not empty, not index |
432 | * LPROPS_HEAP_CNT: number of heaps used for storing categorized LEBs | 432 | * LPROPS_HEAP_CNT: number of heaps used for storing categorized LEBs |
433 | * LPROPS_EMPTY: LEB is empty, not taken | 433 | * LPROPS_EMPTY: LEB is empty, not taken |
434 | * LPROPS_FREEABLE: free + dirty == leb_size, not index, not taken | 434 | * LPROPS_FREEABLE: free + dirty == leb_size, not index, not taken |
@@ -961,7 +961,6 @@ struct ubifs_debug_info; | |||
961 | * @cs_lock: commit state lock | 961 | * @cs_lock: commit state lock |
962 | * @cmt_wq: wait queue to sleep on if the log is full and a commit is running | 962 | * @cmt_wq: wait queue to sleep on if the log is full and a commit is running |
963 | * | 963 | * |
964 | * @fast_unmount: do not run journal commit before un-mounting | ||
965 | * @big_lpt: flag that LPT is too big to write whole during commit | 964 | * @big_lpt: flag that LPT is too big to write whole during commit |
966 | * @no_chk_data_crc: do not check CRCs when reading data nodes (except during | 965 | * @no_chk_data_crc: do not check CRCs when reading data nodes (except during |
967 | * recovery) | 966 | * recovery) |
@@ -1202,7 +1201,6 @@ struct ubifs_info { | |||
1202 | spinlock_t cs_lock; | 1201 | spinlock_t cs_lock; |
1203 | wait_queue_head_t cmt_wq; | 1202 | wait_queue_head_t cmt_wq; |
1204 | 1203 | ||
1205 | unsigned int fast_unmount:1; | ||
1206 | unsigned int big_lpt:1; | 1204 | unsigned int big_lpt:1; |
1207 | unsigned int no_chk_data_crc:1; | 1205 | unsigned int no_chk_data_crc:1; |
1208 | unsigned int bulk_read:1; | 1206 | unsigned int bulk_read:1; |
@@ -1405,13 +1403,13 @@ extern struct list_head ubifs_infos; | |||
1405 | extern spinlock_t ubifs_infos_lock; | 1403 | extern spinlock_t ubifs_infos_lock; |
1406 | extern atomic_long_t ubifs_clean_zn_cnt; | 1404 | extern atomic_long_t ubifs_clean_zn_cnt; |
1407 | extern struct kmem_cache *ubifs_inode_slab; | 1405 | extern struct kmem_cache *ubifs_inode_slab; |
1408 | extern struct super_operations ubifs_super_operations; | 1406 | extern const struct super_operations ubifs_super_operations; |
1409 | extern struct address_space_operations ubifs_file_address_operations; | 1407 | extern const struct address_space_operations ubifs_file_address_operations; |
1410 | extern struct file_operations ubifs_file_operations; | 1408 | extern const struct file_operations ubifs_file_operations; |
1411 | extern struct inode_operations ubifs_file_inode_operations; | 1409 | extern const struct inode_operations ubifs_file_inode_operations; |
1412 | extern struct file_operations ubifs_dir_operations; | 1410 | extern const struct file_operations ubifs_dir_operations; |
1413 | extern struct inode_operations ubifs_dir_inode_operations; | 1411 | extern const struct inode_operations ubifs_dir_inode_operations; |
1414 | extern struct inode_operations ubifs_symlink_inode_operations; | 1412 | extern const struct inode_operations ubifs_symlink_inode_operations; |
1415 | extern struct backing_dev_info ubifs_backing_dev_info; | 1413 | extern struct backing_dev_info ubifs_backing_dev_info; |
1416 | extern struct ubifs_compressor *ubifs_compressors[UBIFS_COMPR_TYPES_CNT]; | 1414 | extern struct ubifs_compressor *ubifs_compressors[UBIFS_COMPR_TYPES_CNT]; |
1417 | 1415 | ||
@@ -1428,7 +1426,7 @@ int ubifs_read_node_wbuf(struct ubifs_wbuf *wbuf, void *buf, int type, int len, | |||
1428 | int ubifs_write_node(struct ubifs_info *c, void *node, int len, int lnum, | 1426 | int ubifs_write_node(struct ubifs_info *c, void *node, int len, int lnum, |
1429 | int offs, int dtype); | 1427 | int offs, int dtype); |
1430 | int ubifs_check_node(const struct ubifs_info *c, const void *buf, int lnum, | 1428 | int ubifs_check_node(const struct ubifs_info *c, const void *buf, int lnum, |
1431 | int offs, int quiet, int chk_crc); | 1429 | int offs, int quiet, int must_chk_crc); |
1432 | void ubifs_prepare_node(struct ubifs_info *c, void *buf, int len, int pad); | 1430 | void ubifs_prepare_node(struct ubifs_info *c, void *buf, int len, int pad); |
1433 | void ubifs_prep_grp_node(struct ubifs_info *c, void *node, int len, int last); | 1431 | void ubifs_prep_grp_node(struct ubifs_info *c, void *node, int len, int last); |
1434 | int ubifs_io_init(struct ubifs_info *c); | 1432 | int ubifs_io_init(struct ubifs_info *c); |
@@ -1495,6 +1493,7 @@ void ubifs_release_ino_dirty(struct ubifs_info *c, struct inode *inode, | |||
1495 | void ubifs_cancel_ino_op(struct ubifs_info *c, struct inode *inode, | 1493 | void ubifs_cancel_ino_op(struct ubifs_info *c, struct inode *inode, |
1496 | struct ubifs_budget_req *req); | 1494 | struct ubifs_budget_req *req); |
1497 | long long ubifs_get_free_space(struct ubifs_info *c); | 1495 | long long ubifs_get_free_space(struct ubifs_info *c); |
1496 | long long ubifs_get_free_space_nolock(struct ubifs_info *c); | ||
1498 | int ubifs_calc_min_idx_lebs(struct ubifs_info *c); | 1497 | int ubifs_calc_min_idx_lebs(struct ubifs_info *c); |
1499 | void ubifs_convert_page_budget(struct ubifs_info *c); | 1498 | void ubifs_convert_page_budget(struct ubifs_info *c); |
1500 | long long ubifs_reported_space(const struct ubifs_info *c, long long free); | 1499 | long long ubifs_reported_space(const struct ubifs_info *c, long long free); |
@@ -1603,6 +1602,7 @@ void ubifs_delete_orphan(struct ubifs_info *c, ino_t inum); | |||
1603 | int ubifs_orphan_start_commit(struct ubifs_info *c); | 1602 | int ubifs_orphan_start_commit(struct ubifs_info *c); |
1604 | int ubifs_orphan_end_commit(struct ubifs_info *c); | 1603 | int ubifs_orphan_end_commit(struct ubifs_info *c); |
1605 | int ubifs_mount_orphans(struct ubifs_info *c, int unclean, int read_only); | 1604 | int ubifs_mount_orphans(struct ubifs_info *c, int unclean, int read_only); |
1605 | int ubifs_clear_orphans(struct ubifs_info *c); | ||
1606 | 1606 | ||
1607 | /* lpt.c */ | 1607 | /* lpt.c */ |
1608 | int ubifs_calc_lpt_geom(struct ubifs_info *c); | 1608 | int ubifs_calc_lpt_geom(struct ubifs_info *c); |
@@ -1646,7 +1646,7 @@ const struct ubifs_lprops *ubifs_change_lp(struct ubifs_info *c, | |||
1646 | const struct ubifs_lprops *lp, | 1646 | const struct ubifs_lprops *lp, |
1647 | int free, int dirty, int flags, | 1647 | int free, int dirty, int flags, |
1648 | int idx_gc_cnt); | 1648 | int idx_gc_cnt); |
1649 | void ubifs_get_lp_stats(struct ubifs_info *c, struct ubifs_lp_stats *stats); | 1649 | void ubifs_get_lp_stats(struct ubifs_info *c, struct ubifs_lp_stats *lst); |
1650 | void ubifs_add_to_cat(struct ubifs_info *c, struct ubifs_lprops *lprops, | 1650 | void ubifs_add_to_cat(struct ubifs_info *c, struct ubifs_lprops *lprops, |
1651 | int cat); | 1651 | int cat); |
1652 | void ubifs_replace_cat(struct ubifs_info *c, struct ubifs_lprops *old_lprops, | 1652 | void ubifs_replace_cat(struct ubifs_info *c, struct ubifs_lprops *old_lprops, |
diff --git a/fs/udf/Kconfig b/fs/udf/Kconfig new file mode 100644 index 000000000000..0e0e99bd6bce --- /dev/null +++ b/fs/udf/Kconfig | |||
@@ -0,0 +1,18 @@ | |||
1 | config UDF_FS | ||
2 | tristate "UDF file system support" | ||
3 | select CRC_ITU_T | ||
4 | help | ||
5 | This is the new file system used on some CD-ROMs and DVDs. Say Y if | ||
6 | you intend to mount DVD discs or CDRW's written in packet mode, or | ||
7 | if written to by other UDF utilities, such as DirectCD. | ||
8 | Please read <file:Documentation/filesystems/udf.txt>. | ||
9 | |||
10 | To compile this file system support as a module, choose M here: the | ||
11 | module will be called udf. | ||
12 | |||
13 | If unsure, say N. | ||
14 | |||
15 | config UDF_NLS | ||
16 | bool | ||
17 | default y | ||
18 | depends on (UDF_FS=m && NLS) || (UDF_FS=y && NLS=y) | ||
diff --git a/fs/ufs/Kconfig b/fs/ufs/Kconfig new file mode 100644 index 000000000000..e4f10a40768a --- /dev/null +++ b/fs/ufs/Kconfig | |||
@@ -0,0 +1,43 @@ | |||
1 | config UFS_FS | ||
2 | tristate "UFS file system support (read only)" | ||
3 | depends on BLOCK | ||
4 | help | ||
5 | BSD and derivate versions of Unix (such as SunOS, FreeBSD, NetBSD, | ||
6 | OpenBSD and NeXTstep) use a file system called UFS. Some System V | ||
7 | Unixes can create and mount hard disk partitions and diskettes using | ||
8 | this file system as well. Saying Y here will allow you to read from | ||
9 | these partitions; if you also want to write to them, say Y to the | ||
10 | experimental "UFS file system write support", below. Please read the | ||
11 | file <file:Documentation/filesystems/ufs.txt> for more information. | ||
12 | |||
13 | The recently released UFS2 variant (used in FreeBSD 5.x) is | ||
14 | READ-ONLY supported. | ||
15 | |||
16 | Note that this option is generally not needed for floppies, since a | ||
17 | good portable way to transport files and directories between unixes | ||
18 | (and even other operating systems) is given by the tar program ("man | ||
19 | tar" or preferably "info tar"). | ||
20 | |||
21 | When accessing NeXTstep files, you may need to convert them from the | ||
22 | NeXT character set to the Latin1 character set; use the program | ||
23 | recode ("info recode") for this purpose. | ||
24 | |||
25 | To compile the UFS file system support as a module, choose M here: the | ||
26 | module will be called ufs. | ||
27 | |||
28 | If you haven't heard about all of this before, it's safe to say N. | ||
29 | |||
30 | config UFS_FS_WRITE | ||
31 | bool "UFS file system write support (DANGEROUS)" | ||
32 | depends on UFS_FS && EXPERIMENTAL | ||
33 | help | ||
34 | Say Y here if you want to try writing to UFS partitions. This is | ||
35 | experimental, so you should back up your UFS partitions beforehand. | ||
36 | |||
37 | config UFS_DEBUG | ||
38 | bool "UFS debugging" | ||
39 | depends on UFS_FS | ||
40 | help | ||
41 | If you are experiencing any problems with the UFS filesystem, say | ||
42 | Y here. This will result in _many_ additional debugging messages to be | ||
43 | written to the system log. | ||
diff --git a/fs/utimes.c b/fs/utimes.c index 6929e3e91d05..e4c75db5d373 100644 --- a/fs/utimes.c +++ b/fs/utimes.c | |||
@@ -24,7 +24,7 @@ | |||
24 | * must be owner or have write permission. | 24 | * must be owner or have write permission. |
25 | * Else, update from *times, must be owner or super user. | 25 | * Else, update from *times, must be owner or super user. |
26 | */ | 26 | */ |
27 | asmlinkage long sys_utime(char __user *filename, struct utimbuf __user *times) | 27 | SYSCALL_DEFINE2(utime, char __user *, filename, struct utimbuf __user *, times) |
28 | { | 28 | { |
29 | struct timespec tv[2]; | 29 | struct timespec tv[2]; |
30 | 30 | ||
@@ -170,7 +170,8 @@ out: | |||
170 | return error; | 170 | return error; |
171 | } | 171 | } |
172 | 172 | ||
173 | asmlinkage long sys_utimensat(int dfd, char __user *filename, struct timespec __user *utimes, int flags) | 173 | SYSCALL_DEFINE4(utimensat, int, dfd, char __user *, filename, |
174 | struct timespec __user *, utimes, int, flags) | ||
174 | { | 175 | { |
175 | struct timespec tstimes[2]; | 176 | struct timespec tstimes[2]; |
176 | 177 | ||
@@ -187,7 +188,8 @@ asmlinkage long sys_utimensat(int dfd, char __user *filename, struct timespec __ | |||
187 | return do_utimes(dfd, filename, utimes ? tstimes : NULL, flags); | 188 | return do_utimes(dfd, filename, utimes ? tstimes : NULL, flags); |
188 | } | 189 | } |
189 | 190 | ||
190 | asmlinkage long sys_futimesat(int dfd, char __user *filename, struct timeval __user *utimes) | 191 | SYSCALL_DEFINE3(futimesat, int, dfd, char __user *, filename, |
192 | struct timeval __user *, utimes) | ||
191 | { | 193 | { |
192 | struct timeval times[2]; | 194 | struct timeval times[2]; |
193 | struct timespec tstimes[2]; | 195 | struct timespec tstimes[2]; |
@@ -214,7 +216,8 @@ asmlinkage long sys_futimesat(int dfd, char __user *filename, struct timeval __u | |||
214 | return do_utimes(dfd, filename, utimes ? tstimes : NULL, 0); | 216 | return do_utimes(dfd, filename, utimes ? tstimes : NULL, 0); |
215 | } | 217 | } |
216 | 218 | ||
217 | asmlinkage long sys_utimes(char __user *filename, struct timeval __user *utimes) | 219 | SYSCALL_DEFINE2(utimes, char __user *, filename, |
220 | struct timeval __user *, utimes) | ||
218 | { | 221 | { |
219 | return sys_futimesat(AT_FDCWD, filename, utimes); | 222 | return sys_futimesat(AT_FDCWD, filename, utimes); |
220 | } | 223 | } |
diff --git a/fs/xattr.c b/fs/xattr.c index 237804cd6b56..197c4fcac032 100644 --- a/fs/xattr.c +++ b/fs/xattr.c | |||
@@ -251,9 +251,9 @@ setxattr(struct dentry *d, const char __user *name, const void __user *value, | |||
251 | return error; | 251 | return error; |
252 | } | 252 | } |
253 | 253 | ||
254 | asmlinkage long | 254 | SYSCALL_DEFINE5(setxattr, const char __user *, pathname, |
255 | sys_setxattr(const char __user *pathname, const char __user *name, | 255 | const char __user *, name, const void __user *, value, |
256 | const void __user *value, size_t size, int flags) | 256 | size_t, size, int, flags) |
257 | { | 257 | { |
258 | struct path path; | 258 | struct path path; |
259 | int error; | 259 | int error; |
@@ -270,9 +270,9 @@ sys_setxattr(const char __user *pathname, const char __user *name, | |||
270 | return error; | 270 | return error; |
271 | } | 271 | } |
272 | 272 | ||
273 | asmlinkage long | 273 | SYSCALL_DEFINE5(lsetxattr, const char __user *, pathname, |
274 | sys_lsetxattr(const char __user *pathname, const char __user *name, | 274 | const char __user *, name, const void __user *, value, |
275 | const void __user *value, size_t size, int flags) | 275 | size_t, size, int, flags) |
276 | { | 276 | { |
277 | struct path path; | 277 | struct path path; |
278 | int error; | 278 | int error; |
@@ -289,9 +289,8 @@ sys_lsetxattr(const char __user *pathname, const char __user *name, | |||
289 | return error; | 289 | return error; |
290 | } | 290 | } |
291 | 291 | ||
292 | asmlinkage long | 292 | SYSCALL_DEFINE5(fsetxattr, int, fd, const char __user *, name, |
293 | sys_fsetxattr(int fd, const char __user *name, const void __user *value, | 293 | const void __user *,value, size_t, size, int, flags) |
294 | size_t size, int flags) | ||
295 | { | 294 | { |
296 | struct file *f; | 295 | struct file *f; |
297 | struct dentry *dentry; | 296 | struct dentry *dentry; |
@@ -349,9 +348,8 @@ getxattr(struct dentry *d, const char __user *name, void __user *value, | |||
349 | return error; | 348 | return error; |
350 | } | 349 | } |
351 | 350 | ||
352 | asmlinkage ssize_t | 351 | SYSCALL_DEFINE4(getxattr, const char __user *, pathname, |
353 | sys_getxattr(const char __user *pathname, const char __user *name, | 352 | const char __user *, name, void __user *, value, size_t, size) |
354 | void __user *value, size_t size) | ||
355 | { | 353 | { |
356 | struct path path; | 354 | struct path path; |
357 | ssize_t error; | 355 | ssize_t error; |
@@ -364,9 +362,8 @@ sys_getxattr(const char __user *pathname, const char __user *name, | |||
364 | return error; | 362 | return error; |
365 | } | 363 | } |
366 | 364 | ||
367 | asmlinkage ssize_t | 365 | SYSCALL_DEFINE4(lgetxattr, const char __user *, pathname, |
368 | sys_lgetxattr(const char __user *pathname, const char __user *name, void __user *value, | 366 | const char __user *, name, void __user *, value, size_t, size) |
369 | size_t size) | ||
370 | { | 367 | { |
371 | struct path path; | 368 | struct path path; |
372 | ssize_t error; | 369 | ssize_t error; |
@@ -379,8 +376,8 @@ sys_lgetxattr(const char __user *pathname, const char __user *name, void __user | |||
379 | return error; | 376 | return error; |
380 | } | 377 | } |
381 | 378 | ||
382 | asmlinkage ssize_t | 379 | SYSCALL_DEFINE4(fgetxattr, int, fd, const char __user *, name, |
383 | sys_fgetxattr(int fd, const char __user *name, void __user *value, size_t size) | 380 | void __user *, value, size_t, size) |
384 | { | 381 | { |
385 | struct file *f; | 382 | struct file *f; |
386 | ssize_t error = -EBADF; | 383 | ssize_t error = -EBADF; |
@@ -424,8 +421,8 @@ listxattr(struct dentry *d, char __user *list, size_t size) | |||
424 | return error; | 421 | return error; |
425 | } | 422 | } |
426 | 423 | ||
427 | asmlinkage ssize_t | 424 | SYSCALL_DEFINE3(listxattr, const char __user *, pathname, char __user *, list, |
428 | sys_listxattr(const char __user *pathname, char __user *list, size_t size) | 425 | size_t, size) |
429 | { | 426 | { |
430 | struct path path; | 427 | struct path path; |
431 | ssize_t error; | 428 | ssize_t error; |
@@ -438,8 +435,8 @@ sys_listxattr(const char __user *pathname, char __user *list, size_t size) | |||
438 | return error; | 435 | return error; |
439 | } | 436 | } |
440 | 437 | ||
441 | asmlinkage ssize_t | 438 | SYSCALL_DEFINE3(llistxattr, const char __user *, pathname, char __user *, list, |
442 | sys_llistxattr(const char __user *pathname, char __user *list, size_t size) | 439 | size_t, size) |
443 | { | 440 | { |
444 | struct path path; | 441 | struct path path; |
445 | ssize_t error; | 442 | ssize_t error; |
@@ -452,8 +449,7 @@ sys_llistxattr(const char __user *pathname, char __user *list, size_t size) | |||
452 | return error; | 449 | return error; |
453 | } | 450 | } |
454 | 451 | ||
455 | asmlinkage ssize_t | 452 | SYSCALL_DEFINE3(flistxattr, int, fd, char __user *, list, size_t, size) |
456 | sys_flistxattr(int fd, char __user *list, size_t size) | ||
457 | { | 453 | { |
458 | struct file *f; | 454 | struct file *f; |
459 | ssize_t error = -EBADF; | 455 | ssize_t error = -EBADF; |
@@ -485,8 +481,8 @@ removexattr(struct dentry *d, const char __user *name) | |||
485 | return vfs_removexattr(d, kname); | 481 | return vfs_removexattr(d, kname); |
486 | } | 482 | } |
487 | 483 | ||
488 | asmlinkage long | 484 | SYSCALL_DEFINE2(removexattr, const char __user *, pathname, |
489 | sys_removexattr(const char __user *pathname, const char __user *name) | 485 | const char __user *, name) |
490 | { | 486 | { |
491 | struct path path; | 487 | struct path path; |
492 | int error; | 488 | int error; |
@@ -503,8 +499,8 @@ sys_removexattr(const char __user *pathname, const char __user *name) | |||
503 | return error; | 499 | return error; |
504 | } | 500 | } |
505 | 501 | ||
506 | asmlinkage long | 502 | SYSCALL_DEFINE2(lremovexattr, const char __user *, pathname, |
507 | sys_lremovexattr(const char __user *pathname, const char __user *name) | 503 | const char __user *, name) |
508 | { | 504 | { |
509 | struct path path; | 505 | struct path path; |
510 | int error; | 506 | int error; |
@@ -521,8 +517,7 @@ sys_lremovexattr(const char __user *pathname, const char __user *name) | |||
521 | return error; | 517 | return error; |
522 | } | 518 | } |
523 | 519 | ||
524 | asmlinkage long | 520 | SYSCALL_DEFINE2(fremovexattr, int, fd, const char __user *, name) |
525 | sys_fremovexattr(int fd, const char __user *name) | ||
526 | { | 521 | { |
527 | struct file *f; | 522 | struct file *f; |
528 | struct dentry *dentry; | 523 | struct dentry *dentry; |
diff --git a/fs/xfs/Kconfig b/fs/xfs/Kconfig index 3f53dd101f99..29228f5899cd 100644 --- a/fs/xfs/Kconfig +++ b/fs/xfs/Kconfig | |||
@@ -1,6 +1,7 @@ | |||
1 | config XFS_FS | 1 | config XFS_FS |
2 | tristate "XFS filesystem support" | 2 | tristate "XFS filesystem support" |
3 | depends on BLOCK | 3 | depends on BLOCK |
4 | select EXPORTFS | ||
4 | help | 5 | help |
5 | XFS is a high performance journaling filesystem which originated | 6 | XFS is a high performance journaling filesystem which originated |
6 | on the SGI IRIX platform. It is completely multi-threaded, can | 7 | on the SGI IRIX platform. It is completely multi-threaded, can |
diff --git a/fs/xfs/linux-2.6/xfs_aops.h b/fs/xfs/linux-2.6/xfs_aops.h index 7b26f5ff9692..1dd528849755 100644 --- a/fs/xfs/linux-2.6/xfs_aops.h +++ b/fs/xfs/linux-2.6/xfs_aops.h | |||
@@ -21,8 +21,6 @@ | |||
21 | extern struct workqueue_struct *xfsdatad_workqueue; | 21 | extern struct workqueue_struct *xfsdatad_workqueue; |
22 | extern mempool_t *xfs_ioend_pool; | 22 | extern mempool_t *xfs_ioend_pool; |
23 | 23 | ||
24 | typedef void (*xfs_ioend_func_t)(void *); | ||
25 | |||
26 | /* | 24 | /* |
27 | * xfs_ioend struct manages large extent writes for XFS. | 25 | * xfs_ioend struct manages large extent writes for XFS. |
28 | * It can manage several multi-page bio's at once. | 26 | * It can manage several multi-page bio's at once. |
diff --git a/fs/xfs/linux-2.6/xfs_export.c b/fs/xfs/linux-2.6/xfs_export.c index 595751f78350..87b8cbd23d4b 100644 --- a/fs/xfs/linux-2.6/xfs_export.c +++ b/fs/xfs/linux-2.6/xfs_export.c | |||
@@ -126,11 +126,26 @@ xfs_nfs_get_inode( | |||
126 | if (ino == 0) | 126 | if (ino == 0) |
127 | return ERR_PTR(-ESTALE); | 127 | return ERR_PTR(-ESTALE); |
128 | 128 | ||
129 | error = xfs_iget(mp, NULL, ino, 0, XFS_ILOCK_SHARED, &ip, 0); | 129 | /* |
130 | if (error) | 130 | * The XFS_IGET_BULKSTAT means that an invalid inode number is just |
131 | * fine and not an indication of a corrupted filesystem. Because | ||
132 | * clients can send any kind of invalid file handle, e.g. after | ||
133 | * a restore on the server we have to deal with this case gracefully. | ||
134 | */ | ||
135 | error = xfs_iget(mp, NULL, ino, XFS_IGET_BULKSTAT, | ||
136 | XFS_ILOCK_SHARED, &ip, 0); | ||
137 | if (error) { | ||
138 | /* | ||
139 | * EINVAL means the inode cluster doesn't exist anymore. | ||
140 | * This implies the filehandle is stale, so we should | ||
141 | * translate it here. | ||
142 | * We don't use ESTALE directly down the chain to not | ||
143 | * confuse applications using bulkstat that expect EINVAL. | ||
144 | */ | ||
145 | if (error == EINVAL) | ||
146 | error = ESTALE; | ||
131 | return ERR_PTR(-error); | 147 | return ERR_PTR(-error); |
132 | if (!ip) | 148 | } |
133 | return ERR_PTR(-EIO); | ||
134 | 149 | ||
135 | if (ip->i_d.di_gen != generation) { | 150 | if (ip->i_d.di_gen != generation) { |
136 | xfs_iput_new(ip, XFS_ILOCK_SHARED); | 151 | xfs_iput_new(ip, XFS_ILOCK_SHARED); |
diff --git a/fs/xfs/linux-2.6/xfs_ioctl.c b/fs/xfs/linux-2.6/xfs_ioctl.c index e5be1e0be802..4bd112313f33 100644 --- a/fs/xfs/linux-2.6/xfs_ioctl.c +++ b/fs/xfs/linux-2.6/xfs_ioctl.c | |||
@@ -50,12 +50,14 @@ | |||
50 | #include "xfs_vnodeops.h" | 50 | #include "xfs_vnodeops.h" |
51 | #include "xfs_quota.h" | 51 | #include "xfs_quota.h" |
52 | #include "xfs_inode_item.h" | 52 | #include "xfs_inode_item.h" |
53 | #include "xfs_export.h" | ||
53 | 54 | ||
54 | #include <linux/capability.h> | 55 | #include <linux/capability.h> |
55 | #include <linux/dcache.h> | 56 | #include <linux/dcache.h> |
56 | #include <linux/mount.h> | 57 | #include <linux/mount.h> |
57 | #include <linux/namei.h> | 58 | #include <linux/namei.h> |
58 | #include <linux/pagemap.h> | 59 | #include <linux/pagemap.h> |
60 | #include <linux/exportfs.h> | ||
59 | 61 | ||
60 | /* | 62 | /* |
61 | * xfs_find_handle maps from userspace xfs_fsop_handlereq structure to | 63 | * xfs_find_handle maps from userspace xfs_fsop_handlereq structure to |
@@ -164,97 +166,69 @@ xfs_find_handle( | |||
164 | return 0; | 166 | return 0; |
165 | } | 167 | } |
166 | 168 | ||
167 | |||
168 | /* | 169 | /* |
169 | * Convert userspace handle data into inode. | 170 | * No need to do permission checks on the various pathname components |
170 | * | 171 | * as the handle operations are privileged. |
171 | * We use the fact that all the fsop_handlereq ioctl calls have a data | ||
172 | * structure argument whose first component is always a xfs_fsop_handlereq_t, | ||
173 | * so we can pass that sub structure into this handy, shared routine. | ||
174 | * | ||
175 | * If no error, caller must always iput the returned inode. | ||
176 | */ | 172 | */ |
177 | STATIC int | 173 | STATIC int |
178 | xfs_vget_fsop_handlereq( | 174 | xfs_handle_acceptable( |
179 | xfs_mount_t *mp, | 175 | void *context, |
180 | struct inode *parinode, /* parent inode pointer */ | 176 | struct dentry *dentry) |
181 | xfs_fsop_handlereq_t *hreq, | 177 | { |
182 | struct inode **inode) | 178 | return 1; |
179 | } | ||
180 | |||
181 | /* | ||
182 | * Convert userspace handle data into a dentry. | ||
183 | */ | ||
184 | struct dentry * | ||
185 | xfs_handle_to_dentry( | ||
186 | struct file *parfilp, | ||
187 | void __user *uhandle, | ||
188 | u32 hlen) | ||
183 | { | 189 | { |
184 | void __user *hanp; | ||
185 | size_t hlen; | ||
186 | xfs_fid_t *xfid; | ||
187 | xfs_handle_t *handlep; | ||
188 | xfs_handle_t handle; | 190 | xfs_handle_t handle; |
189 | xfs_inode_t *ip; | 191 | struct xfs_fid64 fid; |
190 | xfs_ino_t ino; | ||
191 | __u32 igen; | ||
192 | int error; | ||
193 | 192 | ||
194 | /* | 193 | /* |
195 | * Only allow handle opens under a directory. | 194 | * Only allow handle opens under a directory. |
196 | */ | 195 | */ |
197 | if (!S_ISDIR(parinode->i_mode)) | 196 | if (!S_ISDIR(parfilp->f_path.dentry->d_inode->i_mode)) |
198 | return XFS_ERROR(ENOTDIR); | 197 | return ERR_PTR(-ENOTDIR); |
199 | 198 | ||
200 | hanp = hreq->ihandle; | 199 | if (hlen != sizeof(xfs_handle_t)) |
201 | hlen = hreq->ihandlen; | 200 | return ERR_PTR(-EINVAL); |
202 | handlep = &handle; | 201 | if (copy_from_user(&handle, uhandle, hlen)) |
203 | 202 | return ERR_PTR(-EFAULT); | |
204 | if (hlen < sizeof(handlep->ha_fsid) || hlen > sizeof(*handlep)) | 203 | if (handle.ha_fid.fid_len != |
205 | return XFS_ERROR(EINVAL); | 204 | sizeof(handle.ha_fid) - sizeof(handle.ha_fid.fid_len)) |
206 | if (copy_from_user(handlep, hanp, hlen)) | 205 | return ERR_PTR(-EINVAL); |
207 | return XFS_ERROR(EFAULT); | 206 | |
208 | if (hlen < sizeof(*handlep)) | 207 | memset(&fid, 0, sizeof(struct fid)); |
209 | memset(((char *)handlep) + hlen, 0, sizeof(*handlep) - hlen); | 208 | fid.ino = handle.ha_fid.fid_ino; |
210 | if (hlen > sizeof(handlep->ha_fsid)) { | 209 | fid.gen = handle.ha_fid.fid_gen; |
211 | if (handlep->ha_fid.fid_len != | 210 | |
212 | (hlen - sizeof(handlep->ha_fsid) - | 211 | return exportfs_decode_fh(parfilp->f_path.mnt, (struct fid *)&fid, 3, |
213 | sizeof(handlep->ha_fid.fid_len)) || | 212 | FILEID_INO32_GEN | XFS_FILEID_TYPE_64FLAG, |
214 | handlep->ha_fid.fid_pad) | 213 | xfs_handle_acceptable, NULL); |
215 | return XFS_ERROR(EINVAL); | 214 | } |
216 | } | ||
217 | |||
218 | /* | ||
219 | * Crack the handle, obtain the inode # & generation # | ||
220 | */ | ||
221 | xfid = (struct xfs_fid *)&handlep->ha_fid; | ||
222 | if (xfid->fid_len == sizeof(*xfid) - sizeof(xfid->fid_len)) { | ||
223 | ino = xfid->fid_ino; | ||
224 | igen = xfid->fid_gen; | ||
225 | } else { | ||
226 | return XFS_ERROR(EINVAL); | ||
227 | } | ||
228 | |||
229 | /* | ||
230 | * Get the XFS inode, building a Linux inode to go with it. | ||
231 | */ | ||
232 | error = xfs_iget(mp, NULL, ino, 0, XFS_ILOCK_SHARED, &ip, 0); | ||
233 | if (error) | ||
234 | return error; | ||
235 | if (ip == NULL) | ||
236 | return XFS_ERROR(EIO); | ||
237 | if (ip->i_d.di_gen != igen) { | ||
238 | xfs_iput_new(ip, XFS_ILOCK_SHARED); | ||
239 | return XFS_ERROR(ENOENT); | ||
240 | } | ||
241 | |||
242 | xfs_iunlock(ip, XFS_ILOCK_SHARED); | ||
243 | 215 | ||
244 | *inode = VFS_I(ip); | 216 | STATIC struct dentry * |
245 | return 0; | 217 | xfs_handlereq_to_dentry( |
218 | struct file *parfilp, | ||
219 | xfs_fsop_handlereq_t *hreq) | ||
220 | { | ||
221 | return xfs_handle_to_dentry(parfilp, hreq->ihandle, hreq->ihandlen); | ||
246 | } | 222 | } |
247 | 223 | ||
248 | int | 224 | int |
249 | xfs_open_by_handle( | 225 | xfs_open_by_handle( |
250 | xfs_mount_t *mp, | ||
251 | xfs_fsop_handlereq_t *hreq, | ||
252 | struct file *parfilp, | 226 | struct file *parfilp, |
253 | struct inode *parinode) | 227 | xfs_fsop_handlereq_t *hreq) |
254 | { | 228 | { |
255 | const struct cred *cred = current_cred(); | 229 | const struct cred *cred = current_cred(); |
256 | int error; | 230 | int error; |
257 | int new_fd; | 231 | int fd; |
258 | int permflag; | 232 | int permflag; |
259 | struct file *filp; | 233 | struct file *filp; |
260 | struct inode *inode; | 234 | struct inode *inode; |
@@ -263,19 +237,21 @@ xfs_open_by_handle( | |||
263 | if (!capable(CAP_SYS_ADMIN)) | 237 | if (!capable(CAP_SYS_ADMIN)) |
264 | return -XFS_ERROR(EPERM); | 238 | return -XFS_ERROR(EPERM); |
265 | 239 | ||
266 | error = xfs_vget_fsop_handlereq(mp, parinode, hreq, &inode); | 240 | dentry = xfs_handlereq_to_dentry(parfilp, hreq); |
267 | if (error) | 241 | if (IS_ERR(dentry)) |
268 | return -error; | 242 | return PTR_ERR(dentry); |
243 | inode = dentry->d_inode; | ||
269 | 244 | ||
270 | /* Restrict xfs_open_by_handle to directories & regular files. */ | 245 | /* Restrict xfs_open_by_handle to directories & regular files. */ |
271 | if (!(S_ISREG(inode->i_mode) || S_ISDIR(inode->i_mode))) { | 246 | if (!(S_ISREG(inode->i_mode) || S_ISDIR(inode->i_mode))) { |
272 | iput(inode); | 247 | error = -XFS_ERROR(EPERM); |
273 | return -XFS_ERROR(EINVAL); | 248 | goto out_dput; |
274 | } | 249 | } |
275 | 250 | ||
276 | #if BITS_PER_LONG != 32 | 251 | #if BITS_PER_LONG != 32 |
277 | hreq->oflags |= O_LARGEFILE; | 252 | hreq->oflags |= O_LARGEFILE; |
278 | #endif | 253 | #endif |
254 | |||
279 | /* Put open permission in namei format. */ | 255 | /* Put open permission in namei format. */ |
280 | permflag = hreq->oflags; | 256 | permflag = hreq->oflags; |
281 | if ((permflag+1) & O_ACCMODE) | 257 | if ((permflag+1) & O_ACCMODE) |
@@ -285,50 +261,45 @@ xfs_open_by_handle( | |||
285 | 261 | ||
286 | if ((!(permflag & O_APPEND) || (permflag & O_TRUNC)) && | 262 | if ((!(permflag & O_APPEND) || (permflag & O_TRUNC)) && |
287 | (permflag & FMODE_WRITE) && IS_APPEND(inode)) { | 263 | (permflag & FMODE_WRITE) && IS_APPEND(inode)) { |
288 | iput(inode); | 264 | error = -XFS_ERROR(EPERM); |
289 | return -XFS_ERROR(EPERM); | 265 | goto out_dput; |
290 | } | 266 | } |
291 | 267 | ||
292 | if ((permflag & FMODE_WRITE) && IS_IMMUTABLE(inode)) { | 268 | if ((permflag & FMODE_WRITE) && IS_IMMUTABLE(inode)) { |
293 | iput(inode); | 269 | error = -XFS_ERROR(EACCES); |
294 | return -XFS_ERROR(EACCES); | 270 | goto out_dput; |
295 | } | 271 | } |
296 | 272 | ||
297 | /* Can't write directories. */ | 273 | /* Can't write directories. */ |
298 | if ( S_ISDIR(inode->i_mode) && (permflag & FMODE_WRITE)) { | 274 | if (S_ISDIR(inode->i_mode) && (permflag & FMODE_WRITE)) { |
299 | iput(inode); | 275 | error = -XFS_ERROR(EISDIR); |
300 | return -XFS_ERROR(EISDIR); | 276 | goto out_dput; |
301 | } | 277 | } |
302 | 278 | ||
303 | if ((new_fd = get_unused_fd()) < 0) { | 279 | fd = get_unused_fd(); |
304 | iput(inode); | 280 | if (fd < 0) { |
305 | return new_fd; | 281 | error = fd; |
282 | goto out_dput; | ||
306 | } | 283 | } |
307 | 284 | ||
308 | dentry = d_obtain_alias(inode); | 285 | filp = dentry_open(dentry, mntget(parfilp->f_path.mnt), |
309 | if (IS_ERR(dentry)) { | 286 | hreq->oflags, cred); |
310 | put_unused_fd(new_fd); | ||
311 | return PTR_ERR(dentry); | ||
312 | } | ||
313 | |||
314 | /* Ensure umount returns EBUSY on umounts while this file is open. */ | ||
315 | mntget(parfilp->f_path.mnt); | ||
316 | |||
317 | /* Create file pointer. */ | ||
318 | filp = dentry_open(dentry, parfilp->f_path.mnt, hreq->oflags, cred); | ||
319 | if (IS_ERR(filp)) { | 287 | if (IS_ERR(filp)) { |
320 | put_unused_fd(new_fd); | 288 | put_unused_fd(fd); |
321 | return -XFS_ERROR(-PTR_ERR(filp)); | 289 | return PTR_ERR(filp); |
322 | } | 290 | } |
323 | 291 | ||
324 | if (inode->i_mode & S_IFREG) { | 292 | if (inode->i_mode & S_IFREG) { |
325 | /* invisible operation should not change atime */ | ||
326 | filp->f_flags |= O_NOATIME; | 293 | filp->f_flags |= O_NOATIME; |
327 | filp->f_mode |= FMODE_NOCMTIME; | 294 | filp->f_mode |= FMODE_NOCMTIME; |
328 | } | 295 | } |
329 | 296 | ||
330 | fd_install(new_fd, filp); | 297 | fd_install(fd, filp); |
331 | return new_fd; | 298 | return fd; |
299 | |||
300 | out_dput: | ||
301 | dput(dentry); | ||
302 | return error; | ||
332 | } | 303 | } |
333 | 304 | ||
334 | /* | 305 | /* |
@@ -359,11 +330,10 @@ do_readlink( | |||
359 | 330 | ||
360 | int | 331 | int |
361 | xfs_readlink_by_handle( | 332 | xfs_readlink_by_handle( |
362 | xfs_mount_t *mp, | 333 | struct file *parfilp, |
363 | xfs_fsop_handlereq_t *hreq, | 334 | xfs_fsop_handlereq_t *hreq) |
364 | struct inode *parinode) | ||
365 | { | 335 | { |
366 | struct inode *inode; | 336 | struct dentry *dentry; |
367 | __u32 olen; | 337 | __u32 olen; |
368 | void *link; | 338 | void *link; |
369 | int error; | 339 | int error; |
@@ -371,26 +341,28 @@ xfs_readlink_by_handle( | |||
371 | if (!capable(CAP_SYS_ADMIN)) | 341 | if (!capable(CAP_SYS_ADMIN)) |
372 | return -XFS_ERROR(EPERM); | 342 | return -XFS_ERROR(EPERM); |
373 | 343 | ||
374 | error = xfs_vget_fsop_handlereq(mp, parinode, hreq, &inode); | 344 | dentry = xfs_handlereq_to_dentry(parfilp, hreq); |
375 | if (error) | 345 | if (IS_ERR(dentry)) |
376 | return -error; | 346 | return PTR_ERR(dentry); |
377 | 347 | ||
378 | /* Restrict this handle operation to symlinks only. */ | 348 | /* Restrict this handle operation to symlinks only. */ |
379 | if (!S_ISLNK(inode->i_mode)) { | 349 | if (!S_ISLNK(dentry->d_inode->i_mode)) { |
380 | error = -XFS_ERROR(EINVAL); | 350 | error = -XFS_ERROR(EINVAL); |
381 | goto out_iput; | 351 | goto out_dput; |
382 | } | 352 | } |
383 | 353 | ||
384 | if (copy_from_user(&olen, hreq->ohandlen, sizeof(__u32))) { | 354 | if (copy_from_user(&olen, hreq->ohandlen, sizeof(__u32))) { |
385 | error = -XFS_ERROR(EFAULT); | 355 | error = -XFS_ERROR(EFAULT); |
386 | goto out_iput; | 356 | goto out_dput; |
387 | } | 357 | } |
388 | 358 | ||
389 | link = kmalloc(MAXPATHLEN+1, GFP_KERNEL); | 359 | link = kmalloc(MAXPATHLEN+1, GFP_KERNEL); |
390 | if (!link) | 360 | if (!link) { |
391 | goto out_iput; | 361 | error = -XFS_ERROR(ENOMEM); |
362 | goto out_dput; | ||
363 | } | ||
392 | 364 | ||
393 | error = -xfs_readlink(XFS_I(inode), link); | 365 | error = -xfs_readlink(XFS_I(dentry->d_inode), link); |
394 | if (error) | 366 | if (error) |
395 | goto out_kfree; | 367 | goto out_kfree; |
396 | error = do_readlink(hreq->ohandle, olen, link); | 368 | error = do_readlink(hreq->ohandle, olen, link); |
@@ -399,32 +371,31 @@ xfs_readlink_by_handle( | |||
399 | 371 | ||
400 | out_kfree: | 372 | out_kfree: |
401 | kfree(link); | 373 | kfree(link); |
402 | out_iput: | 374 | out_dput: |
403 | iput(inode); | 375 | dput(dentry); |
404 | return error; | 376 | return error; |
405 | } | 377 | } |
406 | 378 | ||
407 | STATIC int | 379 | STATIC int |
408 | xfs_fssetdm_by_handle( | 380 | xfs_fssetdm_by_handle( |
409 | xfs_mount_t *mp, | 381 | struct file *parfilp, |
410 | void __user *arg, | 382 | void __user *arg) |
411 | struct inode *parinode) | ||
412 | { | 383 | { |
413 | int error; | 384 | int error; |
414 | struct fsdmidata fsd; | 385 | struct fsdmidata fsd; |
415 | xfs_fsop_setdm_handlereq_t dmhreq; | 386 | xfs_fsop_setdm_handlereq_t dmhreq; |
416 | struct inode *inode; | 387 | struct dentry *dentry; |
417 | 388 | ||
418 | if (!capable(CAP_MKNOD)) | 389 | if (!capable(CAP_MKNOD)) |
419 | return -XFS_ERROR(EPERM); | 390 | return -XFS_ERROR(EPERM); |
420 | if (copy_from_user(&dmhreq, arg, sizeof(xfs_fsop_setdm_handlereq_t))) | 391 | if (copy_from_user(&dmhreq, arg, sizeof(xfs_fsop_setdm_handlereq_t))) |
421 | return -XFS_ERROR(EFAULT); | 392 | return -XFS_ERROR(EFAULT); |
422 | 393 | ||
423 | error = xfs_vget_fsop_handlereq(mp, parinode, &dmhreq.hreq, &inode); | 394 | dentry = xfs_handlereq_to_dentry(parfilp, &dmhreq.hreq); |
424 | if (error) | 395 | if (IS_ERR(dentry)) |
425 | return -error; | 396 | return PTR_ERR(dentry); |
426 | 397 | ||
427 | if (IS_IMMUTABLE(inode) || IS_APPEND(inode)) { | 398 | if (IS_IMMUTABLE(dentry->d_inode) || IS_APPEND(dentry->d_inode)) { |
428 | error = -XFS_ERROR(EPERM); | 399 | error = -XFS_ERROR(EPERM); |
429 | goto out; | 400 | goto out; |
430 | } | 401 | } |
@@ -434,24 +405,23 @@ xfs_fssetdm_by_handle( | |||
434 | goto out; | 405 | goto out; |
435 | } | 406 | } |
436 | 407 | ||
437 | error = -xfs_set_dmattrs(XFS_I(inode), fsd.fsd_dmevmask, | 408 | error = -xfs_set_dmattrs(XFS_I(dentry->d_inode), fsd.fsd_dmevmask, |
438 | fsd.fsd_dmstate); | 409 | fsd.fsd_dmstate); |
439 | 410 | ||
440 | out: | 411 | out: |
441 | iput(inode); | 412 | dput(dentry); |
442 | return error; | 413 | return error; |
443 | } | 414 | } |
444 | 415 | ||
445 | STATIC int | 416 | STATIC int |
446 | xfs_attrlist_by_handle( | 417 | xfs_attrlist_by_handle( |
447 | xfs_mount_t *mp, | 418 | struct file *parfilp, |
448 | void __user *arg, | 419 | void __user *arg) |
449 | struct inode *parinode) | ||
450 | { | 420 | { |
451 | int error; | 421 | int error = -ENOMEM; |
452 | attrlist_cursor_kern_t *cursor; | 422 | attrlist_cursor_kern_t *cursor; |
453 | xfs_fsop_attrlist_handlereq_t al_hreq; | 423 | xfs_fsop_attrlist_handlereq_t al_hreq; |
454 | struct inode *inode; | 424 | struct dentry *dentry; |
455 | char *kbuf; | 425 | char *kbuf; |
456 | 426 | ||
457 | if (!capable(CAP_SYS_ADMIN)) | 427 | if (!capable(CAP_SYS_ADMIN)) |
@@ -467,16 +437,16 @@ xfs_attrlist_by_handle( | |||
467 | if (al_hreq.flags & ~(ATTR_ROOT | ATTR_SECURE)) | 437 | if (al_hreq.flags & ~(ATTR_ROOT | ATTR_SECURE)) |
468 | return -XFS_ERROR(EINVAL); | 438 | return -XFS_ERROR(EINVAL); |
469 | 439 | ||
470 | error = xfs_vget_fsop_handlereq(mp, parinode, &al_hreq.hreq, &inode); | 440 | dentry = xfs_handlereq_to_dentry(parfilp, &al_hreq.hreq); |
471 | if (error) | 441 | if (IS_ERR(dentry)) |
472 | goto out; | 442 | return PTR_ERR(dentry); |
473 | 443 | ||
474 | kbuf = kmalloc(al_hreq.buflen, GFP_KERNEL); | 444 | kbuf = kmalloc(al_hreq.buflen, GFP_KERNEL); |
475 | if (!kbuf) | 445 | if (!kbuf) |
476 | goto out_vn_rele; | 446 | goto out_dput; |
477 | 447 | ||
478 | cursor = (attrlist_cursor_kern_t *)&al_hreq.pos; | 448 | cursor = (attrlist_cursor_kern_t *)&al_hreq.pos; |
479 | error = xfs_attr_list(XFS_I(inode), kbuf, al_hreq.buflen, | 449 | error = -xfs_attr_list(XFS_I(dentry->d_inode), kbuf, al_hreq.buflen, |
480 | al_hreq.flags, cursor); | 450 | al_hreq.flags, cursor); |
481 | if (error) | 451 | if (error) |
482 | goto out_kfree; | 452 | goto out_kfree; |
@@ -486,10 +456,9 @@ xfs_attrlist_by_handle( | |||
486 | 456 | ||
487 | out_kfree: | 457 | out_kfree: |
488 | kfree(kbuf); | 458 | kfree(kbuf); |
489 | out_vn_rele: | 459 | out_dput: |
490 | iput(inode); | 460 | dput(dentry); |
491 | out: | 461 | return error; |
492 | return -error; | ||
493 | } | 462 | } |
494 | 463 | ||
495 | int | 464 | int |
@@ -564,15 +533,13 @@ xfs_attrmulti_attr_remove( | |||
564 | 533 | ||
565 | STATIC int | 534 | STATIC int |
566 | xfs_attrmulti_by_handle( | 535 | xfs_attrmulti_by_handle( |
567 | xfs_mount_t *mp, | ||
568 | void __user *arg, | ||
569 | struct file *parfilp, | 536 | struct file *parfilp, |
570 | struct inode *parinode) | 537 | void __user *arg) |
571 | { | 538 | { |
572 | int error; | 539 | int error; |
573 | xfs_attr_multiop_t *ops; | 540 | xfs_attr_multiop_t *ops; |
574 | xfs_fsop_attrmulti_handlereq_t am_hreq; | 541 | xfs_fsop_attrmulti_handlereq_t am_hreq; |
575 | struct inode *inode; | 542 | struct dentry *dentry; |
576 | unsigned int i, size; | 543 | unsigned int i, size; |
577 | char *attr_name; | 544 | char *attr_name; |
578 | 545 | ||
@@ -581,19 +548,19 @@ xfs_attrmulti_by_handle( | |||
581 | if (copy_from_user(&am_hreq, arg, sizeof(xfs_fsop_attrmulti_handlereq_t))) | 548 | if (copy_from_user(&am_hreq, arg, sizeof(xfs_fsop_attrmulti_handlereq_t))) |
582 | return -XFS_ERROR(EFAULT); | 549 | return -XFS_ERROR(EFAULT); |
583 | 550 | ||
584 | error = xfs_vget_fsop_handlereq(mp, parinode, &am_hreq.hreq, &inode); | 551 | dentry = xfs_handlereq_to_dentry(parfilp, &am_hreq.hreq); |
585 | if (error) | 552 | if (IS_ERR(dentry)) |
586 | goto out; | 553 | return PTR_ERR(dentry); |
587 | 554 | ||
588 | error = E2BIG; | 555 | error = E2BIG; |
589 | size = am_hreq.opcount * sizeof(xfs_attr_multiop_t); | 556 | size = am_hreq.opcount * sizeof(xfs_attr_multiop_t); |
590 | if (!size || size > 16 * PAGE_SIZE) | 557 | if (!size || size > 16 * PAGE_SIZE) |
591 | goto out_vn_rele; | 558 | goto out_dput; |
592 | 559 | ||
593 | error = ENOMEM; | 560 | error = ENOMEM; |
594 | ops = kmalloc(size, GFP_KERNEL); | 561 | ops = kmalloc(size, GFP_KERNEL); |
595 | if (!ops) | 562 | if (!ops) |
596 | goto out_vn_rele; | 563 | goto out_dput; |
597 | 564 | ||
598 | error = EFAULT; | 565 | error = EFAULT; |
599 | if (copy_from_user(ops, am_hreq.ops, size)) | 566 | if (copy_from_user(ops, am_hreq.ops, size)) |
@@ -615,25 +582,28 @@ xfs_attrmulti_by_handle( | |||
615 | 582 | ||
616 | switch (ops[i].am_opcode) { | 583 | switch (ops[i].am_opcode) { |
617 | case ATTR_OP_GET: | 584 | case ATTR_OP_GET: |
618 | ops[i].am_error = xfs_attrmulti_attr_get(inode, | 585 | ops[i].am_error = xfs_attrmulti_attr_get( |
619 | attr_name, ops[i].am_attrvalue, | 586 | dentry->d_inode, attr_name, |
620 | &ops[i].am_length, ops[i].am_flags); | 587 | ops[i].am_attrvalue, &ops[i].am_length, |
588 | ops[i].am_flags); | ||
621 | break; | 589 | break; |
622 | case ATTR_OP_SET: | 590 | case ATTR_OP_SET: |
623 | ops[i].am_error = mnt_want_write(parfilp->f_path.mnt); | 591 | ops[i].am_error = mnt_want_write(parfilp->f_path.mnt); |
624 | if (ops[i].am_error) | 592 | if (ops[i].am_error) |
625 | break; | 593 | break; |
626 | ops[i].am_error = xfs_attrmulti_attr_set(inode, | 594 | ops[i].am_error = xfs_attrmulti_attr_set( |
627 | attr_name, ops[i].am_attrvalue, | 595 | dentry->d_inode, attr_name, |
628 | ops[i].am_length, ops[i].am_flags); | 596 | ops[i].am_attrvalue, ops[i].am_length, |
597 | ops[i].am_flags); | ||
629 | mnt_drop_write(parfilp->f_path.mnt); | 598 | mnt_drop_write(parfilp->f_path.mnt); |
630 | break; | 599 | break; |
631 | case ATTR_OP_REMOVE: | 600 | case ATTR_OP_REMOVE: |
632 | ops[i].am_error = mnt_want_write(parfilp->f_path.mnt); | 601 | ops[i].am_error = mnt_want_write(parfilp->f_path.mnt); |
633 | if (ops[i].am_error) | 602 | if (ops[i].am_error) |
634 | break; | 603 | break; |
635 | ops[i].am_error = xfs_attrmulti_attr_remove(inode, | 604 | ops[i].am_error = xfs_attrmulti_attr_remove( |
636 | attr_name, ops[i].am_flags); | 605 | dentry->d_inode, attr_name, |
606 | ops[i].am_flags); | ||
637 | mnt_drop_write(parfilp->f_path.mnt); | 607 | mnt_drop_write(parfilp->f_path.mnt); |
638 | break; | 608 | break; |
639 | default: | 609 | default: |
@@ -647,9 +617,8 @@ xfs_attrmulti_by_handle( | |||
647 | kfree(attr_name); | 617 | kfree(attr_name); |
648 | out_kfree_ops: | 618 | out_kfree_ops: |
649 | kfree(ops); | 619 | kfree(ops); |
650 | out_vn_rele: | 620 | out_dput: |
651 | iput(inode); | 621 | dput(dentry); |
652 | out: | ||
653 | return -error; | 622 | return -error; |
654 | } | 623 | } |
655 | 624 | ||
@@ -1440,23 +1409,23 @@ xfs_file_ioctl( | |||
1440 | 1409 | ||
1441 | if (copy_from_user(&hreq, arg, sizeof(xfs_fsop_handlereq_t))) | 1410 | if (copy_from_user(&hreq, arg, sizeof(xfs_fsop_handlereq_t))) |
1442 | return -XFS_ERROR(EFAULT); | 1411 | return -XFS_ERROR(EFAULT); |
1443 | return xfs_open_by_handle(mp, &hreq, filp, inode); | 1412 | return xfs_open_by_handle(filp, &hreq); |
1444 | } | 1413 | } |
1445 | case XFS_IOC_FSSETDM_BY_HANDLE: | 1414 | case XFS_IOC_FSSETDM_BY_HANDLE: |
1446 | return xfs_fssetdm_by_handle(mp, arg, inode); | 1415 | return xfs_fssetdm_by_handle(filp, arg); |
1447 | 1416 | ||
1448 | case XFS_IOC_READLINK_BY_HANDLE: { | 1417 | case XFS_IOC_READLINK_BY_HANDLE: { |
1449 | xfs_fsop_handlereq_t hreq; | 1418 | xfs_fsop_handlereq_t hreq; |
1450 | 1419 | ||
1451 | if (copy_from_user(&hreq, arg, sizeof(xfs_fsop_handlereq_t))) | 1420 | if (copy_from_user(&hreq, arg, sizeof(xfs_fsop_handlereq_t))) |
1452 | return -XFS_ERROR(EFAULT); | 1421 | return -XFS_ERROR(EFAULT); |
1453 | return xfs_readlink_by_handle(mp, &hreq, inode); | 1422 | return xfs_readlink_by_handle(filp, &hreq); |
1454 | } | 1423 | } |
1455 | case XFS_IOC_ATTRLIST_BY_HANDLE: | 1424 | case XFS_IOC_ATTRLIST_BY_HANDLE: |
1456 | return xfs_attrlist_by_handle(mp, arg, inode); | 1425 | return xfs_attrlist_by_handle(filp, arg); |
1457 | 1426 | ||
1458 | case XFS_IOC_ATTRMULTI_BY_HANDLE: | 1427 | case XFS_IOC_ATTRMULTI_BY_HANDLE: |
1459 | return xfs_attrmulti_by_handle(mp, arg, filp, inode); | 1428 | return xfs_attrmulti_by_handle(filp, arg); |
1460 | 1429 | ||
1461 | case XFS_IOC_SWAPEXT: { | 1430 | case XFS_IOC_SWAPEXT: { |
1462 | struct xfs_swapext sxp; | 1431 | struct xfs_swapext sxp; |
diff --git a/fs/xfs/linux-2.6/xfs_ioctl.h b/fs/xfs/linux-2.6/xfs_ioctl.h index 8c16bf2d7e03..7bd7c6afc1eb 100644 --- a/fs/xfs/linux-2.6/xfs_ioctl.h +++ b/fs/xfs/linux-2.6/xfs_ioctl.h | |||
@@ -34,16 +34,13 @@ xfs_find_handle( | |||
34 | 34 | ||
35 | extern int | 35 | extern int |
36 | xfs_open_by_handle( | 36 | xfs_open_by_handle( |
37 | xfs_mount_t *mp, | ||
38 | xfs_fsop_handlereq_t *hreq, | ||
39 | struct file *parfilp, | 37 | struct file *parfilp, |
40 | struct inode *parinode); | 38 | xfs_fsop_handlereq_t *hreq); |
41 | 39 | ||
42 | extern int | 40 | extern int |
43 | xfs_readlink_by_handle( | 41 | xfs_readlink_by_handle( |
44 | xfs_mount_t *mp, | 42 | struct file *parfilp, |
45 | xfs_fsop_handlereq_t *hreq, | 43 | xfs_fsop_handlereq_t *hreq); |
46 | struct inode *parinode); | ||
47 | 44 | ||
48 | extern int | 45 | extern int |
49 | xfs_attrmulti_attr_get( | 46 | xfs_attrmulti_attr_get( |
@@ -67,6 +64,12 @@ xfs_attrmulti_attr_remove( | |||
67 | char *name, | 64 | char *name, |
68 | __uint32_t flags); | 65 | __uint32_t flags); |
69 | 66 | ||
67 | extern struct dentry * | ||
68 | xfs_handle_to_dentry( | ||
69 | struct file *parfilp, | ||
70 | void __user *uhandle, | ||
71 | u32 hlen); | ||
72 | |||
70 | extern long | 73 | extern long |
71 | xfs_file_ioctl( | 74 | xfs_file_ioctl( |
72 | struct file *filp, | 75 | struct file *filp, |
diff --git a/fs/xfs/linux-2.6/xfs_ioctl32.c b/fs/xfs/linux-2.6/xfs_ioctl32.c index 50903ad3182e..c70c4e3db790 100644 --- a/fs/xfs/linux-2.6/xfs_ioctl32.c +++ b/fs/xfs/linux-2.6/xfs_ioctl32.c | |||
@@ -17,6 +17,7 @@ | |||
17 | */ | 17 | */ |
18 | #include <linux/compat.h> | 18 | #include <linux/compat.h> |
19 | #include <linux/ioctl.h> | 19 | #include <linux/ioctl.h> |
20 | #include <linux/mount.h> | ||
20 | #include <asm/uaccess.h> | 21 | #include <asm/uaccess.h> |
21 | #include "xfs.h" | 22 | #include "xfs.h" |
22 | #include "xfs_fs.h" | 23 | #include "xfs_fs.h" |
@@ -340,96 +341,24 @@ xfs_compat_handlereq_copyin( | |||
340 | return 0; | 341 | return 0; |
341 | } | 342 | } |
342 | 343 | ||
343 | /* | 344 | STATIC struct dentry * |
344 | * Convert userspace handle data into inode. | 345 | xfs_compat_handlereq_to_dentry( |
345 | * | 346 | struct file *parfilp, |
346 | * We use the fact that all the fsop_handlereq ioctl calls have a data | 347 | compat_xfs_fsop_handlereq_t *hreq) |
347 | * structure argument whose first component is always a xfs_fsop_handlereq_t, | ||
348 | * so we can pass that sub structure into this handy, shared routine. | ||
349 | * | ||
350 | * If no error, caller must always iput the returned inode. | ||
351 | */ | ||
352 | STATIC int | ||
353 | xfs_vget_fsop_handlereq_compat( | ||
354 | xfs_mount_t *mp, | ||
355 | struct inode *parinode, /* parent inode pointer */ | ||
356 | compat_xfs_fsop_handlereq_t *hreq, | ||
357 | struct inode **inode) | ||
358 | { | 348 | { |
359 | void __user *hanp; | 349 | return xfs_handle_to_dentry(parfilp, |
360 | size_t hlen; | 350 | compat_ptr(hreq->ihandle), hreq->ihandlen); |
361 | xfs_fid_t *xfid; | ||
362 | xfs_handle_t *handlep; | ||
363 | xfs_handle_t handle; | ||
364 | xfs_inode_t *ip; | ||
365 | xfs_ino_t ino; | ||
366 | __u32 igen; | ||
367 | int error; | ||
368 | |||
369 | /* | ||
370 | * Only allow handle opens under a directory. | ||
371 | */ | ||
372 | if (!S_ISDIR(parinode->i_mode)) | ||
373 | return XFS_ERROR(ENOTDIR); | ||
374 | |||
375 | hanp = compat_ptr(hreq->ihandle); | ||
376 | hlen = hreq->ihandlen; | ||
377 | handlep = &handle; | ||
378 | |||
379 | if (hlen < sizeof(handlep->ha_fsid) || hlen > sizeof(*handlep)) | ||
380 | return XFS_ERROR(EINVAL); | ||
381 | if (copy_from_user(handlep, hanp, hlen)) | ||
382 | return XFS_ERROR(EFAULT); | ||
383 | if (hlen < sizeof(*handlep)) | ||
384 | memset(((char *)handlep) + hlen, 0, sizeof(*handlep) - hlen); | ||
385 | if (hlen > sizeof(handlep->ha_fsid)) { | ||
386 | if (handlep->ha_fid.fid_len != | ||
387 | (hlen - sizeof(handlep->ha_fsid) - | ||
388 | sizeof(handlep->ha_fid.fid_len)) || | ||
389 | handlep->ha_fid.fid_pad) | ||
390 | return XFS_ERROR(EINVAL); | ||
391 | } | ||
392 | |||
393 | /* | ||
394 | * Crack the handle, obtain the inode # & generation # | ||
395 | */ | ||
396 | xfid = (struct xfs_fid *)&handlep->ha_fid; | ||
397 | if (xfid->fid_len == sizeof(*xfid) - sizeof(xfid->fid_len)) { | ||
398 | ino = xfid->fid_ino; | ||
399 | igen = xfid->fid_gen; | ||
400 | } else { | ||
401 | return XFS_ERROR(EINVAL); | ||
402 | } | ||
403 | |||
404 | /* | ||
405 | * Get the XFS inode, building a Linux inode to go with it. | ||
406 | */ | ||
407 | error = xfs_iget(mp, NULL, ino, 0, XFS_ILOCK_SHARED, &ip, 0); | ||
408 | if (error) | ||
409 | return error; | ||
410 | if (ip == NULL) | ||
411 | return XFS_ERROR(EIO); | ||
412 | if (ip->i_d.di_gen != igen) { | ||
413 | xfs_iput_new(ip, XFS_ILOCK_SHARED); | ||
414 | return XFS_ERROR(ENOENT); | ||
415 | } | ||
416 | |||
417 | xfs_iunlock(ip, XFS_ILOCK_SHARED); | ||
418 | |||
419 | *inode = VFS_I(ip); | ||
420 | return 0; | ||
421 | } | 351 | } |
422 | 352 | ||
423 | STATIC int | 353 | STATIC int |
424 | xfs_compat_attrlist_by_handle( | 354 | xfs_compat_attrlist_by_handle( |
425 | xfs_mount_t *mp, | 355 | struct file *parfilp, |
426 | void __user *arg, | 356 | void __user *arg) |
427 | struct inode *parinode) | ||
428 | { | 357 | { |
429 | int error; | 358 | int error; |
430 | attrlist_cursor_kern_t *cursor; | 359 | attrlist_cursor_kern_t *cursor; |
431 | compat_xfs_fsop_attrlist_handlereq_t al_hreq; | 360 | compat_xfs_fsop_attrlist_handlereq_t al_hreq; |
432 | struct inode *inode; | 361 | struct dentry *dentry; |
433 | char *kbuf; | 362 | char *kbuf; |
434 | 363 | ||
435 | if (!capable(CAP_SYS_ADMIN)) | 364 | if (!capable(CAP_SYS_ADMIN)) |
@@ -446,17 +375,17 @@ xfs_compat_attrlist_by_handle( | |||
446 | if (al_hreq.flags & ~(ATTR_ROOT | ATTR_SECURE)) | 375 | if (al_hreq.flags & ~(ATTR_ROOT | ATTR_SECURE)) |
447 | return -XFS_ERROR(EINVAL); | 376 | return -XFS_ERROR(EINVAL); |
448 | 377 | ||
449 | error = xfs_vget_fsop_handlereq_compat(mp, parinode, &al_hreq.hreq, | 378 | dentry = xfs_compat_handlereq_to_dentry(parfilp, &al_hreq.hreq); |
450 | &inode); | 379 | if (IS_ERR(dentry)) |
451 | if (error) | 380 | return PTR_ERR(dentry); |
452 | goto out; | ||
453 | 381 | ||
382 | error = -ENOMEM; | ||
454 | kbuf = kmalloc(al_hreq.buflen, GFP_KERNEL); | 383 | kbuf = kmalloc(al_hreq.buflen, GFP_KERNEL); |
455 | if (!kbuf) | 384 | if (!kbuf) |
456 | goto out_vn_rele; | 385 | goto out_dput; |
457 | 386 | ||
458 | cursor = (attrlist_cursor_kern_t *)&al_hreq.pos; | 387 | cursor = (attrlist_cursor_kern_t *)&al_hreq.pos; |
459 | error = xfs_attr_list(XFS_I(inode), kbuf, al_hreq.buflen, | 388 | error = -xfs_attr_list(XFS_I(dentry->d_inode), kbuf, al_hreq.buflen, |
460 | al_hreq.flags, cursor); | 389 | al_hreq.flags, cursor); |
461 | if (error) | 390 | if (error) |
462 | goto out_kfree; | 391 | goto out_kfree; |
@@ -466,22 +395,20 @@ xfs_compat_attrlist_by_handle( | |||
466 | 395 | ||
467 | out_kfree: | 396 | out_kfree: |
468 | kfree(kbuf); | 397 | kfree(kbuf); |
469 | out_vn_rele: | 398 | out_dput: |
470 | iput(inode); | 399 | dput(dentry); |
471 | out: | 400 | return error; |
472 | return -error; | ||
473 | } | 401 | } |
474 | 402 | ||
475 | STATIC int | 403 | STATIC int |
476 | xfs_compat_attrmulti_by_handle( | 404 | xfs_compat_attrmulti_by_handle( |
477 | xfs_mount_t *mp, | 405 | struct file *parfilp, |
478 | void __user *arg, | 406 | void __user *arg) |
479 | struct inode *parinode) | ||
480 | { | 407 | { |
481 | int error; | 408 | int error; |
482 | compat_xfs_attr_multiop_t *ops; | 409 | compat_xfs_attr_multiop_t *ops; |
483 | compat_xfs_fsop_attrmulti_handlereq_t am_hreq; | 410 | compat_xfs_fsop_attrmulti_handlereq_t am_hreq; |
484 | struct inode *inode; | 411 | struct dentry *dentry; |
485 | unsigned int i, size; | 412 | unsigned int i, size; |
486 | char *attr_name; | 413 | char *attr_name; |
487 | 414 | ||
@@ -491,20 +418,19 @@ xfs_compat_attrmulti_by_handle( | |||
491 | sizeof(compat_xfs_fsop_attrmulti_handlereq_t))) | 418 | sizeof(compat_xfs_fsop_attrmulti_handlereq_t))) |
492 | return -XFS_ERROR(EFAULT); | 419 | return -XFS_ERROR(EFAULT); |
493 | 420 | ||
494 | error = xfs_vget_fsop_handlereq_compat(mp, parinode, &am_hreq.hreq, | 421 | dentry = xfs_compat_handlereq_to_dentry(parfilp, &am_hreq.hreq); |
495 | &inode); | 422 | if (IS_ERR(dentry)) |
496 | if (error) | 423 | return PTR_ERR(dentry); |
497 | goto out; | ||
498 | 424 | ||
499 | error = E2BIG; | 425 | error = E2BIG; |
500 | size = am_hreq.opcount * sizeof(compat_xfs_attr_multiop_t); | 426 | size = am_hreq.opcount * sizeof(compat_xfs_attr_multiop_t); |
501 | if (!size || size > 16 * PAGE_SIZE) | 427 | if (!size || size > 16 * PAGE_SIZE) |
502 | goto out_vn_rele; | 428 | goto out_dput; |
503 | 429 | ||
504 | error = ENOMEM; | 430 | error = ENOMEM; |
505 | ops = kmalloc(size, GFP_KERNEL); | 431 | ops = kmalloc(size, GFP_KERNEL); |
506 | if (!ops) | 432 | if (!ops) |
507 | goto out_vn_rele; | 433 | goto out_dput; |
508 | 434 | ||
509 | error = EFAULT; | 435 | error = EFAULT; |
510 | if (copy_from_user(ops, compat_ptr(am_hreq.ops), size)) | 436 | if (copy_from_user(ops, compat_ptr(am_hreq.ops), size)) |
@@ -527,20 +453,29 @@ xfs_compat_attrmulti_by_handle( | |||
527 | 453 | ||
528 | switch (ops[i].am_opcode) { | 454 | switch (ops[i].am_opcode) { |
529 | case ATTR_OP_GET: | 455 | case ATTR_OP_GET: |
530 | ops[i].am_error = xfs_attrmulti_attr_get(inode, | 456 | ops[i].am_error = xfs_attrmulti_attr_get( |
531 | attr_name, | 457 | dentry->d_inode, attr_name, |
532 | compat_ptr(ops[i].am_attrvalue), | 458 | compat_ptr(ops[i].am_attrvalue), |
533 | &ops[i].am_length, ops[i].am_flags); | 459 | &ops[i].am_length, ops[i].am_flags); |
534 | break; | 460 | break; |
535 | case ATTR_OP_SET: | 461 | case ATTR_OP_SET: |
536 | ops[i].am_error = xfs_attrmulti_attr_set(inode, | 462 | ops[i].am_error = mnt_want_write(parfilp->f_path.mnt); |
537 | attr_name, | 463 | if (ops[i].am_error) |
464 | break; | ||
465 | ops[i].am_error = xfs_attrmulti_attr_set( | ||
466 | dentry->d_inode, attr_name, | ||
538 | compat_ptr(ops[i].am_attrvalue), | 467 | compat_ptr(ops[i].am_attrvalue), |
539 | ops[i].am_length, ops[i].am_flags); | 468 | ops[i].am_length, ops[i].am_flags); |
469 | mnt_drop_write(parfilp->f_path.mnt); | ||
540 | break; | 470 | break; |
541 | case ATTR_OP_REMOVE: | 471 | case ATTR_OP_REMOVE: |
542 | ops[i].am_error = xfs_attrmulti_attr_remove(inode, | 472 | ops[i].am_error = mnt_want_write(parfilp->f_path.mnt); |
543 | attr_name, ops[i].am_flags); | 473 | if (ops[i].am_error) |
474 | break; | ||
475 | ops[i].am_error = xfs_attrmulti_attr_remove( | ||
476 | dentry->d_inode, attr_name, | ||
477 | ops[i].am_flags); | ||
478 | mnt_drop_write(parfilp->f_path.mnt); | ||
544 | break; | 479 | break; |
545 | default: | 480 | default: |
546 | ops[i].am_error = EINVAL; | 481 | ops[i].am_error = EINVAL; |
@@ -553,22 +488,20 @@ xfs_compat_attrmulti_by_handle( | |||
553 | kfree(attr_name); | 488 | kfree(attr_name); |
554 | out_kfree_ops: | 489 | out_kfree_ops: |
555 | kfree(ops); | 490 | kfree(ops); |
556 | out_vn_rele: | 491 | out_dput: |
557 | iput(inode); | 492 | dput(dentry); |
558 | out: | ||
559 | return -error; | 493 | return -error; |
560 | } | 494 | } |
561 | 495 | ||
562 | STATIC int | 496 | STATIC int |
563 | xfs_compat_fssetdm_by_handle( | 497 | xfs_compat_fssetdm_by_handle( |
564 | xfs_mount_t *mp, | 498 | struct file *parfilp, |
565 | void __user *arg, | 499 | void __user *arg) |
566 | struct inode *parinode) | ||
567 | { | 500 | { |
568 | int error; | 501 | int error; |
569 | struct fsdmidata fsd; | 502 | struct fsdmidata fsd; |
570 | compat_xfs_fsop_setdm_handlereq_t dmhreq; | 503 | compat_xfs_fsop_setdm_handlereq_t dmhreq; |
571 | struct inode *inode; | 504 | struct dentry *dentry; |
572 | 505 | ||
573 | if (!capable(CAP_MKNOD)) | 506 | if (!capable(CAP_MKNOD)) |
574 | return -XFS_ERROR(EPERM); | 507 | return -XFS_ERROR(EPERM); |
@@ -576,12 +509,11 @@ xfs_compat_fssetdm_by_handle( | |||
576 | sizeof(compat_xfs_fsop_setdm_handlereq_t))) | 509 | sizeof(compat_xfs_fsop_setdm_handlereq_t))) |
577 | return -XFS_ERROR(EFAULT); | 510 | return -XFS_ERROR(EFAULT); |
578 | 511 | ||
579 | error = xfs_vget_fsop_handlereq_compat(mp, parinode, &dmhreq.hreq, | 512 | dentry = xfs_compat_handlereq_to_dentry(parfilp, &dmhreq.hreq); |
580 | &inode); | 513 | if (IS_ERR(dentry)) |
581 | if (error) | 514 | return PTR_ERR(dentry); |
582 | return -error; | ||
583 | 515 | ||
584 | if (IS_IMMUTABLE(inode) || IS_APPEND(inode)) { | 516 | if (IS_IMMUTABLE(dentry->d_inode) || IS_APPEND(dentry->d_inode)) { |
585 | error = -XFS_ERROR(EPERM); | 517 | error = -XFS_ERROR(EPERM); |
586 | goto out; | 518 | goto out; |
587 | } | 519 | } |
@@ -591,11 +523,11 @@ xfs_compat_fssetdm_by_handle( | |||
591 | goto out; | 523 | goto out; |
592 | } | 524 | } |
593 | 525 | ||
594 | error = -xfs_set_dmattrs(XFS_I(inode), fsd.fsd_dmevmask, | 526 | error = -xfs_set_dmattrs(XFS_I(dentry->d_inode), fsd.fsd_dmevmask, |
595 | fsd.fsd_dmstate); | 527 | fsd.fsd_dmstate); |
596 | 528 | ||
597 | out: | 529 | out: |
598 | iput(inode); | 530 | dput(dentry); |
599 | return error; | 531 | return error; |
600 | } | 532 | } |
601 | 533 | ||
@@ -722,21 +654,21 @@ xfs_file_compat_ioctl( | |||
722 | 654 | ||
723 | if (xfs_compat_handlereq_copyin(&hreq, arg)) | 655 | if (xfs_compat_handlereq_copyin(&hreq, arg)) |
724 | return -XFS_ERROR(EFAULT); | 656 | return -XFS_ERROR(EFAULT); |
725 | return xfs_open_by_handle(mp, &hreq, filp, inode); | 657 | return xfs_open_by_handle(filp, &hreq); |
726 | } | 658 | } |
727 | case XFS_IOC_READLINK_BY_HANDLE_32: { | 659 | case XFS_IOC_READLINK_BY_HANDLE_32: { |
728 | struct xfs_fsop_handlereq hreq; | 660 | struct xfs_fsop_handlereq hreq; |
729 | 661 | ||
730 | if (xfs_compat_handlereq_copyin(&hreq, arg)) | 662 | if (xfs_compat_handlereq_copyin(&hreq, arg)) |
731 | return -XFS_ERROR(EFAULT); | 663 | return -XFS_ERROR(EFAULT); |
732 | return xfs_readlink_by_handle(mp, &hreq, inode); | 664 | return xfs_readlink_by_handle(filp, &hreq); |
733 | } | 665 | } |
734 | case XFS_IOC_ATTRLIST_BY_HANDLE_32: | 666 | case XFS_IOC_ATTRLIST_BY_HANDLE_32: |
735 | return xfs_compat_attrlist_by_handle(mp, arg, inode); | 667 | return xfs_compat_attrlist_by_handle(filp, arg); |
736 | case XFS_IOC_ATTRMULTI_BY_HANDLE_32: | 668 | case XFS_IOC_ATTRMULTI_BY_HANDLE_32: |
737 | return xfs_compat_attrmulti_by_handle(mp, arg, inode); | 669 | return xfs_compat_attrmulti_by_handle(filp, arg); |
738 | case XFS_IOC_FSSETDM_BY_HANDLE_32: | 670 | case XFS_IOC_FSSETDM_BY_HANDLE_32: |
739 | return xfs_compat_fssetdm_by_handle(mp, arg, inode); | 671 | return xfs_compat_fssetdm_by_handle(filp, arg); |
740 | default: | 672 | default: |
741 | return -XFS_ERROR(ENOIOCTLCMD); | 673 | return -XFS_ERROR(ENOIOCTLCMD); |
742 | } | 674 | } |
diff --git a/fs/xfs/linux-2.6/xfs_super.c b/fs/xfs/linux-2.6/xfs_super.c index 95a971080368..c71e226da7f5 100644 --- a/fs/xfs/linux-2.6/xfs_super.c +++ b/fs/xfs/linux-2.6/xfs_super.c | |||
@@ -1197,6 +1197,7 @@ xfs_fs_remount( | |||
1197 | struct xfs_mount *mp = XFS_M(sb); | 1197 | struct xfs_mount *mp = XFS_M(sb); |
1198 | substring_t args[MAX_OPT_ARGS]; | 1198 | substring_t args[MAX_OPT_ARGS]; |
1199 | char *p; | 1199 | char *p; |
1200 | int error; | ||
1200 | 1201 | ||
1201 | while ((p = strsep(&options, ",")) != NULL) { | 1202 | while ((p = strsep(&options, ",")) != NULL) { |
1202 | int token; | 1203 | int token; |
@@ -1247,11 +1248,25 @@ xfs_fs_remount( | |||
1247 | } | 1248 | } |
1248 | } | 1249 | } |
1249 | 1250 | ||
1250 | /* rw/ro -> rw */ | 1251 | /* ro -> rw */ |
1251 | if ((mp->m_flags & XFS_MOUNT_RDONLY) && !(*flags & MS_RDONLY)) { | 1252 | if ((mp->m_flags & XFS_MOUNT_RDONLY) && !(*flags & MS_RDONLY)) { |
1252 | mp->m_flags &= ~XFS_MOUNT_RDONLY; | 1253 | mp->m_flags &= ~XFS_MOUNT_RDONLY; |
1253 | if (mp->m_flags & XFS_MOUNT_BARRIER) | 1254 | if (mp->m_flags & XFS_MOUNT_BARRIER) |
1254 | xfs_mountfs_check_barriers(mp); | 1255 | xfs_mountfs_check_barriers(mp); |
1256 | |||
1257 | /* | ||
1258 | * If this is the first remount to writeable state we | ||
1259 | * might have some superblock changes to update. | ||
1260 | */ | ||
1261 | if (mp->m_update_flags) { | ||
1262 | error = xfs_mount_log_sb(mp, mp->m_update_flags); | ||
1263 | if (error) { | ||
1264 | cmn_err(CE_WARN, | ||
1265 | "XFS: failed to write sb changes"); | ||
1266 | return error; | ||
1267 | } | ||
1268 | mp->m_update_flags = 0; | ||
1269 | } | ||
1255 | } | 1270 | } |
1256 | 1271 | ||
1257 | /* rw -> ro */ | 1272 | /* rw -> ro */ |
diff --git a/fs/xfs/linux-2.6/xfs_sync.c b/fs/xfs/linux-2.6/xfs_sync.c index 2ed035354c26..a608e72fa405 100644 --- a/fs/xfs/linux-2.6/xfs_sync.c +++ b/fs/xfs/linux-2.6/xfs_sync.c | |||
@@ -371,7 +371,11 @@ xfs_quiesce_attr( | |||
371 | /* flush inodes and push all remaining buffers out to disk */ | 371 | /* flush inodes and push all remaining buffers out to disk */ |
372 | xfs_quiesce_fs(mp); | 372 | xfs_quiesce_fs(mp); |
373 | 373 | ||
374 | ASSERT_ALWAYS(atomic_read(&mp->m_active_trans) == 0); | 374 | /* |
375 | * Just warn here till VFS can correctly support | ||
376 | * read-only remount without racing. | ||
377 | */ | ||
378 | WARN_ON(atomic_read(&mp->m_active_trans) != 0); | ||
375 | 379 | ||
376 | /* Push the superblock and write an unmount record */ | 380 | /* Push the superblock and write an unmount record */ |
377 | error = xfs_log_sbcount(mp, 1); | 381 | error = xfs_log_sbcount(mp, 1); |
diff --git a/fs/xfs/quota/xfs_dquot.c b/fs/xfs/quota/xfs_dquot.c index 591ca6602bfb..6543c0b29753 100644 --- a/fs/xfs/quota/xfs_dquot.c +++ b/fs/xfs/quota/xfs_dquot.c | |||
@@ -73,6 +73,8 @@ int xfs_dqreq_num; | |||
73 | int xfs_dqerror_mod = 33; | 73 | int xfs_dqerror_mod = 33; |
74 | #endif | 74 | #endif |
75 | 75 | ||
76 | static struct lock_class_key xfs_dquot_other_class; | ||
77 | |||
76 | /* | 78 | /* |
77 | * Allocate and initialize a dquot. We don't always allocate fresh memory; | 79 | * Allocate and initialize a dquot. We don't always allocate fresh memory; |
78 | * we try to reclaim a free dquot if the number of incore dquots are above | 80 | * we try to reclaim a free dquot if the number of incore dquots are above |
@@ -139,7 +141,15 @@ xfs_qm_dqinit( | |||
139 | ASSERT(dqp->q_trace); | 141 | ASSERT(dqp->q_trace); |
140 | xfs_dqtrace_entry(dqp, "DQRECLAIMED_INIT"); | 142 | xfs_dqtrace_entry(dqp, "DQRECLAIMED_INIT"); |
141 | #endif | 143 | #endif |
142 | } | 144 | } |
145 | |||
146 | /* | ||
147 | * In either case we need to make sure group quotas have a different | ||
148 | * lock class than user quotas, to make sure lockdep knows we can | ||
149 | * locks of one of each at the same time. | ||
150 | */ | ||
151 | if (!(type & XFS_DQ_USER)) | ||
152 | lockdep_set_class(&dqp->q_qlock, &xfs_dquot_other_class); | ||
143 | 153 | ||
144 | /* | 154 | /* |
145 | * log item gets initialized later | 155 | * log item gets initialized later |
@@ -421,7 +431,7 @@ xfs_qm_dqalloc( | |||
421 | /* | 431 | /* |
422 | * Initialize the bmap freelist prior to calling bmapi code. | 432 | * Initialize the bmap freelist prior to calling bmapi code. |
423 | */ | 433 | */ |
424 | XFS_BMAP_INIT(&flist, &firstblock); | 434 | xfs_bmap_init(&flist, &firstblock); |
425 | xfs_ilock(quotip, XFS_ILOCK_EXCL); | 435 | xfs_ilock(quotip, XFS_ILOCK_EXCL); |
426 | /* | 436 | /* |
427 | * Return if this type of quotas is turned off while we didn't | 437 | * Return if this type of quotas is turned off while we didn't |
@@ -1383,6 +1393,12 @@ xfs_dqunlock_nonotify( | |||
1383 | mutex_unlock(&(dqp->q_qlock)); | 1393 | mutex_unlock(&(dqp->q_qlock)); |
1384 | } | 1394 | } |
1385 | 1395 | ||
1396 | /* | ||
1397 | * Lock two xfs_dquot structures. | ||
1398 | * | ||
1399 | * To avoid deadlocks we always lock the quota structure with | ||
1400 | * the lowerd id first. | ||
1401 | */ | ||
1386 | void | 1402 | void |
1387 | xfs_dqlock2( | 1403 | xfs_dqlock2( |
1388 | xfs_dquot_t *d1, | 1404 | xfs_dquot_t *d1, |
@@ -1392,18 +1408,16 @@ xfs_dqlock2( | |||
1392 | ASSERT(d1 != d2); | 1408 | ASSERT(d1 != d2); |
1393 | if (be32_to_cpu(d1->q_core.d_id) > | 1409 | if (be32_to_cpu(d1->q_core.d_id) > |
1394 | be32_to_cpu(d2->q_core.d_id)) { | 1410 | be32_to_cpu(d2->q_core.d_id)) { |
1395 | xfs_dqlock(d2); | 1411 | mutex_lock(&d2->q_qlock); |
1396 | xfs_dqlock(d1); | 1412 | mutex_lock_nested(&d1->q_qlock, XFS_QLOCK_NESTED); |
1397 | } else { | 1413 | } else { |
1398 | xfs_dqlock(d1); | 1414 | mutex_lock(&d1->q_qlock); |
1399 | xfs_dqlock(d2); | 1415 | mutex_lock_nested(&d2->q_qlock, XFS_QLOCK_NESTED); |
1400 | } | ||
1401 | } else { | ||
1402 | if (d1) { | ||
1403 | xfs_dqlock(d1); | ||
1404 | } else if (d2) { | ||
1405 | xfs_dqlock(d2); | ||
1406 | } | 1416 | } |
1417 | } else if (d1) { | ||
1418 | mutex_lock(&d1->q_qlock); | ||
1419 | } else if (d2) { | ||
1420 | mutex_lock(&d2->q_qlock); | ||
1407 | } | 1421 | } |
1408 | } | 1422 | } |
1409 | 1423 | ||
diff --git a/fs/xfs/quota/xfs_dquot.h b/fs/xfs/quota/xfs_dquot.h index 7e455337e2ba..d443e93b4331 100644 --- a/fs/xfs/quota/xfs_dquot.h +++ b/fs/xfs/quota/xfs_dquot.h | |||
@@ -97,6 +97,16 @@ typedef struct xfs_dquot { | |||
97 | #define dq_hashlist q_lists.dqm_hashlist | 97 | #define dq_hashlist q_lists.dqm_hashlist |
98 | #define dq_flags q_lists.dqm_flags | 98 | #define dq_flags q_lists.dqm_flags |
99 | 99 | ||
100 | /* | ||
101 | * Lock hierachy for q_qlock: | ||
102 | * XFS_QLOCK_NORMAL is the implicit default, | ||
103 | * XFS_QLOCK_NESTED is the dquot with the higher id in xfs_dqlock2 | ||
104 | */ | ||
105 | enum { | ||
106 | XFS_QLOCK_NORMAL = 0, | ||
107 | XFS_QLOCK_NESTED, | ||
108 | }; | ||
109 | |||
100 | #define XFS_DQHOLD(dqp) ((dqp)->q_nrefs++) | 110 | #define XFS_DQHOLD(dqp) ((dqp)->q_nrefs++) |
101 | 111 | ||
102 | #ifdef DEBUG | 112 | #ifdef DEBUG |
diff --git a/fs/xfs/quota/xfs_qm.c b/fs/xfs/quota/xfs_qm.c index 6b13960cf318..7a2beb64314f 100644 --- a/fs/xfs/quota/xfs_qm.c +++ b/fs/xfs/quota/xfs_qm.c | |||
@@ -1070,6 +1070,13 @@ xfs_qm_sync( | |||
1070 | return 0; | 1070 | return 0; |
1071 | } | 1071 | } |
1072 | 1072 | ||
1073 | /* | ||
1074 | * The hash chains and the mplist use the same xfs_dqhash structure as | ||
1075 | * their list head, but we can take the mplist qh_lock and one of the | ||
1076 | * hash qh_locks at the same time without any problem as they aren't | ||
1077 | * related. | ||
1078 | */ | ||
1079 | static struct lock_class_key xfs_quota_mplist_class; | ||
1073 | 1080 | ||
1074 | /* | 1081 | /* |
1075 | * This initializes all the quota information that's kept in the | 1082 | * This initializes all the quota information that's kept in the |
@@ -1105,6 +1112,8 @@ xfs_qm_init_quotainfo( | |||
1105 | } | 1112 | } |
1106 | 1113 | ||
1107 | xfs_qm_list_init(&qinf->qi_dqlist, "mpdqlist", 0); | 1114 | xfs_qm_list_init(&qinf->qi_dqlist, "mpdqlist", 0); |
1115 | lockdep_set_class(&qinf->qi_dqlist.qh_lock, &xfs_quota_mplist_class); | ||
1116 | |||
1108 | qinf->qi_dqreclaims = 0; | 1117 | qinf->qi_dqreclaims = 0; |
1109 | 1118 | ||
1110 | /* mutex used to serialize quotaoffs */ | 1119 | /* mutex used to serialize quotaoffs */ |
diff --git a/fs/xfs/xfs_acl.h b/fs/xfs/xfs_acl.h index a4e293b93efa..642f1db4def4 100644 --- a/fs/xfs/xfs_acl.h +++ b/fs/xfs/xfs_acl.h | |||
@@ -22,7 +22,6 @@ | |||
22 | * Access Control Lists | 22 | * Access Control Lists |
23 | */ | 23 | */ |
24 | typedef __uint16_t xfs_acl_perm_t; | 24 | typedef __uint16_t xfs_acl_perm_t; |
25 | typedef __int32_t xfs_acl_type_t; | ||
26 | typedef __int32_t xfs_acl_tag_t; | 25 | typedef __int32_t xfs_acl_tag_t; |
27 | typedef __int32_t xfs_acl_id_t; | 26 | typedef __int32_t xfs_acl_id_t; |
28 | 27 | ||
diff --git a/fs/xfs/xfs_ag.h b/fs/xfs/xfs_ag.h index f2e21817a226..143d63ecb20a 100644 --- a/fs/xfs/xfs_ag.h +++ b/fs/xfs/xfs_ag.h | |||
@@ -231,7 +231,7 @@ typedef struct xfs_perag | |||
231 | #define XFS_FSB_TO_AGNO(mp,fsbno) \ | 231 | #define XFS_FSB_TO_AGNO(mp,fsbno) \ |
232 | ((xfs_agnumber_t)((fsbno) >> (mp)->m_sb.sb_agblklog)) | 232 | ((xfs_agnumber_t)((fsbno) >> (mp)->m_sb.sb_agblklog)) |
233 | #define XFS_FSB_TO_AGBNO(mp,fsbno) \ | 233 | #define XFS_FSB_TO_AGBNO(mp,fsbno) \ |
234 | ((xfs_agblock_t)((fsbno) & XFS_MASK32LO((mp)->m_sb.sb_agblklog))) | 234 | ((xfs_agblock_t)((fsbno) & xfs_mask32lo((mp)->m_sb.sb_agblklog))) |
235 | #define XFS_AGB_TO_DADDR(mp,agno,agbno) \ | 235 | #define XFS_AGB_TO_DADDR(mp,agno,agbno) \ |
236 | ((xfs_daddr_t)XFS_FSB_TO_BB(mp, \ | 236 | ((xfs_daddr_t)XFS_FSB_TO_BB(mp, \ |
237 | (xfs_fsblock_t)(agno) * (mp)->m_sb.sb_agblocks + (agbno))) | 237 | (xfs_fsblock_t)(agno) * (mp)->m_sb.sb_agblocks + (agbno))) |
@@ -244,8 +244,8 @@ typedef struct xfs_perag | |||
244 | #define XFS_AG_CHECK_DADDR(mp,d,len) \ | 244 | #define XFS_AG_CHECK_DADDR(mp,d,len) \ |
245 | ((len) == 1 ? \ | 245 | ((len) == 1 ? \ |
246 | ASSERT((d) == XFS_SB_DADDR || \ | 246 | ASSERT((d) == XFS_SB_DADDR || \ |
247 | XFS_DADDR_TO_AGBNO(mp, d) != XFS_SB_DADDR) : \ | 247 | xfs_daddr_to_agbno(mp, d) != XFS_SB_DADDR) : \ |
248 | ASSERT(XFS_DADDR_TO_AGNO(mp, d) == \ | 248 | ASSERT(xfs_daddr_to_agno(mp, d) == \ |
249 | XFS_DADDR_TO_AGNO(mp, (d) + (len) - 1))) | 249 | xfs_daddr_to_agno(mp, (d) + (len) - 1))) |
250 | 250 | ||
251 | #endif /* __XFS_AG_H__ */ | 251 | #endif /* __XFS_AG_H__ */ |
diff --git a/fs/xfs/xfs_alloc_btree.c b/fs/xfs/xfs_alloc_btree.c index 733cb75a8c5d..c10c3a292d30 100644 --- a/fs/xfs/xfs_alloc_btree.c +++ b/fs/xfs/xfs_alloc_btree.c | |||
@@ -115,7 +115,7 @@ xfs_allocbt_free_block( | |||
115 | xfs_agblock_t bno; | 115 | xfs_agblock_t bno; |
116 | int error; | 116 | int error; |
117 | 117 | ||
118 | bno = XFS_DADDR_TO_AGBNO(cur->bc_mp, XFS_BUF_ADDR(bp)); | 118 | bno = xfs_daddr_to_agbno(cur->bc_mp, XFS_BUF_ADDR(bp)); |
119 | error = xfs_alloc_put_freelist(cur->bc_tp, agbp, NULL, bno, 1); | 119 | error = xfs_alloc_put_freelist(cur->bc_tp, agbp, NULL, bno, 1); |
120 | if (error) | 120 | if (error) |
121 | return error; | 121 | return error; |
diff --git a/fs/xfs/xfs_attr.c b/fs/xfs/xfs_attr.c index f7cdc28aff41..5fde1654b430 100644 --- a/fs/xfs/xfs_attr.c +++ b/fs/xfs/xfs_attr.c | |||
@@ -374,7 +374,7 @@ xfs_attr_set_int(xfs_inode_t *dp, struct xfs_name *name, | |||
374 | * It won't fit in the shortform, transform to a leaf block. | 374 | * It won't fit in the shortform, transform to a leaf block. |
375 | * GROT: another possible req'mt for a double-split btree op. | 375 | * GROT: another possible req'mt for a double-split btree op. |
376 | */ | 376 | */ |
377 | XFS_BMAP_INIT(args.flist, args.firstblock); | 377 | xfs_bmap_init(args.flist, args.firstblock); |
378 | error = xfs_attr_shortform_to_leaf(&args); | 378 | error = xfs_attr_shortform_to_leaf(&args); |
379 | if (!error) { | 379 | if (!error) { |
380 | error = xfs_bmap_finish(&args.trans, args.flist, | 380 | error = xfs_bmap_finish(&args.trans, args.flist, |
@@ -956,7 +956,7 @@ xfs_attr_leaf_addname(xfs_da_args_t *args) | |||
956 | * Commit that transaction so that the node_addname() call | 956 | * Commit that transaction so that the node_addname() call |
957 | * can manage its own transactions. | 957 | * can manage its own transactions. |
958 | */ | 958 | */ |
959 | XFS_BMAP_INIT(args->flist, args->firstblock); | 959 | xfs_bmap_init(args->flist, args->firstblock); |
960 | error = xfs_attr_leaf_to_node(args); | 960 | error = xfs_attr_leaf_to_node(args); |
961 | if (!error) { | 961 | if (!error) { |
962 | error = xfs_bmap_finish(&args->trans, args->flist, | 962 | error = xfs_bmap_finish(&args->trans, args->flist, |
@@ -1057,7 +1057,7 @@ xfs_attr_leaf_addname(xfs_da_args_t *args) | |||
1057 | * If the result is small enough, shrink it all into the inode. | 1057 | * If the result is small enough, shrink it all into the inode. |
1058 | */ | 1058 | */ |
1059 | if ((forkoff = xfs_attr_shortform_allfit(bp, dp))) { | 1059 | if ((forkoff = xfs_attr_shortform_allfit(bp, dp))) { |
1060 | XFS_BMAP_INIT(args->flist, args->firstblock); | 1060 | xfs_bmap_init(args->flist, args->firstblock); |
1061 | error = xfs_attr_leaf_to_shortform(bp, args, forkoff); | 1061 | error = xfs_attr_leaf_to_shortform(bp, args, forkoff); |
1062 | /* bp is gone due to xfs_da_shrink_inode */ | 1062 | /* bp is gone due to xfs_da_shrink_inode */ |
1063 | if (!error) { | 1063 | if (!error) { |
@@ -1135,7 +1135,7 @@ xfs_attr_leaf_removename(xfs_da_args_t *args) | |||
1135 | * If the result is small enough, shrink it all into the inode. | 1135 | * If the result is small enough, shrink it all into the inode. |
1136 | */ | 1136 | */ |
1137 | if ((forkoff = xfs_attr_shortform_allfit(bp, dp))) { | 1137 | if ((forkoff = xfs_attr_shortform_allfit(bp, dp))) { |
1138 | XFS_BMAP_INIT(args->flist, args->firstblock); | 1138 | xfs_bmap_init(args->flist, args->firstblock); |
1139 | error = xfs_attr_leaf_to_shortform(bp, args, forkoff); | 1139 | error = xfs_attr_leaf_to_shortform(bp, args, forkoff); |
1140 | /* bp is gone due to xfs_da_shrink_inode */ | 1140 | /* bp is gone due to xfs_da_shrink_inode */ |
1141 | if (!error) { | 1141 | if (!error) { |
@@ -1290,7 +1290,7 @@ restart: | |||
1290 | * have been a b-tree. | 1290 | * have been a b-tree. |
1291 | */ | 1291 | */ |
1292 | xfs_da_state_free(state); | 1292 | xfs_da_state_free(state); |
1293 | XFS_BMAP_INIT(args->flist, args->firstblock); | 1293 | xfs_bmap_init(args->flist, args->firstblock); |
1294 | error = xfs_attr_leaf_to_node(args); | 1294 | error = xfs_attr_leaf_to_node(args); |
1295 | if (!error) { | 1295 | if (!error) { |
1296 | error = xfs_bmap_finish(&args->trans, | 1296 | error = xfs_bmap_finish(&args->trans, |
@@ -1331,7 +1331,7 @@ restart: | |||
1331 | * in the index/blkno/rmtblkno/rmtblkcnt fields and | 1331 | * in the index/blkno/rmtblkno/rmtblkcnt fields and |
1332 | * in the index2/blkno2/rmtblkno2/rmtblkcnt2 fields. | 1332 | * in the index2/blkno2/rmtblkno2/rmtblkcnt2 fields. |
1333 | */ | 1333 | */ |
1334 | XFS_BMAP_INIT(args->flist, args->firstblock); | 1334 | xfs_bmap_init(args->flist, args->firstblock); |
1335 | error = xfs_da_split(state); | 1335 | error = xfs_da_split(state); |
1336 | if (!error) { | 1336 | if (!error) { |
1337 | error = xfs_bmap_finish(&args->trans, args->flist, | 1337 | error = xfs_bmap_finish(&args->trans, args->flist, |
@@ -1443,7 +1443,7 @@ restart: | |||
1443 | * Check to see if the tree needs to be collapsed. | 1443 | * Check to see if the tree needs to be collapsed. |
1444 | */ | 1444 | */ |
1445 | if (retval && (state->path.active > 1)) { | 1445 | if (retval && (state->path.active > 1)) { |
1446 | XFS_BMAP_INIT(args->flist, args->firstblock); | 1446 | xfs_bmap_init(args->flist, args->firstblock); |
1447 | error = xfs_da_join(state); | 1447 | error = xfs_da_join(state); |
1448 | if (!error) { | 1448 | if (!error) { |
1449 | error = xfs_bmap_finish(&args->trans, | 1449 | error = xfs_bmap_finish(&args->trans, |
@@ -1579,7 +1579,7 @@ xfs_attr_node_removename(xfs_da_args_t *args) | |||
1579 | * Check to see if the tree needs to be collapsed. | 1579 | * Check to see if the tree needs to be collapsed. |
1580 | */ | 1580 | */ |
1581 | if (retval && (state->path.active > 1)) { | 1581 | if (retval && (state->path.active > 1)) { |
1582 | XFS_BMAP_INIT(args->flist, args->firstblock); | 1582 | xfs_bmap_init(args->flist, args->firstblock); |
1583 | error = xfs_da_join(state); | 1583 | error = xfs_da_join(state); |
1584 | if (!error) { | 1584 | if (!error) { |
1585 | error = xfs_bmap_finish(&args->trans, args->flist, | 1585 | error = xfs_bmap_finish(&args->trans, args->flist, |
@@ -1630,7 +1630,7 @@ xfs_attr_node_removename(xfs_da_args_t *args) | |||
1630 | == XFS_ATTR_LEAF_MAGIC); | 1630 | == XFS_ATTR_LEAF_MAGIC); |
1631 | 1631 | ||
1632 | if ((forkoff = xfs_attr_shortform_allfit(bp, dp))) { | 1632 | if ((forkoff = xfs_attr_shortform_allfit(bp, dp))) { |
1633 | XFS_BMAP_INIT(args->flist, args->firstblock); | 1633 | xfs_bmap_init(args->flist, args->firstblock); |
1634 | error = xfs_attr_leaf_to_shortform(bp, args, forkoff); | 1634 | error = xfs_attr_leaf_to_shortform(bp, args, forkoff); |
1635 | /* bp is gone due to xfs_da_shrink_inode */ | 1635 | /* bp is gone due to xfs_da_shrink_inode */ |
1636 | if (!error) { | 1636 | if (!error) { |
@@ -2069,7 +2069,7 @@ xfs_attr_rmtval_set(xfs_da_args_t *args) | |||
2069 | /* | 2069 | /* |
2070 | * Allocate a single extent, up to the size of the value. | 2070 | * Allocate a single extent, up to the size of the value. |
2071 | */ | 2071 | */ |
2072 | XFS_BMAP_INIT(args->flist, args->firstblock); | 2072 | xfs_bmap_init(args->flist, args->firstblock); |
2073 | nmap = 1; | 2073 | nmap = 1; |
2074 | error = xfs_bmapi(args->trans, dp, (xfs_fileoff_t)lblkno, | 2074 | error = xfs_bmapi(args->trans, dp, (xfs_fileoff_t)lblkno, |
2075 | blkcnt, | 2075 | blkcnt, |
@@ -2123,7 +2123,7 @@ xfs_attr_rmtval_set(xfs_da_args_t *args) | |||
2123 | /* | 2123 | /* |
2124 | * Try to remember where we decided to put the value. | 2124 | * Try to remember where we decided to put the value. |
2125 | */ | 2125 | */ |
2126 | XFS_BMAP_INIT(args->flist, args->firstblock); | 2126 | xfs_bmap_init(args->flist, args->firstblock); |
2127 | nmap = 1; | 2127 | nmap = 1; |
2128 | error = xfs_bmapi(NULL, dp, (xfs_fileoff_t)lblkno, | 2128 | error = xfs_bmapi(NULL, dp, (xfs_fileoff_t)lblkno, |
2129 | args->rmtblkcnt, | 2129 | args->rmtblkcnt, |
@@ -2188,7 +2188,7 @@ xfs_attr_rmtval_remove(xfs_da_args_t *args) | |||
2188 | /* | 2188 | /* |
2189 | * Try to remember where we decided to put the value. | 2189 | * Try to remember where we decided to put the value. |
2190 | */ | 2190 | */ |
2191 | XFS_BMAP_INIT(args->flist, args->firstblock); | 2191 | xfs_bmap_init(args->flist, args->firstblock); |
2192 | nmap = 1; | 2192 | nmap = 1; |
2193 | error = xfs_bmapi(NULL, args->dp, (xfs_fileoff_t)lblkno, | 2193 | error = xfs_bmapi(NULL, args->dp, (xfs_fileoff_t)lblkno, |
2194 | args->rmtblkcnt, | 2194 | args->rmtblkcnt, |
@@ -2229,7 +2229,7 @@ xfs_attr_rmtval_remove(xfs_da_args_t *args) | |||
2229 | blkcnt = args->rmtblkcnt; | 2229 | blkcnt = args->rmtblkcnt; |
2230 | done = 0; | 2230 | done = 0; |
2231 | while (!done) { | 2231 | while (!done) { |
2232 | XFS_BMAP_INIT(args->flist, args->firstblock); | 2232 | xfs_bmap_init(args->flist, args->firstblock); |
2233 | error = xfs_bunmapi(args->trans, args->dp, lblkno, blkcnt, | 2233 | error = xfs_bunmapi(args->trans, args->dp, lblkno, blkcnt, |
2234 | XFS_BMAPI_ATTRFORK | XFS_BMAPI_METADATA, | 2234 | XFS_BMAPI_ATTRFORK | XFS_BMAPI_METADATA, |
2235 | 1, args->firstblock, args->flist, | 2235 | 1, args->firstblock, args->flist, |
diff --git a/fs/xfs/xfs_attr_leaf.c b/fs/xfs/xfs_attr_leaf.c index 79da6b2ea99e..6c323f8a4cd1 100644 --- a/fs/xfs/xfs_attr_leaf.c +++ b/fs/xfs/xfs_attr_leaf.c | |||
@@ -736,7 +736,7 @@ xfs_attr_shortform_allfit(xfs_dabuf_t *bp, xfs_inode_t *dp) | |||
736 | continue; /* don't copy partial entries */ | 736 | continue; /* don't copy partial entries */ |
737 | if (!(entry->flags & XFS_ATTR_LOCAL)) | 737 | if (!(entry->flags & XFS_ATTR_LOCAL)) |
738 | return(0); | 738 | return(0); |
739 | name_loc = XFS_ATTR_LEAF_NAME_LOCAL(leaf, i); | 739 | name_loc = xfs_attr_leaf_name_local(leaf, i); |
740 | if (name_loc->namelen >= XFS_ATTR_SF_ENTSIZE_MAX) | 740 | if (name_loc->namelen >= XFS_ATTR_SF_ENTSIZE_MAX) |
741 | return(0); | 741 | return(0); |
742 | if (be16_to_cpu(name_loc->valuelen) >= XFS_ATTR_SF_ENTSIZE_MAX) | 742 | if (be16_to_cpu(name_loc->valuelen) >= XFS_ATTR_SF_ENTSIZE_MAX) |
@@ -823,7 +823,7 @@ xfs_attr_leaf_to_shortform(xfs_dabuf_t *bp, xfs_da_args_t *args, int forkoff) | |||
823 | if (!entry->nameidx) | 823 | if (!entry->nameidx) |
824 | continue; | 824 | continue; |
825 | ASSERT(entry->flags & XFS_ATTR_LOCAL); | 825 | ASSERT(entry->flags & XFS_ATTR_LOCAL); |
826 | name_loc = XFS_ATTR_LEAF_NAME_LOCAL(leaf, i); | 826 | name_loc = xfs_attr_leaf_name_local(leaf, i); |
827 | nargs.name = (char *)name_loc->nameval; | 827 | nargs.name = (char *)name_loc->nameval; |
828 | nargs.namelen = name_loc->namelen; | 828 | nargs.namelen = name_loc->namelen; |
829 | nargs.value = (char *)&name_loc->nameval[nargs.namelen]; | 829 | nargs.value = (char *)&name_loc->nameval[nargs.namelen]; |
@@ -1141,14 +1141,14 @@ xfs_attr_leaf_add_work(xfs_dabuf_t *bp, xfs_da_args_t *args, int mapindex) | |||
1141 | * as part of this transaction (a split operation for example). | 1141 | * as part of this transaction (a split operation for example). |
1142 | */ | 1142 | */ |
1143 | if (entry->flags & XFS_ATTR_LOCAL) { | 1143 | if (entry->flags & XFS_ATTR_LOCAL) { |
1144 | name_loc = XFS_ATTR_LEAF_NAME_LOCAL(leaf, args->index); | 1144 | name_loc = xfs_attr_leaf_name_local(leaf, args->index); |
1145 | name_loc->namelen = args->namelen; | 1145 | name_loc->namelen = args->namelen; |
1146 | name_loc->valuelen = cpu_to_be16(args->valuelen); | 1146 | name_loc->valuelen = cpu_to_be16(args->valuelen); |
1147 | memcpy((char *)name_loc->nameval, args->name, args->namelen); | 1147 | memcpy((char *)name_loc->nameval, args->name, args->namelen); |
1148 | memcpy((char *)&name_loc->nameval[args->namelen], args->value, | 1148 | memcpy((char *)&name_loc->nameval[args->namelen], args->value, |
1149 | be16_to_cpu(name_loc->valuelen)); | 1149 | be16_to_cpu(name_loc->valuelen)); |
1150 | } else { | 1150 | } else { |
1151 | name_rmt = XFS_ATTR_LEAF_NAME_REMOTE(leaf, args->index); | 1151 | name_rmt = xfs_attr_leaf_name_remote(leaf, args->index); |
1152 | name_rmt->namelen = args->namelen; | 1152 | name_rmt->namelen = args->namelen; |
1153 | memcpy((char *)name_rmt->name, args->name, args->namelen); | 1153 | memcpy((char *)name_rmt->name, args->name, args->namelen); |
1154 | entry->flags |= XFS_ATTR_INCOMPLETE; | 1154 | entry->flags |= XFS_ATTR_INCOMPLETE; |
@@ -1159,7 +1159,7 @@ xfs_attr_leaf_add_work(xfs_dabuf_t *bp, xfs_da_args_t *args, int mapindex) | |||
1159 | args->rmtblkcnt = XFS_B_TO_FSB(mp, args->valuelen); | 1159 | args->rmtblkcnt = XFS_B_TO_FSB(mp, args->valuelen); |
1160 | } | 1160 | } |
1161 | xfs_da_log_buf(args->trans, bp, | 1161 | xfs_da_log_buf(args->trans, bp, |
1162 | XFS_DA_LOGRANGE(leaf, XFS_ATTR_LEAF_NAME(leaf, args->index), | 1162 | XFS_DA_LOGRANGE(leaf, xfs_attr_leaf_name(leaf, args->index), |
1163 | xfs_attr_leaf_entsize(leaf, args->index))); | 1163 | xfs_attr_leaf_entsize(leaf, args->index))); |
1164 | 1164 | ||
1165 | /* | 1165 | /* |
@@ -1749,10 +1749,10 @@ xfs_attr_leaf_remove(xfs_dabuf_t *bp, xfs_da_args_t *args) | |||
1749 | /* | 1749 | /* |
1750 | * Compress the remaining entries and zero out the removed stuff. | 1750 | * Compress the remaining entries and zero out the removed stuff. |
1751 | */ | 1751 | */ |
1752 | memset(XFS_ATTR_LEAF_NAME(leaf, args->index), 0, entsize); | 1752 | memset(xfs_attr_leaf_name(leaf, args->index), 0, entsize); |
1753 | be16_add_cpu(&hdr->usedbytes, -entsize); | 1753 | be16_add_cpu(&hdr->usedbytes, -entsize); |
1754 | xfs_da_log_buf(args->trans, bp, | 1754 | xfs_da_log_buf(args->trans, bp, |
1755 | XFS_DA_LOGRANGE(leaf, XFS_ATTR_LEAF_NAME(leaf, args->index), | 1755 | XFS_DA_LOGRANGE(leaf, xfs_attr_leaf_name(leaf, args->index), |
1756 | entsize)); | 1756 | entsize)); |
1757 | 1757 | ||
1758 | tmp = (be16_to_cpu(hdr->count) - args->index) | 1758 | tmp = (be16_to_cpu(hdr->count) - args->index) |
@@ -1985,7 +1985,7 @@ xfs_attr_leaf_lookup_int(xfs_dabuf_t *bp, xfs_da_args_t *args) | |||
1985 | continue; | 1985 | continue; |
1986 | } | 1986 | } |
1987 | if (entry->flags & XFS_ATTR_LOCAL) { | 1987 | if (entry->flags & XFS_ATTR_LOCAL) { |
1988 | name_loc = XFS_ATTR_LEAF_NAME_LOCAL(leaf, probe); | 1988 | name_loc = xfs_attr_leaf_name_local(leaf, probe); |
1989 | if (name_loc->namelen != args->namelen) | 1989 | if (name_loc->namelen != args->namelen) |
1990 | continue; | 1990 | continue; |
1991 | if (memcmp(args->name, (char *)name_loc->nameval, args->namelen) != 0) | 1991 | if (memcmp(args->name, (char *)name_loc->nameval, args->namelen) != 0) |
@@ -1995,7 +1995,7 @@ xfs_attr_leaf_lookup_int(xfs_dabuf_t *bp, xfs_da_args_t *args) | |||
1995 | args->index = probe; | 1995 | args->index = probe; |
1996 | return(XFS_ERROR(EEXIST)); | 1996 | return(XFS_ERROR(EEXIST)); |
1997 | } else { | 1997 | } else { |
1998 | name_rmt = XFS_ATTR_LEAF_NAME_REMOTE(leaf, probe); | 1998 | name_rmt = xfs_attr_leaf_name_remote(leaf, probe); |
1999 | if (name_rmt->namelen != args->namelen) | 1999 | if (name_rmt->namelen != args->namelen) |
2000 | continue; | 2000 | continue; |
2001 | if (memcmp(args->name, (char *)name_rmt->name, | 2001 | if (memcmp(args->name, (char *)name_rmt->name, |
@@ -2035,7 +2035,7 @@ xfs_attr_leaf_getvalue(xfs_dabuf_t *bp, xfs_da_args_t *args) | |||
2035 | 2035 | ||
2036 | entry = &leaf->entries[args->index]; | 2036 | entry = &leaf->entries[args->index]; |
2037 | if (entry->flags & XFS_ATTR_LOCAL) { | 2037 | if (entry->flags & XFS_ATTR_LOCAL) { |
2038 | name_loc = XFS_ATTR_LEAF_NAME_LOCAL(leaf, args->index); | 2038 | name_loc = xfs_attr_leaf_name_local(leaf, args->index); |
2039 | ASSERT(name_loc->namelen == args->namelen); | 2039 | ASSERT(name_loc->namelen == args->namelen); |
2040 | ASSERT(memcmp(args->name, name_loc->nameval, args->namelen) == 0); | 2040 | ASSERT(memcmp(args->name, name_loc->nameval, args->namelen) == 0); |
2041 | valuelen = be16_to_cpu(name_loc->valuelen); | 2041 | valuelen = be16_to_cpu(name_loc->valuelen); |
@@ -2050,7 +2050,7 @@ xfs_attr_leaf_getvalue(xfs_dabuf_t *bp, xfs_da_args_t *args) | |||
2050 | args->valuelen = valuelen; | 2050 | args->valuelen = valuelen; |
2051 | memcpy(args->value, &name_loc->nameval[args->namelen], valuelen); | 2051 | memcpy(args->value, &name_loc->nameval[args->namelen], valuelen); |
2052 | } else { | 2052 | } else { |
2053 | name_rmt = XFS_ATTR_LEAF_NAME_REMOTE(leaf, args->index); | 2053 | name_rmt = xfs_attr_leaf_name_remote(leaf, args->index); |
2054 | ASSERT(name_rmt->namelen == args->namelen); | 2054 | ASSERT(name_rmt->namelen == args->namelen); |
2055 | ASSERT(memcmp(args->name, name_rmt->name, args->namelen) == 0); | 2055 | ASSERT(memcmp(args->name, name_rmt->name, args->namelen) == 0); |
2056 | valuelen = be32_to_cpu(name_rmt->valuelen); | 2056 | valuelen = be32_to_cpu(name_rmt->valuelen); |
@@ -2143,7 +2143,7 @@ xfs_attr_leaf_moveents(xfs_attr_leafblock_t *leaf_s, int start_s, | |||
2143 | * off for 6.2, should be revisited later. | 2143 | * off for 6.2, should be revisited later. |
2144 | */ | 2144 | */ |
2145 | if (entry_s->flags & XFS_ATTR_INCOMPLETE) { /* skip partials? */ | 2145 | if (entry_s->flags & XFS_ATTR_INCOMPLETE) { /* skip partials? */ |
2146 | memset(XFS_ATTR_LEAF_NAME(leaf_s, start_s + i), 0, tmp); | 2146 | memset(xfs_attr_leaf_name(leaf_s, start_s + i), 0, tmp); |
2147 | be16_add_cpu(&hdr_s->usedbytes, -tmp); | 2147 | be16_add_cpu(&hdr_s->usedbytes, -tmp); |
2148 | be16_add_cpu(&hdr_s->count, -1); | 2148 | be16_add_cpu(&hdr_s->count, -1); |
2149 | entry_d--; /* to compensate for ++ in loop hdr */ | 2149 | entry_d--; /* to compensate for ++ in loop hdr */ |
@@ -2160,11 +2160,11 @@ xfs_attr_leaf_moveents(xfs_attr_leafblock_t *leaf_s, int start_s, | |||
2160 | entry_d->flags = entry_s->flags; | 2160 | entry_d->flags = entry_s->flags; |
2161 | ASSERT(be16_to_cpu(entry_d->nameidx) + tmp | 2161 | ASSERT(be16_to_cpu(entry_d->nameidx) + tmp |
2162 | <= XFS_LBSIZE(mp)); | 2162 | <= XFS_LBSIZE(mp)); |
2163 | memmove(XFS_ATTR_LEAF_NAME(leaf_d, desti), | 2163 | memmove(xfs_attr_leaf_name(leaf_d, desti), |
2164 | XFS_ATTR_LEAF_NAME(leaf_s, start_s + i), tmp); | 2164 | xfs_attr_leaf_name(leaf_s, start_s + i), tmp); |
2165 | ASSERT(be16_to_cpu(entry_s->nameidx) + tmp | 2165 | ASSERT(be16_to_cpu(entry_s->nameidx) + tmp |
2166 | <= XFS_LBSIZE(mp)); | 2166 | <= XFS_LBSIZE(mp)); |
2167 | memset(XFS_ATTR_LEAF_NAME(leaf_s, start_s + i), 0, tmp); | 2167 | memset(xfs_attr_leaf_name(leaf_s, start_s + i), 0, tmp); |
2168 | be16_add_cpu(&hdr_s->usedbytes, -tmp); | 2168 | be16_add_cpu(&hdr_s->usedbytes, -tmp); |
2169 | be16_add_cpu(&hdr_d->usedbytes, tmp); | 2169 | be16_add_cpu(&hdr_d->usedbytes, tmp); |
2170 | be16_add_cpu(&hdr_s->count, -1); | 2170 | be16_add_cpu(&hdr_s->count, -1); |
@@ -2276,12 +2276,12 @@ xfs_attr_leaf_entsize(xfs_attr_leafblock_t *leaf, int index) | |||
2276 | 2276 | ||
2277 | ASSERT(be16_to_cpu(leaf->hdr.info.magic) == XFS_ATTR_LEAF_MAGIC); | 2277 | ASSERT(be16_to_cpu(leaf->hdr.info.magic) == XFS_ATTR_LEAF_MAGIC); |
2278 | if (leaf->entries[index].flags & XFS_ATTR_LOCAL) { | 2278 | if (leaf->entries[index].flags & XFS_ATTR_LOCAL) { |
2279 | name_loc = XFS_ATTR_LEAF_NAME_LOCAL(leaf, index); | 2279 | name_loc = xfs_attr_leaf_name_local(leaf, index); |
2280 | size = XFS_ATTR_LEAF_ENTSIZE_LOCAL(name_loc->namelen, | 2280 | size = xfs_attr_leaf_entsize_local(name_loc->namelen, |
2281 | be16_to_cpu(name_loc->valuelen)); | 2281 | be16_to_cpu(name_loc->valuelen)); |
2282 | } else { | 2282 | } else { |
2283 | name_rmt = XFS_ATTR_LEAF_NAME_REMOTE(leaf, index); | 2283 | name_rmt = xfs_attr_leaf_name_remote(leaf, index); |
2284 | size = XFS_ATTR_LEAF_ENTSIZE_REMOTE(name_rmt->namelen); | 2284 | size = xfs_attr_leaf_entsize_remote(name_rmt->namelen); |
2285 | } | 2285 | } |
2286 | return(size); | 2286 | return(size); |
2287 | } | 2287 | } |
@@ -2297,13 +2297,13 @@ xfs_attr_leaf_newentsize(int namelen, int valuelen, int blocksize, int *local) | |||
2297 | { | 2297 | { |
2298 | int size; | 2298 | int size; |
2299 | 2299 | ||
2300 | size = XFS_ATTR_LEAF_ENTSIZE_LOCAL(namelen, valuelen); | 2300 | size = xfs_attr_leaf_entsize_local(namelen, valuelen); |
2301 | if (size < XFS_ATTR_LEAF_ENTSIZE_LOCAL_MAX(blocksize)) { | 2301 | if (size < xfs_attr_leaf_entsize_local_max(blocksize)) { |
2302 | if (local) { | 2302 | if (local) { |
2303 | *local = 1; | 2303 | *local = 1; |
2304 | } | 2304 | } |
2305 | } else { | 2305 | } else { |
2306 | size = XFS_ATTR_LEAF_ENTSIZE_REMOTE(namelen); | 2306 | size = xfs_attr_leaf_entsize_remote(namelen); |
2307 | if (local) { | 2307 | if (local) { |
2308 | *local = 0; | 2308 | *local = 0; |
2309 | } | 2309 | } |
@@ -2372,7 +2372,7 @@ xfs_attr_leaf_list_int(xfs_dabuf_t *bp, xfs_attr_list_context_t *context) | |||
2372 | 2372 | ||
2373 | if (entry->flags & XFS_ATTR_LOCAL) { | 2373 | if (entry->flags & XFS_ATTR_LOCAL) { |
2374 | xfs_attr_leaf_name_local_t *name_loc = | 2374 | xfs_attr_leaf_name_local_t *name_loc = |
2375 | XFS_ATTR_LEAF_NAME_LOCAL(leaf, i); | 2375 | xfs_attr_leaf_name_local(leaf, i); |
2376 | 2376 | ||
2377 | retval = context->put_listent(context, | 2377 | retval = context->put_listent(context, |
2378 | entry->flags, | 2378 | entry->flags, |
@@ -2384,7 +2384,7 @@ xfs_attr_leaf_list_int(xfs_dabuf_t *bp, xfs_attr_list_context_t *context) | |||
2384 | return retval; | 2384 | return retval; |
2385 | } else { | 2385 | } else { |
2386 | xfs_attr_leaf_name_remote_t *name_rmt = | 2386 | xfs_attr_leaf_name_remote_t *name_rmt = |
2387 | XFS_ATTR_LEAF_NAME_REMOTE(leaf, i); | 2387 | xfs_attr_leaf_name_remote(leaf, i); |
2388 | 2388 | ||
2389 | int valuelen = be32_to_cpu(name_rmt->valuelen); | 2389 | int valuelen = be32_to_cpu(name_rmt->valuelen); |
2390 | 2390 | ||
@@ -2468,11 +2468,11 @@ xfs_attr_leaf_clearflag(xfs_da_args_t *args) | |||
2468 | 2468 | ||
2469 | #ifdef DEBUG | 2469 | #ifdef DEBUG |
2470 | if (entry->flags & XFS_ATTR_LOCAL) { | 2470 | if (entry->flags & XFS_ATTR_LOCAL) { |
2471 | name_loc = XFS_ATTR_LEAF_NAME_LOCAL(leaf, args->index); | 2471 | name_loc = xfs_attr_leaf_name_local(leaf, args->index); |
2472 | namelen = name_loc->namelen; | 2472 | namelen = name_loc->namelen; |
2473 | name = (char *)name_loc->nameval; | 2473 | name = (char *)name_loc->nameval; |
2474 | } else { | 2474 | } else { |
2475 | name_rmt = XFS_ATTR_LEAF_NAME_REMOTE(leaf, args->index); | 2475 | name_rmt = xfs_attr_leaf_name_remote(leaf, args->index); |
2476 | namelen = name_rmt->namelen; | 2476 | namelen = name_rmt->namelen; |
2477 | name = (char *)name_rmt->name; | 2477 | name = (char *)name_rmt->name; |
2478 | } | 2478 | } |
@@ -2487,7 +2487,7 @@ xfs_attr_leaf_clearflag(xfs_da_args_t *args) | |||
2487 | 2487 | ||
2488 | if (args->rmtblkno) { | 2488 | if (args->rmtblkno) { |
2489 | ASSERT((entry->flags & XFS_ATTR_LOCAL) == 0); | 2489 | ASSERT((entry->flags & XFS_ATTR_LOCAL) == 0); |
2490 | name_rmt = XFS_ATTR_LEAF_NAME_REMOTE(leaf, args->index); | 2490 | name_rmt = xfs_attr_leaf_name_remote(leaf, args->index); |
2491 | name_rmt->valueblk = cpu_to_be32(args->rmtblkno); | 2491 | name_rmt->valueblk = cpu_to_be32(args->rmtblkno); |
2492 | name_rmt->valuelen = cpu_to_be32(args->valuelen); | 2492 | name_rmt->valuelen = cpu_to_be32(args->valuelen); |
2493 | xfs_da_log_buf(args->trans, bp, | 2493 | xfs_da_log_buf(args->trans, bp, |
@@ -2534,7 +2534,7 @@ xfs_attr_leaf_setflag(xfs_da_args_t *args) | |||
2534 | xfs_da_log_buf(args->trans, bp, | 2534 | xfs_da_log_buf(args->trans, bp, |
2535 | XFS_DA_LOGRANGE(leaf, entry, sizeof(*entry))); | 2535 | XFS_DA_LOGRANGE(leaf, entry, sizeof(*entry))); |
2536 | if ((entry->flags & XFS_ATTR_LOCAL) == 0) { | 2536 | if ((entry->flags & XFS_ATTR_LOCAL) == 0) { |
2537 | name_rmt = XFS_ATTR_LEAF_NAME_REMOTE(leaf, args->index); | 2537 | name_rmt = xfs_attr_leaf_name_remote(leaf, args->index); |
2538 | name_rmt->valueblk = 0; | 2538 | name_rmt->valueblk = 0; |
2539 | name_rmt->valuelen = 0; | 2539 | name_rmt->valuelen = 0; |
2540 | xfs_da_log_buf(args->trans, bp, | 2540 | xfs_da_log_buf(args->trans, bp, |
@@ -2607,20 +2607,20 @@ xfs_attr_leaf_flipflags(xfs_da_args_t *args) | |||
2607 | 2607 | ||
2608 | #ifdef DEBUG | 2608 | #ifdef DEBUG |
2609 | if (entry1->flags & XFS_ATTR_LOCAL) { | 2609 | if (entry1->flags & XFS_ATTR_LOCAL) { |
2610 | name_loc = XFS_ATTR_LEAF_NAME_LOCAL(leaf1, args->index); | 2610 | name_loc = xfs_attr_leaf_name_local(leaf1, args->index); |
2611 | namelen1 = name_loc->namelen; | 2611 | namelen1 = name_loc->namelen; |
2612 | name1 = (char *)name_loc->nameval; | 2612 | name1 = (char *)name_loc->nameval; |
2613 | } else { | 2613 | } else { |
2614 | name_rmt = XFS_ATTR_LEAF_NAME_REMOTE(leaf1, args->index); | 2614 | name_rmt = xfs_attr_leaf_name_remote(leaf1, args->index); |
2615 | namelen1 = name_rmt->namelen; | 2615 | namelen1 = name_rmt->namelen; |
2616 | name1 = (char *)name_rmt->name; | 2616 | name1 = (char *)name_rmt->name; |
2617 | } | 2617 | } |
2618 | if (entry2->flags & XFS_ATTR_LOCAL) { | 2618 | if (entry2->flags & XFS_ATTR_LOCAL) { |
2619 | name_loc = XFS_ATTR_LEAF_NAME_LOCAL(leaf2, args->index2); | 2619 | name_loc = xfs_attr_leaf_name_local(leaf2, args->index2); |
2620 | namelen2 = name_loc->namelen; | 2620 | namelen2 = name_loc->namelen; |
2621 | name2 = (char *)name_loc->nameval; | 2621 | name2 = (char *)name_loc->nameval; |
2622 | } else { | 2622 | } else { |
2623 | name_rmt = XFS_ATTR_LEAF_NAME_REMOTE(leaf2, args->index2); | 2623 | name_rmt = xfs_attr_leaf_name_remote(leaf2, args->index2); |
2624 | namelen2 = name_rmt->namelen; | 2624 | namelen2 = name_rmt->namelen; |
2625 | name2 = (char *)name_rmt->name; | 2625 | name2 = (char *)name_rmt->name; |
2626 | } | 2626 | } |
@@ -2637,7 +2637,7 @@ xfs_attr_leaf_flipflags(xfs_da_args_t *args) | |||
2637 | XFS_DA_LOGRANGE(leaf1, entry1, sizeof(*entry1))); | 2637 | XFS_DA_LOGRANGE(leaf1, entry1, sizeof(*entry1))); |
2638 | if (args->rmtblkno) { | 2638 | if (args->rmtblkno) { |
2639 | ASSERT((entry1->flags & XFS_ATTR_LOCAL) == 0); | 2639 | ASSERT((entry1->flags & XFS_ATTR_LOCAL) == 0); |
2640 | name_rmt = XFS_ATTR_LEAF_NAME_REMOTE(leaf1, args->index); | 2640 | name_rmt = xfs_attr_leaf_name_remote(leaf1, args->index); |
2641 | name_rmt->valueblk = cpu_to_be32(args->rmtblkno); | 2641 | name_rmt->valueblk = cpu_to_be32(args->rmtblkno); |
2642 | name_rmt->valuelen = cpu_to_be32(args->valuelen); | 2642 | name_rmt->valuelen = cpu_to_be32(args->valuelen); |
2643 | xfs_da_log_buf(args->trans, bp1, | 2643 | xfs_da_log_buf(args->trans, bp1, |
@@ -2648,7 +2648,7 @@ xfs_attr_leaf_flipflags(xfs_da_args_t *args) | |||
2648 | xfs_da_log_buf(args->trans, bp2, | 2648 | xfs_da_log_buf(args->trans, bp2, |
2649 | XFS_DA_LOGRANGE(leaf2, entry2, sizeof(*entry2))); | 2649 | XFS_DA_LOGRANGE(leaf2, entry2, sizeof(*entry2))); |
2650 | if ((entry2->flags & XFS_ATTR_LOCAL) == 0) { | 2650 | if ((entry2->flags & XFS_ATTR_LOCAL) == 0) { |
2651 | name_rmt = XFS_ATTR_LEAF_NAME_REMOTE(leaf2, args->index2); | 2651 | name_rmt = xfs_attr_leaf_name_remote(leaf2, args->index2); |
2652 | name_rmt->valueblk = 0; | 2652 | name_rmt->valueblk = 0; |
2653 | name_rmt->valuelen = 0; | 2653 | name_rmt->valuelen = 0; |
2654 | xfs_da_log_buf(args->trans, bp2, | 2654 | xfs_da_log_buf(args->trans, bp2, |
@@ -2855,7 +2855,7 @@ xfs_attr_leaf_inactive(xfs_trans_t **trans, xfs_inode_t *dp, xfs_dabuf_t *bp) | |||
2855 | for (i = 0; i < be16_to_cpu(leaf->hdr.count); entry++, i++) { | 2855 | for (i = 0; i < be16_to_cpu(leaf->hdr.count); entry++, i++) { |
2856 | if (be16_to_cpu(entry->nameidx) && | 2856 | if (be16_to_cpu(entry->nameidx) && |
2857 | ((entry->flags & XFS_ATTR_LOCAL) == 0)) { | 2857 | ((entry->flags & XFS_ATTR_LOCAL) == 0)) { |
2858 | name_rmt = XFS_ATTR_LEAF_NAME_REMOTE(leaf, i); | 2858 | name_rmt = xfs_attr_leaf_name_remote(leaf, i); |
2859 | if (name_rmt->valueblk) | 2859 | if (name_rmt->valueblk) |
2860 | count++; | 2860 | count++; |
2861 | } | 2861 | } |
@@ -2883,7 +2883,7 @@ xfs_attr_leaf_inactive(xfs_trans_t **trans, xfs_inode_t *dp, xfs_dabuf_t *bp) | |||
2883 | for (i = 0; i < be16_to_cpu(leaf->hdr.count); entry++, i++) { | 2883 | for (i = 0; i < be16_to_cpu(leaf->hdr.count); entry++, i++) { |
2884 | if (be16_to_cpu(entry->nameidx) && | 2884 | if (be16_to_cpu(entry->nameidx) && |
2885 | ((entry->flags & XFS_ATTR_LOCAL) == 0)) { | 2885 | ((entry->flags & XFS_ATTR_LOCAL) == 0)) { |
2886 | name_rmt = XFS_ATTR_LEAF_NAME_REMOTE(leaf, i); | 2886 | name_rmt = xfs_attr_leaf_name_remote(leaf, i); |
2887 | if (name_rmt->valueblk) { | 2887 | if (name_rmt->valueblk) { |
2888 | lp->valueblk = be32_to_cpu(name_rmt->valueblk); | 2888 | lp->valueblk = be32_to_cpu(name_rmt->valueblk); |
2889 | lp->valuelen = XFS_B_TO_FSB(dp->i_mount, | 2889 | lp->valuelen = XFS_B_TO_FSB(dp->i_mount, |
diff --git a/fs/xfs/xfs_attr_leaf.h b/fs/xfs/xfs_attr_leaf.h index 83e9af417ca2..9c7d22fdcf4d 100644 --- a/fs/xfs/xfs_attr_leaf.h +++ b/fs/xfs/xfs_attr_leaf.h | |||
@@ -151,8 +151,6 @@ typedef struct xfs_attr_leafblock { | |||
151 | /* | 151 | /* |
152 | * Cast typed pointers for "local" and "remote" name/value structs. | 152 | * Cast typed pointers for "local" and "remote" name/value structs. |
153 | */ | 153 | */ |
154 | #define XFS_ATTR_LEAF_NAME_REMOTE(leafp,idx) \ | ||
155 | xfs_attr_leaf_name_remote(leafp,idx) | ||
156 | static inline xfs_attr_leaf_name_remote_t * | 154 | static inline xfs_attr_leaf_name_remote_t * |
157 | xfs_attr_leaf_name_remote(xfs_attr_leafblock_t *leafp, int idx) | 155 | xfs_attr_leaf_name_remote(xfs_attr_leafblock_t *leafp, int idx) |
158 | { | 156 | { |
@@ -160,8 +158,6 @@ xfs_attr_leaf_name_remote(xfs_attr_leafblock_t *leafp, int idx) | |||
160 | &((char *)leafp)[be16_to_cpu(leafp->entries[idx].nameidx)]; | 158 | &((char *)leafp)[be16_to_cpu(leafp->entries[idx].nameidx)]; |
161 | } | 159 | } |
162 | 160 | ||
163 | #define XFS_ATTR_LEAF_NAME_LOCAL(leafp,idx) \ | ||
164 | xfs_attr_leaf_name_local(leafp,idx) | ||
165 | static inline xfs_attr_leaf_name_local_t * | 161 | static inline xfs_attr_leaf_name_local_t * |
166 | xfs_attr_leaf_name_local(xfs_attr_leafblock_t *leafp, int idx) | 162 | xfs_attr_leaf_name_local(xfs_attr_leafblock_t *leafp, int idx) |
167 | { | 163 | { |
@@ -169,8 +165,6 @@ xfs_attr_leaf_name_local(xfs_attr_leafblock_t *leafp, int idx) | |||
169 | &((char *)leafp)[be16_to_cpu(leafp->entries[idx].nameidx)]; | 165 | &((char *)leafp)[be16_to_cpu(leafp->entries[idx].nameidx)]; |
170 | } | 166 | } |
171 | 167 | ||
172 | #define XFS_ATTR_LEAF_NAME(leafp,idx) \ | ||
173 | xfs_attr_leaf_name(leafp,idx) | ||
174 | static inline char *xfs_attr_leaf_name(xfs_attr_leafblock_t *leafp, int idx) | 168 | static inline char *xfs_attr_leaf_name(xfs_attr_leafblock_t *leafp, int idx) |
175 | { | 169 | { |
176 | return &((char *)leafp)[be16_to_cpu(leafp->entries[idx].nameidx)]; | 170 | return &((char *)leafp)[be16_to_cpu(leafp->entries[idx].nameidx)]; |
@@ -181,24 +175,18 @@ static inline char *xfs_attr_leaf_name(xfs_attr_leafblock_t *leafp, int idx) | |||
181 | * a "local" name/value structure, a "remote" name/value structure, and | 175 | * a "local" name/value structure, a "remote" name/value structure, and |
182 | * a pointer which might be either. | 176 | * a pointer which might be either. |
183 | */ | 177 | */ |
184 | #define XFS_ATTR_LEAF_ENTSIZE_REMOTE(nlen) \ | ||
185 | xfs_attr_leaf_entsize_remote(nlen) | ||
186 | static inline int xfs_attr_leaf_entsize_remote(int nlen) | 178 | static inline int xfs_attr_leaf_entsize_remote(int nlen) |
187 | { | 179 | { |
188 | return ((uint)sizeof(xfs_attr_leaf_name_remote_t) - 1 + (nlen) + \ | 180 | return ((uint)sizeof(xfs_attr_leaf_name_remote_t) - 1 + (nlen) + \ |
189 | XFS_ATTR_LEAF_NAME_ALIGN - 1) & ~(XFS_ATTR_LEAF_NAME_ALIGN - 1); | 181 | XFS_ATTR_LEAF_NAME_ALIGN - 1) & ~(XFS_ATTR_LEAF_NAME_ALIGN - 1); |
190 | } | 182 | } |
191 | 183 | ||
192 | #define XFS_ATTR_LEAF_ENTSIZE_LOCAL(nlen,vlen) \ | ||
193 | xfs_attr_leaf_entsize_local(nlen,vlen) | ||
194 | static inline int xfs_attr_leaf_entsize_local(int nlen, int vlen) | 184 | static inline int xfs_attr_leaf_entsize_local(int nlen, int vlen) |
195 | { | 185 | { |
196 | return ((uint)sizeof(xfs_attr_leaf_name_local_t) - 1 + (nlen) + (vlen) + | 186 | return ((uint)sizeof(xfs_attr_leaf_name_local_t) - 1 + (nlen) + (vlen) + |
197 | XFS_ATTR_LEAF_NAME_ALIGN - 1) & ~(XFS_ATTR_LEAF_NAME_ALIGN - 1); | 187 | XFS_ATTR_LEAF_NAME_ALIGN - 1) & ~(XFS_ATTR_LEAF_NAME_ALIGN - 1); |
198 | } | 188 | } |
199 | 189 | ||
200 | #define XFS_ATTR_LEAF_ENTSIZE_LOCAL_MAX(bsize) \ | ||
201 | xfs_attr_leaf_entsize_local_max(bsize) | ||
202 | static inline int xfs_attr_leaf_entsize_local_max(int bsize) | 190 | static inline int xfs_attr_leaf_entsize_local_max(int bsize) |
203 | { | 191 | { |
204 | return (((bsize) >> 1) + ((bsize) >> 2)); | 192 | return (((bsize) >> 1) + ((bsize) >> 2)); |
diff --git a/fs/xfs/xfs_bit.h b/fs/xfs/xfs_bit.h index bca7b243c319..f1e3c907044d 100644 --- a/fs/xfs/xfs_bit.h +++ b/fs/xfs/xfs_bit.h | |||
@@ -23,24 +23,16 @@ | |||
23 | */ | 23 | */ |
24 | 24 | ||
25 | /* | 25 | /* |
26 | * masks with n high/low bits set, 32-bit values & 64-bit values | 26 | * masks with n high/low bits set, 64-bit values |
27 | */ | 27 | */ |
28 | #define XFS_MASK32HI(n) xfs_mask32hi(n) | ||
29 | static inline __uint32_t xfs_mask32hi(int n) | ||
30 | { | ||
31 | return (__uint32_t)-1 << (32 - (n)); | ||
32 | } | ||
33 | #define XFS_MASK64HI(n) xfs_mask64hi(n) | ||
34 | static inline __uint64_t xfs_mask64hi(int n) | 28 | static inline __uint64_t xfs_mask64hi(int n) |
35 | { | 29 | { |
36 | return (__uint64_t)-1 << (64 - (n)); | 30 | return (__uint64_t)-1 << (64 - (n)); |
37 | } | 31 | } |
38 | #define XFS_MASK32LO(n) xfs_mask32lo(n) | ||
39 | static inline __uint32_t xfs_mask32lo(int n) | 32 | static inline __uint32_t xfs_mask32lo(int n) |
40 | { | 33 | { |
41 | return ((__uint32_t)1 << (n)) - 1; | 34 | return ((__uint32_t)1 << (n)) - 1; |
42 | } | 35 | } |
43 | #define XFS_MASK64LO(n) xfs_mask64lo(n) | ||
44 | static inline __uint64_t xfs_mask64lo(int n) | 36 | static inline __uint64_t xfs_mask64lo(int n) |
45 | { | 37 | { |
46 | return ((__uint64_t)1 << (n)) - 1; | 38 | return ((__uint64_t)1 << (n)) - 1; |
diff --git a/fs/xfs/xfs_bmap.c b/fs/xfs/xfs_bmap.c index 138308e70d14..c852cd65aaea 100644 --- a/fs/xfs/xfs_bmap.c +++ b/fs/xfs/xfs_bmap.c | |||
@@ -595,9 +595,9 @@ xfs_bmap_add_extent( | |||
595 | xfs_iext_insert(ifp, 0, 1, new); | 595 | xfs_iext_insert(ifp, 0, 1, new); |
596 | ASSERT(cur == NULL); | 596 | ASSERT(cur == NULL); |
597 | ifp->if_lastex = 0; | 597 | ifp->if_lastex = 0; |
598 | if (!ISNULLSTARTBLOCK(new->br_startblock)) { | 598 | if (!isnullstartblock(new->br_startblock)) { |
599 | XFS_IFORK_NEXT_SET(ip, whichfork, 1); | 599 | XFS_IFORK_NEXT_SET(ip, whichfork, 1); |
600 | logflags = XFS_ILOG_CORE | XFS_ILOG_FEXT(whichfork); | 600 | logflags = XFS_ILOG_CORE | xfs_ilog_fext(whichfork); |
601 | } else | 601 | } else |
602 | logflags = 0; | 602 | logflags = 0; |
603 | /* DELTA: single new extent */ | 603 | /* DELTA: single new extent */ |
@@ -613,7 +613,7 @@ xfs_bmap_add_extent( | |||
613 | /* | 613 | /* |
614 | * Any kind of new delayed allocation goes here. | 614 | * Any kind of new delayed allocation goes here. |
615 | */ | 615 | */ |
616 | else if (ISNULLSTARTBLOCK(new->br_startblock)) { | 616 | else if (isnullstartblock(new->br_startblock)) { |
617 | if (cur) | 617 | if (cur) |
618 | ASSERT((cur->bc_private.b.flags & | 618 | ASSERT((cur->bc_private.b.flags & |
619 | XFS_BTCUR_BPRV_WASDEL) == 0); | 619 | XFS_BTCUR_BPRV_WASDEL) == 0); |
@@ -644,11 +644,11 @@ xfs_bmap_add_extent( | |||
644 | * in a delayed or unwritten allocation with a real one, or | 644 | * in a delayed or unwritten allocation with a real one, or |
645 | * converting real back to unwritten. | 645 | * converting real back to unwritten. |
646 | */ | 646 | */ |
647 | if (!ISNULLSTARTBLOCK(new->br_startblock) && | 647 | if (!isnullstartblock(new->br_startblock) && |
648 | new->br_startoff + new->br_blockcount > prev.br_startoff) { | 648 | new->br_startoff + new->br_blockcount > prev.br_startoff) { |
649 | if (prev.br_state != XFS_EXT_UNWRITTEN && | 649 | if (prev.br_state != XFS_EXT_UNWRITTEN && |
650 | ISNULLSTARTBLOCK(prev.br_startblock)) { | 650 | isnullstartblock(prev.br_startblock)) { |
651 | da_old = STARTBLOCKVAL(prev.br_startblock); | 651 | da_old = startblockval(prev.br_startblock); |
652 | if (cur) | 652 | if (cur) |
653 | ASSERT(cur->bc_private.b.flags & | 653 | ASSERT(cur->bc_private.b.flags & |
654 | XFS_BTCUR_BPRV_WASDEL); | 654 | XFS_BTCUR_BPRV_WASDEL); |
@@ -803,7 +803,7 @@ xfs_bmap_add_extent_delay_real( | |||
803 | */ | 803 | */ |
804 | if (STATE_SET_TEST(LEFT_VALID, idx > 0)) { | 804 | if (STATE_SET_TEST(LEFT_VALID, idx > 0)) { |
805 | xfs_bmbt_get_all(xfs_iext_get_ext(ifp, idx - 1), &LEFT); | 805 | xfs_bmbt_get_all(xfs_iext_get_ext(ifp, idx - 1), &LEFT); |
806 | STATE_SET(LEFT_DELAY, ISNULLSTARTBLOCK(LEFT.br_startblock)); | 806 | STATE_SET(LEFT_DELAY, isnullstartblock(LEFT.br_startblock)); |
807 | } | 807 | } |
808 | STATE_SET(LEFT_CONTIG, | 808 | STATE_SET(LEFT_CONTIG, |
809 | STATE_TEST(LEFT_VALID) && !STATE_TEST(LEFT_DELAY) && | 809 | STATE_TEST(LEFT_VALID) && !STATE_TEST(LEFT_DELAY) && |
@@ -820,7 +820,7 @@ xfs_bmap_add_extent_delay_real( | |||
820 | idx < | 820 | idx < |
821 | ip->i_df.if_bytes / (uint)sizeof(xfs_bmbt_rec_t) - 1)) { | 821 | ip->i_df.if_bytes / (uint)sizeof(xfs_bmbt_rec_t) - 1)) { |
822 | xfs_bmbt_get_all(xfs_iext_get_ext(ifp, idx + 1), &RIGHT); | 822 | xfs_bmbt_get_all(xfs_iext_get_ext(ifp, idx + 1), &RIGHT); |
823 | STATE_SET(RIGHT_DELAY, ISNULLSTARTBLOCK(RIGHT.br_startblock)); | 823 | STATE_SET(RIGHT_DELAY, isnullstartblock(RIGHT.br_startblock)); |
824 | } | 824 | } |
825 | STATE_SET(RIGHT_CONTIG, | 825 | STATE_SET(RIGHT_CONTIG, |
826 | STATE_TEST(RIGHT_VALID) && !STATE_TEST(RIGHT_DELAY) && | 826 | STATE_TEST(RIGHT_VALID) && !STATE_TEST(RIGHT_DELAY) && |
@@ -1019,8 +1019,8 @@ xfs_bmap_add_extent_delay_real( | |||
1019 | goto done; | 1019 | goto done; |
1020 | } | 1020 | } |
1021 | temp = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), | 1021 | temp = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), |
1022 | STARTBLOCKVAL(PREV.br_startblock)); | 1022 | startblockval(PREV.br_startblock)); |
1023 | xfs_bmbt_set_startblock(ep, NULLSTARTBLOCK((int)temp)); | 1023 | xfs_bmbt_set_startblock(ep, nullstartblock((int)temp)); |
1024 | XFS_BMAP_TRACE_POST_UPDATE("LF|LC", ip, idx, XFS_DATA_FORK); | 1024 | XFS_BMAP_TRACE_POST_UPDATE("LF|LC", ip, idx, XFS_DATA_FORK); |
1025 | *dnew = temp; | 1025 | *dnew = temp; |
1026 | /* DELTA: The boundary between two in-core extents moved. */ | 1026 | /* DELTA: The boundary between two in-core extents moved. */ |
@@ -1067,10 +1067,10 @@ xfs_bmap_add_extent_delay_real( | |||
1067 | goto done; | 1067 | goto done; |
1068 | } | 1068 | } |
1069 | temp = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), | 1069 | temp = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), |
1070 | STARTBLOCKVAL(PREV.br_startblock) - | 1070 | startblockval(PREV.br_startblock) - |
1071 | (cur ? cur->bc_private.b.allocated : 0)); | 1071 | (cur ? cur->bc_private.b.allocated : 0)); |
1072 | ep = xfs_iext_get_ext(ifp, idx + 1); | 1072 | ep = xfs_iext_get_ext(ifp, idx + 1); |
1073 | xfs_bmbt_set_startblock(ep, NULLSTARTBLOCK((int)temp)); | 1073 | xfs_bmbt_set_startblock(ep, nullstartblock((int)temp)); |
1074 | XFS_BMAP_TRACE_POST_UPDATE("LF", ip, idx + 1, XFS_DATA_FORK); | 1074 | XFS_BMAP_TRACE_POST_UPDATE("LF", ip, idx + 1, XFS_DATA_FORK); |
1075 | *dnew = temp; | 1075 | *dnew = temp; |
1076 | /* DELTA: One in-core extent is split in two. */ | 1076 | /* DELTA: One in-core extent is split in two. */ |
@@ -1110,8 +1110,8 @@ xfs_bmap_add_extent_delay_real( | |||
1110 | goto done; | 1110 | goto done; |
1111 | } | 1111 | } |
1112 | temp = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), | 1112 | temp = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), |
1113 | STARTBLOCKVAL(PREV.br_startblock)); | 1113 | startblockval(PREV.br_startblock)); |
1114 | xfs_bmbt_set_startblock(ep, NULLSTARTBLOCK((int)temp)); | 1114 | xfs_bmbt_set_startblock(ep, nullstartblock((int)temp)); |
1115 | XFS_BMAP_TRACE_POST_UPDATE("RF|RC", ip, idx, XFS_DATA_FORK); | 1115 | XFS_BMAP_TRACE_POST_UPDATE("RF|RC", ip, idx, XFS_DATA_FORK); |
1116 | *dnew = temp; | 1116 | *dnew = temp; |
1117 | /* DELTA: The boundary between two in-core extents moved. */ | 1117 | /* DELTA: The boundary between two in-core extents moved. */ |
@@ -1157,10 +1157,10 @@ xfs_bmap_add_extent_delay_real( | |||
1157 | goto done; | 1157 | goto done; |
1158 | } | 1158 | } |
1159 | temp = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), | 1159 | temp = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), |
1160 | STARTBLOCKVAL(PREV.br_startblock) - | 1160 | startblockval(PREV.br_startblock) - |
1161 | (cur ? cur->bc_private.b.allocated : 0)); | 1161 | (cur ? cur->bc_private.b.allocated : 0)); |
1162 | ep = xfs_iext_get_ext(ifp, idx); | 1162 | ep = xfs_iext_get_ext(ifp, idx); |
1163 | xfs_bmbt_set_startblock(ep, NULLSTARTBLOCK((int)temp)); | 1163 | xfs_bmbt_set_startblock(ep, nullstartblock((int)temp)); |
1164 | XFS_BMAP_TRACE_POST_UPDATE("RF", ip, idx, XFS_DATA_FORK); | 1164 | XFS_BMAP_TRACE_POST_UPDATE("RF", ip, idx, XFS_DATA_FORK); |
1165 | *dnew = temp; | 1165 | *dnew = temp; |
1166 | /* DELTA: One in-core extent is split in two. */ | 1166 | /* DELTA: One in-core extent is split in two. */ |
@@ -1213,7 +1213,7 @@ xfs_bmap_add_extent_delay_real( | |||
1213 | } | 1213 | } |
1214 | temp = xfs_bmap_worst_indlen(ip, temp); | 1214 | temp = xfs_bmap_worst_indlen(ip, temp); |
1215 | temp2 = xfs_bmap_worst_indlen(ip, temp2); | 1215 | temp2 = xfs_bmap_worst_indlen(ip, temp2); |
1216 | diff = (int)(temp + temp2 - STARTBLOCKVAL(PREV.br_startblock) - | 1216 | diff = (int)(temp + temp2 - startblockval(PREV.br_startblock) - |
1217 | (cur ? cur->bc_private.b.allocated : 0)); | 1217 | (cur ? cur->bc_private.b.allocated : 0)); |
1218 | if (diff > 0 && | 1218 | if (diff > 0 && |
1219 | xfs_mod_incore_sb(ip->i_mount, XFS_SBS_FDBLOCKS, -((int64_t)diff), rsvd)) { | 1219 | xfs_mod_incore_sb(ip->i_mount, XFS_SBS_FDBLOCKS, -((int64_t)diff), rsvd)) { |
@@ -1241,11 +1241,11 @@ xfs_bmap_add_extent_delay_real( | |||
1241 | } | 1241 | } |
1242 | } | 1242 | } |
1243 | ep = xfs_iext_get_ext(ifp, idx); | 1243 | ep = xfs_iext_get_ext(ifp, idx); |
1244 | xfs_bmbt_set_startblock(ep, NULLSTARTBLOCK((int)temp)); | 1244 | xfs_bmbt_set_startblock(ep, nullstartblock((int)temp)); |
1245 | XFS_BMAP_TRACE_POST_UPDATE("0", ip, idx, XFS_DATA_FORK); | 1245 | XFS_BMAP_TRACE_POST_UPDATE("0", ip, idx, XFS_DATA_FORK); |
1246 | XFS_BMAP_TRACE_PRE_UPDATE("0", ip, idx + 2, XFS_DATA_FORK); | 1246 | XFS_BMAP_TRACE_PRE_UPDATE("0", ip, idx + 2, XFS_DATA_FORK); |
1247 | xfs_bmbt_set_startblock(xfs_iext_get_ext(ifp, idx + 2), | 1247 | xfs_bmbt_set_startblock(xfs_iext_get_ext(ifp, idx + 2), |
1248 | NULLSTARTBLOCK((int)temp2)); | 1248 | nullstartblock((int)temp2)); |
1249 | XFS_BMAP_TRACE_POST_UPDATE("0", ip, idx + 2, XFS_DATA_FORK); | 1249 | XFS_BMAP_TRACE_POST_UPDATE("0", ip, idx + 2, XFS_DATA_FORK); |
1250 | *dnew = temp + temp2; | 1250 | *dnew = temp + temp2; |
1251 | /* DELTA: One in-core extent is split in three. */ | 1251 | /* DELTA: One in-core extent is split in three. */ |
@@ -1365,7 +1365,7 @@ xfs_bmap_add_extent_unwritten_real( | |||
1365 | */ | 1365 | */ |
1366 | if (STATE_SET_TEST(LEFT_VALID, idx > 0)) { | 1366 | if (STATE_SET_TEST(LEFT_VALID, idx > 0)) { |
1367 | xfs_bmbt_get_all(xfs_iext_get_ext(ifp, idx - 1), &LEFT); | 1367 | xfs_bmbt_get_all(xfs_iext_get_ext(ifp, idx - 1), &LEFT); |
1368 | STATE_SET(LEFT_DELAY, ISNULLSTARTBLOCK(LEFT.br_startblock)); | 1368 | STATE_SET(LEFT_DELAY, isnullstartblock(LEFT.br_startblock)); |
1369 | } | 1369 | } |
1370 | STATE_SET(LEFT_CONTIG, | 1370 | STATE_SET(LEFT_CONTIG, |
1371 | STATE_TEST(LEFT_VALID) && !STATE_TEST(LEFT_DELAY) && | 1371 | STATE_TEST(LEFT_VALID) && !STATE_TEST(LEFT_DELAY) && |
@@ -1382,7 +1382,7 @@ xfs_bmap_add_extent_unwritten_real( | |||
1382 | idx < | 1382 | idx < |
1383 | ip->i_df.if_bytes / (uint)sizeof(xfs_bmbt_rec_t) - 1)) { | 1383 | ip->i_df.if_bytes / (uint)sizeof(xfs_bmbt_rec_t) - 1)) { |
1384 | xfs_bmbt_get_all(xfs_iext_get_ext(ifp, idx + 1), &RIGHT); | 1384 | xfs_bmbt_get_all(xfs_iext_get_ext(ifp, idx + 1), &RIGHT); |
1385 | STATE_SET(RIGHT_DELAY, ISNULLSTARTBLOCK(RIGHT.br_startblock)); | 1385 | STATE_SET(RIGHT_DELAY, isnullstartblock(RIGHT.br_startblock)); |
1386 | } | 1386 | } |
1387 | STATE_SET(RIGHT_CONTIG, | 1387 | STATE_SET(RIGHT_CONTIG, |
1388 | STATE_TEST(RIGHT_VALID) && !STATE_TEST(RIGHT_DELAY) && | 1388 | STATE_TEST(RIGHT_VALID) && !STATE_TEST(RIGHT_DELAY) && |
@@ -1889,13 +1889,13 @@ xfs_bmap_add_extent_hole_delay( | |||
1889 | ifp = XFS_IFORK_PTR(ip, XFS_DATA_FORK); | 1889 | ifp = XFS_IFORK_PTR(ip, XFS_DATA_FORK); |
1890 | ep = xfs_iext_get_ext(ifp, idx); | 1890 | ep = xfs_iext_get_ext(ifp, idx); |
1891 | state = 0; | 1891 | state = 0; |
1892 | ASSERT(ISNULLSTARTBLOCK(new->br_startblock)); | 1892 | ASSERT(isnullstartblock(new->br_startblock)); |
1893 | /* | 1893 | /* |
1894 | * Check and set flags if this segment has a left neighbor | 1894 | * Check and set flags if this segment has a left neighbor |
1895 | */ | 1895 | */ |
1896 | if (STATE_SET_TEST(LEFT_VALID, idx > 0)) { | 1896 | if (STATE_SET_TEST(LEFT_VALID, idx > 0)) { |
1897 | xfs_bmbt_get_all(xfs_iext_get_ext(ifp, idx - 1), &left); | 1897 | xfs_bmbt_get_all(xfs_iext_get_ext(ifp, idx - 1), &left); |
1898 | STATE_SET(LEFT_DELAY, ISNULLSTARTBLOCK(left.br_startblock)); | 1898 | STATE_SET(LEFT_DELAY, isnullstartblock(left.br_startblock)); |
1899 | } | 1899 | } |
1900 | /* | 1900 | /* |
1901 | * Check and set flags if the current (right) segment exists. | 1901 | * Check and set flags if the current (right) segment exists. |
@@ -1905,7 +1905,7 @@ xfs_bmap_add_extent_hole_delay( | |||
1905 | idx < | 1905 | idx < |
1906 | ip->i_df.if_bytes / (uint)sizeof(xfs_bmbt_rec_t))) { | 1906 | ip->i_df.if_bytes / (uint)sizeof(xfs_bmbt_rec_t))) { |
1907 | xfs_bmbt_get_all(ep, &right); | 1907 | xfs_bmbt_get_all(ep, &right); |
1908 | STATE_SET(RIGHT_DELAY, ISNULLSTARTBLOCK(right.br_startblock)); | 1908 | STATE_SET(RIGHT_DELAY, isnullstartblock(right.br_startblock)); |
1909 | } | 1909 | } |
1910 | /* | 1910 | /* |
1911 | * Set contiguity flags on the left and right neighbors. | 1911 | * Set contiguity flags on the left and right neighbors. |
@@ -1938,12 +1938,12 @@ xfs_bmap_add_extent_hole_delay( | |||
1938 | XFS_BMAP_TRACE_PRE_UPDATE("LC|RC", ip, idx - 1, | 1938 | XFS_BMAP_TRACE_PRE_UPDATE("LC|RC", ip, idx - 1, |
1939 | XFS_DATA_FORK); | 1939 | XFS_DATA_FORK); |
1940 | xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, idx - 1), temp); | 1940 | xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, idx - 1), temp); |
1941 | oldlen = STARTBLOCKVAL(left.br_startblock) + | 1941 | oldlen = startblockval(left.br_startblock) + |
1942 | STARTBLOCKVAL(new->br_startblock) + | 1942 | startblockval(new->br_startblock) + |
1943 | STARTBLOCKVAL(right.br_startblock); | 1943 | startblockval(right.br_startblock); |
1944 | newlen = xfs_bmap_worst_indlen(ip, temp); | 1944 | newlen = xfs_bmap_worst_indlen(ip, temp); |
1945 | xfs_bmbt_set_startblock(xfs_iext_get_ext(ifp, idx - 1), | 1945 | xfs_bmbt_set_startblock(xfs_iext_get_ext(ifp, idx - 1), |
1946 | NULLSTARTBLOCK((int)newlen)); | 1946 | nullstartblock((int)newlen)); |
1947 | XFS_BMAP_TRACE_POST_UPDATE("LC|RC", ip, idx - 1, | 1947 | XFS_BMAP_TRACE_POST_UPDATE("LC|RC", ip, idx - 1, |
1948 | XFS_DATA_FORK); | 1948 | XFS_DATA_FORK); |
1949 | XFS_BMAP_TRACE_DELETE("LC|RC", ip, idx, 1, XFS_DATA_FORK); | 1949 | XFS_BMAP_TRACE_DELETE("LC|RC", ip, idx, 1, XFS_DATA_FORK); |
@@ -1964,11 +1964,11 @@ xfs_bmap_add_extent_hole_delay( | |||
1964 | XFS_BMAP_TRACE_PRE_UPDATE("LC", ip, idx - 1, | 1964 | XFS_BMAP_TRACE_PRE_UPDATE("LC", ip, idx - 1, |
1965 | XFS_DATA_FORK); | 1965 | XFS_DATA_FORK); |
1966 | xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, idx - 1), temp); | 1966 | xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, idx - 1), temp); |
1967 | oldlen = STARTBLOCKVAL(left.br_startblock) + | 1967 | oldlen = startblockval(left.br_startblock) + |
1968 | STARTBLOCKVAL(new->br_startblock); | 1968 | startblockval(new->br_startblock); |
1969 | newlen = xfs_bmap_worst_indlen(ip, temp); | 1969 | newlen = xfs_bmap_worst_indlen(ip, temp); |
1970 | xfs_bmbt_set_startblock(xfs_iext_get_ext(ifp, idx - 1), | 1970 | xfs_bmbt_set_startblock(xfs_iext_get_ext(ifp, idx - 1), |
1971 | NULLSTARTBLOCK((int)newlen)); | 1971 | nullstartblock((int)newlen)); |
1972 | XFS_BMAP_TRACE_POST_UPDATE("LC", ip, idx - 1, | 1972 | XFS_BMAP_TRACE_POST_UPDATE("LC", ip, idx - 1, |
1973 | XFS_DATA_FORK); | 1973 | XFS_DATA_FORK); |
1974 | ip->i_df.if_lastex = idx - 1; | 1974 | ip->i_df.if_lastex = idx - 1; |
@@ -1985,11 +1985,11 @@ xfs_bmap_add_extent_hole_delay( | |||
1985 | */ | 1985 | */ |
1986 | XFS_BMAP_TRACE_PRE_UPDATE("RC", ip, idx, XFS_DATA_FORK); | 1986 | XFS_BMAP_TRACE_PRE_UPDATE("RC", ip, idx, XFS_DATA_FORK); |
1987 | temp = new->br_blockcount + right.br_blockcount; | 1987 | temp = new->br_blockcount + right.br_blockcount; |
1988 | oldlen = STARTBLOCKVAL(new->br_startblock) + | 1988 | oldlen = startblockval(new->br_startblock) + |
1989 | STARTBLOCKVAL(right.br_startblock); | 1989 | startblockval(right.br_startblock); |
1990 | newlen = xfs_bmap_worst_indlen(ip, temp); | 1990 | newlen = xfs_bmap_worst_indlen(ip, temp); |
1991 | xfs_bmbt_set_allf(ep, new->br_startoff, | 1991 | xfs_bmbt_set_allf(ep, new->br_startoff, |
1992 | NULLSTARTBLOCK((int)newlen), temp, right.br_state); | 1992 | nullstartblock((int)newlen), temp, right.br_state); |
1993 | XFS_BMAP_TRACE_POST_UPDATE("RC", ip, idx, XFS_DATA_FORK); | 1993 | XFS_BMAP_TRACE_POST_UPDATE("RC", ip, idx, XFS_DATA_FORK); |
1994 | ip->i_df.if_lastex = idx; | 1994 | ip->i_df.if_lastex = idx; |
1995 | /* DELTA: One in-core extent grew into a hole. */ | 1995 | /* DELTA: One in-core extent grew into a hole. */ |
@@ -2085,7 +2085,7 @@ xfs_bmap_add_extent_hole_real( | |||
2085 | */ | 2085 | */ |
2086 | if (STATE_SET_TEST(LEFT_VALID, idx > 0)) { | 2086 | if (STATE_SET_TEST(LEFT_VALID, idx > 0)) { |
2087 | xfs_bmbt_get_all(xfs_iext_get_ext(ifp, idx - 1), &left); | 2087 | xfs_bmbt_get_all(xfs_iext_get_ext(ifp, idx - 1), &left); |
2088 | STATE_SET(LEFT_DELAY, ISNULLSTARTBLOCK(left.br_startblock)); | 2088 | STATE_SET(LEFT_DELAY, isnullstartblock(left.br_startblock)); |
2089 | } | 2089 | } |
2090 | /* | 2090 | /* |
2091 | * Check and set flags if this segment has a current value. | 2091 | * Check and set flags if this segment has a current value. |
@@ -2095,7 +2095,7 @@ xfs_bmap_add_extent_hole_real( | |||
2095 | idx < | 2095 | idx < |
2096 | ifp->if_bytes / (uint)sizeof(xfs_bmbt_rec_t))) { | 2096 | ifp->if_bytes / (uint)sizeof(xfs_bmbt_rec_t))) { |
2097 | xfs_bmbt_get_all(ep, &right); | 2097 | xfs_bmbt_get_all(ep, &right); |
2098 | STATE_SET(RIGHT_DELAY, ISNULLSTARTBLOCK(right.br_startblock)); | 2098 | STATE_SET(RIGHT_DELAY, isnullstartblock(right.br_startblock)); |
2099 | } | 2099 | } |
2100 | /* | 2100 | /* |
2101 | * We're inserting a real allocation between "left" and "right". | 2101 | * We're inserting a real allocation between "left" and "right". |
@@ -2143,7 +2143,7 @@ xfs_bmap_add_extent_hole_real( | |||
2143 | XFS_IFORK_NEXT_SET(ip, whichfork, | 2143 | XFS_IFORK_NEXT_SET(ip, whichfork, |
2144 | XFS_IFORK_NEXTENTS(ip, whichfork) - 1); | 2144 | XFS_IFORK_NEXTENTS(ip, whichfork) - 1); |
2145 | if (cur == NULL) { | 2145 | if (cur == NULL) { |
2146 | rval = XFS_ILOG_CORE | XFS_ILOG_FEXT(whichfork); | 2146 | rval = XFS_ILOG_CORE | xfs_ilog_fext(whichfork); |
2147 | } else { | 2147 | } else { |
2148 | rval = XFS_ILOG_CORE; | 2148 | rval = XFS_ILOG_CORE; |
2149 | if ((error = xfs_bmbt_lookup_eq(cur, | 2149 | if ((error = xfs_bmbt_lookup_eq(cur, |
@@ -2185,7 +2185,7 @@ xfs_bmap_add_extent_hole_real( | |||
2185 | XFS_BMAP_TRACE_POST_UPDATE("LC", ip, idx - 1, whichfork); | 2185 | XFS_BMAP_TRACE_POST_UPDATE("LC", ip, idx - 1, whichfork); |
2186 | ifp->if_lastex = idx - 1; | 2186 | ifp->if_lastex = idx - 1; |
2187 | if (cur == NULL) { | 2187 | if (cur == NULL) { |
2188 | rval = XFS_ILOG_FEXT(whichfork); | 2188 | rval = xfs_ilog_fext(whichfork); |
2189 | } else { | 2189 | } else { |
2190 | rval = 0; | 2190 | rval = 0; |
2191 | if ((error = xfs_bmbt_lookup_eq(cur, | 2191 | if ((error = xfs_bmbt_lookup_eq(cur, |
@@ -2220,7 +2220,7 @@ xfs_bmap_add_extent_hole_real( | |||
2220 | XFS_BMAP_TRACE_POST_UPDATE("RC", ip, idx, whichfork); | 2220 | XFS_BMAP_TRACE_POST_UPDATE("RC", ip, idx, whichfork); |
2221 | ifp->if_lastex = idx; | 2221 | ifp->if_lastex = idx; |
2222 | if (cur == NULL) { | 2222 | if (cur == NULL) { |
2223 | rval = XFS_ILOG_FEXT(whichfork); | 2223 | rval = xfs_ilog_fext(whichfork); |
2224 | } else { | 2224 | } else { |
2225 | rval = 0; | 2225 | rval = 0; |
2226 | if ((error = xfs_bmbt_lookup_eq(cur, | 2226 | if ((error = xfs_bmbt_lookup_eq(cur, |
@@ -2254,7 +2254,7 @@ xfs_bmap_add_extent_hole_real( | |||
2254 | XFS_IFORK_NEXT_SET(ip, whichfork, | 2254 | XFS_IFORK_NEXT_SET(ip, whichfork, |
2255 | XFS_IFORK_NEXTENTS(ip, whichfork) + 1); | 2255 | XFS_IFORK_NEXTENTS(ip, whichfork) + 1); |
2256 | if (cur == NULL) { | 2256 | if (cur == NULL) { |
2257 | rval = XFS_ILOG_CORE | XFS_ILOG_FEXT(whichfork); | 2257 | rval = XFS_ILOG_CORE | xfs_ilog_fext(whichfork); |
2258 | } else { | 2258 | } else { |
2259 | rval = XFS_ILOG_CORE; | 2259 | rval = XFS_ILOG_CORE; |
2260 | if ((error = xfs_bmbt_lookup_eq(cur, | 2260 | if ((error = xfs_bmbt_lookup_eq(cur, |
@@ -2482,7 +2482,7 @@ xfs_bmap_adjacent( | |||
2482 | * try to use it's last block as our starting point. | 2482 | * try to use it's last block as our starting point. |
2483 | */ | 2483 | */ |
2484 | if (ap->eof && ap->prevp->br_startoff != NULLFILEOFF && | 2484 | if (ap->eof && ap->prevp->br_startoff != NULLFILEOFF && |
2485 | !ISNULLSTARTBLOCK(ap->prevp->br_startblock) && | 2485 | !isnullstartblock(ap->prevp->br_startblock) && |
2486 | ISVALID(ap->prevp->br_startblock + ap->prevp->br_blockcount, | 2486 | ISVALID(ap->prevp->br_startblock + ap->prevp->br_blockcount, |
2487 | ap->prevp->br_startblock)) { | 2487 | ap->prevp->br_startblock)) { |
2488 | ap->rval = ap->prevp->br_startblock + ap->prevp->br_blockcount; | 2488 | ap->rval = ap->prevp->br_startblock + ap->prevp->br_blockcount; |
@@ -2511,7 +2511,7 @@ xfs_bmap_adjacent( | |||
2511 | * start block based on it. | 2511 | * start block based on it. |
2512 | */ | 2512 | */ |
2513 | if (ap->prevp->br_startoff != NULLFILEOFF && | 2513 | if (ap->prevp->br_startoff != NULLFILEOFF && |
2514 | !ISNULLSTARTBLOCK(ap->prevp->br_startblock) && | 2514 | !isnullstartblock(ap->prevp->br_startblock) && |
2515 | (prevbno = ap->prevp->br_startblock + | 2515 | (prevbno = ap->prevp->br_startblock + |
2516 | ap->prevp->br_blockcount) && | 2516 | ap->prevp->br_blockcount) && |
2517 | ISVALID(prevbno, ap->prevp->br_startblock)) { | 2517 | ISVALID(prevbno, ap->prevp->br_startblock)) { |
@@ -2552,7 +2552,7 @@ xfs_bmap_adjacent( | |||
2552 | * If there's a following (right) block, select a requested | 2552 | * If there's a following (right) block, select a requested |
2553 | * start block based on it. | 2553 | * start block based on it. |
2554 | */ | 2554 | */ |
2555 | if (!ISNULLSTARTBLOCK(ap->gotp->br_startblock)) { | 2555 | if (!isnullstartblock(ap->gotp->br_startblock)) { |
2556 | /* | 2556 | /* |
2557 | * Calculate gap to start of next block. | 2557 | * Calculate gap to start of next block. |
2558 | */ | 2558 | */ |
@@ -3082,7 +3082,7 @@ xfs_bmap_btree_to_extents( | |||
3082 | ASSERT(ifp->if_broot == NULL); | 3082 | ASSERT(ifp->if_broot == NULL); |
3083 | ASSERT((ifp->if_flags & XFS_IFBROOT) == 0); | 3083 | ASSERT((ifp->if_flags & XFS_IFBROOT) == 0); |
3084 | XFS_IFORK_FMT_SET(ip, whichfork, XFS_DINODE_FMT_EXTENTS); | 3084 | XFS_IFORK_FMT_SET(ip, whichfork, XFS_DINODE_FMT_EXTENTS); |
3085 | *logflagsp = XFS_ILOG_CORE | XFS_ILOG_FEXT(whichfork); | 3085 | *logflagsp = XFS_ILOG_CORE | xfs_ilog_fext(whichfork); |
3086 | return 0; | 3086 | return 0; |
3087 | } | 3087 | } |
3088 | 3088 | ||
@@ -3136,8 +3136,8 @@ xfs_bmap_del_extent( | |||
3136 | del_endoff = del->br_startoff + del->br_blockcount; | 3136 | del_endoff = del->br_startoff + del->br_blockcount; |
3137 | got_endoff = got.br_startoff + got.br_blockcount; | 3137 | got_endoff = got.br_startoff + got.br_blockcount; |
3138 | ASSERT(got_endoff >= del_endoff); | 3138 | ASSERT(got_endoff >= del_endoff); |
3139 | delay = ISNULLSTARTBLOCK(got.br_startblock); | 3139 | delay = isnullstartblock(got.br_startblock); |
3140 | ASSERT(ISNULLSTARTBLOCK(del->br_startblock) == delay); | 3140 | ASSERT(isnullstartblock(del->br_startblock) == delay); |
3141 | flags = 0; | 3141 | flags = 0; |
3142 | qfield = 0; | 3142 | qfield = 0; |
3143 | error = 0; | 3143 | error = 0; |
@@ -3189,7 +3189,7 @@ xfs_bmap_del_extent( | |||
3189 | } | 3189 | } |
3190 | da_old = da_new = 0; | 3190 | da_old = da_new = 0; |
3191 | } else { | 3191 | } else { |
3192 | da_old = STARTBLOCKVAL(got.br_startblock); | 3192 | da_old = startblockval(got.br_startblock); |
3193 | da_new = 0; | 3193 | da_new = 0; |
3194 | nblks = 0; | 3194 | nblks = 0; |
3195 | do_fx = 0; | 3195 | do_fx = 0; |
@@ -3213,7 +3213,7 @@ xfs_bmap_del_extent( | |||
3213 | XFS_IFORK_NEXTENTS(ip, whichfork) - 1); | 3213 | XFS_IFORK_NEXTENTS(ip, whichfork) - 1); |
3214 | flags |= XFS_ILOG_CORE; | 3214 | flags |= XFS_ILOG_CORE; |
3215 | if (!cur) { | 3215 | if (!cur) { |
3216 | flags |= XFS_ILOG_FEXT(whichfork); | 3216 | flags |= xfs_ilog_fext(whichfork); |
3217 | break; | 3217 | break; |
3218 | } | 3218 | } |
3219 | if ((error = xfs_btree_delete(cur, &i))) | 3219 | if ((error = xfs_btree_delete(cur, &i))) |
@@ -3233,7 +3233,7 @@ xfs_bmap_del_extent( | |||
3233 | if (delay) { | 3233 | if (delay) { |
3234 | temp = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), | 3234 | temp = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), |
3235 | da_old); | 3235 | da_old); |
3236 | xfs_bmbt_set_startblock(ep, NULLSTARTBLOCK((int)temp)); | 3236 | xfs_bmbt_set_startblock(ep, nullstartblock((int)temp)); |
3237 | XFS_BMAP_TRACE_POST_UPDATE("2", ip, idx, | 3237 | XFS_BMAP_TRACE_POST_UPDATE("2", ip, idx, |
3238 | whichfork); | 3238 | whichfork); |
3239 | da_new = temp; | 3239 | da_new = temp; |
@@ -3242,7 +3242,7 @@ xfs_bmap_del_extent( | |||
3242 | xfs_bmbt_set_startblock(ep, del_endblock); | 3242 | xfs_bmbt_set_startblock(ep, del_endblock); |
3243 | XFS_BMAP_TRACE_POST_UPDATE("2", ip, idx, whichfork); | 3243 | XFS_BMAP_TRACE_POST_UPDATE("2", ip, idx, whichfork); |
3244 | if (!cur) { | 3244 | if (!cur) { |
3245 | flags |= XFS_ILOG_FEXT(whichfork); | 3245 | flags |= xfs_ilog_fext(whichfork); |
3246 | break; | 3246 | break; |
3247 | } | 3247 | } |
3248 | if ((error = xfs_bmbt_update(cur, del_endoff, del_endblock, | 3248 | if ((error = xfs_bmbt_update(cur, del_endoff, del_endblock, |
@@ -3262,7 +3262,7 @@ xfs_bmap_del_extent( | |||
3262 | if (delay) { | 3262 | if (delay) { |
3263 | temp = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), | 3263 | temp = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), |
3264 | da_old); | 3264 | da_old); |
3265 | xfs_bmbt_set_startblock(ep, NULLSTARTBLOCK((int)temp)); | 3265 | xfs_bmbt_set_startblock(ep, nullstartblock((int)temp)); |
3266 | XFS_BMAP_TRACE_POST_UPDATE("1", ip, idx, | 3266 | XFS_BMAP_TRACE_POST_UPDATE("1", ip, idx, |
3267 | whichfork); | 3267 | whichfork); |
3268 | da_new = temp; | 3268 | da_new = temp; |
@@ -3270,7 +3270,7 @@ xfs_bmap_del_extent( | |||
3270 | } | 3270 | } |
3271 | XFS_BMAP_TRACE_POST_UPDATE("1", ip, idx, whichfork); | 3271 | XFS_BMAP_TRACE_POST_UPDATE("1", ip, idx, whichfork); |
3272 | if (!cur) { | 3272 | if (!cur) { |
3273 | flags |= XFS_ILOG_FEXT(whichfork); | 3273 | flags |= xfs_ilog_fext(whichfork); |
3274 | break; | 3274 | break; |
3275 | } | 3275 | } |
3276 | if ((error = xfs_bmbt_update(cur, got.br_startoff, | 3276 | if ((error = xfs_bmbt_update(cur, got.br_startoff, |
@@ -3345,22 +3345,22 @@ xfs_bmap_del_extent( | |||
3345 | } | 3345 | } |
3346 | XFS_WANT_CORRUPTED_GOTO(i == 1, done); | 3346 | XFS_WANT_CORRUPTED_GOTO(i == 1, done); |
3347 | } else | 3347 | } else |
3348 | flags |= XFS_ILOG_FEXT(whichfork); | 3348 | flags |= xfs_ilog_fext(whichfork); |
3349 | XFS_IFORK_NEXT_SET(ip, whichfork, | 3349 | XFS_IFORK_NEXT_SET(ip, whichfork, |
3350 | XFS_IFORK_NEXTENTS(ip, whichfork) + 1); | 3350 | XFS_IFORK_NEXTENTS(ip, whichfork) + 1); |
3351 | } else { | 3351 | } else { |
3352 | ASSERT(whichfork == XFS_DATA_FORK); | 3352 | ASSERT(whichfork == XFS_DATA_FORK); |
3353 | temp = xfs_bmap_worst_indlen(ip, temp); | 3353 | temp = xfs_bmap_worst_indlen(ip, temp); |
3354 | xfs_bmbt_set_startblock(ep, NULLSTARTBLOCK((int)temp)); | 3354 | xfs_bmbt_set_startblock(ep, nullstartblock((int)temp)); |
3355 | temp2 = xfs_bmap_worst_indlen(ip, temp2); | 3355 | temp2 = xfs_bmap_worst_indlen(ip, temp2); |
3356 | new.br_startblock = NULLSTARTBLOCK((int)temp2); | 3356 | new.br_startblock = nullstartblock((int)temp2); |
3357 | da_new = temp + temp2; | 3357 | da_new = temp + temp2; |
3358 | while (da_new > da_old) { | 3358 | while (da_new > da_old) { |
3359 | if (temp) { | 3359 | if (temp) { |
3360 | temp--; | 3360 | temp--; |
3361 | da_new--; | 3361 | da_new--; |
3362 | xfs_bmbt_set_startblock(ep, | 3362 | xfs_bmbt_set_startblock(ep, |
3363 | NULLSTARTBLOCK((int)temp)); | 3363 | nullstartblock((int)temp)); |
3364 | } | 3364 | } |
3365 | if (da_new == da_old) | 3365 | if (da_new == da_old) |
3366 | break; | 3366 | break; |
@@ -3368,7 +3368,7 @@ xfs_bmap_del_extent( | |||
3368 | temp2--; | 3368 | temp2--; |
3369 | da_new--; | 3369 | da_new--; |
3370 | new.br_startblock = | 3370 | new.br_startblock = |
3371 | NULLSTARTBLOCK((int)temp2); | 3371 | nullstartblock((int)temp2); |
3372 | } | 3372 | } |
3373 | } | 3373 | } |
3374 | } | 3374 | } |
@@ -3545,7 +3545,7 @@ xfs_bmap_extents_to_btree( | |||
3545 | nextents = ifp->if_bytes / (uint)sizeof(xfs_bmbt_rec_t); | 3545 | nextents = ifp->if_bytes / (uint)sizeof(xfs_bmbt_rec_t); |
3546 | for (cnt = i = 0; i < nextents; i++) { | 3546 | for (cnt = i = 0; i < nextents; i++) { |
3547 | ep = xfs_iext_get_ext(ifp, i); | 3547 | ep = xfs_iext_get_ext(ifp, i); |
3548 | if (!ISNULLSTARTBLOCK(xfs_bmbt_get_startblock(ep))) { | 3548 | if (!isnullstartblock(xfs_bmbt_get_startblock(ep))) { |
3549 | arp->l0 = cpu_to_be64(ep->l0); | 3549 | arp->l0 = cpu_to_be64(ep->l0); |
3550 | arp->l1 = cpu_to_be64(ep->l1); | 3550 | arp->l1 = cpu_to_be64(ep->l1); |
3551 | arp++; cnt++; | 3551 | arp++; cnt++; |
@@ -3572,7 +3572,7 @@ xfs_bmap_extents_to_btree( | |||
3572 | xfs_btree_log_recs(cur, abp, 1, be16_to_cpu(ablock->bb_numrecs)); | 3572 | xfs_btree_log_recs(cur, abp, 1, be16_to_cpu(ablock->bb_numrecs)); |
3573 | ASSERT(*curp == NULL); | 3573 | ASSERT(*curp == NULL); |
3574 | *curp = cur; | 3574 | *curp = cur; |
3575 | *logflagsp = XFS_ILOG_CORE | XFS_ILOG_FBROOT(whichfork); | 3575 | *logflagsp = XFS_ILOG_CORE | xfs_ilog_fbroot(whichfork); |
3576 | return 0; | 3576 | return 0; |
3577 | } | 3577 | } |
3578 | 3578 | ||
@@ -3676,7 +3676,7 @@ xfs_bmap_local_to_extents( | |||
3676 | ip->i_d.di_nblocks = 1; | 3676 | ip->i_d.di_nblocks = 1; |
3677 | XFS_TRANS_MOD_DQUOT_BYINO(args.mp, tp, ip, | 3677 | XFS_TRANS_MOD_DQUOT_BYINO(args.mp, tp, ip, |
3678 | XFS_TRANS_DQ_BCOUNT, 1L); | 3678 | XFS_TRANS_DQ_BCOUNT, 1L); |
3679 | flags |= XFS_ILOG_FEXT(whichfork); | 3679 | flags |= xfs_ilog_fext(whichfork); |
3680 | } else { | 3680 | } else { |
3681 | ASSERT(XFS_IFORK_NEXTENTS(ip, whichfork) == 0); | 3681 | ASSERT(XFS_IFORK_NEXTENTS(ip, whichfork) == 0); |
3682 | xfs_bmap_forkoff_reset(ip->i_mount, ip, whichfork); | 3682 | xfs_bmap_forkoff_reset(ip->i_mount, ip, whichfork); |
@@ -4082,7 +4082,7 @@ xfs_bmap_add_attrfork( | |||
4082 | XFS_IFORK_ASIZE(ip) / (uint)sizeof(xfs_bmbt_rec_t); | 4082 | XFS_IFORK_ASIZE(ip) / (uint)sizeof(xfs_bmbt_rec_t); |
4083 | ip->i_afp->if_flags = XFS_IFEXTENTS; | 4083 | ip->i_afp->if_flags = XFS_IFEXTENTS; |
4084 | logflags = 0; | 4084 | logflags = 0; |
4085 | XFS_BMAP_INIT(&flist, &firstblock); | 4085 | xfs_bmap_init(&flist, &firstblock); |
4086 | switch (ip->i_d.di_format) { | 4086 | switch (ip->i_d.di_format) { |
4087 | case XFS_DINODE_FMT_LOCAL: | 4087 | case XFS_DINODE_FMT_LOCAL: |
4088 | error = xfs_bmap_add_attrfork_local(tp, ip, &firstblock, &flist, | 4088 | error = xfs_bmap_add_attrfork_local(tp, ip, &firstblock, &flist, |
@@ -4162,7 +4162,7 @@ xfs_bmap_add_free( | |||
4162 | ASSERT(bno != NULLFSBLOCK); | 4162 | ASSERT(bno != NULLFSBLOCK); |
4163 | ASSERT(len > 0); | 4163 | ASSERT(len > 0); |
4164 | ASSERT(len <= MAXEXTLEN); | 4164 | ASSERT(len <= MAXEXTLEN); |
4165 | ASSERT(!ISNULLSTARTBLOCK(bno)); | 4165 | ASSERT(!isnullstartblock(bno)); |
4166 | agno = XFS_FSB_TO_AGNO(mp, bno); | 4166 | agno = XFS_FSB_TO_AGNO(mp, bno); |
4167 | agbno = XFS_FSB_TO_AGBNO(mp, bno); | 4167 | agbno = XFS_FSB_TO_AGBNO(mp, bno); |
4168 | ASSERT(agno < mp->m_sb.sb_agcount); | 4168 | ASSERT(agno < mp->m_sb.sb_agcount); |
@@ -4909,7 +4909,7 @@ xfs_bmapi( | |||
4909 | got.br_startoff = end; | 4909 | got.br_startoff = end; |
4910 | inhole = eof || got.br_startoff > bno; | 4910 | inhole = eof || got.br_startoff > bno; |
4911 | wasdelay = wr && !inhole && !(flags & XFS_BMAPI_DELAY) && | 4911 | wasdelay = wr && !inhole && !(flags & XFS_BMAPI_DELAY) && |
4912 | ISNULLSTARTBLOCK(got.br_startblock); | 4912 | isnullstartblock(got.br_startblock); |
4913 | /* | 4913 | /* |
4914 | * First, deal with the hole before the allocated space | 4914 | * First, deal with the hole before the allocated space |
4915 | * that we found, if any. | 4915 | * that we found, if any. |
@@ -5028,7 +5028,7 @@ xfs_bmapi( | |||
5028 | } | 5028 | } |
5029 | 5029 | ||
5030 | ip->i_delayed_blks += alen; | 5030 | ip->i_delayed_blks += alen; |
5031 | abno = NULLSTARTBLOCK(indlen); | 5031 | abno = nullstartblock(indlen); |
5032 | } else { | 5032 | } else { |
5033 | /* | 5033 | /* |
5034 | * If first time, allocate and fill in | 5034 | * If first time, allocate and fill in |
@@ -5144,8 +5144,8 @@ xfs_bmapi( | |||
5144 | aoff + alen); | 5144 | aoff + alen); |
5145 | #ifdef DEBUG | 5145 | #ifdef DEBUG |
5146 | if (flags & XFS_BMAPI_DELAY) { | 5146 | if (flags & XFS_BMAPI_DELAY) { |
5147 | ASSERT(ISNULLSTARTBLOCK(got.br_startblock)); | 5147 | ASSERT(isnullstartblock(got.br_startblock)); |
5148 | ASSERT(STARTBLOCKVAL(got.br_startblock) > 0); | 5148 | ASSERT(startblockval(got.br_startblock) > 0); |
5149 | } | 5149 | } |
5150 | ASSERT(got.br_state == XFS_EXT_NORM || | 5150 | ASSERT(got.br_state == XFS_EXT_NORM || |
5151 | got.br_state == XFS_EXT_UNWRITTEN); | 5151 | got.br_state == XFS_EXT_UNWRITTEN); |
@@ -5179,7 +5179,7 @@ xfs_bmapi( | |||
5179 | ASSERT((bno >= obno) || (n == 0)); | 5179 | ASSERT((bno >= obno) || (n == 0)); |
5180 | ASSERT(bno < end); | 5180 | ASSERT(bno < end); |
5181 | mval->br_startoff = bno; | 5181 | mval->br_startoff = bno; |
5182 | if (ISNULLSTARTBLOCK(got.br_startblock)) { | 5182 | if (isnullstartblock(got.br_startblock)) { |
5183 | ASSERT(!wr || (flags & XFS_BMAPI_DELAY)); | 5183 | ASSERT(!wr || (flags & XFS_BMAPI_DELAY)); |
5184 | mval->br_startblock = DELAYSTARTBLOCK; | 5184 | mval->br_startblock = DELAYSTARTBLOCK; |
5185 | } else | 5185 | } else |
@@ -5201,7 +5201,7 @@ xfs_bmapi( | |||
5201 | ASSERT(mval->br_blockcount <= len); | 5201 | ASSERT(mval->br_blockcount <= len); |
5202 | } else { | 5202 | } else { |
5203 | *mval = got; | 5203 | *mval = got; |
5204 | if (ISNULLSTARTBLOCK(mval->br_startblock)) { | 5204 | if (isnullstartblock(mval->br_startblock)) { |
5205 | ASSERT(!wr || (flags & XFS_BMAPI_DELAY)); | 5205 | ASSERT(!wr || (flags & XFS_BMAPI_DELAY)); |
5206 | mval->br_startblock = DELAYSTARTBLOCK; | 5206 | mval->br_startblock = DELAYSTARTBLOCK; |
5207 | } | 5207 | } |
@@ -5329,12 +5329,12 @@ error0: | |||
5329 | * Log everything. Do this after conversion, there's no point in | 5329 | * Log everything. Do this after conversion, there's no point in |
5330 | * logging the extent records if we've converted to btree format. | 5330 | * logging the extent records if we've converted to btree format. |
5331 | */ | 5331 | */ |
5332 | if ((logflags & XFS_ILOG_FEXT(whichfork)) && | 5332 | if ((logflags & xfs_ilog_fext(whichfork)) && |
5333 | XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_EXTENTS) | 5333 | XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_EXTENTS) |
5334 | logflags &= ~XFS_ILOG_FEXT(whichfork); | 5334 | logflags &= ~xfs_ilog_fext(whichfork); |
5335 | else if ((logflags & XFS_ILOG_FBROOT(whichfork)) && | 5335 | else if ((logflags & xfs_ilog_fbroot(whichfork)) && |
5336 | XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_BTREE) | 5336 | XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_BTREE) |
5337 | logflags &= ~XFS_ILOG_FBROOT(whichfork); | 5337 | logflags &= ~xfs_ilog_fbroot(whichfork); |
5338 | /* | 5338 | /* |
5339 | * Log whatever the flags say, even if error. Otherwise we might miss | 5339 | * Log whatever the flags say, even if error. Otherwise we might miss |
5340 | * detecting a case where the data is changed, there's an error, | 5340 | * detecting a case where the data is changed, there's an error, |
@@ -5411,7 +5411,7 @@ xfs_bmapi_single( | |||
5411 | *fsb = NULLFSBLOCK; | 5411 | *fsb = NULLFSBLOCK; |
5412 | return 0; | 5412 | return 0; |
5413 | } | 5413 | } |
5414 | ASSERT(!ISNULLSTARTBLOCK(got.br_startblock)); | 5414 | ASSERT(!isnullstartblock(got.br_startblock)); |
5415 | ASSERT(bno < got.br_startoff + got.br_blockcount); | 5415 | ASSERT(bno < got.br_startoff + got.br_blockcount); |
5416 | *fsb = got.br_startblock + (bno - got.br_startoff); | 5416 | *fsb = got.br_startblock + (bno - got.br_startoff); |
5417 | ifp->if_lastex = lastx; | 5417 | ifp->if_lastex = lastx; |
@@ -5543,7 +5543,7 @@ xfs_bunmapi( | |||
5543 | */ | 5543 | */ |
5544 | ASSERT(ep != NULL); | 5544 | ASSERT(ep != NULL); |
5545 | del = got; | 5545 | del = got; |
5546 | wasdel = ISNULLSTARTBLOCK(del.br_startblock); | 5546 | wasdel = isnullstartblock(del.br_startblock); |
5547 | if (got.br_startoff < start) { | 5547 | if (got.br_startoff < start) { |
5548 | del.br_startoff = start; | 5548 | del.br_startoff = start; |
5549 | del.br_blockcount -= start - got.br_startoff; | 5549 | del.br_blockcount -= start - got.br_startoff; |
@@ -5638,7 +5638,7 @@ xfs_bunmapi( | |||
5638 | xfs_bmbt_get_all(xfs_iext_get_ext(ifp, | 5638 | xfs_bmbt_get_all(xfs_iext_get_ext(ifp, |
5639 | lastx - 1), &prev); | 5639 | lastx - 1), &prev); |
5640 | ASSERT(prev.br_state == XFS_EXT_NORM); | 5640 | ASSERT(prev.br_state == XFS_EXT_NORM); |
5641 | ASSERT(!ISNULLSTARTBLOCK(prev.br_startblock)); | 5641 | ASSERT(!isnullstartblock(prev.br_startblock)); |
5642 | ASSERT(del.br_startblock == | 5642 | ASSERT(del.br_startblock == |
5643 | prev.br_startblock + prev.br_blockcount); | 5643 | prev.br_startblock + prev.br_blockcount); |
5644 | if (prev.br_startoff < start) { | 5644 | if (prev.br_startoff < start) { |
@@ -5666,7 +5666,7 @@ xfs_bunmapi( | |||
5666 | } | 5666 | } |
5667 | } | 5667 | } |
5668 | if (wasdel) { | 5668 | if (wasdel) { |
5669 | ASSERT(STARTBLOCKVAL(del.br_startblock) > 0); | 5669 | ASSERT(startblockval(del.br_startblock) > 0); |
5670 | /* Update realtime/data freespace, unreserve quota */ | 5670 | /* Update realtime/data freespace, unreserve quota */ |
5671 | if (isrt) { | 5671 | if (isrt) { |
5672 | xfs_filblks_t rtexts; | 5672 | xfs_filblks_t rtexts; |
@@ -5782,12 +5782,12 @@ error0: | |||
5782 | * Log everything. Do this after conversion, there's no point in | 5782 | * Log everything. Do this after conversion, there's no point in |
5783 | * logging the extent records if we've converted to btree format. | 5783 | * logging the extent records if we've converted to btree format. |
5784 | */ | 5784 | */ |
5785 | if ((logflags & XFS_ILOG_FEXT(whichfork)) && | 5785 | if ((logflags & xfs_ilog_fext(whichfork)) && |
5786 | XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_EXTENTS) | 5786 | XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_EXTENTS) |
5787 | logflags &= ~XFS_ILOG_FEXT(whichfork); | 5787 | logflags &= ~xfs_ilog_fext(whichfork); |
5788 | else if ((logflags & XFS_ILOG_FBROOT(whichfork)) && | 5788 | else if ((logflags & xfs_ilog_fbroot(whichfork)) && |
5789 | XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_BTREE) | 5789 | XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_BTREE) |
5790 | logflags &= ~XFS_ILOG_FBROOT(whichfork); | 5790 | logflags &= ~xfs_ilog_fbroot(whichfork); |
5791 | /* | 5791 | /* |
5792 | * Log inode even in the error case, if the transaction | 5792 | * Log inode even in the error case, if the transaction |
5793 | * is dirty we'll need to shut down the filesystem. | 5793 | * is dirty we'll need to shut down the filesystem. |
@@ -5838,7 +5838,7 @@ xfs_getbmapx_fix_eof_hole( | |||
5838 | if (startblock == DELAYSTARTBLOCK) | 5838 | if (startblock == DELAYSTARTBLOCK) |
5839 | out->bmv_block = -2; | 5839 | out->bmv_block = -2; |
5840 | else | 5840 | else |
5841 | out->bmv_block = XFS_FSB_TO_DB(ip, startblock); | 5841 | out->bmv_block = xfs_fsb_to_db(ip, startblock); |
5842 | fileblock = XFS_BB_TO_FSB(ip->i_mount, out->bmv_offset); | 5842 | fileblock = XFS_BB_TO_FSB(ip->i_mount, out->bmv_offset); |
5843 | ifp = XFS_IFORK_PTR(ip, XFS_DATA_FORK); | 5843 | ifp = XFS_IFORK_PTR(ip, XFS_DATA_FORK); |
5844 | if (xfs_iext_bno_to_ext(ifp, fileblock, &lastx) && | 5844 | if (xfs_iext_bno_to_ext(ifp, fileblock, &lastx) && |
@@ -5979,7 +5979,7 @@ xfs_getbmap( | |||
5979 | if (nex > XFS_IFORK_NEXTENTS(ip, whichfork) * 2 + 1) | 5979 | if (nex > XFS_IFORK_NEXTENTS(ip, whichfork) * 2 + 1) |
5980 | nex = XFS_IFORK_NEXTENTS(ip, whichfork) * 2 + 1; | 5980 | nex = XFS_IFORK_NEXTENTS(ip, whichfork) * 2 + 1; |
5981 | 5981 | ||
5982 | bmapi_flags = XFS_BMAPI_AFLAG(whichfork) | | 5982 | bmapi_flags = xfs_bmapi_aflag(whichfork) | |
5983 | ((iflags & BMV_IF_PREALLOC) ? 0 : XFS_BMAPI_IGSTATE); | 5983 | ((iflags & BMV_IF_PREALLOC) ? 0 : XFS_BMAPI_IGSTATE); |
5984 | 5984 | ||
5985 | /* | 5985 | /* |
@@ -6098,7 +6098,7 @@ xfs_bmap_isaeof( | |||
6098 | */ | 6098 | */ |
6099 | *aeof = (off >= s.br_startoff && | 6099 | *aeof = (off >= s.br_startoff && |
6100 | off < s.br_startoff + s.br_blockcount && | 6100 | off < s.br_startoff + s.br_blockcount && |
6101 | ISNULLSTARTBLOCK(s.br_startblock)) || | 6101 | isnullstartblock(s.br_startblock)) || |
6102 | off >= s.br_startoff + s.br_blockcount; | 6102 | off >= s.br_startoff + s.br_blockcount; |
6103 | return 0; | 6103 | return 0; |
6104 | } | 6104 | } |
diff --git a/fs/xfs/xfs_bmap.h b/fs/xfs/xfs_bmap.h index 284571c05ed0..be2979d88d32 100644 --- a/fs/xfs/xfs_bmap.h +++ b/fs/xfs/xfs_bmap.h | |||
@@ -95,7 +95,6 @@ typedef struct xfs_bmap_free | |||
95 | /* need write cache flushing and no */ | 95 | /* need write cache flushing and no */ |
96 | /* additional allocation alignments */ | 96 | /* additional allocation alignments */ |
97 | 97 | ||
98 | #define XFS_BMAPI_AFLAG(w) xfs_bmapi_aflag(w) | ||
99 | static inline int xfs_bmapi_aflag(int w) | 98 | static inline int xfs_bmapi_aflag(int w) |
100 | { | 99 | { |
101 | return (w == XFS_ATTR_FORK ? XFS_BMAPI_ATTRFORK : 0); | 100 | return (w == XFS_ATTR_FORK ? XFS_BMAPI_ATTRFORK : 0); |
@@ -107,7 +106,6 @@ static inline int xfs_bmapi_aflag(int w) | |||
107 | #define DELAYSTARTBLOCK ((xfs_fsblock_t)-1LL) | 106 | #define DELAYSTARTBLOCK ((xfs_fsblock_t)-1LL) |
108 | #define HOLESTARTBLOCK ((xfs_fsblock_t)-2LL) | 107 | #define HOLESTARTBLOCK ((xfs_fsblock_t)-2LL) |
109 | 108 | ||
110 | #define XFS_BMAP_INIT(flp,fbp) xfs_bmap_init(flp,fbp) | ||
111 | static inline void xfs_bmap_init(xfs_bmap_free_t *flp, xfs_fsblock_t *fbp) | 109 | static inline void xfs_bmap_init(xfs_bmap_free_t *flp, xfs_fsblock_t *fbp) |
112 | { | 110 | { |
113 | ((flp)->xbf_first = NULL, (flp)->xbf_count = 0, \ | 111 | ((flp)->xbf_first = NULL, (flp)->xbf_count = 0, \ |
diff --git a/fs/xfs/xfs_bmap_btree.c b/fs/xfs/xfs_bmap_btree.c index 8f1ec73725d3..0760d352586f 100644 --- a/fs/xfs/xfs_bmap_btree.c +++ b/fs/xfs/xfs_bmap_btree.c | |||
@@ -110,25 +110,25 @@ __xfs_bmbt_get_all( | |||
110 | 110 | ||
111 | ext_flag = (int)(l0 >> (64 - BMBT_EXNTFLAG_BITLEN)); | 111 | ext_flag = (int)(l0 >> (64 - BMBT_EXNTFLAG_BITLEN)); |
112 | s->br_startoff = ((xfs_fileoff_t)l0 & | 112 | s->br_startoff = ((xfs_fileoff_t)l0 & |
113 | XFS_MASK64LO(64 - BMBT_EXNTFLAG_BITLEN)) >> 9; | 113 | xfs_mask64lo(64 - BMBT_EXNTFLAG_BITLEN)) >> 9; |
114 | #if XFS_BIG_BLKNOS | 114 | #if XFS_BIG_BLKNOS |
115 | s->br_startblock = (((xfs_fsblock_t)l0 & XFS_MASK64LO(9)) << 43) | | 115 | s->br_startblock = (((xfs_fsblock_t)l0 & xfs_mask64lo(9)) << 43) | |
116 | (((xfs_fsblock_t)l1) >> 21); | 116 | (((xfs_fsblock_t)l1) >> 21); |
117 | #else | 117 | #else |
118 | #ifdef DEBUG | 118 | #ifdef DEBUG |
119 | { | 119 | { |
120 | xfs_dfsbno_t b; | 120 | xfs_dfsbno_t b; |
121 | 121 | ||
122 | b = (((xfs_dfsbno_t)l0 & XFS_MASK64LO(9)) << 43) | | 122 | b = (((xfs_dfsbno_t)l0 & xfs_mask64lo(9)) << 43) | |
123 | (((xfs_dfsbno_t)l1) >> 21); | 123 | (((xfs_dfsbno_t)l1) >> 21); |
124 | ASSERT((b >> 32) == 0 || ISNULLDSTARTBLOCK(b)); | 124 | ASSERT((b >> 32) == 0 || isnulldstartblock(b)); |
125 | s->br_startblock = (xfs_fsblock_t)b; | 125 | s->br_startblock = (xfs_fsblock_t)b; |
126 | } | 126 | } |
127 | #else /* !DEBUG */ | 127 | #else /* !DEBUG */ |
128 | s->br_startblock = (xfs_fsblock_t)(((xfs_dfsbno_t)l1) >> 21); | 128 | s->br_startblock = (xfs_fsblock_t)(((xfs_dfsbno_t)l1) >> 21); |
129 | #endif /* DEBUG */ | 129 | #endif /* DEBUG */ |
130 | #endif /* XFS_BIG_BLKNOS */ | 130 | #endif /* XFS_BIG_BLKNOS */ |
131 | s->br_blockcount = (xfs_filblks_t)(l1 & XFS_MASK64LO(21)); | 131 | s->br_blockcount = (xfs_filblks_t)(l1 & xfs_mask64lo(21)); |
132 | /* This is xfs_extent_state() in-line */ | 132 | /* This is xfs_extent_state() in-line */ |
133 | if (ext_flag) { | 133 | if (ext_flag) { |
134 | ASSERT(s->br_blockcount != 0); /* saved for DMIG */ | 134 | ASSERT(s->br_blockcount != 0); /* saved for DMIG */ |
@@ -153,7 +153,7 @@ xfs_filblks_t | |||
153 | xfs_bmbt_get_blockcount( | 153 | xfs_bmbt_get_blockcount( |
154 | xfs_bmbt_rec_host_t *r) | 154 | xfs_bmbt_rec_host_t *r) |
155 | { | 155 | { |
156 | return (xfs_filblks_t)(r->l1 & XFS_MASK64LO(21)); | 156 | return (xfs_filblks_t)(r->l1 & xfs_mask64lo(21)); |
157 | } | 157 | } |
158 | 158 | ||
159 | /* | 159 | /* |
@@ -164,15 +164,15 @@ xfs_bmbt_get_startblock( | |||
164 | xfs_bmbt_rec_host_t *r) | 164 | xfs_bmbt_rec_host_t *r) |
165 | { | 165 | { |
166 | #if XFS_BIG_BLKNOS | 166 | #if XFS_BIG_BLKNOS |
167 | return (((xfs_fsblock_t)r->l0 & XFS_MASK64LO(9)) << 43) | | 167 | return (((xfs_fsblock_t)r->l0 & xfs_mask64lo(9)) << 43) | |
168 | (((xfs_fsblock_t)r->l1) >> 21); | 168 | (((xfs_fsblock_t)r->l1) >> 21); |
169 | #else | 169 | #else |
170 | #ifdef DEBUG | 170 | #ifdef DEBUG |
171 | xfs_dfsbno_t b; | 171 | xfs_dfsbno_t b; |
172 | 172 | ||
173 | b = (((xfs_dfsbno_t)r->l0 & XFS_MASK64LO(9)) << 43) | | 173 | b = (((xfs_dfsbno_t)r->l0 & xfs_mask64lo(9)) << 43) | |
174 | (((xfs_dfsbno_t)r->l1) >> 21); | 174 | (((xfs_dfsbno_t)r->l1) >> 21); |
175 | ASSERT((b >> 32) == 0 || ISNULLDSTARTBLOCK(b)); | 175 | ASSERT((b >> 32) == 0 || isnulldstartblock(b)); |
176 | return (xfs_fsblock_t)b; | 176 | return (xfs_fsblock_t)b; |
177 | #else /* !DEBUG */ | 177 | #else /* !DEBUG */ |
178 | return (xfs_fsblock_t)(((xfs_dfsbno_t)r->l1) >> 21); | 178 | return (xfs_fsblock_t)(((xfs_dfsbno_t)r->l1) >> 21); |
@@ -188,7 +188,7 @@ xfs_bmbt_get_startoff( | |||
188 | xfs_bmbt_rec_host_t *r) | 188 | xfs_bmbt_rec_host_t *r) |
189 | { | 189 | { |
190 | return ((xfs_fileoff_t)r->l0 & | 190 | return ((xfs_fileoff_t)r->l0 & |
191 | XFS_MASK64LO(64 - BMBT_EXNTFLAG_BITLEN)) >> 9; | 191 | xfs_mask64lo(64 - BMBT_EXNTFLAG_BITLEN)) >> 9; |
192 | } | 192 | } |
193 | 193 | ||
194 | xfs_exntst_t | 194 | xfs_exntst_t |
@@ -219,7 +219,7 @@ xfs_filblks_t | |||
219 | xfs_bmbt_disk_get_blockcount( | 219 | xfs_bmbt_disk_get_blockcount( |
220 | xfs_bmbt_rec_t *r) | 220 | xfs_bmbt_rec_t *r) |
221 | { | 221 | { |
222 | return (xfs_filblks_t)(be64_to_cpu(r->l1) & XFS_MASK64LO(21)); | 222 | return (xfs_filblks_t)(be64_to_cpu(r->l1) & xfs_mask64lo(21)); |
223 | } | 223 | } |
224 | 224 | ||
225 | /* | 225 | /* |
@@ -230,7 +230,7 @@ xfs_bmbt_disk_get_startoff( | |||
230 | xfs_bmbt_rec_t *r) | 230 | xfs_bmbt_rec_t *r) |
231 | { | 231 | { |
232 | return ((xfs_fileoff_t)be64_to_cpu(r->l0) & | 232 | return ((xfs_fileoff_t)be64_to_cpu(r->l0) & |
233 | XFS_MASK64LO(64 - BMBT_EXNTFLAG_BITLEN)) >> 9; | 233 | xfs_mask64lo(64 - BMBT_EXNTFLAG_BITLEN)) >> 9; |
234 | } | 234 | } |
235 | 235 | ||
236 | 236 | ||
@@ -248,33 +248,33 @@ xfs_bmbt_set_allf( | |||
248 | int extent_flag = (state == XFS_EXT_NORM) ? 0 : 1; | 248 | int extent_flag = (state == XFS_EXT_NORM) ? 0 : 1; |
249 | 249 | ||
250 | ASSERT(state == XFS_EXT_NORM || state == XFS_EXT_UNWRITTEN); | 250 | ASSERT(state == XFS_EXT_NORM || state == XFS_EXT_UNWRITTEN); |
251 | ASSERT((startoff & XFS_MASK64HI(64-BMBT_STARTOFF_BITLEN)) == 0); | 251 | ASSERT((startoff & xfs_mask64hi(64-BMBT_STARTOFF_BITLEN)) == 0); |
252 | ASSERT((blockcount & XFS_MASK64HI(64-BMBT_BLOCKCOUNT_BITLEN)) == 0); | 252 | ASSERT((blockcount & xfs_mask64hi(64-BMBT_BLOCKCOUNT_BITLEN)) == 0); |
253 | 253 | ||
254 | #if XFS_BIG_BLKNOS | 254 | #if XFS_BIG_BLKNOS |
255 | ASSERT((startblock & XFS_MASK64HI(64-BMBT_STARTBLOCK_BITLEN)) == 0); | 255 | ASSERT((startblock & xfs_mask64hi(64-BMBT_STARTBLOCK_BITLEN)) == 0); |
256 | 256 | ||
257 | r->l0 = ((xfs_bmbt_rec_base_t)extent_flag << 63) | | 257 | r->l0 = ((xfs_bmbt_rec_base_t)extent_flag << 63) | |
258 | ((xfs_bmbt_rec_base_t)startoff << 9) | | 258 | ((xfs_bmbt_rec_base_t)startoff << 9) | |
259 | ((xfs_bmbt_rec_base_t)startblock >> 43); | 259 | ((xfs_bmbt_rec_base_t)startblock >> 43); |
260 | r->l1 = ((xfs_bmbt_rec_base_t)startblock << 21) | | 260 | r->l1 = ((xfs_bmbt_rec_base_t)startblock << 21) | |
261 | ((xfs_bmbt_rec_base_t)blockcount & | 261 | ((xfs_bmbt_rec_base_t)blockcount & |
262 | (xfs_bmbt_rec_base_t)XFS_MASK64LO(21)); | 262 | (xfs_bmbt_rec_base_t)xfs_mask64lo(21)); |
263 | #else /* !XFS_BIG_BLKNOS */ | 263 | #else /* !XFS_BIG_BLKNOS */ |
264 | if (ISNULLSTARTBLOCK(startblock)) { | 264 | if (isnullstartblock(startblock)) { |
265 | r->l0 = ((xfs_bmbt_rec_base_t)extent_flag << 63) | | 265 | r->l0 = ((xfs_bmbt_rec_base_t)extent_flag << 63) | |
266 | ((xfs_bmbt_rec_base_t)startoff << 9) | | 266 | ((xfs_bmbt_rec_base_t)startoff << 9) | |
267 | (xfs_bmbt_rec_base_t)XFS_MASK64LO(9); | 267 | (xfs_bmbt_rec_base_t)xfs_mask64lo(9); |
268 | r->l1 = XFS_MASK64HI(11) | | 268 | r->l1 = xfs_mask64hi(11) | |
269 | ((xfs_bmbt_rec_base_t)startblock << 21) | | 269 | ((xfs_bmbt_rec_base_t)startblock << 21) | |
270 | ((xfs_bmbt_rec_base_t)blockcount & | 270 | ((xfs_bmbt_rec_base_t)blockcount & |
271 | (xfs_bmbt_rec_base_t)XFS_MASK64LO(21)); | 271 | (xfs_bmbt_rec_base_t)xfs_mask64lo(21)); |
272 | } else { | 272 | } else { |
273 | r->l0 = ((xfs_bmbt_rec_base_t)extent_flag << 63) | | 273 | r->l0 = ((xfs_bmbt_rec_base_t)extent_flag << 63) | |
274 | ((xfs_bmbt_rec_base_t)startoff << 9); | 274 | ((xfs_bmbt_rec_base_t)startoff << 9); |
275 | r->l1 = ((xfs_bmbt_rec_base_t)startblock << 21) | | 275 | r->l1 = ((xfs_bmbt_rec_base_t)startblock << 21) | |
276 | ((xfs_bmbt_rec_base_t)blockcount & | 276 | ((xfs_bmbt_rec_base_t)blockcount & |
277 | (xfs_bmbt_rec_base_t)XFS_MASK64LO(21)); | 277 | (xfs_bmbt_rec_base_t)xfs_mask64lo(21)); |
278 | } | 278 | } |
279 | #endif /* XFS_BIG_BLKNOS */ | 279 | #endif /* XFS_BIG_BLKNOS */ |
280 | } | 280 | } |
@@ -306,11 +306,11 @@ xfs_bmbt_disk_set_allf( | |||
306 | int extent_flag = (state == XFS_EXT_NORM) ? 0 : 1; | 306 | int extent_flag = (state == XFS_EXT_NORM) ? 0 : 1; |
307 | 307 | ||
308 | ASSERT(state == XFS_EXT_NORM || state == XFS_EXT_UNWRITTEN); | 308 | ASSERT(state == XFS_EXT_NORM || state == XFS_EXT_UNWRITTEN); |
309 | ASSERT((startoff & XFS_MASK64HI(64-BMBT_STARTOFF_BITLEN)) == 0); | 309 | ASSERT((startoff & xfs_mask64hi(64-BMBT_STARTOFF_BITLEN)) == 0); |
310 | ASSERT((blockcount & XFS_MASK64HI(64-BMBT_BLOCKCOUNT_BITLEN)) == 0); | 310 | ASSERT((blockcount & xfs_mask64hi(64-BMBT_BLOCKCOUNT_BITLEN)) == 0); |
311 | 311 | ||
312 | #if XFS_BIG_BLKNOS | 312 | #if XFS_BIG_BLKNOS |
313 | ASSERT((startblock & XFS_MASK64HI(64-BMBT_STARTBLOCK_BITLEN)) == 0); | 313 | ASSERT((startblock & xfs_mask64hi(64-BMBT_STARTBLOCK_BITLEN)) == 0); |
314 | 314 | ||
315 | r->l0 = cpu_to_be64( | 315 | r->l0 = cpu_to_be64( |
316 | ((xfs_bmbt_rec_base_t)extent_flag << 63) | | 316 | ((xfs_bmbt_rec_base_t)extent_flag << 63) | |
@@ -319,17 +319,17 @@ xfs_bmbt_disk_set_allf( | |||
319 | r->l1 = cpu_to_be64( | 319 | r->l1 = cpu_to_be64( |
320 | ((xfs_bmbt_rec_base_t)startblock << 21) | | 320 | ((xfs_bmbt_rec_base_t)startblock << 21) | |
321 | ((xfs_bmbt_rec_base_t)blockcount & | 321 | ((xfs_bmbt_rec_base_t)blockcount & |
322 | (xfs_bmbt_rec_base_t)XFS_MASK64LO(21))); | 322 | (xfs_bmbt_rec_base_t)xfs_mask64lo(21))); |
323 | #else /* !XFS_BIG_BLKNOS */ | 323 | #else /* !XFS_BIG_BLKNOS */ |
324 | if (ISNULLSTARTBLOCK(startblock)) { | 324 | if (isnullstartblock(startblock)) { |
325 | r->l0 = cpu_to_be64( | 325 | r->l0 = cpu_to_be64( |
326 | ((xfs_bmbt_rec_base_t)extent_flag << 63) | | 326 | ((xfs_bmbt_rec_base_t)extent_flag << 63) | |
327 | ((xfs_bmbt_rec_base_t)startoff << 9) | | 327 | ((xfs_bmbt_rec_base_t)startoff << 9) | |
328 | (xfs_bmbt_rec_base_t)XFS_MASK64LO(9)); | 328 | (xfs_bmbt_rec_base_t)xfs_mask64lo(9)); |
329 | r->l1 = cpu_to_be64(XFS_MASK64HI(11) | | 329 | r->l1 = cpu_to_be64(xfs_mask64hi(11) | |
330 | ((xfs_bmbt_rec_base_t)startblock << 21) | | 330 | ((xfs_bmbt_rec_base_t)startblock << 21) | |
331 | ((xfs_bmbt_rec_base_t)blockcount & | 331 | ((xfs_bmbt_rec_base_t)blockcount & |
332 | (xfs_bmbt_rec_base_t)XFS_MASK64LO(21))); | 332 | (xfs_bmbt_rec_base_t)xfs_mask64lo(21))); |
333 | } else { | 333 | } else { |
334 | r->l0 = cpu_to_be64( | 334 | r->l0 = cpu_to_be64( |
335 | ((xfs_bmbt_rec_base_t)extent_flag << 63) | | 335 | ((xfs_bmbt_rec_base_t)extent_flag << 63) | |
@@ -337,7 +337,7 @@ xfs_bmbt_disk_set_allf( | |||
337 | r->l1 = cpu_to_be64( | 337 | r->l1 = cpu_to_be64( |
338 | ((xfs_bmbt_rec_base_t)startblock << 21) | | 338 | ((xfs_bmbt_rec_base_t)startblock << 21) | |
339 | ((xfs_bmbt_rec_base_t)blockcount & | 339 | ((xfs_bmbt_rec_base_t)blockcount & |
340 | (xfs_bmbt_rec_base_t)XFS_MASK64LO(21))); | 340 | (xfs_bmbt_rec_base_t)xfs_mask64lo(21))); |
341 | } | 341 | } |
342 | #endif /* XFS_BIG_BLKNOS */ | 342 | #endif /* XFS_BIG_BLKNOS */ |
343 | } | 343 | } |
@@ -362,9 +362,9 @@ xfs_bmbt_set_blockcount( | |||
362 | xfs_bmbt_rec_host_t *r, | 362 | xfs_bmbt_rec_host_t *r, |
363 | xfs_filblks_t v) | 363 | xfs_filblks_t v) |
364 | { | 364 | { |
365 | ASSERT((v & XFS_MASK64HI(43)) == 0); | 365 | ASSERT((v & xfs_mask64hi(43)) == 0); |
366 | r->l1 = (r->l1 & (xfs_bmbt_rec_base_t)XFS_MASK64HI(43)) | | 366 | r->l1 = (r->l1 & (xfs_bmbt_rec_base_t)xfs_mask64hi(43)) | |
367 | (xfs_bmbt_rec_base_t)(v & XFS_MASK64LO(21)); | 367 | (xfs_bmbt_rec_base_t)(v & xfs_mask64lo(21)); |
368 | } | 368 | } |
369 | 369 | ||
370 | /* | 370 | /* |
@@ -376,21 +376,21 @@ xfs_bmbt_set_startblock( | |||
376 | xfs_fsblock_t v) | 376 | xfs_fsblock_t v) |
377 | { | 377 | { |
378 | #if XFS_BIG_BLKNOS | 378 | #if XFS_BIG_BLKNOS |
379 | ASSERT((v & XFS_MASK64HI(12)) == 0); | 379 | ASSERT((v & xfs_mask64hi(12)) == 0); |
380 | r->l0 = (r->l0 & (xfs_bmbt_rec_base_t)XFS_MASK64HI(55)) | | 380 | r->l0 = (r->l0 & (xfs_bmbt_rec_base_t)xfs_mask64hi(55)) | |
381 | (xfs_bmbt_rec_base_t)(v >> 43); | 381 | (xfs_bmbt_rec_base_t)(v >> 43); |
382 | r->l1 = (r->l1 & (xfs_bmbt_rec_base_t)XFS_MASK64LO(21)) | | 382 | r->l1 = (r->l1 & (xfs_bmbt_rec_base_t)xfs_mask64lo(21)) | |
383 | (xfs_bmbt_rec_base_t)(v << 21); | 383 | (xfs_bmbt_rec_base_t)(v << 21); |
384 | #else /* !XFS_BIG_BLKNOS */ | 384 | #else /* !XFS_BIG_BLKNOS */ |
385 | if (ISNULLSTARTBLOCK(v)) { | 385 | if (isnullstartblock(v)) { |
386 | r->l0 |= (xfs_bmbt_rec_base_t)XFS_MASK64LO(9); | 386 | r->l0 |= (xfs_bmbt_rec_base_t)xfs_mask64lo(9); |
387 | r->l1 = (xfs_bmbt_rec_base_t)XFS_MASK64HI(11) | | 387 | r->l1 = (xfs_bmbt_rec_base_t)xfs_mask64hi(11) | |
388 | ((xfs_bmbt_rec_base_t)v << 21) | | 388 | ((xfs_bmbt_rec_base_t)v << 21) | |
389 | (r->l1 & (xfs_bmbt_rec_base_t)XFS_MASK64LO(21)); | 389 | (r->l1 & (xfs_bmbt_rec_base_t)xfs_mask64lo(21)); |
390 | } else { | 390 | } else { |
391 | r->l0 &= ~(xfs_bmbt_rec_base_t)XFS_MASK64LO(9); | 391 | r->l0 &= ~(xfs_bmbt_rec_base_t)xfs_mask64lo(9); |
392 | r->l1 = ((xfs_bmbt_rec_base_t)v << 21) | | 392 | r->l1 = ((xfs_bmbt_rec_base_t)v << 21) | |
393 | (r->l1 & (xfs_bmbt_rec_base_t)XFS_MASK64LO(21)); | 393 | (r->l1 & (xfs_bmbt_rec_base_t)xfs_mask64lo(21)); |
394 | } | 394 | } |
395 | #endif /* XFS_BIG_BLKNOS */ | 395 | #endif /* XFS_BIG_BLKNOS */ |
396 | } | 396 | } |
@@ -403,10 +403,10 @@ xfs_bmbt_set_startoff( | |||
403 | xfs_bmbt_rec_host_t *r, | 403 | xfs_bmbt_rec_host_t *r, |
404 | xfs_fileoff_t v) | 404 | xfs_fileoff_t v) |
405 | { | 405 | { |
406 | ASSERT((v & XFS_MASK64HI(9)) == 0); | 406 | ASSERT((v & xfs_mask64hi(9)) == 0); |
407 | r->l0 = (r->l0 & (xfs_bmbt_rec_base_t) XFS_MASK64HI(1)) | | 407 | r->l0 = (r->l0 & (xfs_bmbt_rec_base_t) xfs_mask64hi(1)) | |
408 | ((xfs_bmbt_rec_base_t)v << 9) | | 408 | ((xfs_bmbt_rec_base_t)v << 9) | |
409 | (r->l0 & (xfs_bmbt_rec_base_t)XFS_MASK64LO(9)); | 409 | (r->l0 & (xfs_bmbt_rec_base_t)xfs_mask64lo(9)); |
410 | } | 410 | } |
411 | 411 | ||
412 | /* | 412 | /* |
@@ -419,9 +419,9 @@ xfs_bmbt_set_state( | |||
419 | { | 419 | { |
420 | ASSERT(v == XFS_EXT_NORM || v == XFS_EXT_UNWRITTEN); | 420 | ASSERT(v == XFS_EXT_NORM || v == XFS_EXT_UNWRITTEN); |
421 | if (v == XFS_EXT_NORM) | 421 | if (v == XFS_EXT_NORM) |
422 | r->l0 &= XFS_MASK64LO(64 - BMBT_EXNTFLAG_BITLEN); | 422 | r->l0 &= xfs_mask64lo(64 - BMBT_EXNTFLAG_BITLEN); |
423 | else | 423 | else |
424 | r->l0 |= XFS_MASK64HI(BMBT_EXNTFLAG_BITLEN); | 424 | r->l0 |= xfs_mask64hi(BMBT_EXNTFLAG_BITLEN); |
425 | } | 425 | } |
426 | 426 | ||
427 | /* | 427 | /* |
diff --git a/fs/xfs/xfs_bmap_btree.h b/fs/xfs/xfs_bmap_btree.h index a4555abb6622..0e8df007615e 100644 --- a/fs/xfs/xfs_bmap_btree.h +++ b/fs/xfs/xfs_bmap_btree.h | |||
@@ -76,26 +76,22 @@ typedef struct xfs_bmbt_rec_host { | |||
76 | #define DSTARTBLOCKMASK \ | 76 | #define DSTARTBLOCKMASK \ |
77 | (((((xfs_dfsbno_t)1) << DSTARTBLOCKMASKBITS) - 1) << STARTBLOCKVALBITS) | 77 | (((((xfs_dfsbno_t)1) << DSTARTBLOCKMASKBITS) - 1) << STARTBLOCKVALBITS) |
78 | 78 | ||
79 | #define ISNULLSTARTBLOCK(x) isnullstartblock(x) | ||
80 | static inline int isnullstartblock(xfs_fsblock_t x) | 79 | static inline int isnullstartblock(xfs_fsblock_t x) |
81 | { | 80 | { |
82 | return ((x) & STARTBLOCKMASK) == STARTBLOCKMASK; | 81 | return ((x) & STARTBLOCKMASK) == STARTBLOCKMASK; |
83 | } | 82 | } |
84 | 83 | ||
85 | #define ISNULLDSTARTBLOCK(x) isnulldstartblock(x) | ||
86 | static inline int isnulldstartblock(xfs_dfsbno_t x) | 84 | static inline int isnulldstartblock(xfs_dfsbno_t x) |
87 | { | 85 | { |
88 | return ((x) & DSTARTBLOCKMASK) == DSTARTBLOCKMASK; | 86 | return ((x) & DSTARTBLOCKMASK) == DSTARTBLOCKMASK; |
89 | } | 87 | } |
90 | 88 | ||
91 | #define NULLSTARTBLOCK(k) nullstartblock(k) | ||
92 | static inline xfs_fsblock_t nullstartblock(int k) | 89 | static inline xfs_fsblock_t nullstartblock(int k) |
93 | { | 90 | { |
94 | ASSERT(k < (1 << STARTBLOCKVALBITS)); | 91 | ASSERT(k < (1 << STARTBLOCKVALBITS)); |
95 | return STARTBLOCKMASK | (k); | 92 | return STARTBLOCKMASK | (k); |
96 | } | 93 | } |
97 | 94 | ||
98 | #define STARTBLOCKVAL(x) startblockval(x) | ||
99 | static inline xfs_filblks_t startblockval(xfs_fsblock_t x) | 95 | static inline xfs_filblks_t startblockval(xfs_fsblock_t x) |
100 | { | 96 | { |
101 | return (xfs_filblks_t)((x) & ~STARTBLOCKMASK); | 97 | return (xfs_filblks_t)((x) & ~STARTBLOCKMASK); |
diff --git a/fs/xfs/xfs_btree.c b/fs/xfs/xfs_btree.c index 7ed59267420d..e73c332eb23f 100644 --- a/fs/xfs/xfs_btree.c +++ b/fs/xfs/xfs_btree.c | |||
@@ -730,8 +730,8 @@ xfs_btree_readahead_lblock( | |||
730 | struct xfs_btree_block *block) | 730 | struct xfs_btree_block *block) |
731 | { | 731 | { |
732 | int rval = 0; | 732 | int rval = 0; |
733 | xfs_fsblock_t left = be64_to_cpu(block->bb_u.l.bb_leftsib); | 733 | xfs_dfsbno_t left = be64_to_cpu(block->bb_u.l.bb_leftsib); |
734 | xfs_fsblock_t right = be64_to_cpu(block->bb_u.l.bb_rightsib); | 734 | xfs_dfsbno_t right = be64_to_cpu(block->bb_u.l.bb_rightsib); |
735 | 735 | ||
736 | if ((lr & XFS_BTCUR_LEFTRA) && left != NULLDFSBNO) { | 736 | if ((lr & XFS_BTCUR_LEFTRA) && left != NULLDFSBNO) { |
737 | xfs_btree_reada_bufl(cur->bc_mp, left, 1); | 737 | xfs_btree_reada_bufl(cur->bc_mp, left, 1); |
@@ -843,7 +843,7 @@ xfs_btree_ptr_is_null( | |||
843 | union xfs_btree_ptr *ptr) | 843 | union xfs_btree_ptr *ptr) |
844 | { | 844 | { |
845 | if (cur->bc_flags & XFS_BTREE_LONG_PTRS) | 845 | if (cur->bc_flags & XFS_BTREE_LONG_PTRS) |
846 | return be64_to_cpu(ptr->l) == NULLFSBLOCK; | 846 | return be64_to_cpu(ptr->l) == NULLDFSBNO; |
847 | else | 847 | else |
848 | return be32_to_cpu(ptr->s) == NULLAGBLOCK; | 848 | return be32_to_cpu(ptr->s) == NULLAGBLOCK; |
849 | } | 849 | } |
@@ -854,7 +854,7 @@ xfs_btree_set_ptr_null( | |||
854 | union xfs_btree_ptr *ptr) | 854 | union xfs_btree_ptr *ptr) |
855 | { | 855 | { |
856 | if (cur->bc_flags & XFS_BTREE_LONG_PTRS) | 856 | if (cur->bc_flags & XFS_BTREE_LONG_PTRS) |
857 | ptr->l = cpu_to_be64(NULLFSBLOCK); | 857 | ptr->l = cpu_to_be64(NULLDFSBNO); |
858 | else | 858 | else |
859 | ptr->s = cpu_to_be32(NULLAGBLOCK); | 859 | ptr->s = cpu_to_be32(NULLAGBLOCK); |
860 | } | 860 | } |
@@ -918,8 +918,8 @@ xfs_btree_init_block( | |||
918 | new->bb_numrecs = cpu_to_be16(numrecs); | 918 | new->bb_numrecs = cpu_to_be16(numrecs); |
919 | 919 | ||
920 | if (cur->bc_flags & XFS_BTREE_LONG_PTRS) { | 920 | if (cur->bc_flags & XFS_BTREE_LONG_PTRS) { |
921 | new->bb_u.l.bb_leftsib = cpu_to_be64(NULLFSBLOCK); | 921 | new->bb_u.l.bb_leftsib = cpu_to_be64(NULLDFSBNO); |
922 | new->bb_u.l.bb_rightsib = cpu_to_be64(NULLFSBLOCK); | 922 | new->bb_u.l.bb_rightsib = cpu_to_be64(NULLDFSBNO); |
923 | } else { | 923 | } else { |
924 | new->bb_u.s.bb_leftsib = cpu_to_be32(NULLAGBLOCK); | 924 | new->bb_u.s.bb_leftsib = cpu_to_be32(NULLAGBLOCK); |
925 | new->bb_u.s.bb_rightsib = cpu_to_be32(NULLAGBLOCK); | 925 | new->bb_u.s.bb_rightsib = cpu_to_be32(NULLAGBLOCK); |
@@ -960,7 +960,7 @@ xfs_btree_buf_to_ptr( | |||
960 | ptr->l = cpu_to_be64(XFS_DADDR_TO_FSB(cur->bc_mp, | 960 | ptr->l = cpu_to_be64(XFS_DADDR_TO_FSB(cur->bc_mp, |
961 | XFS_BUF_ADDR(bp))); | 961 | XFS_BUF_ADDR(bp))); |
962 | else { | 962 | else { |
963 | ptr->s = cpu_to_be32(XFS_DADDR_TO_AGBNO(cur->bc_mp, | 963 | ptr->s = cpu_to_be32(xfs_daddr_to_agbno(cur->bc_mp, |
964 | XFS_BUF_ADDR(bp))); | 964 | XFS_BUF_ADDR(bp))); |
965 | } | 965 | } |
966 | } | 966 | } |
@@ -971,7 +971,7 @@ xfs_btree_ptr_to_daddr( | |||
971 | union xfs_btree_ptr *ptr) | 971 | union xfs_btree_ptr *ptr) |
972 | { | 972 | { |
973 | if (cur->bc_flags & XFS_BTREE_LONG_PTRS) { | 973 | if (cur->bc_flags & XFS_BTREE_LONG_PTRS) { |
974 | ASSERT(be64_to_cpu(ptr->l) != NULLFSBLOCK); | 974 | ASSERT(be64_to_cpu(ptr->l) != NULLDFSBNO); |
975 | 975 | ||
976 | return XFS_FSB_TO_DADDR(cur->bc_mp, be64_to_cpu(ptr->l)); | 976 | return XFS_FSB_TO_DADDR(cur->bc_mp, be64_to_cpu(ptr->l)); |
977 | } else { | 977 | } else { |
@@ -2454,7 +2454,7 @@ xfs_btree_new_iroot( | |||
2454 | xfs_btree_log_ptrs(cur, cbp, 1, be16_to_cpu(cblock->bb_numrecs)); | 2454 | xfs_btree_log_ptrs(cur, cbp, 1, be16_to_cpu(cblock->bb_numrecs)); |
2455 | 2455 | ||
2456 | *logflags |= | 2456 | *logflags |= |
2457 | XFS_ILOG_CORE | XFS_ILOG_FBROOT(cur->bc_private.b.whichfork); | 2457 | XFS_ILOG_CORE | xfs_ilog_fbroot(cur->bc_private.b.whichfork); |
2458 | *stat = 1; | 2458 | *stat = 1; |
2459 | XFS_BTREE_TRACE_CURSOR(cur, XBT_EXIT); | 2459 | XFS_BTREE_TRACE_CURSOR(cur, XBT_EXIT); |
2460 | return 0; | 2460 | return 0; |
@@ -3048,7 +3048,7 @@ xfs_btree_kill_iroot( | |||
3048 | cur->bc_bufs[level - 1] = NULL; | 3048 | cur->bc_bufs[level - 1] = NULL; |
3049 | be16_add_cpu(&block->bb_level, -1); | 3049 | be16_add_cpu(&block->bb_level, -1); |
3050 | xfs_trans_log_inode(cur->bc_tp, ip, | 3050 | xfs_trans_log_inode(cur->bc_tp, ip, |
3051 | XFS_ILOG_CORE | XFS_ILOG_FBROOT(cur->bc_private.b.whichfork)); | 3051 | XFS_ILOG_CORE | xfs_ilog_fbroot(cur->bc_private.b.whichfork)); |
3052 | cur->bc_nlevels--; | 3052 | cur->bc_nlevels--; |
3053 | out0: | 3053 | out0: |
3054 | XFS_BTREE_TRACE_CURSOR(cur, XBT_EXIT); | 3054 | XFS_BTREE_TRACE_CURSOR(cur, XBT_EXIT); |
diff --git a/fs/xfs/xfs_da_btree.c b/fs/xfs/xfs_da_btree.c index a11a8390bf6c..c45f74ff1a5b 100644 --- a/fs/xfs/xfs_da_btree.c +++ b/fs/xfs/xfs_da_btree.c | |||
@@ -1597,7 +1597,7 @@ xfs_da_grow_inode(xfs_da_args_t *args, xfs_dablk_t *new_blkno) | |||
1597 | nmap = 1; | 1597 | nmap = 1; |
1598 | ASSERT(args->firstblock != NULL); | 1598 | ASSERT(args->firstblock != NULL); |
1599 | if ((error = xfs_bmapi(tp, dp, bno, count, | 1599 | if ((error = xfs_bmapi(tp, dp, bno, count, |
1600 | XFS_BMAPI_AFLAG(w)|XFS_BMAPI_WRITE|XFS_BMAPI_METADATA| | 1600 | xfs_bmapi_aflag(w)|XFS_BMAPI_WRITE|XFS_BMAPI_METADATA| |
1601 | XFS_BMAPI_CONTIG, | 1601 | XFS_BMAPI_CONTIG, |
1602 | args->firstblock, args->total, &map, &nmap, | 1602 | args->firstblock, args->total, &map, &nmap, |
1603 | args->flist, NULL))) { | 1603 | args->flist, NULL))) { |
@@ -1618,7 +1618,7 @@ xfs_da_grow_inode(xfs_da_args_t *args, xfs_dablk_t *new_blkno) | |||
1618 | nmap = MIN(XFS_BMAP_MAX_NMAP, count); | 1618 | nmap = MIN(XFS_BMAP_MAX_NMAP, count); |
1619 | c = (int)(bno + count - b); | 1619 | c = (int)(bno + count - b); |
1620 | if ((error = xfs_bmapi(tp, dp, b, c, | 1620 | if ((error = xfs_bmapi(tp, dp, b, c, |
1621 | XFS_BMAPI_AFLAG(w)|XFS_BMAPI_WRITE| | 1621 | xfs_bmapi_aflag(w)|XFS_BMAPI_WRITE| |
1622 | XFS_BMAPI_METADATA, | 1622 | XFS_BMAPI_METADATA, |
1623 | args->firstblock, args->total, | 1623 | args->firstblock, args->total, |
1624 | &mapp[mapi], &nmap, args->flist, | 1624 | &mapp[mapi], &nmap, args->flist, |
@@ -1882,7 +1882,7 @@ xfs_da_shrink_inode(xfs_da_args_t *args, xfs_dablk_t dead_blkno, | |||
1882 | * the last block to the place we want to kill. | 1882 | * the last block to the place we want to kill. |
1883 | */ | 1883 | */ |
1884 | if ((error = xfs_bunmapi(tp, dp, dead_blkno, count, | 1884 | if ((error = xfs_bunmapi(tp, dp, dead_blkno, count, |
1885 | XFS_BMAPI_AFLAG(w)|XFS_BMAPI_METADATA, | 1885 | xfs_bmapi_aflag(w)|XFS_BMAPI_METADATA, |
1886 | 0, args->firstblock, args->flist, NULL, | 1886 | 0, args->firstblock, args->flist, NULL, |
1887 | &done)) == ENOSPC) { | 1887 | &done)) == ENOSPC) { |
1888 | if (w != XFS_DATA_FORK) | 1888 | if (w != XFS_DATA_FORK) |
@@ -1987,7 +1987,7 @@ xfs_da_do_buf( | |||
1987 | if ((error = xfs_bmapi(trans, dp, (xfs_fileoff_t)bno, | 1987 | if ((error = xfs_bmapi(trans, dp, (xfs_fileoff_t)bno, |
1988 | nfsb, | 1988 | nfsb, |
1989 | XFS_BMAPI_METADATA | | 1989 | XFS_BMAPI_METADATA | |
1990 | XFS_BMAPI_AFLAG(whichfork), | 1990 | xfs_bmapi_aflag(whichfork), |
1991 | NULL, 0, mapp, &nmap, NULL, NULL))) | 1991 | NULL, 0, mapp, &nmap, NULL, NULL))) |
1992 | goto exit0; | 1992 | goto exit0; |
1993 | } | 1993 | } |
diff --git a/fs/xfs/xfs_dfrag.c b/fs/xfs/xfs_dfrag.c index b4c1ee713492..f8278cfcc1d3 100644 --- a/fs/xfs/xfs_dfrag.c +++ b/fs/xfs/xfs_dfrag.c | |||
@@ -55,17 +55,11 @@ xfs_swapext( | |||
55 | struct file *file, *target_file; | 55 | struct file *file, *target_file; |
56 | int error = 0; | 56 | int error = 0; |
57 | 57 | ||
58 | sxp = kmem_alloc(sizeof(xfs_swapext_t), KM_MAYFAIL); | ||
59 | if (!sxp) { | ||
60 | error = XFS_ERROR(ENOMEM); | ||
61 | goto out; | ||
62 | } | ||
63 | |||
64 | /* Pull information for the target fd */ | 58 | /* Pull information for the target fd */ |
65 | file = fget((int)sxp->sx_fdtarget); | 59 | file = fget((int)sxp->sx_fdtarget); |
66 | if (!file) { | 60 | if (!file) { |
67 | error = XFS_ERROR(EINVAL); | 61 | error = XFS_ERROR(EINVAL); |
68 | goto out_free_sxp; | 62 | goto out; |
69 | } | 63 | } |
70 | 64 | ||
71 | if (!(file->f_mode & FMODE_WRITE) || (file->f_flags & O_APPEND)) { | 65 | if (!(file->f_mode & FMODE_WRITE) || (file->f_flags & O_APPEND)) { |
@@ -109,8 +103,6 @@ xfs_swapext( | |||
109 | fput(target_file); | 103 | fput(target_file); |
110 | out_put_file: | 104 | out_put_file: |
111 | fput(file); | 105 | fput(file); |
112 | out_free_sxp: | ||
113 | kmem_free(sxp); | ||
114 | out: | 106 | out: |
115 | return error; | 107 | return error; |
116 | } | 108 | } |
diff --git a/fs/xfs/xfs_dir2_block.c b/fs/xfs/xfs_dir2_block.c index e2fa0a1d8e96..e1f0a06aaf04 100644 --- a/fs/xfs/xfs_dir2_block.c +++ b/fs/xfs/xfs_dir2_block.c | |||
@@ -517,9 +517,9 @@ xfs_dir2_block_getdents( | |||
517 | /* | 517 | /* |
518 | * If it didn't fit, set the final offset to here & return. | 518 | * If it didn't fit, set the final offset to here & return. |
519 | */ | 519 | */ |
520 | if (filldir(dirent, dep->name, dep->namelen, cook, | 520 | if (filldir(dirent, dep->name, dep->namelen, cook & 0x7fffffff, |
521 | ino, DT_UNKNOWN)) { | 521 | ino, DT_UNKNOWN)) { |
522 | *offset = cook; | 522 | *offset = cook & 0x7fffffff; |
523 | xfs_da_brelse(NULL, bp); | 523 | xfs_da_brelse(NULL, bp); |
524 | return 0; | 524 | return 0; |
525 | } | 525 | } |
@@ -529,7 +529,8 @@ xfs_dir2_block_getdents( | |||
529 | * Reached the end of the block. | 529 | * Reached the end of the block. |
530 | * Set the offset to a non-existent block 1 and return. | 530 | * Set the offset to a non-existent block 1 and return. |
531 | */ | 531 | */ |
532 | *offset = xfs_dir2_db_off_to_dataptr(mp, mp->m_dirdatablk + 1, 0); | 532 | *offset = xfs_dir2_db_off_to_dataptr(mp, mp->m_dirdatablk + 1, 0) & |
533 | 0x7fffffff; | ||
533 | xfs_da_brelse(NULL, bp); | 534 | xfs_da_brelse(NULL, bp); |
534 | return 0; | 535 | return 0; |
535 | } | 536 | } |
diff --git a/fs/xfs/xfs_dir2_leaf.c b/fs/xfs/xfs_dir2_leaf.c index 93535992cb60..ef805a374eec 100644 --- a/fs/xfs/xfs_dir2_leaf.c +++ b/fs/xfs/xfs_dir2_leaf.c | |||
@@ -1092,7 +1092,7 @@ xfs_dir2_leaf_getdents( | |||
1092 | * Won't fit. Return to caller. | 1092 | * Won't fit. Return to caller. |
1093 | */ | 1093 | */ |
1094 | if (filldir(dirent, dep->name, dep->namelen, | 1094 | if (filldir(dirent, dep->name, dep->namelen, |
1095 | xfs_dir2_byte_to_dataptr(mp, curoff), | 1095 | xfs_dir2_byte_to_dataptr(mp, curoff) & 0x7fffffff, |
1096 | ino, DT_UNKNOWN)) | 1096 | ino, DT_UNKNOWN)) |
1097 | break; | 1097 | break; |
1098 | 1098 | ||
@@ -1108,9 +1108,9 @@ xfs_dir2_leaf_getdents( | |||
1108 | * All done. Set output offset value to current offset. | 1108 | * All done. Set output offset value to current offset. |
1109 | */ | 1109 | */ |
1110 | if (curoff > xfs_dir2_dataptr_to_byte(mp, XFS_DIR2_MAX_DATAPTR)) | 1110 | if (curoff > xfs_dir2_dataptr_to_byte(mp, XFS_DIR2_MAX_DATAPTR)) |
1111 | *offset = XFS_DIR2_MAX_DATAPTR; | 1111 | *offset = XFS_DIR2_MAX_DATAPTR & 0x7fffffff; |
1112 | else | 1112 | else |
1113 | *offset = xfs_dir2_byte_to_dataptr(mp, curoff); | 1113 | *offset = xfs_dir2_byte_to_dataptr(mp, curoff) & 0x7fffffff; |
1114 | kmem_free(map); | 1114 | kmem_free(map); |
1115 | if (bp) | 1115 | if (bp) |
1116 | xfs_da_brelse(NULL, bp); | 1116 | xfs_da_brelse(NULL, bp); |
diff --git a/fs/xfs/xfs_dir2_sf.c b/fs/xfs/xfs_dir2_sf.c index b46af0013ec9..a8a8a6efad5b 100644 --- a/fs/xfs/xfs_dir2_sf.c +++ b/fs/xfs/xfs_dir2_sf.c | |||
@@ -752,8 +752,8 @@ xfs_dir2_sf_getdents( | |||
752 | #if XFS_BIG_INUMS | 752 | #if XFS_BIG_INUMS |
753 | ino += mp->m_inoadd; | 753 | ino += mp->m_inoadd; |
754 | #endif | 754 | #endif |
755 | if (filldir(dirent, ".", 1, dot_offset, ino, DT_DIR)) { | 755 | if (filldir(dirent, ".", 1, dot_offset & 0x7fffffff, ino, DT_DIR)) { |
756 | *offset = dot_offset; | 756 | *offset = dot_offset & 0x7fffffff; |
757 | return 0; | 757 | return 0; |
758 | } | 758 | } |
759 | } | 759 | } |
@@ -766,8 +766,8 @@ xfs_dir2_sf_getdents( | |||
766 | #if XFS_BIG_INUMS | 766 | #if XFS_BIG_INUMS |
767 | ino += mp->m_inoadd; | 767 | ino += mp->m_inoadd; |
768 | #endif | 768 | #endif |
769 | if (filldir(dirent, "..", 2, dotdot_offset, ino, DT_DIR)) { | 769 | if (filldir(dirent, "..", 2, dotdot_offset & 0x7fffffff, ino, DT_DIR)) { |
770 | *offset = dotdot_offset; | 770 | *offset = dotdot_offset & 0x7fffffff; |
771 | return 0; | 771 | return 0; |
772 | } | 772 | } |
773 | } | 773 | } |
@@ -791,14 +791,15 @@ xfs_dir2_sf_getdents( | |||
791 | #endif | 791 | #endif |
792 | 792 | ||
793 | if (filldir(dirent, sfep->name, sfep->namelen, | 793 | if (filldir(dirent, sfep->name, sfep->namelen, |
794 | off, ino, DT_UNKNOWN)) { | 794 | off & 0x7fffffff, ino, DT_UNKNOWN)) { |
795 | *offset = off; | 795 | *offset = off & 0x7fffffff; |
796 | return 0; | 796 | return 0; |
797 | } | 797 | } |
798 | sfep = xfs_dir2_sf_nextentry(sfp, sfep); | 798 | sfep = xfs_dir2_sf_nextentry(sfp, sfep); |
799 | } | 799 | } |
800 | 800 | ||
801 | *offset = xfs_dir2_db_off_to_dataptr(mp, mp->m_dirdatablk + 1, 0); | 801 | *offset = xfs_dir2_db_off_to_dataptr(mp, mp->m_dirdatablk + 1, 0) & |
802 | 0x7fffffff; | ||
802 | return 0; | 803 | return 0; |
803 | } | 804 | } |
804 | 805 | ||
diff --git a/fs/xfs/xfs_ialloc.c b/fs/xfs/xfs_ialloc.c index e6ebbaeb4dc6..ab016e5ae7be 100644 --- a/fs/xfs/xfs_ialloc.c +++ b/fs/xfs/xfs_ialloc.c | |||
@@ -357,7 +357,7 @@ xfs_ialloc_ag_alloc( | |||
357 | int ioffset = i << args.mp->m_sb.sb_inodelog; | 357 | int ioffset = i << args.mp->m_sb.sb_inodelog; |
358 | uint isize = sizeof(struct xfs_dinode); | 358 | uint isize = sizeof(struct xfs_dinode); |
359 | 359 | ||
360 | free = XFS_MAKE_IPTR(args.mp, fbuf, i); | 360 | free = xfs_make_iptr(args.mp, fbuf, i); |
361 | free->di_magic = cpu_to_be16(XFS_DINODE_MAGIC); | 361 | free->di_magic = cpu_to_be16(XFS_DINODE_MAGIC); |
362 | free->di_version = version; | 362 | free->di_version = version; |
363 | free->di_gen = cpu_to_be32(gen); | 363 | free->di_gen = cpu_to_be32(gen); |
@@ -937,7 +937,7 @@ nextag: | |||
937 | } | 937 | } |
938 | } | 938 | } |
939 | } | 939 | } |
940 | offset = XFS_IALLOC_FIND_FREE(&rec.ir_free); | 940 | offset = xfs_ialloc_find_free(&rec.ir_free); |
941 | ASSERT(offset >= 0); | 941 | ASSERT(offset >= 0); |
942 | ASSERT(offset < XFS_INODES_PER_CHUNK); | 942 | ASSERT(offset < XFS_INODES_PER_CHUNK); |
943 | ASSERT((XFS_AGINO_TO_OFFSET(mp, rec.ir_startino) % | 943 | ASSERT((XFS_AGINO_TO_OFFSET(mp, rec.ir_startino) % |
@@ -1279,7 +1279,7 @@ xfs_imap( | |||
1279 | offset = XFS_INO_TO_OFFSET(mp, ino); | 1279 | offset = XFS_INO_TO_OFFSET(mp, ino); |
1280 | ASSERT(offset < mp->m_sb.sb_inopblock); | 1280 | ASSERT(offset < mp->m_sb.sb_inopblock); |
1281 | 1281 | ||
1282 | cluster_agbno = XFS_DADDR_TO_AGBNO(mp, imap->im_blkno); | 1282 | cluster_agbno = xfs_daddr_to_agbno(mp, imap->im_blkno); |
1283 | offset += (agbno - cluster_agbno) * mp->m_sb.sb_inopblock; | 1283 | offset += (agbno - cluster_agbno) * mp->m_sb.sb_inopblock; |
1284 | 1284 | ||
1285 | imap->im_len = XFS_FSB_TO_BB(mp, blks_per_cluster); | 1285 | imap->im_len = XFS_FSB_TO_BB(mp, blks_per_cluster); |
diff --git a/fs/xfs/xfs_ialloc.h b/fs/xfs/xfs_ialloc.h index 50f558a4e0a8..aeee8278f92c 100644 --- a/fs/xfs/xfs_ialloc.h +++ b/fs/xfs/xfs_ialloc.h | |||
@@ -39,7 +39,6 @@ struct xfs_trans; | |||
39 | /* | 39 | /* |
40 | * Make an inode pointer out of the buffer/offset. | 40 | * Make an inode pointer out of the buffer/offset. |
41 | */ | 41 | */ |
42 | #define XFS_MAKE_IPTR(mp,b,o) xfs_make_iptr(mp,b,o) | ||
43 | static inline struct xfs_dinode * | 42 | static inline struct xfs_dinode * |
44 | xfs_make_iptr(struct xfs_mount *mp, struct xfs_buf *b, int o) | 43 | xfs_make_iptr(struct xfs_mount *mp, struct xfs_buf *b, int o) |
45 | { | 44 | { |
@@ -50,7 +49,6 @@ xfs_make_iptr(struct xfs_mount *mp, struct xfs_buf *b, int o) | |||
50 | /* | 49 | /* |
51 | * Find a free (set) bit in the inode bitmask. | 50 | * Find a free (set) bit in the inode bitmask. |
52 | */ | 51 | */ |
53 | #define XFS_IALLOC_FIND_FREE(fp) xfs_ialloc_find_free(fp) | ||
54 | static inline int xfs_ialloc_find_free(xfs_inofree_t *fp) | 52 | static inline int xfs_ialloc_find_free(xfs_inofree_t *fp) |
55 | { | 53 | { |
56 | return xfs_lowbit64(*fp); | 54 | return xfs_lowbit64(*fp); |
diff --git a/fs/xfs/xfs_ialloc_btree.h b/fs/xfs/xfs_ialloc_btree.h index 37e5dd01a577..5580e255ff06 100644 --- a/fs/xfs/xfs_ialloc_btree.h +++ b/fs/xfs/xfs_ialloc_btree.h | |||
@@ -36,7 +36,6 @@ typedef __uint64_t xfs_inofree_t; | |||
36 | #define XFS_INODES_PER_CHUNK_LOG (XFS_NBBYLOG + 3) | 36 | #define XFS_INODES_PER_CHUNK_LOG (XFS_NBBYLOG + 3) |
37 | #define XFS_INOBT_ALL_FREE ((xfs_inofree_t)-1) | 37 | #define XFS_INOBT_ALL_FREE ((xfs_inofree_t)-1) |
38 | 38 | ||
39 | #define XFS_INOBT_MASKN(i,n) xfs_inobt_maskn(i,n) | ||
40 | static inline xfs_inofree_t xfs_inobt_maskn(int i, int n) | 39 | static inline xfs_inofree_t xfs_inobt_maskn(int i, int n) |
41 | { | 40 | { |
42 | return (((n) >= XFS_INODES_PER_CHUNK ? \ | 41 | return (((n) >= XFS_INODES_PER_CHUNK ? \ |
diff --git a/fs/xfs/xfs_inode.c b/fs/xfs/xfs_inode.c index 5a5e035e5d38..e7ae08d1df48 100644 --- a/fs/xfs/xfs_inode.c +++ b/fs/xfs/xfs_inode.c | |||
@@ -424,6 +424,19 @@ xfs_iformat( | |||
424 | case XFS_DINODE_FMT_LOCAL: | 424 | case XFS_DINODE_FMT_LOCAL: |
425 | atp = (xfs_attr_shortform_t *)XFS_DFORK_APTR(dip); | 425 | atp = (xfs_attr_shortform_t *)XFS_DFORK_APTR(dip); |
426 | size = be16_to_cpu(atp->hdr.totsize); | 426 | size = be16_to_cpu(atp->hdr.totsize); |
427 | |||
428 | if (unlikely(size < sizeof(struct xfs_attr_sf_hdr))) { | ||
429 | xfs_fs_repair_cmn_err(CE_WARN, ip->i_mount, | ||
430 | "corrupt inode %Lu " | ||
431 | "(bad attr fork size %Ld).", | ||
432 | (unsigned long long) ip->i_ino, | ||
433 | (long long) size); | ||
434 | XFS_CORRUPTION_ERROR("xfs_iformat(8)", | ||
435 | XFS_ERRLEVEL_LOW, | ||
436 | ip->i_mount, dip); | ||
437 | return XFS_ERROR(EFSCORRUPTED); | ||
438 | } | ||
439 | |||
427 | error = xfs_iformat_local(ip, dip, XFS_ATTR_FORK, size); | 440 | error = xfs_iformat_local(ip, dip, XFS_ATTR_FORK, size); |
428 | break; | 441 | break; |
429 | case XFS_DINODE_FMT_EXTENTS: | 442 | case XFS_DINODE_FMT_EXTENTS: |
@@ -1601,10 +1614,10 @@ xfs_itruncate_finish( | |||
1601 | * in this file with garbage in them once recovery | 1614 | * in this file with garbage in them once recovery |
1602 | * runs. | 1615 | * runs. |
1603 | */ | 1616 | */ |
1604 | XFS_BMAP_INIT(&free_list, &first_block); | 1617 | xfs_bmap_init(&free_list, &first_block); |
1605 | error = xfs_bunmapi(ntp, ip, | 1618 | error = xfs_bunmapi(ntp, ip, |
1606 | first_unmap_block, unmap_len, | 1619 | first_unmap_block, unmap_len, |
1607 | XFS_BMAPI_AFLAG(fork) | | 1620 | xfs_bmapi_aflag(fork) | |
1608 | (sync ? 0 : XFS_BMAPI_ASYNC), | 1621 | (sync ? 0 : XFS_BMAPI_ASYNC), |
1609 | XFS_ITRUNC_MAX_EXTENTS, | 1622 | XFS_ITRUNC_MAX_EXTENTS, |
1610 | &first_block, &free_list, | 1623 | &first_block, &free_list, |
@@ -2557,7 +2570,7 @@ xfs_iextents_copy( | |||
2557 | for (i = 0; i < nrecs; i++) { | 2570 | for (i = 0; i < nrecs; i++) { |
2558 | xfs_bmbt_rec_host_t *ep = xfs_iext_get_ext(ifp, i); | 2571 | xfs_bmbt_rec_host_t *ep = xfs_iext_get_ext(ifp, i); |
2559 | start_block = xfs_bmbt_get_startblock(ep); | 2572 | start_block = xfs_bmbt_get_startblock(ep); |
2560 | if (ISNULLSTARTBLOCK(start_block)) { | 2573 | if (isnullstartblock(start_block)) { |
2561 | /* | 2574 | /* |
2562 | * It's a delayed allocation extent, so skip it. | 2575 | * It's a delayed allocation extent, so skip it. |
2563 | */ | 2576 | */ |
diff --git a/fs/xfs/xfs_inode_item.h b/fs/xfs/xfs_inode_item.h index 1ff04cc323ad..9957d0602d54 100644 --- a/fs/xfs/xfs_inode_item.h +++ b/fs/xfs/xfs_inode_item.h | |||
@@ -111,20 +111,16 @@ typedef struct xfs_inode_log_format_64 { | |||
111 | 111 | ||
112 | #define XFS_ILI_IOLOCKED_ANY (XFS_ILI_IOLOCKED_EXCL | XFS_ILI_IOLOCKED_SHARED) | 112 | #define XFS_ILI_IOLOCKED_ANY (XFS_ILI_IOLOCKED_EXCL | XFS_ILI_IOLOCKED_SHARED) |
113 | 113 | ||
114 | |||
115 | #define XFS_ILOG_FBROOT(w) xfs_ilog_fbroot(w) | ||
116 | static inline int xfs_ilog_fbroot(int w) | 114 | static inline int xfs_ilog_fbroot(int w) |
117 | { | 115 | { |
118 | return (w == XFS_DATA_FORK ? XFS_ILOG_DBROOT : XFS_ILOG_ABROOT); | 116 | return (w == XFS_DATA_FORK ? XFS_ILOG_DBROOT : XFS_ILOG_ABROOT); |
119 | } | 117 | } |
120 | 118 | ||
121 | #define XFS_ILOG_FEXT(w) xfs_ilog_fext(w) | ||
122 | static inline int xfs_ilog_fext(int w) | 119 | static inline int xfs_ilog_fext(int w) |
123 | { | 120 | { |
124 | return (w == XFS_DATA_FORK ? XFS_ILOG_DEXT : XFS_ILOG_AEXT); | 121 | return (w == XFS_DATA_FORK ? XFS_ILOG_DEXT : XFS_ILOG_AEXT); |
125 | } | 122 | } |
126 | 123 | ||
127 | #define XFS_ILOG_FDATA(w) xfs_ilog_fdata(w) | ||
128 | static inline int xfs_ilog_fdata(int w) | 124 | static inline int xfs_ilog_fdata(int w) |
129 | { | 125 | { |
130 | return (w == XFS_DATA_FORK ? XFS_ILOG_DDATA : XFS_ILOG_ADATA); | 126 | return (w == XFS_DATA_FORK ? XFS_ILOG_DDATA : XFS_ILOG_ADATA); |
diff --git a/fs/xfs/xfs_iomap.c b/fs/xfs/xfs_iomap.c index 911062cf73a6..08ce72316bfe 100644 --- a/fs/xfs/xfs_iomap.c +++ b/fs/xfs/xfs_iomap.c | |||
@@ -155,7 +155,7 @@ xfs_imap_to_bmap( | |||
155 | iomapp->iomap_bn = IOMAP_DADDR_NULL; | 155 | iomapp->iomap_bn = IOMAP_DADDR_NULL; |
156 | iomapp->iomap_flags |= IOMAP_DELAY; | 156 | iomapp->iomap_flags |= IOMAP_DELAY; |
157 | } else { | 157 | } else { |
158 | iomapp->iomap_bn = XFS_FSB_TO_DB(ip, start_block); | 158 | iomapp->iomap_bn = xfs_fsb_to_db(ip, start_block); |
159 | if (ISUNWRITTEN(imap)) | 159 | if (ISUNWRITTEN(imap)) |
160 | iomapp->iomap_flags |= IOMAP_UNWRITTEN; | 160 | iomapp->iomap_flags |= IOMAP_UNWRITTEN; |
161 | } | 161 | } |
@@ -261,7 +261,7 @@ xfs_iomap( | |||
261 | xfs_iunlock(ip, lockmode); | 261 | xfs_iunlock(ip, lockmode); |
262 | lockmode = 0; | 262 | lockmode = 0; |
263 | 263 | ||
264 | if (nimaps && !ISNULLSTARTBLOCK(imap.br_startblock)) { | 264 | if (nimaps && !isnullstartblock(imap.br_startblock)) { |
265 | xfs_iomap_map_trace(XFS_IOMAP_WRITE_MAP, ip, | 265 | xfs_iomap_map_trace(XFS_IOMAP_WRITE_MAP, ip, |
266 | offset, count, iomapp, &imap, flags); | 266 | offset, count, iomapp, &imap, flags); |
267 | break; | 267 | break; |
@@ -491,7 +491,7 @@ xfs_iomap_write_direct( | |||
491 | /* | 491 | /* |
492 | * Issue the xfs_bmapi() call to allocate the blocks | 492 | * Issue the xfs_bmapi() call to allocate the blocks |
493 | */ | 493 | */ |
494 | XFS_BMAP_INIT(&free_list, &firstfsb); | 494 | xfs_bmap_init(&free_list, &firstfsb); |
495 | nimaps = 1; | 495 | nimaps = 1; |
496 | error = xfs_bmapi(tp, ip, offset_fsb, count_fsb, bmapi_flag, | 496 | error = xfs_bmapi(tp, ip, offset_fsb, count_fsb, bmapi_flag, |
497 | &firstfsb, 0, &imap, &nimaps, &free_list, NULL); | 497 | &firstfsb, 0, &imap, &nimaps, &free_list, NULL); |
@@ -751,7 +751,7 @@ xfs_iomap_write_allocate( | |||
751 | xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL); | 751 | xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL); |
752 | xfs_trans_ihold(tp, ip); | 752 | xfs_trans_ihold(tp, ip); |
753 | 753 | ||
754 | XFS_BMAP_INIT(&free_list, &first_block); | 754 | xfs_bmap_init(&free_list, &first_block); |
755 | 755 | ||
756 | /* | 756 | /* |
757 | * it is possible that the extents have changed since | 757 | * it is possible that the extents have changed since |
@@ -911,7 +911,7 @@ xfs_iomap_write_unwritten( | |||
911 | /* | 911 | /* |
912 | * Modify the unwritten extent state of the buffer. | 912 | * Modify the unwritten extent state of the buffer. |
913 | */ | 913 | */ |
914 | XFS_BMAP_INIT(&free_list, &firstfsb); | 914 | xfs_bmap_init(&free_list, &firstfsb); |
915 | nimaps = 1; | 915 | nimaps = 1; |
916 | error = xfs_bmapi(tp, ip, offset_fsb, count_fsb, | 916 | error = xfs_bmapi(tp, ip, offset_fsb, count_fsb, |
917 | XFS_BMAPI_WRITE|XFS_BMAPI_CONVERT, &firstfsb, | 917 | XFS_BMAPI_WRITE|XFS_BMAPI_CONVERT, &firstfsb, |
diff --git a/fs/xfs/xfs_itable.c b/fs/xfs/xfs_itable.c index e19d0a8d5618..cf98a805ec90 100644 --- a/fs/xfs/xfs_itable.c +++ b/fs/xfs/xfs_itable.c | |||
@@ -453,7 +453,7 @@ xfs_bulkstat( | |||
453 | (chunkidx = agino - gino + 1) < | 453 | (chunkidx = agino - gino + 1) < |
454 | XFS_INODES_PER_CHUNK && | 454 | XFS_INODES_PER_CHUNK && |
455 | /* there are some left allocated */ | 455 | /* there are some left allocated */ |
456 | XFS_INOBT_MASKN(chunkidx, | 456 | xfs_inobt_maskn(chunkidx, |
457 | XFS_INODES_PER_CHUNK - chunkidx) & ~gfree) { | 457 | XFS_INODES_PER_CHUNK - chunkidx) & ~gfree) { |
458 | /* | 458 | /* |
459 | * Grab the chunk record. Mark all the | 459 | * Grab the chunk record. Mark all the |
@@ -464,7 +464,7 @@ xfs_bulkstat( | |||
464 | if (XFS_INOBT_MASK(i) & ~gfree) | 464 | if (XFS_INOBT_MASK(i) & ~gfree) |
465 | gcnt++; | 465 | gcnt++; |
466 | } | 466 | } |
467 | gfree |= XFS_INOBT_MASKN(0, chunkidx); | 467 | gfree |= xfs_inobt_maskn(0, chunkidx); |
468 | irbp->ir_startino = gino; | 468 | irbp->ir_startino = gino; |
469 | irbp->ir_freecount = gcnt; | 469 | irbp->ir_freecount = gcnt; |
470 | irbp->ir_free = gfree; | 470 | irbp->ir_free = gfree; |
@@ -535,7 +535,7 @@ xfs_bulkstat( | |||
535 | chunkidx < XFS_INODES_PER_CHUNK; | 535 | chunkidx < XFS_INODES_PER_CHUNK; |
536 | chunkidx += nicluster, | 536 | chunkidx += nicluster, |
537 | agbno += nbcluster) { | 537 | agbno += nbcluster) { |
538 | if (XFS_INOBT_MASKN(chunkidx, | 538 | if (xfs_inobt_maskn(chunkidx, |
539 | nicluster) & ~gfree) | 539 | nicluster) & ~gfree) |
540 | xfs_btree_reada_bufs(mp, agno, | 540 | xfs_btree_reada_bufs(mp, agno, |
541 | agbno, nbcluster); | 541 | agbno, nbcluster); |
diff --git a/fs/xfs/xfs_log_recover.c b/fs/xfs/xfs_log_recover.c index 35cca98bd94c..b1047de2fffd 100644 --- a/fs/xfs/xfs_log_recover.c +++ b/fs/xfs/xfs_log_recover.c | |||
@@ -70,16 +70,21 @@ STATIC void xlog_recover_check_summary(xlog_t *); | |||
70 | xfs_buf_t * | 70 | xfs_buf_t * |
71 | xlog_get_bp( | 71 | xlog_get_bp( |
72 | xlog_t *log, | 72 | xlog_t *log, |
73 | int num_bblks) | 73 | int nbblks) |
74 | { | 74 | { |
75 | ASSERT(num_bblks > 0); | 75 | if (nbblks <= 0 || nbblks > log->l_logBBsize) { |
76 | xlog_warn("XFS: Invalid block length (0x%x) given for buffer", nbblks); | ||
77 | XFS_ERROR_REPORT("xlog_get_bp(1)", | ||
78 | XFS_ERRLEVEL_HIGH, log->l_mp); | ||
79 | return NULL; | ||
80 | } | ||
76 | 81 | ||
77 | if (log->l_sectbb_log) { | 82 | if (log->l_sectbb_log) { |
78 | if (num_bblks > 1) | 83 | if (nbblks > 1) |
79 | num_bblks += XLOG_SECTOR_ROUNDUP_BBCOUNT(log, 1); | 84 | nbblks += XLOG_SECTOR_ROUNDUP_BBCOUNT(log, 1); |
80 | num_bblks = XLOG_SECTOR_ROUNDUP_BBCOUNT(log, num_bblks); | 85 | nbblks = XLOG_SECTOR_ROUNDUP_BBCOUNT(log, nbblks); |
81 | } | 86 | } |
82 | return xfs_buf_get_noaddr(BBTOB(num_bblks), log->l_mp->m_logdev_targp); | 87 | return xfs_buf_get_noaddr(BBTOB(nbblks), log->l_mp->m_logdev_targp); |
83 | } | 88 | } |
84 | 89 | ||
85 | void | 90 | void |
@@ -102,6 +107,13 @@ xlog_bread( | |||
102 | { | 107 | { |
103 | int error; | 108 | int error; |
104 | 109 | ||
110 | if (nbblks <= 0 || nbblks > log->l_logBBsize) { | ||
111 | xlog_warn("XFS: Invalid block length (0x%x) given for buffer", nbblks); | ||
112 | XFS_ERROR_REPORT("xlog_bread(1)", | ||
113 | XFS_ERRLEVEL_HIGH, log->l_mp); | ||
114 | return EFSCORRUPTED; | ||
115 | } | ||
116 | |||
105 | if (log->l_sectbb_log) { | 117 | if (log->l_sectbb_log) { |
106 | blk_no = XLOG_SECTOR_ROUNDDOWN_BLKNO(log, blk_no); | 118 | blk_no = XLOG_SECTOR_ROUNDDOWN_BLKNO(log, blk_no); |
107 | nbblks = XLOG_SECTOR_ROUNDUP_BBCOUNT(log, nbblks); | 119 | nbblks = XLOG_SECTOR_ROUNDUP_BBCOUNT(log, nbblks); |
@@ -139,6 +151,13 @@ xlog_bwrite( | |||
139 | { | 151 | { |
140 | int error; | 152 | int error; |
141 | 153 | ||
154 | if (nbblks <= 0 || nbblks > log->l_logBBsize) { | ||
155 | xlog_warn("XFS: Invalid block length (0x%x) given for buffer", nbblks); | ||
156 | XFS_ERROR_REPORT("xlog_bwrite(1)", | ||
157 | XFS_ERRLEVEL_HIGH, log->l_mp); | ||
158 | return EFSCORRUPTED; | ||
159 | } | ||
160 | |||
142 | if (log->l_sectbb_log) { | 161 | if (log->l_sectbb_log) { |
143 | blk_no = XLOG_SECTOR_ROUNDDOWN_BLKNO(log, blk_no); | 162 | blk_no = XLOG_SECTOR_ROUNDDOWN_BLKNO(log, blk_no); |
144 | nbblks = XLOG_SECTOR_ROUNDUP_BBCOUNT(log, nbblks); | 163 | nbblks = XLOG_SECTOR_ROUNDUP_BBCOUNT(log, nbblks); |
diff --git a/fs/xfs/xfs_mount.c b/fs/xfs/xfs_mount.c index 3c97c6463a4e..35300250e86d 100644 --- a/fs/xfs/xfs_mount.c +++ b/fs/xfs/xfs_mount.c | |||
@@ -45,7 +45,6 @@ | |||
45 | #include "xfs_fsops.h" | 45 | #include "xfs_fsops.h" |
46 | #include "xfs_utils.h" | 46 | #include "xfs_utils.h" |
47 | 47 | ||
48 | STATIC int xfs_mount_log_sb(xfs_mount_t *, __int64_t); | ||
49 | STATIC int xfs_uuid_mount(xfs_mount_t *); | 48 | STATIC int xfs_uuid_mount(xfs_mount_t *); |
50 | STATIC void xfs_unmountfs_wait(xfs_mount_t *); | 49 | STATIC void xfs_unmountfs_wait(xfs_mount_t *); |
51 | 50 | ||
@@ -682,7 +681,7 @@ xfs_initialize_perag_data(xfs_mount_t *mp, xfs_agnumber_t agcount) | |||
682 | * Update alignment values based on mount options and sb values | 681 | * Update alignment values based on mount options and sb values |
683 | */ | 682 | */ |
684 | STATIC int | 683 | STATIC int |
685 | xfs_update_alignment(xfs_mount_t *mp, __uint64_t *update_flags) | 684 | xfs_update_alignment(xfs_mount_t *mp) |
686 | { | 685 | { |
687 | xfs_sb_t *sbp = &(mp->m_sb); | 686 | xfs_sb_t *sbp = &(mp->m_sb); |
688 | 687 | ||
@@ -736,11 +735,11 @@ xfs_update_alignment(xfs_mount_t *mp, __uint64_t *update_flags) | |||
736 | if (xfs_sb_version_hasdalign(sbp)) { | 735 | if (xfs_sb_version_hasdalign(sbp)) { |
737 | if (sbp->sb_unit != mp->m_dalign) { | 736 | if (sbp->sb_unit != mp->m_dalign) { |
738 | sbp->sb_unit = mp->m_dalign; | 737 | sbp->sb_unit = mp->m_dalign; |
739 | *update_flags |= XFS_SB_UNIT; | 738 | mp->m_update_flags |= XFS_SB_UNIT; |
740 | } | 739 | } |
741 | if (sbp->sb_width != mp->m_swidth) { | 740 | if (sbp->sb_width != mp->m_swidth) { |
742 | sbp->sb_width = mp->m_swidth; | 741 | sbp->sb_width = mp->m_swidth; |
743 | *update_flags |= XFS_SB_WIDTH; | 742 | mp->m_update_flags |= XFS_SB_WIDTH; |
744 | } | 743 | } |
745 | } | 744 | } |
746 | } else if ((mp->m_flags & XFS_MOUNT_NOALIGN) != XFS_MOUNT_NOALIGN && | 745 | } else if ((mp->m_flags & XFS_MOUNT_NOALIGN) != XFS_MOUNT_NOALIGN && |
@@ -905,7 +904,6 @@ xfs_mountfs( | |||
905 | xfs_sb_t *sbp = &(mp->m_sb); | 904 | xfs_sb_t *sbp = &(mp->m_sb); |
906 | xfs_inode_t *rip; | 905 | xfs_inode_t *rip; |
907 | __uint64_t resblks; | 906 | __uint64_t resblks; |
908 | __int64_t update_flags = 0LL; | ||
909 | uint quotamount, quotaflags; | 907 | uint quotamount, quotaflags; |
910 | int uuid_mounted = 0; | 908 | int uuid_mounted = 0; |
911 | int error = 0; | 909 | int error = 0; |
@@ -933,7 +931,7 @@ xfs_mountfs( | |||
933 | "XFS: correcting sb_features alignment problem"); | 931 | "XFS: correcting sb_features alignment problem"); |
934 | sbp->sb_features2 |= sbp->sb_bad_features2; | 932 | sbp->sb_features2 |= sbp->sb_bad_features2; |
935 | sbp->sb_bad_features2 = sbp->sb_features2; | 933 | sbp->sb_bad_features2 = sbp->sb_features2; |
936 | update_flags |= XFS_SB_FEATURES2 | XFS_SB_BAD_FEATURES2; | 934 | mp->m_update_flags |= XFS_SB_FEATURES2 | XFS_SB_BAD_FEATURES2; |
937 | 935 | ||
938 | /* | 936 | /* |
939 | * Re-check for ATTR2 in case it was found in bad_features2 | 937 | * Re-check for ATTR2 in case it was found in bad_features2 |
@@ -947,11 +945,11 @@ xfs_mountfs( | |||
947 | if (xfs_sb_version_hasattr2(&mp->m_sb) && | 945 | if (xfs_sb_version_hasattr2(&mp->m_sb) && |
948 | (mp->m_flags & XFS_MOUNT_NOATTR2)) { | 946 | (mp->m_flags & XFS_MOUNT_NOATTR2)) { |
949 | xfs_sb_version_removeattr2(&mp->m_sb); | 947 | xfs_sb_version_removeattr2(&mp->m_sb); |
950 | update_flags |= XFS_SB_FEATURES2; | 948 | mp->m_update_flags |= XFS_SB_FEATURES2; |
951 | 949 | ||
952 | /* update sb_versionnum for the clearing of the morebits */ | 950 | /* update sb_versionnum for the clearing of the morebits */ |
953 | if (!sbp->sb_features2) | 951 | if (!sbp->sb_features2) |
954 | update_flags |= XFS_SB_VERSIONNUM; | 952 | mp->m_update_flags |= XFS_SB_VERSIONNUM; |
955 | } | 953 | } |
956 | 954 | ||
957 | /* | 955 | /* |
@@ -960,7 +958,7 @@ xfs_mountfs( | |||
960 | * allocator alignment is within an ag, therefore ag has | 958 | * allocator alignment is within an ag, therefore ag has |
961 | * to be aligned at stripe boundary. | 959 | * to be aligned at stripe boundary. |
962 | */ | 960 | */ |
963 | error = xfs_update_alignment(mp, &update_flags); | 961 | error = xfs_update_alignment(mp); |
964 | if (error) | 962 | if (error) |
965 | goto error1; | 963 | goto error1; |
966 | 964 | ||
@@ -1137,10 +1135,12 @@ xfs_mountfs( | |||
1137 | } | 1135 | } |
1138 | 1136 | ||
1139 | /* | 1137 | /* |
1140 | * If fs is not mounted readonly, then update the superblock changes. | 1138 | * If this is a read-only mount defer the superblock updates until |
1139 | * the next remount into writeable mode. Otherwise we would never | ||
1140 | * perform the update e.g. for the root filesystem. | ||
1141 | */ | 1141 | */ |
1142 | if (update_flags && !(mp->m_flags & XFS_MOUNT_RDONLY)) { | 1142 | if (mp->m_update_flags && !(mp->m_flags & XFS_MOUNT_RDONLY)) { |
1143 | error = xfs_mount_log_sb(mp, update_flags); | 1143 | error = xfs_mount_log_sb(mp, mp->m_update_flags); |
1144 | if (error) { | 1144 | if (error) { |
1145 | cmn_err(CE_WARN, "XFS: failed to write sb changes"); | 1145 | cmn_err(CE_WARN, "XFS: failed to write sb changes"); |
1146 | goto error4; | 1146 | goto error4; |
@@ -1820,7 +1820,7 @@ xfs_uuid_mount( | |||
1820 | * be altered by the mount options, as well as any potential sb_features2 | 1820 | * be altered by the mount options, as well as any potential sb_features2 |
1821 | * fixup. Only the first superblock is updated. | 1821 | * fixup. Only the first superblock is updated. |
1822 | */ | 1822 | */ |
1823 | STATIC int | 1823 | int |
1824 | xfs_mount_log_sb( | 1824 | xfs_mount_log_sb( |
1825 | xfs_mount_t *mp, | 1825 | xfs_mount_t *mp, |
1826 | __int64_t fields) | 1826 | __int64_t fields) |
diff --git a/fs/xfs/xfs_mount.h b/fs/xfs/xfs_mount.h index c1e028467327..f5e9937f9bdb 100644 --- a/fs/xfs/xfs_mount.h +++ b/fs/xfs/xfs_mount.h | |||
@@ -44,9 +44,9 @@ typedef struct xfs_trans_reservations { | |||
44 | 44 | ||
45 | #ifndef __KERNEL__ | 45 | #ifndef __KERNEL__ |
46 | 46 | ||
47 | #define XFS_DADDR_TO_AGNO(mp,d) \ | 47 | #define xfs_daddr_to_agno(mp,d) \ |
48 | ((xfs_agnumber_t)(XFS_BB_TO_FSBT(mp, d) / (mp)->m_sb.sb_agblocks)) | 48 | ((xfs_agnumber_t)(XFS_BB_TO_FSBT(mp, d) / (mp)->m_sb.sb_agblocks)) |
49 | #define XFS_DADDR_TO_AGBNO(mp,d) \ | 49 | #define xfs_daddr_to_agbno(mp,d) \ |
50 | ((xfs_agblock_t)(XFS_BB_TO_FSBT(mp, d) % (mp)->m_sb.sb_agblocks)) | 50 | ((xfs_agblock_t)(XFS_BB_TO_FSBT(mp, d) % (mp)->m_sb.sb_agblocks)) |
51 | 51 | ||
52 | #else /* __KERNEL__ */ | 52 | #else /* __KERNEL__ */ |
@@ -327,6 +327,8 @@ typedef struct xfs_mount { | |||
327 | spinlock_t m_sync_lock; /* work item list lock */ | 327 | spinlock_t m_sync_lock; /* work item list lock */ |
328 | int m_sync_seq; /* sync thread generation no. */ | 328 | int m_sync_seq; /* sync thread generation no. */ |
329 | wait_queue_head_t m_wait_single_sync_task; | 329 | wait_queue_head_t m_wait_single_sync_task; |
330 | __int64_t m_update_flags; /* sb flags we need to update | ||
331 | on the next remount,rw */ | ||
330 | } xfs_mount_t; | 332 | } xfs_mount_t; |
331 | 333 | ||
332 | /* | 334 | /* |
@@ -439,7 +441,6 @@ void xfs_do_force_shutdown(struct xfs_mount *mp, int flags, char *fname, | |||
439 | */ | 441 | */ |
440 | #define XFS_MFSI_QUIET 0x40 /* Be silent if mount errors found */ | 442 | #define XFS_MFSI_QUIET 0x40 /* Be silent if mount errors found */ |
441 | 443 | ||
442 | #define XFS_DADDR_TO_AGNO(mp,d) xfs_daddr_to_agno(mp,d) | ||
443 | static inline xfs_agnumber_t | 444 | static inline xfs_agnumber_t |
444 | xfs_daddr_to_agno(struct xfs_mount *mp, xfs_daddr_t d) | 445 | xfs_daddr_to_agno(struct xfs_mount *mp, xfs_daddr_t d) |
445 | { | 446 | { |
@@ -448,7 +449,6 @@ xfs_daddr_to_agno(struct xfs_mount *mp, xfs_daddr_t d) | |||
448 | return (xfs_agnumber_t) ld; | 449 | return (xfs_agnumber_t) ld; |
449 | } | 450 | } |
450 | 451 | ||
451 | #define XFS_DADDR_TO_AGBNO(mp,d) xfs_daddr_to_agbno(mp,d) | ||
452 | static inline xfs_agblock_t | 452 | static inline xfs_agblock_t |
453 | xfs_daddr_to_agbno(struct xfs_mount *mp, xfs_daddr_t d) | 453 | xfs_daddr_to_agbno(struct xfs_mount *mp, xfs_daddr_t d) |
454 | { | 454 | { |
@@ -514,6 +514,7 @@ extern int xfs_mod_incore_sb_unlocked(xfs_mount_t *, xfs_sb_field_t, | |||
514 | int64_t, int); | 514 | int64_t, int); |
515 | extern int xfs_mod_incore_sb_batch(xfs_mount_t *, xfs_mod_sb_t *, | 515 | extern int xfs_mod_incore_sb_batch(xfs_mount_t *, xfs_mod_sb_t *, |
516 | uint, int); | 516 | uint, int); |
517 | extern int xfs_mount_log_sb(xfs_mount_t *, __int64_t); | ||
517 | extern struct xfs_buf *xfs_getsb(xfs_mount_t *, int); | 518 | extern struct xfs_buf *xfs_getsb(xfs_mount_t *, int); |
518 | extern int xfs_readsb(xfs_mount_t *, int); | 519 | extern int xfs_readsb(xfs_mount_t *, int); |
519 | extern void xfs_freesb(xfs_mount_t *); | 520 | extern void xfs_freesb(xfs_mount_t *); |
diff --git a/fs/xfs/xfs_rename.c b/fs/xfs/xfs_rename.c index 86471bb40fd4..58f85e9cd11d 100644 --- a/fs/xfs/xfs_rename.c +++ b/fs/xfs/xfs_rename.c | |||
@@ -147,7 +147,7 @@ xfs_rename( | |||
147 | xfs_sort_for_rename(src_dp, target_dp, src_ip, target_ip, | 147 | xfs_sort_for_rename(src_dp, target_dp, src_ip, target_ip, |
148 | inodes, &num_inodes); | 148 | inodes, &num_inodes); |
149 | 149 | ||
150 | XFS_BMAP_INIT(&free_list, &first_block); | 150 | xfs_bmap_init(&free_list, &first_block); |
151 | tp = xfs_trans_alloc(mp, XFS_TRANS_RENAME); | 151 | tp = xfs_trans_alloc(mp, XFS_TRANS_RENAME); |
152 | cancel_flags = XFS_TRANS_RELEASE_LOG_RES; | 152 | cancel_flags = XFS_TRANS_RELEASE_LOG_RES; |
153 | spaceres = XFS_RENAME_SPACE_RES(mp, target_name->len); | 153 | spaceres = XFS_RENAME_SPACE_RES(mp, target_name->len); |
diff --git a/fs/xfs/xfs_rtalloc.c b/fs/xfs/xfs_rtalloc.c index edf12c7b834c..c5bb86f3ec05 100644 --- a/fs/xfs/xfs_rtalloc.c +++ b/fs/xfs/xfs_rtalloc.c | |||
@@ -120,7 +120,7 @@ xfs_growfs_rt_alloc( | |||
120 | if ((error = xfs_trans_iget(mp, tp, ino, 0, | 120 | if ((error = xfs_trans_iget(mp, tp, ino, 0, |
121 | XFS_ILOCK_EXCL, &ip))) | 121 | XFS_ILOCK_EXCL, &ip))) |
122 | goto error_cancel; | 122 | goto error_cancel; |
123 | XFS_BMAP_INIT(&flist, &firstblock); | 123 | xfs_bmap_init(&flist, &firstblock); |
124 | /* | 124 | /* |
125 | * Allocate blocks to the bitmap file. | 125 | * Allocate blocks to the bitmap file. |
126 | */ | 126 | */ |
diff --git a/fs/xfs/xfs_rw.h b/fs/xfs/xfs_rw.h index f87db5344ce6..f76c003ec55d 100644 --- a/fs/xfs/xfs_rw.h +++ b/fs/xfs/xfs_rw.h | |||
@@ -28,7 +28,6 @@ struct xfs_mount; | |||
28 | * file is a real time file or not, because the bmap code | 28 | * file is a real time file or not, because the bmap code |
29 | * does. | 29 | * does. |
30 | */ | 30 | */ |
31 | #define XFS_FSB_TO_DB(ip,fsb) xfs_fsb_to_db(ip,fsb) | ||
32 | static inline xfs_daddr_t | 31 | static inline xfs_daddr_t |
33 | xfs_fsb_to_db(struct xfs_inode *ip, xfs_fsblock_t fsb) | 32 | xfs_fsb_to_db(struct xfs_inode *ip, xfs_fsblock_t fsb) |
34 | { | 33 | { |
diff --git a/fs/xfs/xfs_sb.h b/fs/xfs/xfs_sb.h index 1ed71916e4c9..1b017c657494 100644 --- a/fs/xfs/xfs_sb.h +++ b/fs/xfs/xfs_sb.h | |||
@@ -505,7 +505,7 @@ static inline void xfs_sb_version_removeattr2(xfs_sb_t *sbp) | |||
505 | 505 | ||
506 | #define XFS_HDR_BLOCK(mp,d) ((xfs_agblock_t)XFS_BB_TO_FSBT(mp,d)) | 506 | #define XFS_HDR_BLOCK(mp,d) ((xfs_agblock_t)XFS_BB_TO_FSBT(mp,d)) |
507 | #define XFS_DADDR_TO_FSB(mp,d) XFS_AGB_TO_FSB(mp, \ | 507 | #define XFS_DADDR_TO_FSB(mp,d) XFS_AGB_TO_FSB(mp, \ |
508 | XFS_DADDR_TO_AGNO(mp,d), XFS_DADDR_TO_AGBNO(mp,d)) | 508 | xfs_daddr_to_agno(mp,d), xfs_daddr_to_agbno(mp,d)) |
509 | #define XFS_FSB_TO_DADDR(mp,fsbno) XFS_AGB_TO_DADDR(mp, \ | 509 | #define XFS_FSB_TO_DADDR(mp,fsbno) XFS_AGB_TO_DADDR(mp, \ |
510 | XFS_FSB_TO_AGNO(mp,fsbno), XFS_FSB_TO_AGBNO(mp,fsbno)) | 510 | XFS_FSB_TO_AGNO(mp,fsbno), XFS_FSB_TO_AGBNO(mp,fsbno)) |
511 | 511 | ||
diff --git a/fs/xfs/xfs_types.h b/fs/xfs/xfs_types.h index 0f5191644ab2..b2f724502f1b 100644 --- a/fs/xfs/xfs_types.h +++ b/fs/xfs/xfs_types.h | |||
@@ -45,7 +45,7 @@ typedef __uint32_t prid_t; /* project ID */ | |||
45 | typedef __uint32_t inst_t; /* an instruction */ | 45 | typedef __uint32_t inst_t; /* an instruction */ |
46 | 46 | ||
47 | typedef __s64 xfs_off_t; /* <file offset> type */ | 47 | typedef __s64 xfs_off_t; /* <file offset> type */ |
48 | typedef __u64 xfs_ino_t; /* <inode> type */ | 48 | typedef unsigned long long xfs_ino_t; /* <inode> type */ |
49 | typedef __s64 xfs_daddr_t; /* <disk address> type */ | 49 | typedef __s64 xfs_daddr_t; /* <disk address> type */ |
50 | typedef char * xfs_caddr_t; /* <core address> type */ | 50 | typedef char * xfs_caddr_t; /* <core address> type */ |
51 | typedef __u32 xfs_dev_t; | 51 | typedef __u32 xfs_dev_t; |
@@ -111,8 +111,6 @@ typedef __uint64_t xfs_fileoff_t; /* block number in a file */ | |||
111 | typedef __int64_t xfs_sfiloff_t; /* signed block number in a file */ | 111 | typedef __int64_t xfs_sfiloff_t; /* signed block number in a file */ |
112 | typedef __uint64_t xfs_filblks_t; /* number of blocks in a file */ | 112 | typedef __uint64_t xfs_filblks_t; /* number of blocks in a file */ |
113 | 113 | ||
114 | typedef __uint8_t xfs_arch_t; /* architecture of an xfs fs */ | ||
115 | |||
116 | /* | 114 | /* |
117 | * Null values for the types. | 115 | * Null values for the types. |
118 | */ | 116 | */ |
diff --git a/fs/xfs/xfs_vnodeops.c b/fs/xfs/xfs_vnodeops.c index f07bf8768c3a..0e55c5d7db5f 100644 --- a/fs/xfs/xfs_vnodeops.c +++ b/fs/xfs/xfs_vnodeops.c | |||
@@ -862,7 +862,7 @@ xfs_inactive_symlink_rmt( | |||
862 | * Find the block(s) so we can inval and unmap them. | 862 | * Find the block(s) so we can inval and unmap them. |
863 | */ | 863 | */ |
864 | done = 0; | 864 | done = 0; |
865 | XFS_BMAP_INIT(&free_list, &first_block); | 865 | xfs_bmap_init(&free_list, &first_block); |
866 | nmaps = ARRAY_SIZE(mval); | 866 | nmaps = ARRAY_SIZE(mval); |
867 | if ((error = xfs_bmapi(tp, ip, 0, XFS_B_TO_FSB(mp, size), | 867 | if ((error = xfs_bmapi(tp, ip, 0, XFS_B_TO_FSB(mp, size), |
868 | XFS_BMAPI_METADATA, &first_block, 0, mval, &nmaps, | 868 | XFS_BMAPI_METADATA, &first_block, 0, mval, &nmaps, |
@@ -1288,7 +1288,7 @@ xfs_inactive( | |||
1288 | /* | 1288 | /* |
1289 | * Free the inode. | 1289 | * Free the inode. |
1290 | */ | 1290 | */ |
1291 | XFS_BMAP_INIT(&free_list, &first_block); | 1291 | xfs_bmap_init(&free_list, &first_block); |
1292 | error = xfs_ifree(tp, ip, &free_list); | 1292 | error = xfs_ifree(tp, ip, &free_list); |
1293 | if (error) { | 1293 | if (error) { |
1294 | /* | 1294 | /* |
@@ -1461,7 +1461,7 @@ xfs_create( | |||
1461 | xfs_ilock(dp, XFS_ILOCK_EXCL | XFS_ILOCK_PARENT); | 1461 | xfs_ilock(dp, XFS_ILOCK_EXCL | XFS_ILOCK_PARENT); |
1462 | unlock_dp_on_error = B_TRUE; | 1462 | unlock_dp_on_error = B_TRUE; |
1463 | 1463 | ||
1464 | XFS_BMAP_INIT(&free_list, &first_block); | 1464 | xfs_bmap_init(&free_list, &first_block); |
1465 | 1465 | ||
1466 | ASSERT(ip == NULL); | 1466 | ASSERT(ip == NULL); |
1467 | 1467 | ||
@@ -1879,7 +1879,7 @@ xfs_remove( | |||
1879 | } | 1879 | } |
1880 | } | 1880 | } |
1881 | 1881 | ||
1882 | XFS_BMAP_INIT(&free_list, &first_block); | 1882 | xfs_bmap_init(&free_list, &first_block); |
1883 | error = xfs_dir_removename(tp, dp, name, ip->i_ino, | 1883 | error = xfs_dir_removename(tp, dp, name, ip->i_ino, |
1884 | &first_block, &free_list, resblks); | 1884 | &first_block, &free_list, resblks); |
1885 | if (error) { | 1885 | if (error) { |
@@ -2059,7 +2059,7 @@ xfs_link( | |||
2059 | if (error) | 2059 | if (error) |
2060 | goto error_return; | 2060 | goto error_return; |
2061 | 2061 | ||
2062 | XFS_BMAP_INIT(&free_list, &first_block); | 2062 | xfs_bmap_init(&free_list, &first_block); |
2063 | 2063 | ||
2064 | error = xfs_dir_createname(tp, tdp, target_name, sip->i_ino, | 2064 | error = xfs_dir_createname(tp, tdp, target_name, sip->i_ino, |
2065 | &first_block, &free_list, resblks); | 2065 | &first_block, &free_list, resblks); |
@@ -2231,7 +2231,7 @@ xfs_mkdir( | |||
2231 | xfs_trans_ijoin(tp, dp, XFS_ILOCK_EXCL); | 2231 | xfs_trans_ijoin(tp, dp, XFS_ILOCK_EXCL); |
2232 | unlock_dp_on_error = B_FALSE; | 2232 | unlock_dp_on_error = B_FALSE; |
2233 | 2233 | ||
2234 | XFS_BMAP_INIT(&free_list, &first_block); | 2234 | xfs_bmap_init(&free_list, &first_block); |
2235 | 2235 | ||
2236 | error = xfs_dir_createname(tp, dp, dir_name, cdp->i_ino, | 2236 | error = xfs_dir_createname(tp, dp, dir_name, cdp->i_ino, |
2237 | &first_block, &free_list, resblks ? | 2237 | &first_block, &free_list, resblks ? |
@@ -2438,7 +2438,7 @@ xfs_symlink( | |||
2438 | * Initialize the bmap freelist prior to calling either | 2438 | * Initialize the bmap freelist prior to calling either |
2439 | * bmapi or the directory create code. | 2439 | * bmapi or the directory create code. |
2440 | */ | 2440 | */ |
2441 | XFS_BMAP_INIT(&free_list, &first_block); | 2441 | xfs_bmap_init(&free_list, &first_block); |
2442 | 2442 | ||
2443 | /* | 2443 | /* |
2444 | * Allocate an inode for the symlink. | 2444 | * Allocate an inode for the symlink. |
@@ -2860,7 +2860,7 @@ retry: | |||
2860 | /* | 2860 | /* |
2861 | * Issue the xfs_bmapi() call to allocate the blocks | 2861 | * Issue the xfs_bmapi() call to allocate the blocks |
2862 | */ | 2862 | */ |
2863 | XFS_BMAP_INIT(&free_list, &firstfsb); | 2863 | xfs_bmap_init(&free_list, &firstfsb); |
2864 | error = xfs_bmapi(tp, ip, startoffset_fsb, | 2864 | error = xfs_bmapi(tp, ip, startoffset_fsb, |
2865 | allocatesize_fsb, bmapi_flag, | 2865 | allocatesize_fsb, bmapi_flag, |
2866 | &firstfsb, 0, imapp, &nimaps, | 2866 | &firstfsb, 0, imapp, &nimaps, |
@@ -2980,7 +2980,7 @@ xfs_zero_remaining_bytes( | |||
2980 | XFS_BUF_UNDONE(bp); | 2980 | XFS_BUF_UNDONE(bp); |
2981 | XFS_BUF_UNWRITE(bp); | 2981 | XFS_BUF_UNWRITE(bp); |
2982 | XFS_BUF_READ(bp); | 2982 | XFS_BUF_READ(bp); |
2983 | XFS_BUF_SET_ADDR(bp, XFS_FSB_TO_DB(ip, imap.br_startblock)); | 2983 | XFS_BUF_SET_ADDR(bp, xfs_fsb_to_db(ip, imap.br_startblock)); |
2984 | xfsbdstrat(mp, bp); | 2984 | xfsbdstrat(mp, bp); |
2985 | error = xfs_iowait(bp); | 2985 | error = xfs_iowait(bp); |
2986 | if (error) { | 2986 | if (error) { |
@@ -3186,7 +3186,7 @@ xfs_free_file_space( | |||
3186 | /* | 3186 | /* |
3187 | * issue the bunmapi() call to free the blocks | 3187 | * issue the bunmapi() call to free the blocks |
3188 | */ | 3188 | */ |
3189 | XFS_BMAP_INIT(&free_list, &firstfsb); | 3189 | xfs_bmap_init(&free_list, &firstfsb); |
3190 | error = xfs_bunmapi(tp, ip, startoffset_fsb, | 3190 | error = xfs_bunmapi(tp, ip, startoffset_fsb, |
3191 | endoffset_fsb - startoffset_fsb, | 3191 | endoffset_fsb - startoffset_fsb, |
3192 | 0, 2, &firstfsb, &free_list, NULL, &done); | 3192 | 0, 2, &firstfsb, &free_list, NULL, &done); |