diff options
| -rw-r--r-- | Documentation/device-mapper/dm-crypt.txt | 25 | ||||
| -rw-r--r-- | Documentation/device-mapper/dm-raid.txt | 4 | ||||
| -rw-r--r-- | drivers/md/Kconfig | 10 | ||||
| -rw-r--r-- | drivers/md/dm-bufio.c | 28 | ||||
| -rw-r--r-- | drivers/md/dm-cache-metadata.c | 3 | ||||
| -rw-r--r-- | drivers/md/dm-cache-policy-smq.c | 2 | ||||
| -rw-r--r-- | drivers/md/dm-cache-target.c | 3 | ||||
| -rw-r--r-- | drivers/md/dm-crypt.c | 214 | ||||
| -rw-r--r-- | drivers/md/dm-flakey.c | 53 | ||||
| -rw-r--r-- | drivers/md/dm-io.c | 34 | ||||
| -rw-r--r-- | drivers/md/dm-ioctl.c | 2 | ||||
| -rw-r--r-- | drivers/md/dm-mpath.c | 42 | ||||
| -rw-r--r-- | drivers/md/dm-raid.c | 82 | ||||
| -rw-r--r-- | drivers/md/dm-rq.c | 18 | ||||
| -rw-r--r-- | drivers/md/dm-table.c | 43 | ||||
| -rw-r--r-- | drivers/md/dm-verity-target.c | 2 | ||||
| -rw-r--r-- | drivers/md/dm.c | 4 | ||||
| -rw-r--r-- | drivers/md/persistent-data/dm-array.c | 2 | ||||
| -rw-r--r-- | drivers/md/persistent-data/dm-block-manager.c | 19 | ||||
| -rw-r--r-- | drivers/md/persistent-data/dm-space-map-common.c | 4 | ||||
| -rw-r--r-- | drivers/md/persistent-data/dm-space-map-metadata.c | 14 | ||||
| -rw-r--r-- | include/uapi/linux/dm-log-userspace.h | 53 |
22 files changed, 461 insertions, 200 deletions
diff --git a/Documentation/device-mapper/dm-crypt.txt b/Documentation/device-mapper/dm-crypt.txt index 692171fe9da0..6f15fcea9566 100644 --- a/Documentation/device-mapper/dm-crypt.txt +++ b/Documentation/device-mapper/dm-crypt.txt | |||
| @@ -21,13 +21,30 @@ Parameters: <cipher> <key> <iv_offset> <device path> \ | |||
| 21 | /proc/crypto contains supported crypto modes | 21 | /proc/crypto contains supported crypto modes |
| 22 | 22 | ||
| 23 | <key> | 23 | <key> |
| 24 | Key used for encryption. It is encoded as a hexadecimal number. | 24 | Key used for encryption. It is encoded either as a hexadecimal number |
| 25 | or it can be passed as <key_string> prefixed with single colon | ||
| 26 | character (':') for keys residing in kernel keyring service. | ||
| 25 | You can only use key sizes that are valid for the selected cipher | 27 | You can only use key sizes that are valid for the selected cipher |
| 26 | in combination with the selected iv mode. | 28 | in combination with the selected iv mode. |
| 27 | Note that for some iv modes the key string can contain additional | 29 | Note that for some iv modes the key string can contain additional |
| 28 | keys (for example IV seed) so the key contains more parts concatenated | 30 | keys (for example IV seed) so the key contains more parts concatenated |
| 29 | into a single string. | 31 | into a single string. |
| 30 | 32 | ||
| 33 | <key_string> | ||
| 34 | The kernel keyring key is identified by string in following format: | ||
| 35 | <key_size>:<key_type>:<key_description>. | ||
| 36 | |||
| 37 | <key_size> | ||
| 38 | The encryption key size in bytes. The kernel key payload size must match | ||
| 39 | the value passed in <key_size>. | ||
| 40 | |||
| 41 | <key_type> | ||
| 42 | Either 'logon' or 'user' kernel key type. | ||
| 43 | |||
| 44 | <key_description> | ||
| 45 | The kernel keyring key description crypt target should look for | ||
| 46 | when loading key of <key_type>. | ||
| 47 | |||
| 31 | <keycount> | 48 | <keycount> |
| 32 | Multi-key compatibility mode. You can define <keycount> keys and | 49 | Multi-key compatibility mode. You can define <keycount> keys and |
| 33 | then sectors are encrypted according to their offsets (sector 0 uses key0; | 50 | then sectors are encrypted according to their offsets (sector 0 uses key0; |
| @@ -90,6 +107,12 @@ dmsetup create crypt1 --table "0 `blockdev --getsize $1` crypt aes-cbc-essiv:sha | |||
| 90 | 107 | ||
| 91 | [[ | 108 | [[ |
| 92 | #!/bin/sh | 109 | #!/bin/sh |
| 110 | # Create a crypt device using dmsetup when encryption key is stored in keyring service | ||
| 111 | dmsetup create crypt2 --table "0 `blockdev --getsize $1` crypt aes-cbc-essiv:sha256 :32:logon:my_prefix:my_key 0 $1 0" | ||
| 112 | ]] | ||
| 113 | |||
| 114 | [[ | ||
| 115 | #!/bin/sh | ||
| 93 | # Create a crypt device using cryptsetup and LUKS header with default cipher | 116 | # Create a crypt device using cryptsetup and LUKS header with default cipher |
| 94 | cryptsetup luksFormat $1 | 117 | cryptsetup luksFormat $1 |
| 95 | cryptsetup luksOpen $1 crypt1 | 118 | cryptsetup luksOpen $1 crypt1 |
diff --git a/Documentation/device-mapper/dm-raid.txt b/Documentation/device-mapper/dm-raid.txt index 9bd531aa2279..5e3786fd9ea7 100644 --- a/Documentation/device-mapper/dm-raid.txt +++ b/Documentation/device-mapper/dm-raid.txt | |||
| @@ -242,6 +242,10 @@ recovery. Here is a fuller description of the individual fields: | |||
| 242 | in RAID1/10 or wrong parity values found in RAID4/5/6. | 242 | in RAID1/10 or wrong parity values found in RAID4/5/6. |
| 243 | This value is valid only after a "check" of the array | 243 | This value is valid only after a "check" of the array |
| 244 | is performed. A healthy array has a 'mismatch_cnt' of 0. | 244 | is performed. A healthy array has a 'mismatch_cnt' of 0. |
| 245 | <data_offset> The current data offset to the start of the user data on | ||
| 246 | each component device of a raid set (see the respective | ||
| 247 | raid parameter to support out-of-place reshaping). | ||
| 248 | |||
| 245 | 249 | ||
| 246 | Message Interface | 250 | Message Interface |
| 247 | ----------------- | 251 | ----------------- |
diff --git a/drivers/md/Kconfig b/drivers/md/Kconfig index 02a5345a44a6..b7767da50c26 100644 --- a/drivers/md/Kconfig +++ b/drivers/md/Kconfig | |||
| @@ -240,9 +240,17 @@ config DM_BUFIO | |||
| 240 | as a cache, holding recently-read blocks in memory and performing | 240 | as a cache, holding recently-read blocks in memory and performing |
| 241 | delayed writes. | 241 | delayed writes. |
| 242 | 242 | ||
| 243 | config DM_DEBUG_BLOCK_MANAGER_LOCKING | ||
| 244 | bool "Block manager locking" | ||
| 245 | depends on DM_BUFIO | ||
| 246 | ---help--- | ||
| 247 | Block manager locking can catch various metadata corruption issues. | ||
| 248 | |||
| 249 | If unsure, say N. | ||
| 250 | |||
| 243 | config DM_DEBUG_BLOCK_STACK_TRACING | 251 | config DM_DEBUG_BLOCK_STACK_TRACING |
| 244 | bool "Keep stack trace of persistent data block lock holders" | 252 | bool "Keep stack trace of persistent data block lock holders" |
| 245 | depends on STACKTRACE_SUPPORT && DM_BUFIO | 253 | depends on STACKTRACE_SUPPORT && DM_DEBUG_BLOCK_MANAGER_LOCKING |
| 246 | select STACKTRACE | 254 | select STACKTRACE |
| 247 | ---help--- | 255 | ---help--- |
| 248 | Enable this for messages that may help debug problems with the | 256 | Enable this for messages that may help debug problems with the |
diff --git a/drivers/md/dm-bufio.c b/drivers/md/dm-bufio.c index 262e75365cc0..84d2f0e4c754 100644 --- a/drivers/md/dm-bufio.c +++ b/drivers/md/dm-bufio.c | |||
| @@ -820,12 +820,14 @@ enum new_flag { | |||
| 820 | static struct dm_buffer *__alloc_buffer_wait_no_callback(struct dm_bufio_client *c, enum new_flag nf) | 820 | static struct dm_buffer *__alloc_buffer_wait_no_callback(struct dm_bufio_client *c, enum new_flag nf) |
| 821 | { | 821 | { |
| 822 | struct dm_buffer *b; | 822 | struct dm_buffer *b; |
| 823 | bool tried_noio_alloc = false; | ||
| 823 | 824 | ||
| 824 | /* | 825 | /* |
| 825 | * dm-bufio is resistant to allocation failures (it just keeps | 826 | * dm-bufio is resistant to allocation failures (it just keeps |
| 826 | * one buffer reserved in cases all the allocations fail). | 827 | * one buffer reserved in cases all the allocations fail). |
| 827 | * So set flags to not try too hard: | 828 | * So set flags to not try too hard: |
| 828 | * GFP_NOIO: don't recurse into the I/O layer | 829 | * GFP_NOWAIT: don't wait; if we need to sleep we'll release our |
| 830 | * mutex and wait ourselves. | ||
| 829 | * __GFP_NORETRY: don't retry and rather return failure | 831 | * __GFP_NORETRY: don't retry and rather return failure |
| 830 | * __GFP_NOMEMALLOC: don't use emergency reserves | 832 | * __GFP_NOMEMALLOC: don't use emergency reserves |
| 831 | * __GFP_NOWARN: don't print a warning in case of failure | 833 | * __GFP_NOWARN: don't print a warning in case of failure |
| @@ -835,7 +837,7 @@ static struct dm_buffer *__alloc_buffer_wait_no_callback(struct dm_bufio_client | |||
| 835 | */ | 837 | */ |
| 836 | while (1) { | 838 | while (1) { |
| 837 | if (dm_bufio_cache_size_latch != 1) { | 839 | if (dm_bufio_cache_size_latch != 1) { |
| 838 | b = alloc_buffer(c, GFP_NOIO | __GFP_NORETRY | __GFP_NOMEMALLOC | __GFP_NOWARN); | 840 | b = alloc_buffer(c, GFP_NOWAIT | __GFP_NORETRY | __GFP_NOMEMALLOC | __GFP_NOWARN); |
| 839 | if (b) | 841 | if (b) |
| 840 | return b; | 842 | return b; |
| 841 | } | 843 | } |
| @@ -843,6 +845,15 @@ static struct dm_buffer *__alloc_buffer_wait_no_callback(struct dm_bufio_client | |||
| 843 | if (nf == NF_PREFETCH) | 845 | if (nf == NF_PREFETCH) |
| 844 | return NULL; | 846 | return NULL; |
| 845 | 847 | ||
| 848 | if (dm_bufio_cache_size_latch != 1 && !tried_noio_alloc) { | ||
| 849 | dm_bufio_unlock(c); | ||
| 850 | b = alloc_buffer(c, GFP_NOIO | __GFP_NORETRY | __GFP_NOMEMALLOC | __GFP_NOWARN); | ||
| 851 | dm_bufio_lock(c); | ||
| 852 | if (b) | ||
| 853 | return b; | ||
| 854 | tried_noio_alloc = true; | ||
| 855 | } | ||
| 856 | |||
| 846 | if (!list_empty(&c->reserved_buffers)) { | 857 | if (!list_empty(&c->reserved_buffers)) { |
| 847 | b = list_entry(c->reserved_buffers.next, | 858 | b = list_entry(c->reserved_buffers.next, |
| 848 | struct dm_buffer, lru_list); | 859 | struct dm_buffer, lru_list); |
| @@ -1585,18 +1596,9 @@ dm_bufio_shrink_scan(struct shrinker *shrink, struct shrink_control *sc) | |||
| 1585 | static unsigned long | 1596 | static unsigned long |
| 1586 | dm_bufio_shrink_count(struct shrinker *shrink, struct shrink_control *sc) | 1597 | dm_bufio_shrink_count(struct shrinker *shrink, struct shrink_control *sc) |
| 1587 | { | 1598 | { |
| 1588 | struct dm_bufio_client *c; | 1599 | struct dm_bufio_client *c = container_of(shrink, struct dm_bufio_client, shrinker); |
| 1589 | unsigned long count; | ||
| 1590 | |||
| 1591 | c = container_of(shrink, struct dm_bufio_client, shrinker); | ||
| 1592 | if (sc->gfp_mask & __GFP_FS) | ||
| 1593 | dm_bufio_lock(c); | ||
| 1594 | else if (!dm_bufio_trylock(c)) | ||
| 1595 | return 0; | ||
| 1596 | 1600 | ||
| 1597 | count = c->n_buffers[LIST_CLEAN] + c->n_buffers[LIST_DIRTY]; | 1601 | return ACCESS_ONCE(c->n_buffers[LIST_CLEAN]) + ACCESS_ONCE(c->n_buffers[LIST_DIRTY]); |
| 1598 | dm_bufio_unlock(c); | ||
| 1599 | return count; | ||
| 1600 | } | 1602 | } |
| 1601 | 1603 | ||
| 1602 | /* | 1604 | /* |
diff --git a/drivers/md/dm-cache-metadata.c b/drivers/md/dm-cache-metadata.c index 695577812cf6..624fe4319b24 100644 --- a/drivers/md/dm-cache-metadata.c +++ b/drivers/md/dm-cache-metadata.c | |||
| @@ -383,7 +383,6 @@ static int __format_metadata(struct dm_cache_metadata *cmd) | |||
| 383 | goto bad; | 383 | goto bad; |
| 384 | 384 | ||
| 385 | dm_disk_bitset_init(cmd->tm, &cmd->discard_info); | 385 | dm_disk_bitset_init(cmd->tm, &cmd->discard_info); |
| 386 | |||
| 387 | r = dm_bitset_empty(&cmd->discard_info, &cmd->discard_root); | 386 | r = dm_bitset_empty(&cmd->discard_info, &cmd->discard_root); |
| 388 | if (r < 0) | 387 | if (r < 0) |
| 389 | goto bad; | 388 | goto bad; |
| @@ -789,7 +788,7 @@ static struct dm_cache_metadata *lookup_or_open(struct block_device *bdev, | |||
| 789 | static bool same_params(struct dm_cache_metadata *cmd, sector_t data_block_size) | 788 | static bool same_params(struct dm_cache_metadata *cmd, sector_t data_block_size) |
| 790 | { | 789 | { |
| 791 | if (cmd->data_block_size != data_block_size) { | 790 | if (cmd->data_block_size != data_block_size) { |
| 792 | DMERR("data_block_size (%llu) different from that in metadata (%llu)\n", | 791 | DMERR("data_block_size (%llu) different from that in metadata (%llu)", |
| 793 | (unsigned long long) data_block_size, | 792 | (unsigned long long) data_block_size, |
| 794 | (unsigned long long) cmd->data_block_size); | 793 | (unsigned long long) cmd->data_block_size); |
| 795 | return false; | 794 | return false; |
diff --git a/drivers/md/dm-cache-policy-smq.c b/drivers/md/dm-cache-policy-smq.c index c33f4a6e1d7d..f19c6930a67c 100644 --- a/drivers/md/dm-cache-policy-smq.c +++ b/drivers/md/dm-cache-policy-smq.c | |||
| @@ -1361,7 +1361,7 @@ static void smq_clear_dirty(struct dm_cache_policy *p, dm_oblock_t oblock) | |||
| 1361 | 1361 | ||
| 1362 | static unsigned random_level(dm_cblock_t cblock) | 1362 | static unsigned random_level(dm_cblock_t cblock) |
| 1363 | { | 1363 | { |
| 1364 | return hash_32_generic(from_cblock(cblock), 9) & (NR_CACHE_LEVELS - 1); | 1364 | return hash_32(from_cblock(cblock), 9) & (NR_CACHE_LEVELS - 1); |
| 1365 | } | 1365 | } |
| 1366 | 1366 | ||
| 1367 | static int smq_load_mapping(struct dm_cache_policy *p, | 1367 | static int smq_load_mapping(struct dm_cache_policy *p, |
diff --git a/drivers/md/dm-cache-target.c b/drivers/md/dm-cache-target.c index 59b2c50562e4..e04c61e0839e 100644 --- a/drivers/md/dm-cache-target.c +++ b/drivers/md/dm-cache-target.c | |||
| @@ -989,7 +989,8 @@ static void set_cache_mode(struct cache *cache, enum cache_metadata_mode new_mod | |||
| 989 | enum cache_metadata_mode old_mode = get_cache_mode(cache); | 989 | enum cache_metadata_mode old_mode = get_cache_mode(cache); |
| 990 | 990 | ||
| 991 | if (dm_cache_metadata_needs_check(cache->cmd, &needs_check)) { | 991 | if (dm_cache_metadata_needs_check(cache->cmd, &needs_check)) { |
| 992 | DMERR("unable to read needs_check flag, setting failure mode"); | 992 | DMERR("%s: unable to read needs_check flag, setting failure mode.", |
| 993 | cache_device_name(cache)); | ||
| 993 | new_mode = CM_FAIL; | 994 | new_mode = CM_FAIL; |
| 994 | } | 995 | } |
| 995 | 996 | ||
diff --git a/drivers/md/dm-crypt.c b/drivers/md/dm-crypt.c index 68a9eb4f3f36..7c6c57216bf2 100644 --- a/drivers/md/dm-crypt.c +++ b/drivers/md/dm-crypt.c | |||
| @@ -12,6 +12,7 @@ | |||
| 12 | #include <linux/module.h> | 12 | #include <linux/module.h> |
| 13 | #include <linux/init.h> | 13 | #include <linux/init.h> |
| 14 | #include <linux/kernel.h> | 14 | #include <linux/kernel.h> |
| 15 | #include <linux/key.h> | ||
| 15 | #include <linux/bio.h> | 16 | #include <linux/bio.h> |
| 16 | #include <linux/blkdev.h> | 17 | #include <linux/blkdev.h> |
| 17 | #include <linux/mempool.h> | 18 | #include <linux/mempool.h> |
| @@ -23,12 +24,14 @@ | |||
| 23 | #include <linux/atomic.h> | 24 | #include <linux/atomic.h> |
| 24 | #include <linux/scatterlist.h> | 25 | #include <linux/scatterlist.h> |
| 25 | #include <linux/rbtree.h> | 26 | #include <linux/rbtree.h> |
| 27 | #include <linux/ctype.h> | ||
| 26 | #include <asm/page.h> | 28 | #include <asm/page.h> |
| 27 | #include <asm/unaligned.h> | 29 | #include <asm/unaligned.h> |
| 28 | #include <crypto/hash.h> | 30 | #include <crypto/hash.h> |
| 29 | #include <crypto/md5.h> | 31 | #include <crypto/md5.h> |
| 30 | #include <crypto/algapi.h> | 32 | #include <crypto/algapi.h> |
| 31 | #include <crypto/skcipher.h> | 33 | #include <crypto/skcipher.h> |
| 34 | #include <keys/user-type.h> | ||
| 32 | 35 | ||
| 33 | #include <linux/device-mapper.h> | 36 | #include <linux/device-mapper.h> |
| 34 | 37 | ||
| @@ -140,8 +143,9 @@ struct crypt_config { | |||
| 140 | 143 | ||
| 141 | char *cipher; | 144 | char *cipher; |
| 142 | char *cipher_string; | 145 | char *cipher_string; |
| 146 | char *key_string; | ||
| 143 | 147 | ||
| 144 | struct crypt_iv_operations *iv_gen_ops; | 148 | const struct crypt_iv_operations *iv_gen_ops; |
| 145 | union { | 149 | union { |
| 146 | struct iv_essiv_private essiv; | 150 | struct iv_essiv_private essiv; |
| 147 | struct iv_benbi_private benbi; | 151 | struct iv_benbi_private benbi; |
| @@ -758,15 +762,15 @@ static int crypt_iv_tcw_post(struct crypt_config *cc, u8 *iv, | |||
| 758 | return r; | 762 | return r; |
| 759 | } | 763 | } |
| 760 | 764 | ||
| 761 | static struct crypt_iv_operations crypt_iv_plain_ops = { | 765 | static const struct crypt_iv_operations crypt_iv_plain_ops = { |
| 762 | .generator = crypt_iv_plain_gen | 766 | .generator = crypt_iv_plain_gen |
| 763 | }; | 767 | }; |
| 764 | 768 | ||
| 765 | static struct crypt_iv_operations crypt_iv_plain64_ops = { | 769 | static const struct crypt_iv_operations crypt_iv_plain64_ops = { |
| 766 | .generator = crypt_iv_plain64_gen | 770 | .generator = crypt_iv_plain64_gen |
| 767 | }; | 771 | }; |
| 768 | 772 | ||
| 769 | static struct crypt_iv_operations crypt_iv_essiv_ops = { | 773 | static const struct crypt_iv_operations crypt_iv_essiv_ops = { |
| 770 | .ctr = crypt_iv_essiv_ctr, | 774 | .ctr = crypt_iv_essiv_ctr, |
| 771 | .dtr = crypt_iv_essiv_dtr, | 775 | .dtr = crypt_iv_essiv_dtr, |
| 772 | .init = crypt_iv_essiv_init, | 776 | .init = crypt_iv_essiv_init, |
| @@ -774,17 +778,17 @@ static struct crypt_iv_operations crypt_iv_essiv_ops = { | |||
| 774 | .generator = crypt_iv_essiv_gen | 778 | .generator = crypt_iv_essiv_gen |
| 775 | }; | 779 | }; |
| 776 | 780 | ||
| 777 | static struct crypt_iv_operations crypt_iv_benbi_ops = { | 781 | static const struct crypt_iv_operations crypt_iv_benbi_ops = { |
| 778 | .ctr = crypt_iv_benbi_ctr, | 782 | .ctr = crypt_iv_benbi_ctr, |
| 779 | .dtr = crypt_iv_benbi_dtr, | 783 | .dtr = crypt_iv_benbi_dtr, |
| 780 | .generator = crypt_iv_benbi_gen | 784 | .generator = crypt_iv_benbi_gen |
| 781 | }; | 785 | }; |
| 782 | 786 | ||
| 783 | static struct crypt_iv_operations crypt_iv_null_ops = { | 787 | static const struct crypt_iv_operations crypt_iv_null_ops = { |
| 784 | .generator = crypt_iv_null_gen | 788 | .generator = crypt_iv_null_gen |
| 785 | }; | 789 | }; |
| 786 | 790 | ||
| 787 | static struct crypt_iv_operations crypt_iv_lmk_ops = { | 791 | static const struct crypt_iv_operations crypt_iv_lmk_ops = { |
| 788 | .ctr = crypt_iv_lmk_ctr, | 792 | .ctr = crypt_iv_lmk_ctr, |
| 789 | .dtr = crypt_iv_lmk_dtr, | 793 | .dtr = crypt_iv_lmk_dtr, |
| 790 | .init = crypt_iv_lmk_init, | 794 | .init = crypt_iv_lmk_init, |
| @@ -793,7 +797,7 @@ static struct crypt_iv_operations crypt_iv_lmk_ops = { | |||
| 793 | .post = crypt_iv_lmk_post | 797 | .post = crypt_iv_lmk_post |
| 794 | }; | 798 | }; |
| 795 | 799 | ||
| 796 | static struct crypt_iv_operations crypt_iv_tcw_ops = { | 800 | static const struct crypt_iv_operations crypt_iv_tcw_ops = { |
| 797 | .ctr = crypt_iv_tcw_ctr, | 801 | .ctr = crypt_iv_tcw_ctr, |
| 798 | .dtr = crypt_iv_tcw_dtr, | 802 | .dtr = crypt_iv_tcw_dtr, |
| 799 | .init = crypt_iv_tcw_init, | 803 | .init = crypt_iv_tcw_init, |
| @@ -994,7 +998,6 @@ static struct bio *crypt_alloc_buffer(struct dm_crypt_io *io, unsigned size) | |||
| 994 | gfp_t gfp_mask = GFP_NOWAIT | __GFP_HIGHMEM; | 998 | gfp_t gfp_mask = GFP_NOWAIT | __GFP_HIGHMEM; |
| 995 | unsigned i, len, remaining_size; | 999 | unsigned i, len, remaining_size; |
| 996 | struct page *page; | 1000 | struct page *page; |
| 997 | struct bio_vec *bvec; | ||
| 998 | 1001 | ||
| 999 | retry: | 1002 | retry: |
| 1000 | if (unlikely(gfp_mask & __GFP_DIRECT_RECLAIM)) | 1003 | if (unlikely(gfp_mask & __GFP_DIRECT_RECLAIM)) |
| @@ -1019,12 +1022,7 @@ retry: | |||
| 1019 | 1022 | ||
| 1020 | len = (remaining_size > PAGE_SIZE) ? PAGE_SIZE : remaining_size; | 1023 | len = (remaining_size > PAGE_SIZE) ? PAGE_SIZE : remaining_size; |
| 1021 | 1024 | ||
| 1022 | bvec = &clone->bi_io_vec[clone->bi_vcnt++]; | 1025 | bio_add_page(clone, page, len, 0); |
| 1023 | bvec->bv_page = page; | ||
| 1024 | bvec->bv_len = len; | ||
| 1025 | bvec->bv_offset = 0; | ||
| 1026 | |||
| 1027 | clone->bi_iter.bi_size += len; | ||
| 1028 | 1026 | ||
| 1029 | remaining_size -= len; | 1027 | remaining_size -= len; |
| 1030 | } | 1028 | } |
| @@ -1471,7 +1469,7 @@ static int crypt_alloc_tfms(struct crypt_config *cc, char *ciphermode) | |||
| 1471 | return 0; | 1469 | return 0; |
| 1472 | } | 1470 | } |
| 1473 | 1471 | ||
| 1474 | static int crypt_setkey_allcpus(struct crypt_config *cc) | 1472 | static int crypt_setkey(struct crypt_config *cc) |
| 1475 | { | 1473 | { |
| 1476 | unsigned subkey_size; | 1474 | unsigned subkey_size; |
| 1477 | int err = 0, i, r; | 1475 | int err = 0, i, r; |
| @@ -1490,25 +1488,157 @@ static int crypt_setkey_allcpus(struct crypt_config *cc) | |||
| 1490 | return err; | 1488 | return err; |
| 1491 | } | 1489 | } |
| 1492 | 1490 | ||
| 1491 | #ifdef CONFIG_KEYS | ||
| 1492 | |||
| 1493 | static bool contains_whitespace(const char *str) | ||
| 1494 | { | ||
| 1495 | while (*str) | ||
| 1496 | if (isspace(*str++)) | ||
| 1497 | return true; | ||
| 1498 | return false; | ||
| 1499 | } | ||
| 1500 | |||
| 1501 | static int crypt_set_keyring_key(struct crypt_config *cc, const char *key_string) | ||
| 1502 | { | ||
| 1503 | char *new_key_string, *key_desc; | ||
| 1504 | int ret; | ||
| 1505 | struct key *key; | ||
| 1506 | const struct user_key_payload *ukp; | ||
| 1507 | |||
| 1508 | /* | ||
| 1509 | * Reject key_string with whitespace. dm core currently lacks code for | ||
| 1510 | * proper whitespace escaping in arguments on DM_TABLE_STATUS path. | ||
| 1511 | */ | ||
| 1512 | if (contains_whitespace(key_string)) { | ||
| 1513 | DMERR("whitespace chars not allowed in key string"); | ||
| 1514 | return -EINVAL; | ||
| 1515 | } | ||
| 1516 | |||
| 1517 | /* look for next ':' separating key_type from key_description */ | ||
| 1518 | key_desc = strpbrk(key_string, ":"); | ||
| 1519 | if (!key_desc || key_desc == key_string || !strlen(key_desc + 1)) | ||
| 1520 | return -EINVAL; | ||
| 1521 | |||
| 1522 | if (strncmp(key_string, "logon:", key_desc - key_string + 1) && | ||
| 1523 | strncmp(key_string, "user:", key_desc - key_string + 1)) | ||
| 1524 | return -EINVAL; | ||
| 1525 | |||
| 1526 | new_key_string = kstrdup(key_string, GFP_KERNEL); | ||
| 1527 | if (!new_key_string) | ||
| 1528 | return -ENOMEM; | ||
| 1529 | |||
| 1530 | key = request_key(key_string[0] == 'l' ? &key_type_logon : &key_type_user, | ||
| 1531 | key_desc + 1, NULL); | ||
| 1532 | if (IS_ERR(key)) { | ||
| 1533 | kzfree(new_key_string); | ||
| 1534 | return PTR_ERR(key); | ||
| 1535 | } | ||
| 1536 | |||
| 1537 | rcu_read_lock(); | ||
| 1538 | |||
| 1539 | ukp = user_key_payload(key); | ||
| 1540 | if (!ukp) { | ||
| 1541 | rcu_read_unlock(); | ||
| 1542 | key_put(key); | ||
| 1543 | kzfree(new_key_string); | ||
| 1544 | return -EKEYREVOKED; | ||
| 1545 | } | ||
| 1546 | |||
| 1547 | if (cc->key_size != ukp->datalen) { | ||
| 1548 | rcu_read_unlock(); | ||
| 1549 | key_put(key); | ||
| 1550 | kzfree(new_key_string); | ||
| 1551 | return -EINVAL; | ||
| 1552 | } | ||
| 1553 | |||
| 1554 | memcpy(cc->key, ukp->data, cc->key_size); | ||
| 1555 | |||
| 1556 | rcu_read_unlock(); | ||
| 1557 | key_put(key); | ||
| 1558 | |||
| 1559 | /* clear the flag since following operations may invalidate previously valid key */ | ||
| 1560 | clear_bit(DM_CRYPT_KEY_VALID, &cc->flags); | ||
| 1561 | |||
| 1562 | ret = crypt_setkey(cc); | ||
| 1563 | |||
| 1564 | /* wipe the kernel key payload copy in each case */ | ||
| 1565 | memset(cc->key, 0, cc->key_size * sizeof(u8)); | ||
| 1566 | |||
| 1567 | if (!ret) { | ||
| 1568 | set_bit(DM_CRYPT_KEY_VALID, &cc->flags); | ||
| 1569 | kzfree(cc->key_string); | ||
| 1570 | cc->key_string = new_key_string; | ||
| 1571 | } else | ||
| 1572 | kzfree(new_key_string); | ||
| 1573 | |||
| 1574 | return ret; | ||
| 1575 | } | ||
| 1576 | |||
| 1577 | static int get_key_size(char **key_string) | ||
| 1578 | { | ||
| 1579 | char *colon, dummy; | ||
| 1580 | int ret; | ||
| 1581 | |||
| 1582 | if (*key_string[0] != ':') | ||
| 1583 | return strlen(*key_string) >> 1; | ||
| 1584 | |||
| 1585 | /* look for next ':' in key string */ | ||
| 1586 | colon = strpbrk(*key_string + 1, ":"); | ||
| 1587 | if (!colon) | ||
| 1588 | return -EINVAL; | ||
| 1589 | |||
| 1590 | if (sscanf(*key_string + 1, "%u%c", &ret, &dummy) != 2 || dummy != ':') | ||
| 1591 | return -EINVAL; | ||
| 1592 | |||
| 1593 | *key_string = colon; | ||
| 1594 | |||
| 1595 | /* remaining key string should be :<logon|user>:<key_desc> */ | ||
| 1596 | |||
| 1597 | return ret; | ||
| 1598 | } | ||
| 1599 | |||
| 1600 | #else | ||
| 1601 | |||
| 1602 | static int crypt_set_keyring_key(struct crypt_config *cc, const char *key_string) | ||
| 1603 | { | ||
| 1604 | return -EINVAL; | ||
| 1605 | } | ||
| 1606 | |||
| 1607 | static int get_key_size(char **key_string) | ||
| 1608 | { | ||
| 1609 | return (*key_string[0] == ':') ? -EINVAL : strlen(*key_string) >> 1; | ||
| 1610 | } | ||
| 1611 | |||
| 1612 | #endif | ||
| 1613 | |||
| 1493 | static int crypt_set_key(struct crypt_config *cc, char *key) | 1614 | static int crypt_set_key(struct crypt_config *cc, char *key) |
| 1494 | { | 1615 | { |
| 1495 | int r = -EINVAL; | 1616 | int r = -EINVAL; |
| 1496 | int key_string_len = strlen(key); | 1617 | int key_string_len = strlen(key); |
| 1497 | 1618 | ||
| 1498 | /* The key size may not be changed. */ | ||
| 1499 | if (cc->key_size != (key_string_len >> 1)) | ||
| 1500 | goto out; | ||
| 1501 | |||
| 1502 | /* Hyphen (which gives a key_size of zero) means there is no key. */ | 1619 | /* Hyphen (which gives a key_size of zero) means there is no key. */ |
| 1503 | if (!cc->key_size && strcmp(key, "-")) | 1620 | if (!cc->key_size && strcmp(key, "-")) |
| 1504 | goto out; | 1621 | goto out; |
| 1505 | 1622 | ||
| 1506 | if (cc->key_size && crypt_decode_key(cc->key, key, cc->key_size) < 0) | 1623 | /* ':' means the key is in kernel keyring, short-circuit normal key processing */ |
| 1624 | if (key[0] == ':') { | ||
| 1625 | r = crypt_set_keyring_key(cc, key + 1); | ||
| 1507 | goto out; | 1626 | goto out; |
| 1627 | } | ||
| 1508 | 1628 | ||
| 1509 | set_bit(DM_CRYPT_KEY_VALID, &cc->flags); | 1629 | /* clear the flag since following operations may invalidate previously valid key */ |
| 1630 | clear_bit(DM_CRYPT_KEY_VALID, &cc->flags); | ||
| 1510 | 1631 | ||
| 1511 | r = crypt_setkey_allcpus(cc); | 1632 | /* wipe references to any kernel keyring key */ |
| 1633 | kzfree(cc->key_string); | ||
| 1634 | cc->key_string = NULL; | ||
| 1635 | |||
| 1636 | if (cc->key_size && crypt_decode_key(cc->key, key, cc->key_size) < 0) | ||
| 1637 | goto out; | ||
| 1638 | |||
| 1639 | r = crypt_setkey(cc); | ||
| 1640 | if (!r) | ||
| 1641 | set_bit(DM_CRYPT_KEY_VALID, &cc->flags); | ||
| 1512 | 1642 | ||
| 1513 | out: | 1643 | out: |
| 1514 | /* Hex key string not needed after here, so wipe it. */ | 1644 | /* Hex key string not needed after here, so wipe it. */ |
| @@ -1521,8 +1651,10 @@ static int crypt_wipe_key(struct crypt_config *cc) | |||
| 1521 | { | 1651 | { |
| 1522 | clear_bit(DM_CRYPT_KEY_VALID, &cc->flags); | 1652 | clear_bit(DM_CRYPT_KEY_VALID, &cc->flags); |
| 1523 | memset(&cc->key, 0, cc->key_size * sizeof(u8)); | 1653 | memset(&cc->key, 0, cc->key_size * sizeof(u8)); |
| 1654 | kzfree(cc->key_string); | ||
| 1655 | cc->key_string = NULL; | ||
| 1524 | 1656 | ||
| 1525 | return crypt_setkey_allcpus(cc); | 1657 | return crypt_setkey(cc); |
| 1526 | } | 1658 | } |
| 1527 | 1659 | ||
| 1528 | static void crypt_dtr(struct dm_target *ti) | 1660 | static void crypt_dtr(struct dm_target *ti) |
| @@ -1558,6 +1690,7 @@ static void crypt_dtr(struct dm_target *ti) | |||
| 1558 | 1690 | ||
| 1559 | kzfree(cc->cipher); | 1691 | kzfree(cc->cipher); |
| 1560 | kzfree(cc->cipher_string); | 1692 | kzfree(cc->cipher_string); |
| 1693 | kzfree(cc->key_string); | ||
| 1561 | 1694 | ||
| 1562 | /* Must zero key material before freeing */ | 1695 | /* Must zero key material before freeing */ |
| 1563 | kzfree(cc); | 1696 | kzfree(cc); |
| @@ -1726,12 +1859,13 @@ bad_mem: | |||
| 1726 | 1859 | ||
| 1727 | /* | 1860 | /* |
| 1728 | * Construct an encryption mapping: | 1861 | * Construct an encryption mapping: |
| 1729 | * <cipher> <key> <iv_offset> <dev_path> <start> | 1862 | * <cipher> [<key>|:<key_size>:<user|logon>:<key_description>] <iv_offset> <dev_path> <start> |
| 1730 | */ | 1863 | */ |
| 1731 | static int crypt_ctr(struct dm_target *ti, unsigned int argc, char **argv) | 1864 | static int crypt_ctr(struct dm_target *ti, unsigned int argc, char **argv) |
| 1732 | { | 1865 | { |
| 1733 | struct crypt_config *cc; | 1866 | struct crypt_config *cc; |
| 1734 | unsigned int key_size, opt_params; | 1867 | int key_size; |
| 1868 | unsigned int opt_params; | ||
| 1735 | unsigned long long tmpll; | 1869 | unsigned long long tmpll; |
| 1736 | int ret; | 1870 | int ret; |
| 1737 | size_t iv_size_padding; | 1871 | size_t iv_size_padding; |
| @@ -1748,7 +1882,11 @@ static int crypt_ctr(struct dm_target *ti, unsigned int argc, char **argv) | |||
| 1748 | return -EINVAL; | 1882 | return -EINVAL; |
| 1749 | } | 1883 | } |
| 1750 | 1884 | ||
| 1751 | key_size = strlen(argv[1]) >> 1; | 1885 | key_size = get_key_size(&argv[1]); |
| 1886 | if (key_size < 0) { | ||
| 1887 | ti->error = "Cannot parse key size"; | ||
| 1888 | return -EINVAL; | ||
| 1889 | } | ||
| 1752 | 1890 | ||
| 1753 | cc = kzalloc(sizeof(*cc) + key_size * sizeof(u8), GFP_KERNEL); | 1891 | cc = kzalloc(sizeof(*cc) + key_size * sizeof(u8), GFP_KERNEL); |
| 1754 | if (!cc) { | 1892 | if (!cc) { |
| @@ -1955,10 +2093,13 @@ static void crypt_status(struct dm_target *ti, status_type_t type, | |||
| 1955 | case STATUSTYPE_TABLE: | 2093 | case STATUSTYPE_TABLE: |
| 1956 | DMEMIT("%s ", cc->cipher_string); | 2094 | DMEMIT("%s ", cc->cipher_string); |
| 1957 | 2095 | ||
| 1958 | if (cc->key_size > 0) | 2096 | if (cc->key_size > 0) { |
| 1959 | for (i = 0; i < cc->key_size; i++) | 2097 | if (cc->key_string) |
| 1960 | DMEMIT("%02x", cc->key[i]); | 2098 | DMEMIT(":%u:%s", cc->key_size, cc->key_string); |
| 1961 | else | 2099 | else |
| 2100 | for (i = 0; i < cc->key_size; i++) | ||
| 2101 | DMEMIT("%02x", cc->key[i]); | ||
| 2102 | } else | ||
| 1962 | DMEMIT("-"); | 2103 | DMEMIT("-"); |
| 1963 | 2104 | ||
| 1964 | DMEMIT(" %llu %s %llu", (unsigned long long)cc->iv_offset, | 2105 | DMEMIT(" %llu %s %llu", (unsigned long long)cc->iv_offset, |
| @@ -2014,7 +2155,7 @@ static void crypt_resume(struct dm_target *ti) | |||
| 2014 | static int crypt_message(struct dm_target *ti, unsigned argc, char **argv) | 2155 | static int crypt_message(struct dm_target *ti, unsigned argc, char **argv) |
| 2015 | { | 2156 | { |
| 2016 | struct crypt_config *cc = ti->private; | 2157 | struct crypt_config *cc = ti->private; |
| 2017 | int ret = -EINVAL; | 2158 | int key_size, ret = -EINVAL; |
| 2018 | 2159 | ||
| 2019 | if (argc < 2) | 2160 | if (argc < 2) |
| 2020 | goto error; | 2161 | goto error; |
| @@ -2025,6 +2166,13 @@ static int crypt_message(struct dm_target *ti, unsigned argc, char **argv) | |||
| 2025 | return -EINVAL; | 2166 | return -EINVAL; |
| 2026 | } | 2167 | } |
| 2027 | if (argc == 3 && !strcasecmp(argv[1], "set")) { | 2168 | if (argc == 3 && !strcasecmp(argv[1], "set")) { |
| 2169 | /* The key size may not be changed. */ | ||
| 2170 | key_size = get_key_size(&argv[2]); | ||
| 2171 | if (key_size < 0 || cc->key_size != key_size) { | ||
| 2172 | memset(argv[2], '0', strlen(argv[2])); | ||
| 2173 | return -EINVAL; | ||
| 2174 | } | ||
| 2175 | |||
| 2028 | ret = crypt_set_key(cc, argv[2]); | 2176 | ret = crypt_set_key(cc, argv[2]); |
| 2029 | if (ret) | 2177 | if (ret) |
| 2030 | return ret; | 2178 | return ret; |
| @@ -2068,7 +2216,7 @@ static void crypt_io_hints(struct dm_target *ti, struct queue_limits *limits) | |||
| 2068 | 2216 | ||
| 2069 | static struct target_type crypt_target = { | 2217 | static struct target_type crypt_target = { |
| 2070 | .name = "crypt", | 2218 | .name = "crypt", |
| 2071 | .version = {1, 14, 1}, | 2219 | .version = {1, 15, 0}, |
| 2072 | .module = THIS_MODULE, | 2220 | .module = THIS_MODULE, |
| 2073 | .ctr = crypt_ctr, | 2221 | .ctr = crypt_ctr, |
| 2074 | .dtr = crypt_dtr, | 2222 | .dtr = crypt_dtr, |
diff --git a/drivers/md/dm-flakey.c b/drivers/md/dm-flakey.c index 6a2e8dd44a1b..13305a182611 100644 --- a/drivers/md/dm-flakey.c +++ b/drivers/md/dm-flakey.c | |||
| @@ -36,7 +36,8 @@ struct flakey_c { | |||
| 36 | }; | 36 | }; |
| 37 | 37 | ||
| 38 | enum feature_flag_bits { | 38 | enum feature_flag_bits { |
| 39 | DROP_WRITES | 39 | DROP_WRITES, |
| 40 | ERROR_WRITES | ||
| 40 | }; | 41 | }; |
| 41 | 42 | ||
| 42 | struct per_bio_data { | 43 | struct per_bio_data { |
| @@ -76,6 +77,25 @@ static int parse_features(struct dm_arg_set *as, struct flakey_c *fc, | |||
| 76 | if (test_and_set_bit(DROP_WRITES, &fc->flags)) { | 77 | if (test_and_set_bit(DROP_WRITES, &fc->flags)) { |
| 77 | ti->error = "Feature drop_writes duplicated"; | 78 | ti->error = "Feature drop_writes duplicated"; |
| 78 | return -EINVAL; | 79 | return -EINVAL; |
| 80 | } else if (test_bit(ERROR_WRITES, &fc->flags)) { | ||
| 81 | ti->error = "Feature drop_writes conflicts with feature error_writes"; | ||
| 82 | return -EINVAL; | ||
| 83 | } | ||
| 84 | |||
| 85 | continue; | ||
| 86 | } | ||
| 87 | |||
| 88 | /* | ||
| 89 | * error_writes | ||
| 90 | */ | ||
| 91 | if (!strcasecmp(arg_name, "error_writes")) { | ||
| 92 | if (test_and_set_bit(ERROR_WRITES, &fc->flags)) { | ||
| 93 | ti->error = "Feature error_writes duplicated"; | ||
| 94 | return -EINVAL; | ||
| 95 | |||
| 96 | } else if (test_bit(DROP_WRITES, &fc->flags)) { | ||
| 97 | ti->error = "Feature error_writes conflicts with feature drop_writes"; | ||
| 98 | return -EINVAL; | ||
| 79 | } | 99 | } |
| 80 | 100 | ||
| 81 | continue; | 101 | continue; |
| @@ -135,6 +155,10 @@ static int parse_features(struct dm_arg_set *as, struct flakey_c *fc, | |||
| 135 | if (test_bit(DROP_WRITES, &fc->flags) && (fc->corrupt_bio_rw == WRITE)) { | 155 | if (test_bit(DROP_WRITES, &fc->flags) && (fc->corrupt_bio_rw == WRITE)) { |
| 136 | ti->error = "drop_writes is incompatible with corrupt_bio_byte with the WRITE flag set"; | 156 | ti->error = "drop_writes is incompatible with corrupt_bio_byte with the WRITE flag set"; |
| 137 | return -EINVAL; | 157 | return -EINVAL; |
| 158 | |||
| 159 | } else if (test_bit(ERROR_WRITES, &fc->flags) && (fc->corrupt_bio_rw == WRITE)) { | ||
| 160 | ti->error = "error_writes is incompatible with corrupt_bio_byte with the WRITE flag set"; | ||
| 161 | return -EINVAL; | ||
| 138 | } | 162 | } |
| 139 | 163 | ||
| 140 | return 0; | 164 | return 0; |
| @@ -200,11 +224,13 @@ static int flakey_ctr(struct dm_target *ti, unsigned int argc, char **argv) | |||
| 200 | 224 | ||
| 201 | if (!(fc->up_interval + fc->down_interval)) { | 225 | if (!(fc->up_interval + fc->down_interval)) { |
| 202 | ti->error = "Total (up + down) interval is zero"; | 226 | ti->error = "Total (up + down) interval is zero"; |
| 227 | r = -EINVAL; | ||
| 203 | goto bad; | 228 | goto bad; |
| 204 | } | 229 | } |
| 205 | 230 | ||
| 206 | if (fc->up_interval + fc->down_interval < fc->up_interval) { | 231 | if (fc->up_interval + fc->down_interval < fc->up_interval) { |
| 207 | ti->error = "Interval overflow"; | 232 | ti->error = "Interval overflow"; |
| 233 | r = -EINVAL; | ||
| 208 | goto bad; | 234 | goto bad; |
| 209 | } | 235 | } |
| 210 | 236 | ||
| @@ -289,22 +315,27 @@ static int flakey_map(struct dm_target *ti, struct bio *bio) | |||
| 289 | pb->bio_submitted = true; | 315 | pb->bio_submitted = true; |
| 290 | 316 | ||
| 291 | /* | 317 | /* |
| 292 | * Error reads if neither corrupt_bio_byte or drop_writes are set. | 318 | * Error reads if neither corrupt_bio_byte or drop_writes or error_writes are set. |
| 293 | * Otherwise, flakey_end_io() will decide if the reads should be modified. | 319 | * Otherwise, flakey_end_io() will decide if the reads should be modified. |
| 294 | */ | 320 | */ |
| 295 | if (bio_data_dir(bio) == READ) { | 321 | if (bio_data_dir(bio) == READ) { |
| 296 | if (!fc->corrupt_bio_byte && !test_bit(DROP_WRITES, &fc->flags)) | 322 | if (!fc->corrupt_bio_byte && !test_bit(DROP_WRITES, &fc->flags) && |
| 323 | !test_bit(ERROR_WRITES, &fc->flags)) | ||
| 297 | return -EIO; | 324 | return -EIO; |
| 298 | goto map_bio; | 325 | goto map_bio; |
| 299 | } | 326 | } |
| 300 | 327 | ||
| 301 | /* | 328 | /* |
| 302 | * Drop writes? | 329 | * Drop or error writes? |
| 303 | */ | 330 | */ |
| 304 | if (test_bit(DROP_WRITES, &fc->flags)) { | 331 | if (test_bit(DROP_WRITES, &fc->flags)) { |
| 305 | bio_endio(bio); | 332 | bio_endio(bio); |
| 306 | return DM_MAPIO_SUBMITTED; | 333 | return DM_MAPIO_SUBMITTED; |
| 307 | } | 334 | } |
| 335 | else if (test_bit(ERROR_WRITES, &fc->flags)) { | ||
| 336 | bio_io_error(bio); | ||
| 337 | return DM_MAPIO_SUBMITTED; | ||
| 338 | } | ||
| 308 | 339 | ||
| 309 | /* | 340 | /* |
| 310 | * Corrupt matching writes. | 341 | * Corrupt matching writes. |
| @@ -340,10 +371,11 @@ static int flakey_end_io(struct dm_target *ti, struct bio *bio, int error) | |||
| 340 | */ | 371 | */ |
| 341 | corrupt_bio_data(bio, fc); | 372 | corrupt_bio_data(bio, fc); |
| 342 | 373 | ||
| 343 | } else if (!test_bit(DROP_WRITES, &fc->flags)) { | 374 | } else if (!test_bit(DROP_WRITES, &fc->flags) && |
| 375 | !test_bit(ERROR_WRITES, &fc->flags)) { | ||
| 344 | /* | 376 | /* |
| 345 | * Error read during the down_interval if drop_writes | 377 | * Error read during the down_interval if drop_writes |
| 346 | * wasn't configured. | 378 | * and error_writes were not configured. |
| 347 | */ | 379 | */ |
| 348 | return -EIO; | 380 | return -EIO; |
| 349 | } | 381 | } |
| @@ -357,7 +389,7 @@ static void flakey_status(struct dm_target *ti, status_type_t type, | |||
| 357 | { | 389 | { |
| 358 | unsigned sz = 0; | 390 | unsigned sz = 0; |
| 359 | struct flakey_c *fc = ti->private; | 391 | struct flakey_c *fc = ti->private; |
| 360 | unsigned drop_writes; | 392 | unsigned drop_writes, error_writes; |
| 361 | 393 | ||
| 362 | switch (type) { | 394 | switch (type) { |
| 363 | case STATUSTYPE_INFO: | 395 | case STATUSTYPE_INFO: |
| @@ -370,10 +402,13 @@ static void flakey_status(struct dm_target *ti, status_type_t type, | |||
| 370 | fc->down_interval); | 402 | fc->down_interval); |
| 371 | 403 | ||
| 372 | drop_writes = test_bit(DROP_WRITES, &fc->flags); | 404 | drop_writes = test_bit(DROP_WRITES, &fc->flags); |
| 373 | DMEMIT("%u ", drop_writes + (fc->corrupt_bio_byte > 0) * 5); | 405 | error_writes = test_bit(ERROR_WRITES, &fc->flags); |
| 406 | DMEMIT("%u ", drop_writes + error_writes + (fc->corrupt_bio_byte > 0) * 5); | ||
| 374 | 407 | ||
| 375 | if (drop_writes) | 408 | if (drop_writes) |
| 376 | DMEMIT("drop_writes "); | 409 | DMEMIT("drop_writes "); |
| 410 | else if (error_writes) | ||
| 411 | DMEMIT("error_writes "); | ||
| 377 | 412 | ||
| 378 | if (fc->corrupt_bio_byte) | 413 | if (fc->corrupt_bio_byte) |
| 379 | DMEMIT("corrupt_bio_byte %u %c %u %u ", | 414 | DMEMIT("corrupt_bio_byte %u %c %u %u ", |
| @@ -410,7 +445,7 @@ static int flakey_iterate_devices(struct dm_target *ti, iterate_devices_callout_ | |||
| 410 | 445 | ||
| 411 | static struct target_type flakey_target = { | 446 | static struct target_type flakey_target = { |
| 412 | .name = "flakey", | 447 | .name = "flakey", |
| 413 | .version = {1, 3, 1}, | 448 | .version = {1, 4, 0}, |
| 414 | .module = THIS_MODULE, | 449 | .module = THIS_MODULE, |
| 415 | .ctr = flakey_ctr, | 450 | .ctr = flakey_ctr, |
| 416 | .dtr = flakey_dtr, | 451 | .dtr = flakey_dtr, |
diff --git a/drivers/md/dm-io.c b/drivers/md/dm-io.c index 0bf1a12e35fe..03940bf36f6c 100644 --- a/drivers/md/dm-io.c +++ b/drivers/md/dm-io.c | |||
| @@ -162,7 +162,10 @@ struct dpages { | |||
| 162 | struct page **p, unsigned long *len, unsigned *offset); | 162 | struct page **p, unsigned long *len, unsigned *offset); |
| 163 | void (*next_page)(struct dpages *dp); | 163 | void (*next_page)(struct dpages *dp); |
| 164 | 164 | ||
| 165 | unsigned context_u; | 165 | union { |
| 166 | unsigned context_u; | ||
| 167 | struct bvec_iter context_bi; | ||
| 168 | }; | ||
| 166 | void *context_ptr; | 169 | void *context_ptr; |
| 167 | 170 | ||
| 168 | void *vma_invalidate_address; | 171 | void *vma_invalidate_address; |
| @@ -204,25 +207,36 @@ static void list_dp_init(struct dpages *dp, struct page_list *pl, unsigned offse | |||
| 204 | static void bio_get_page(struct dpages *dp, struct page **p, | 207 | static void bio_get_page(struct dpages *dp, struct page **p, |
| 205 | unsigned long *len, unsigned *offset) | 208 | unsigned long *len, unsigned *offset) |
| 206 | { | 209 | { |
| 207 | struct bio_vec *bvec = dp->context_ptr; | 210 | struct bio_vec bvec = bvec_iter_bvec((struct bio_vec *)dp->context_ptr, |
| 208 | *p = bvec->bv_page; | 211 | dp->context_bi); |
| 209 | *len = bvec->bv_len - dp->context_u; | 212 | |
| 210 | *offset = bvec->bv_offset + dp->context_u; | 213 | *p = bvec.bv_page; |
| 214 | *len = bvec.bv_len; | ||
| 215 | *offset = bvec.bv_offset; | ||
| 216 | |||
| 217 | /* avoid figuring it out again in bio_next_page() */ | ||
| 218 | dp->context_bi.bi_sector = (sector_t)bvec.bv_len; | ||
| 211 | } | 219 | } |
| 212 | 220 | ||
| 213 | static void bio_next_page(struct dpages *dp) | 221 | static void bio_next_page(struct dpages *dp) |
| 214 | { | 222 | { |
| 215 | struct bio_vec *bvec = dp->context_ptr; | 223 | unsigned int len = (unsigned int)dp->context_bi.bi_sector; |
| 216 | dp->context_ptr = bvec + 1; | 224 | |
| 217 | dp->context_u = 0; | 225 | bvec_iter_advance((struct bio_vec *)dp->context_ptr, |
| 226 | &dp->context_bi, len); | ||
| 218 | } | 227 | } |
| 219 | 228 | ||
| 220 | static void bio_dp_init(struct dpages *dp, struct bio *bio) | 229 | static void bio_dp_init(struct dpages *dp, struct bio *bio) |
| 221 | { | 230 | { |
| 222 | dp->get_page = bio_get_page; | 231 | dp->get_page = bio_get_page; |
| 223 | dp->next_page = bio_next_page; | 232 | dp->next_page = bio_next_page; |
| 224 | dp->context_ptr = __bvec_iter_bvec(bio->bi_io_vec, bio->bi_iter); | 233 | |
| 225 | dp->context_u = bio->bi_iter.bi_bvec_done; | 234 | /* |
| 235 | * We just use bvec iterator to retrieve pages, so it is ok to | ||
| 236 | * access the bvec table directly here | ||
| 237 | */ | ||
| 238 | dp->context_ptr = bio->bi_io_vec; | ||
| 239 | dp->context_bi = bio->bi_iter; | ||
| 226 | } | 240 | } |
| 227 | 241 | ||
| 228 | /* | 242 | /* |
diff --git a/drivers/md/dm-ioctl.c b/drivers/md/dm-ioctl.c index 966eb4b61aed..c72a77048b73 100644 --- a/drivers/md/dm-ioctl.c +++ b/drivers/md/dm-ioctl.c | |||
| @@ -1697,7 +1697,7 @@ static int copy_params(struct dm_ioctl __user *user, struct dm_ioctl *param_kern | |||
| 1697 | { | 1697 | { |
| 1698 | struct dm_ioctl *dmi; | 1698 | struct dm_ioctl *dmi; |
| 1699 | int secure_data; | 1699 | int secure_data; |
| 1700 | const size_t minimum_data_size = sizeof(*param_kernel) - sizeof(param_kernel->data); | 1700 | const size_t minimum_data_size = offsetof(struct dm_ioctl, data); |
| 1701 | 1701 | ||
| 1702 | if (copy_from_user(param_kernel, user, minimum_data_size)) | 1702 | if (copy_from_user(param_kernel, user, minimum_data_size)) |
| 1703 | return -EFAULT; | 1703 | return -EFAULT; |
diff --git a/drivers/md/dm-mpath.c b/drivers/md/dm-mpath.c index e477af8596e2..6400cffb986d 100644 --- a/drivers/md/dm-mpath.c +++ b/drivers/md/dm-mpath.c | |||
| @@ -372,16 +372,13 @@ static int __pg_init_all_paths(struct multipath *m) | |||
| 372 | return atomic_read(&m->pg_init_in_progress); | 372 | return atomic_read(&m->pg_init_in_progress); |
| 373 | } | 373 | } |
| 374 | 374 | ||
| 375 | static int pg_init_all_paths(struct multipath *m) | 375 | static void pg_init_all_paths(struct multipath *m) |
| 376 | { | 376 | { |
| 377 | int r; | ||
| 378 | unsigned long flags; | 377 | unsigned long flags; |
| 379 | 378 | ||
| 380 | spin_lock_irqsave(&m->lock, flags); | 379 | spin_lock_irqsave(&m->lock, flags); |
| 381 | r = __pg_init_all_paths(m); | 380 | __pg_init_all_paths(m); |
| 382 | spin_unlock_irqrestore(&m->lock, flags); | 381 | spin_unlock_irqrestore(&m->lock, flags); |
| 383 | |||
| 384 | return r; | ||
| 385 | } | 382 | } |
| 386 | 383 | ||
| 387 | static void __switch_pg(struct multipath *m, struct priority_group *pg) | 384 | static void __switch_pg(struct multipath *m, struct priority_group *pg) |
| @@ -583,16 +580,17 @@ static int __multipath_map(struct dm_target *ti, struct request *clone, | |||
| 583 | * .request_fn stacked on blk-mq path(s) and | 580 | * .request_fn stacked on blk-mq path(s) and |
| 584 | * blk-mq stacked on blk-mq path(s). | 581 | * blk-mq stacked on blk-mq path(s). |
| 585 | */ | 582 | */ |
| 586 | *__clone = blk_mq_alloc_request(bdev_get_queue(bdev), | 583 | clone = blk_mq_alloc_request(bdev_get_queue(bdev), |
| 587 | rq_data_dir(rq), BLK_MQ_REQ_NOWAIT); | 584 | rq_data_dir(rq), BLK_MQ_REQ_NOWAIT); |
| 588 | if (IS_ERR(*__clone)) { | 585 | if (IS_ERR(clone)) { |
| 589 | /* ENOMEM, requeue */ | 586 | /* EBUSY, ENODEV or EWOULDBLOCK: requeue */ |
| 590 | clear_request_fn_mpio(m, map_context); | 587 | clear_request_fn_mpio(m, map_context); |
| 591 | return r; | 588 | return r; |
| 592 | } | 589 | } |
| 593 | (*__clone)->bio = (*__clone)->biotail = NULL; | 590 | clone->bio = clone->biotail = NULL; |
| 594 | (*__clone)->rq_disk = bdev->bd_disk; | 591 | clone->rq_disk = bdev->bd_disk; |
| 595 | (*__clone)->cmd_flags |= REQ_FAILFAST_TRANSPORT; | 592 | clone->cmd_flags |= REQ_FAILFAST_TRANSPORT; |
| 593 | *__clone = clone; | ||
| 596 | } | 594 | } |
| 597 | 595 | ||
| 598 | if (pgpath->pg->ps.type->start_io) | 596 | if (pgpath->pg->ps.type->start_io) |
| @@ -852,18 +850,22 @@ retain: | |||
| 852 | attached_handler_name = scsi_dh_attached_handler_name(q, GFP_KERNEL); | 850 | attached_handler_name = scsi_dh_attached_handler_name(q, GFP_KERNEL); |
| 853 | if (attached_handler_name) { | 851 | if (attached_handler_name) { |
| 854 | /* | 852 | /* |
| 853 | * Clear any hw_handler_params associated with a | ||
| 854 | * handler that isn't already attached. | ||
| 855 | */ | ||
| 856 | if (m->hw_handler_name && strcmp(attached_handler_name, m->hw_handler_name)) { | ||
| 857 | kfree(m->hw_handler_params); | ||
| 858 | m->hw_handler_params = NULL; | ||
| 859 | } | ||
| 860 | |||
| 861 | /* | ||
| 855 | * Reset hw_handler_name to match the attached handler | 862 | * Reset hw_handler_name to match the attached handler |
| 856 | * and clear any hw_handler_params associated with the | ||
| 857 | * ignored handler. | ||
| 858 | * | 863 | * |
| 859 | * NB. This modifies the table line to show the actual | 864 | * NB. This modifies the table line to show the actual |
| 860 | * handler instead of the original table passed in. | 865 | * handler instead of the original table passed in. |
| 861 | */ | 866 | */ |
| 862 | kfree(m->hw_handler_name); | 867 | kfree(m->hw_handler_name); |
| 863 | m->hw_handler_name = attached_handler_name; | 868 | m->hw_handler_name = attached_handler_name; |
| 864 | |||
| 865 | kfree(m->hw_handler_params); | ||
| 866 | m->hw_handler_params = NULL; | ||
| 867 | } | 869 | } |
| 868 | } | 870 | } |
| 869 | 871 | ||
| @@ -1002,6 +1004,8 @@ static int parse_hw_handler(struct dm_arg_set *as, struct multipath *m) | |||
| 1002 | } | 1004 | } |
| 1003 | 1005 | ||
| 1004 | m->hw_handler_name = kstrdup(dm_shift_arg(as), GFP_KERNEL); | 1006 | m->hw_handler_name = kstrdup(dm_shift_arg(as), GFP_KERNEL); |
| 1007 | if (!m->hw_handler_name) | ||
| 1008 | return -EINVAL; | ||
| 1005 | 1009 | ||
| 1006 | if (hw_argc > 1) { | 1010 | if (hw_argc > 1) { |
| 1007 | char *p; | 1011 | char *p; |
| @@ -1362,7 +1366,7 @@ static int switch_pg_num(struct multipath *m, const char *pgstr) | |||
| 1362 | char dummy; | 1366 | char dummy; |
| 1363 | 1367 | ||
| 1364 | if (!pgstr || (sscanf(pgstr, "%u%c", &pgnum, &dummy) != 1) || !pgnum || | 1368 | if (!pgstr || (sscanf(pgstr, "%u%c", &pgnum, &dummy) != 1) || !pgnum || |
| 1365 | (pgnum > m->nr_priority_groups)) { | 1369 | !m->nr_priority_groups || (pgnum > m->nr_priority_groups)) { |
| 1366 | DMWARN("invalid PG number supplied to switch_pg_num"); | 1370 | DMWARN("invalid PG number supplied to switch_pg_num"); |
| 1367 | return -EINVAL; | 1371 | return -EINVAL; |
| 1368 | } | 1372 | } |
| @@ -1394,7 +1398,7 @@ static int bypass_pg_num(struct multipath *m, const char *pgstr, bool bypassed) | |||
| 1394 | char dummy; | 1398 | char dummy; |
| 1395 | 1399 | ||
| 1396 | if (!pgstr || (sscanf(pgstr, "%u%c", &pgnum, &dummy) != 1) || !pgnum || | 1400 | if (!pgstr || (sscanf(pgstr, "%u%c", &pgnum, &dummy) != 1) || !pgnum || |
| 1397 | (pgnum > m->nr_priority_groups)) { | 1401 | !m->nr_priority_groups || (pgnum > m->nr_priority_groups)) { |
| 1398 | DMWARN("invalid PG number supplied to bypass_pg"); | 1402 | DMWARN("invalid PG number supplied to bypass_pg"); |
| 1399 | return -EINVAL; | 1403 | return -EINVAL; |
| 1400 | } | 1404 | } |
diff --git a/drivers/md/dm-raid.c b/drivers/md/dm-raid.c index 953159d9a825..b8f978e551d7 100644 --- a/drivers/md/dm-raid.c +++ b/drivers/md/dm-raid.c | |||
| @@ -160,7 +160,6 @@ struct raid_dev { | |||
| 160 | CTR_FLAG_DAEMON_SLEEP | \ | 160 | CTR_FLAG_DAEMON_SLEEP | \ |
| 161 | CTR_FLAG_MIN_RECOVERY_RATE | \ | 161 | CTR_FLAG_MIN_RECOVERY_RATE | \ |
| 162 | CTR_FLAG_MAX_RECOVERY_RATE | \ | 162 | CTR_FLAG_MAX_RECOVERY_RATE | \ |
| 163 | CTR_FLAG_MAX_WRITE_BEHIND | \ | ||
| 164 | CTR_FLAG_STRIPE_CACHE | \ | 163 | CTR_FLAG_STRIPE_CACHE | \ |
| 165 | CTR_FLAG_REGION_SIZE | \ | 164 | CTR_FLAG_REGION_SIZE | \ |
| 166 | CTR_FLAG_DELTA_DISKS | \ | 165 | CTR_FLAG_DELTA_DISKS | \ |
| @@ -171,7 +170,6 @@ struct raid_dev { | |||
| 171 | CTR_FLAG_DAEMON_SLEEP | \ | 170 | CTR_FLAG_DAEMON_SLEEP | \ |
| 172 | CTR_FLAG_MIN_RECOVERY_RATE | \ | 171 | CTR_FLAG_MIN_RECOVERY_RATE | \ |
| 173 | CTR_FLAG_MAX_RECOVERY_RATE | \ | 172 | CTR_FLAG_MAX_RECOVERY_RATE | \ |
| 174 | CTR_FLAG_MAX_WRITE_BEHIND | \ | ||
| 175 | CTR_FLAG_STRIPE_CACHE | \ | 173 | CTR_FLAG_STRIPE_CACHE | \ |
| 176 | CTR_FLAG_REGION_SIZE | \ | 174 | CTR_FLAG_REGION_SIZE | \ |
| 177 | CTR_FLAG_DELTA_DISKS | \ | 175 | CTR_FLAG_DELTA_DISKS | \ |
| @@ -2050,16 +2048,17 @@ static int super_init_validation(struct raid_set *rs, struct md_rdev *rdev) | |||
| 2050 | 2048 | ||
| 2051 | mddev->reshape_position = MaxSector; | 2049 | mddev->reshape_position = MaxSector; |
| 2052 | 2050 | ||
| 2051 | mddev->raid_disks = le32_to_cpu(sb->num_devices); | ||
| 2052 | mddev->level = le32_to_cpu(sb->level); | ||
| 2053 | mddev->layout = le32_to_cpu(sb->layout); | ||
| 2054 | mddev->chunk_sectors = le32_to_cpu(sb->stripe_sectors); | ||
| 2055 | |||
| 2053 | /* | 2056 | /* |
| 2054 | * Reshaping is supported, e.g. reshape_position is valid | 2057 | * Reshaping is supported, e.g. reshape_position is valid |
| 2055 | * in superblock and superblock content is authoritative. | 2058 | * in superblock and superblock content is authoritative. |
| 2056 | */ | 2059 | */ |
| 2057 | if (le32_to_cpu(sb->compat_features) & FEATURE_FLAG_SUPPORTS_V190) { | 2060 | if (le32_to_cpu(sb->compat_features) & FEATURE_FLAG_SUPPORTS_V190) { |
| 2058 | /* Superblock is authoritative wrt given raid set layout! */ | 2061 | /* Superblock is authoritative wrt given raid set layout! */ |
| 2059 | mddev->raid_disks = le32_to_cpu(sb->num_devices); | ||
| 2060 | mddev->level = le32_to_cpu(sb->level); | ||
| 2061 | mddev->layout = le32_to_cpu(sb->layout); | ||
| 2062 | mddev->chunk_sectors = le32_to_cpu(sb->stripe_sectors); | ||
| 2063 | mddev->new_level = le32_to_cpu(sb->new_level); | 2062 | mddev->new_level = le32_to_cpu(sb->new_level); |
| 2064 | mddev->new_layout = le32_to_cpu(sb->new_layout); | 2063 | mddev->new_layout = le32_to_cpu(sb->new_layout); |
| 2065 | mddev->new_chunk_sectors = le32_to_cpu(sb->new_stripe_sectors); | 2064 | mddev->new_chunk_sectors = le32_to_cpu(sb->new_stripe_sectors); |
| @@ -2087,38 +2086,44 @@ static int super_init_validation(struct raid_set *rs, struct md_rdev *rdev) | |||
| 2087 | /* | 2086 | /* |
| 2088 | * No takeover/reshaping, because we don't have the extended v1.9.0 metadata | 2087 | * No takeover/reshaping, because we don't have the extended v1.9.0 metadata |
| 2089 | */ | 2088 | */ |
| 2090 | if (le32_to_cpu(sb->level) != mddev->new_level) { | 2089 | struct raid_type *rt_cur = get_raid_type_by_ll(mddev->level, mddev->layout); |
| 2091 | DMERR("Reshaping/takeover raid sets not yet supported. (raid level/stripes/size change)"); | 2090 | struct raid_type *rt_new = get_raid_type_by_ll(mddev->new_level, mddev->new_layout); |
| 2092 | return -EINVAL; | ||
| 2093 | } | ||
| 2094 | if (le32_to_cpu(sb->layout) != mddev->new_layout) { | ||
| 2095 | DMERR("Reshaping raid sets not yet supported. (raid layout change)"); | ||
| 2096 | DMERR(" 0x%X vs 0x%X", le32_to_cpu(sb->layout), mddev->layout); | ||
| 2097 | DMERR(" Old layout: %s w/ %d copies", | ||
| 2098 | raid10_md_layout_to_format(le32_to_cpu(sb->layout)), | ||
| 2099 | raid10_md_layout_to_copies(le32_to_cpu(sb->layout))); | ||
| 2100 | DMERR(" New layout: %s w/ %d copies", | ||
| 2101 | raid10_md_layout_to_format(mddev->layout), | ||
| 2102 | raid10_md_layout_to_copies(mddev->layout)); | ||
| 2103 | return -EINVAL; | ||
| 2104 | } | ||
| 2105 | if (le32_to_cpu(sb->stripe_sectors) != mddev->new_chunk_sectors) { | ||
| 2106 | DMERR("Reshaping raid sets not yet supported. (stripe sectors change)"); | ||
| 2107 | return -EINVAL; | ||
| 2108 | } | ||
| 2109 | 2091 | ||
| 2110 | /* We can only change the number of devices in raid1 with old (i.e. pre 1.0.7) metadata */ | 2092 | if (rs_takeover_requested(rs)) { |
| 2111 | if (!rt_is_raid1(rs->raid_type) && | 2093 | if (rt_cur && rt_new) |
| 2112 | (le32_to_cpu(sb->num_devices) != mddev->raid_disks)) { | 2094 | DMERR("Takeover raid sets from %s to %s not yet supported by metadata. (raid level change)", |
| 2113 | DMERR("Reshaping raid sets not yet supported. (device count change from %u to %u)", | 2095 | rt_cur->name, rt_new->name); |
| 2114 | sb->num_devices, mddev->raid_disks); | 2096 | else |
| 2097 | DMERR("Takeover raid sets not yet supported by metadata. (raid level change)"); | ||
| 2098 | return -EINVAL; | ||
| 2099 | } else if (rs_reshape_requested(rs)) { | ||
| 2100 | DMERR("Reshaping raid sets not yet supported by metadata. (raid layout change keeping level)"); | ||
| 2101 | if (mddev->layout != mddev->new_layout) { | ||
| 2102 | if (rt_cur && rt_new) | ||
| 2103 | DMERR(" current layout %s vs new layout %s", | ||
| 2104 | rt_cur->name, rt_new->name); | ||
| 2105 | else | ||
| 2106 | DMERR(" current layout 0x%X vs new layout 0x%X", | ||
| 2107 | le32_to_cpu(sb->layout), mddev->new_layout); | ||
| 2108 | } | ||
| 2109 | if (mddev->chunk_sectors != mddev->new_chunk_sectors) | ||
| 2110 | DMERR(" current stripe sectors %u vs new stripe sectors %u", | ||
| 2111 | mddev->chunk_sectors, mddev->new_chunk_sectors); | ||
| 2112 | if (rs->delta_disks) | ||
| 2113 | DMERR(" current %u disks vs new %u disks", | ||
| 2114 | mddev->raid_disks, mddev->raid_disks + rs->delta_disks); | ||
| 2115 | if (rs_is_raid10(rs)) { | ||
| 2116 | DMERR(" Old layout: %s w/ %u copies", | ||
| 2117 | raid10_md_layout_to_format(mddev->layout), | ||
| 2118 | raid10_md_layout_to_copies(mddev->layout)); | ||
| 2119 | DMERR(" New layout: %s w/ %u copies", | ||
| 2120 | raid10_md_layout_to_format(mddev->new_layout), | ||
| 2121 | raid10_md_layout_to_copies(mddev->new_layout)); | ||
| 2122 | } | ||
| 2115 | return -EINVAL; | 2123 | return -EINVAL; |
| 2116 | } | 2124 | } |
| 2117 | 2125 | ||
| 2118 | DMINFO("Discovered old metadata format; upgrading to extended metadata format"); | 2126 | DMINFO("Discovered old metadata format; upgrading to extended metadata format"); |
| 2119 | |||
| 2120 | /* Table line is checked vs. authoritative superblock */ | ||
| 2121 | rs_set_new(rs); | ||
| 2122 | } | 2127 | } |
| 2123 | 2128 | ||
| 2124 | if (!test_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags)) | 2129 | if (!test_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags)) |
| @@ -2211,7 +2216,7 @@ static int super_init_validation(struct raid_set *rs, struct md_rdev *rdev) | |||
| 2211 | continue; | 2216 | continue; |
| 2212 | 2217 | ||
| 2213 | if (role != r->raid_disk) { | 2218 | if (role != r->raid_disk) { |
| 2214 | if (__is_raid10_near(mddev->layout)) { | 2219 | if (rs_is_raid10(rs) && __is_raid10_near(mddev->layout)) { |
| 2215 | if (mddev->raid_disks % __raid10_near_copies(mddev->layout) || | 2220 | if (mddev->raid_disks % __raid10_near_copies(mddev->layout) || |
| 2216 | rs->raid_disks % rs->raid10_copies) { | 2221 | rs->raid_disks % rs->raid10_copies) { |
| 2217 | rs->ti->error = | 2222 | rs->ti->error = |
| @@ -2994,6 +2999,9 @@ static int raid_ctr(struct dm_target *ti, unsigned int argc, char **argv) | |||
| 2994 | } | 2999 | } |
| 2995 | } | 3000 | } |
| 2996 | 3001 | ||
| 3002 | /* Disable/enable discard support on raid set. */ | ||
| 3003 | configure_discard_support(rs); | ||
| 3004 | |||
| 2997 | mddev_unlock(&rs->md); | 3005 | mddev_unlock(&rs->md); |
| 2998 | return 0; | 3006 | return 0; |
| 2999 | 3007 | ||
| @@ -3580,12 +3588,6 @@ static int raid_preresume(struct dm_target *ti) | |||
| 3580 | if (test_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags)) | 3588 | if (test_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags)) |
| 3581 | rs_update_sbs(rs); | 3589 | rs_update_sbs(rs); |
| 3582 | 3590 | ||
| 3583 | /* | ||
| 3584 | * Disable/enable discard support on raid set after any | ||
| 3585 | * conversion, because devices can have been added | ||
| 3586 | */ | ||
| 3587 | configure_discard_support(rs); | ||
| 3588 | |||
| 3589 | /* Load the bitmap from disk unless raid0 */ | 3591 | /* Load the bitmap from disk unless raid0 */ |
| 3590 | r = __load_dirty_region_bitmap(rs); | 3592 | r = __load_dirty_region_bitmap(rs); |
| 3591 | if (r) | 3593 | if (r) |
diff --git a/drivers/md/dm-rq.c b/drivers/md/dm-rq.c index b2a9e2d161e4..9d7275fb541a 100644 --- a/drivers/md/dm-rq.c +++ b/drivers/md/dm-rq.c | |||
| @@ -23,11 +23,7 @@ static unsigned dm_mq_queue_depth = DM_MQ_QUEUE_DEPTH; | |||
| 23 | #define RESERVED_REQUEST_BASED_IOS 256 | 23 | #define RESERVED_REQUEST_BASED_IOS 256 |
| 24 | static unsigned reserved_rq_based_ios = RESERVED_REQUEST_BASED_IOS; | 24 | static unsigned reserved_rq_based_ios = RESERVED_REQUEST_BASED_IOS; |
| 25 | 25 | ||
| 26 | #ifdef CONFIG_DM_MQ_DEFAULT | 26 | static bool use_blk_mq = IS_ENABLED(CONFIG_DM_MQ_DEFAULT); |
| 27 | static bool use_blk_mq = true; | ||
| 28 | #else | ||
| 29 | static bool use_blk_mq = false; | ||
| 30 | #endif | ||
| 31 | 27 | ||
| 32 | bool dm_use_blk_mq_default(void) | 28 | bool dm_use_blk_mq_default(void) |
| 33 | { | 29 | { |
| @@ -210,6 +206,9 @@ static void rq_end_stats(struct mapped_device *md, struct request *orig) | |||
| 210 | */ | 206 | */ |
| 211 | static void rq_completed(struct mapped_device *md, int rw, bool run_queue) | 207 | static void rq_completed(struct mapped_device *md, int rw, bool run_queue) |
| 212 | { | 208 | { |
| 209 | struct request_queue *q = md->queue; | ||
| 210 | unsigned long flags; | ||
| 211 | |||
| 213 | atomic_dec(&md->pending[rw]); | 212 | atomic_dec(&md->pending[rw]); |
| 214 | 213 | ||
| 215 | /* nudge anyone waiting on suspend queue */ | 214 | /* nudge anyone waiting on suspend queue */ |
| @@ -222,8 +221,11 @@ static void rq_completed(struct mapped_device *md, int rw, bool run_queue) | |||
| 222 | * back into ->request_fn() could deadlock attempting to grab the | 221 | * back into ->request_fn() could deadlock attempting to grab the |
| 223 | * queue lock again. | 222 | * queue lock again. |
| 224 | */ | 223 | */ |
| 225 | if (!md->queue->mq_ops && run_queue) | 224 | if (!q->mq_ops && run_queue) { |
| 226 | blk_run_queue_async(md->queue); | 225 | spin_lock_irqsave(q->queue_lock, flags); |
| 226 | blk_run_queue_async(q); | ||
| 227 | spin_unlock_irqrestore(q->queue_lock, flags); | ||
| 228 | } | ||
| 227 | 229 | ||
| 228 | /* | 230 | /* |
| 229 | * dm_put() must be at the end of this function. See the comment above | 231 | * dm_put() must be at the end of this function. See the comment above |
| @@ -798,7 +800,7 @@ static void dm_old_request_fn(struct request_queue *q) | |||
| 798 | pos = blk_rq_pos(rq); | 800 | pos = blk_rq_pos(rq); |
| 799 | 801 | ||
| 800 | if ((dm_old_request_peeked_before_merge_deadline(md) && | 802 | if ((dm_old_request_peeked_before_merge_deadline(md) && |
| 801 | md_in_flight(md) && rq->bio && rq->bio->bi_vcnt == 1 && | 803 | md_in_flight(md) && rq->bio && !bio_multiple_segments(rq->bio) && |
| 802 | md->last_rq_pos == pos && md->last_rq_rw == rq_data_dir(rq)) || | 804 | md->last_rq_pos == pos && md->last_rq_rw == rq_data_dir(rq)) || |
| 803 | (ti->type->busy && ti->type->busy(ti))) { | 805 | (ti->type->busy && ti->type->busy(ti))) { |
| 804 | blk_delay_queue(q, 10); | 806 | blk_delay_queue(q, 10); |
diff --git a/drivers/md/dm-table.c b/drivers/md/dm-table.c index c4b53b332607..0a427de23ed2 100644 --- a/drivers/md/dm-table.c +++ b/drivers/md/dm-table.c | |||
| @@ -871,7 +871,7 @@ static int dm_table_determine_type(struct dm_table *t) | |||
| 871 | { | 871 | { |
| 872 | unsigned i; | 872 | unsigned i; |
| 873 | unsigned bio_based = 0, request_based = 0, hybrid = 0; | 873 | unsigned bio_based = 0, request_based = 0, hybrid = 0; |
| 874 | bool verify_blk_mq = false; | 874 | unsigned sq_count = 0, mq_count = 0; |
| 875 | struct dm_target *tgt; | 875 | struct dm_target *tgt; |
| 876 | struct dm_dev_internal *dd; | 876 | struct dm_dev_internal *dd; |
| 877 | struct list_head *devices = dm_table_get_devices(t); | 877 | struct list_head *devices = dm_table_get_devices(t); |
| @@ -924,12 +924,6 @@ static int dm_table_determine_type(struct dm_table *t) | |||
| 924 | 924 | ||
| 925 | BUG_ON(!request_based); /* No targets in this table */ | 925 | BUG_ON(!request_based); /* No targets in this table */ |
| 926 | 926 | ||
| 927 | if (list_empty(devices) && __table_type_request_based(live_md_type)) { | ||
| 928 | /* inherit live MD type */ | ||
| 929 | t->type = live_md_type; | ||
| 930 | return 0; | ||
| 931 | } | ||
| 932 | |||
| 933 | /* | 927 | /* |
| 934 | * The only way to establish DM_TYPE_MQ_REQUEST_BASED is by | 928 | * The only way to establish DM_TYPE_MQ_REQUEST_BASED is by |
| 935 | * having a compatible target use dm_table_set_type. | 929 | * having a compatible target use dm_table_set_type. |
| @@ -948,6 +942,19 @@ verify_rq_based: | |||
| 948 | return -EINVAL; | 942 | return -EINVAL; |
| 949 | } | 943 | } |
| 950 | 944 | ||
| 945 | if (list_empty(devices)) { | ||
| 946 | int srcu_idx; | ||
| 947 | struct dm_table *live_table = dm_get_live_table(t->md, &srcu_idx); | ||
| 948 | |||
| 949 | /* inherit live table's type and all_blk_mq */ | ||
| 950 | if (live_table) { | ||
| 951 | t->type = live_table->type; | ||
| 952 | t->all_blk_mq = live_table->all_blk_mq; | ||
| 953 | } | ||
| 954 | dm_put_live_table(t->md, srcu_idx); | ||
| 955 | return 0; | ||
| 956 | } | ||
| 957 | |||
| 951 | /* Non-request-stackable devices can't be used for request-based dm */ | 958 | /* Non-request-stackable devices can't be used for request-based dm */ |
| 952 | list_for_each_entry(dd, devices, list) { | 959 | list_for_each_entry(dd, devices, list) { |
| 953 | struct request_queue *q = bdev_get_queue(dd->dm_dev->bdev); | 960 | struct request_queue *q = bdev_get_queue(dd->dm_dev->bdev); |
| @@ -959,19 +966,19 @@ verify_rq_based: | |||
| 959 | } | 966 | } |
| 960 | 967 | ||
| 961 | if (q->mq_ops) | 968 | if (q->mq_ops) |
| 962 | verify_blk_mq = true; | 969 | mq_count++; |
| 970 | else | ||
| 971 | sq_count++; | ||
| 963 | } | 972 | } |
| 973 | if (sq_count && mq_count) { | ||
| 974 | DMERR("table load rejected: not all devices are blk-mq request-stackable"); | ||
| 975 | return -EINVAL; | ||
| 976 | } | ||
| 977 | t->all_blk_mq = mq_count > 0; | ||
| 964 | 978 | ||
| 965 | if (verify_blk_mq) { | 979 | if (t->type == DM_TYPE_MQ_REQUEST_BASED && !t->all_blk_mq) { |
| 966 | /* verify _all_ devices in the table are blk-mq devices */ | 980 | DMERR("table load rejected: all devices are not blk-mq request-stackable"); |
| 967 | list_for_each_entry(dd, devices, list) | 981 | return -EINVAL; |
| 968 | if (!bdev_get_queue(dd->dm_dev->bdev)->mq_ops) { | ||
| 969 | DMERR("table load rejected: not all devices" | ||
| 970 | " are blk-mq request-stackable"); | ||
| 971 | return -EINVAL; | ||
| 972 | } | ||
| 973 | |||
| 974 | t->all_blk_mq = true; | ||
| 975 | } | 982 | } |
| 976 | 983 | ||
| 977 | return 0; | 984 | return 0; |
diff --git a/drivers/md/dm-verity-target.c b/drivers/md/dm-verity-target.c index 0aba34a7b3b3..7335d8a3fc47 100644 --- a/drivers/md/dm-verity-target.c +++ b/drivers/md/dm-verity-target.c | |||
| @@ -868,7 +868,7 @@ static int verity_ctr(struct dm_target *ti, unsigned argc, char **argv) | |||
| 868 | 868 | ||
| 869 | r = dm_get_device(ti, argv[2], FMODE_READ, &v->hash_dev); | 869 | r = dm_get_device(ti, argv[2], FMODE_READ, &v->hash_dev); |
| 870 | if (r) { | 870 | if (r) { |
| 871 | ti->error = "Data device lookup failed"; | 871 | ti->error = "Hash device lookup failed"; |
| 872 | goto bad; | 872 | goto bad; |
| 873 | } | 873 | } |
| 874 | 874 | ||
diff --git a/drivers/md/dm.c b/drivers/md/dm.c index ffa97b742a68..3086da5664f3 100644 --- a/drivers/md/dm.c +++ b/drivers/md/dm.c | |||
| @@ -1886,9 +1886,7 @@ static void __dm_destroy(struct mapped_device *md, bool wait) | |||
| 1886 | set_bit(DMF_FREEING, &md->flags); | 1886 | set_bit(DMF_FREEING, &md->flags); |
| 1887 | spin_unlock(&_minor_lock); | 1887 | spin_unlock(&_minor_lock); |
| 1888 | 1888 | ||
| 1889 | spin_lock_irq(q->queue_lock); | 1889 | blk_set_queue_dying(q); |
| 1890 | queue_flag_set(QUEUE_FLAG_DYING, q); | ||
| 1891 | spin_unlock_irq(q->queue_lock); | ||
| 1892 | 1890 | ||
| 1893 | if (dm_request_based(md) && md->kworker_task) | 1891 | if (dm_request_based(md) && md->kworker_task) |
| 1894 | kthread_flush_worker(&md->kworker); | 1892 | kthread_flush_worker(&md->kworker); |
diff --git a/drivers/md/persistent-data/dm-array.c b/drivers/md/persistent-data/dm-array.c index e83047cbb2da..7938cd21fa4c 100644 --- a/drivers/md/persistent-data/dm-array.c +++ b/drivers/md/persistent-data/dm-array.c | |||
| @@ -700,13 +700,11 @@ static int populate_ablock_with_values(struct dm_array_info *info, struct array_ | |||
| 700 | { | 700 | { |
| 701 | int r; | 701 | int r; |
| 702 | unsigned i; | 702 | unsigned i; |
| 703 | uint32_t nr_entries; | ||
| 704 | struct dm_btree_value_type *vt = &info->value_type; | 703 | struct dm_btree_value_type *vt = &info->value_type; |
| 705 | 704 | ||
| 706 | BUG_ON(le32_to_cpu(ab->nr_entries)); | 705 | BUG_ON(le32_to_cpu(ab->nr_entries)); |
| 707 | BUG_ON(new_nr > le32_to_cpu(ab->max_entries)); | 706 | BUG_ON(new_nr > le32_to_cpu(ab->max_entries)); |
| 708 | 707 | ||
| 709 | nr_entries = le32_to_cpu(ab->nr_entries); | ||
| 710 | for (i = 0; i < new_nr; i++) { | 708 | for (i = 0; i < new_nr; i++) { |
| 711 | r = fn(base + i, element_at(info, ab, i), context); | 709 | r = fn(base + i, element_at(info, ab, i), context); |
| 712 | if (r) | 710 | if (r) |
diff --git a/drivers/md/persistent-data/dm-block-manager.c b/drivers/md/persistent-data/dm-block-manager.c index 1e33dd51c21f..a6dde7cab458 100644 --- a/drivers/md/persistent-data/dm-block-manager.c +++ b/drivers/md/persistent-data/dm-block-manager.c | |||
| @@ -18,6 +18,8 @@ | |||
| 18 | 18 | ||
| 19 | /*----------------------------------------------------------------*/ | 19 | /*----------------------------------------------------------------*/ |
| 20 | 20 | ||
| 21 | #ifdef CONFIG_DM_DEBUG_BLOCK_MANAGER_LOCKING | ||
| 22 | |||
| 21 | /* | 23 | /* |
| 22 | * This is a read/write semaphore with a couple of differences. | 24 | * This is a read/write semaphore with a couple of differences. |
| 23 | * | 25 | * |
| @@ -302,6 +304,18 @@ static void report_recursive_bug(dm_block_t b, int r) | |||
| 302 | (unsigned long long) b); | 304 | (unsigned long long) b); |
| 303 | } | 305 | } |
| 304 | 306 | ||
| 307 | #else /* !CONFIG_DM_DEBUG_BLOCK_MANAGER_LOCKING */ | ||
| 308 | |||
| 309 | #define bl_init(x) do { } while (0) | ||
| 310 | #define bl_down_read(x) 0 | ||
| 311 | #define bl_down_read_nonblock(x) 0 | ||
| 312 | #define bl_up_read(x) do { } while (0) | ||
| 313 | #define bl_down_write(x) 0 | ||
| 314 | #define bl_up_write(x) do { } while (0) | ||
| 315 | #define report_recursive_bug(x, y) do { } while (0) | ||
| 316 | |||
| 317 | #endif /* CONFIG_DM_DEBUG_BLOCK_MANAGER_LOCKING */ | ||
| 318 | |||
| 305 | /*----------------------------------------------------------------*/ | 319 | /*----------------------------------------------------------------*/ |
| 306 | 320 | ||
| 307 | /* | 321 | /* |
| @@ -330,8 +344,11 @@ EXPORT_SYMBOL_GPL(dm_block_data); | |||
| 330 | 344 | ||
| 331 | struct buffer_aux { | 345 | struct buffer_aux { |
| 332 | struct dm_block_validator *validator; | 346 | struct dm_block_validator *validator; |
| 333 | struct block_lock lock; | ||
| 334 | int write_locked; | 347 | int write_locked; |
| 348 | |||
| 349 | #ifdef CONFIG_DM_DEBUG_BLOCK_MANAGER_LOCKING | ||
| 350 | struct block_lock lock; | ||
| 351 | #endif | ||
| 335 | }; | 352 | }; |
| 336 | 353 | ||
| 337 | static void dm_block_manager_alloc_callback(struct dm_buffer *buf) | 354 | static void dm_block_manager_alloc_callback(struct dm_buffer *buf) |
diff --git a/drivers/md/persistent-data/dm-space-map-common.c b/drivers/md/persistent-data/dm-space-map-common.c index 306d2e4502c4..4c28608a0c94 100644 --- a/drivers/md/persistent-data/dm-space-map-common.c +++ b/drivers/md/persistent-data/dm-space-map-common.c | |||
| @@ -464,7 +464,8 @@ static int sm_ll_mutate(struct ll_disk *ll, dm_block_t b, | |||
| 464 | ll->nr_allocated--; | 464 | ll->nr_allocated--; |
| 465 | le32_add_cpu(&ie_disk.nr_free, 1); | 465 | le32_add_cpu(&ie_disk.nr_free, 1); |
| 466 | ie_disk.none_free_before = cpu_to_le32(min(le32_to_cpu(ie_disk.none_free_before), bit)); | 466 | ie_disk.none_free_before = cpu_to_le32(min(le32_to_cpu(ie_disk.none_free_before), bit)); |
| 467 | } | 467 | } else |
| 468 | *ev = SM_NONE; | ||
| 468 | 469 | ||
| 469 | return ll->save_ie(ll, index, &ie_disk); | 470 | return ll->save_ie(ll, index, &ie_disk); |
| 470 | } | 471 | } |
| @@ -547,7 +548,6 @@ static int metadata_ll_init_index(struct ll_disk *ll) | |||
| 547 | if (r < 0) | 548 | if (r < 0) |
| 548 | return r; | 549 | return r; |
| 549 | 550 | ||
| 550 | memcpy(dm_block_data(b), &ll->mi_le, sizeof(ll->mi_le)); | ||
| 551 | ll->bitmap_root = dm_block_location(b); | 551 | ll->bitmap_root = dm_block_location(b); |
| 552 | 552 | ||
| 553 | dm_tm_unlock(ll->tm, b); | 553 | dm_tm_unlock(ll->tm, b); |
diff --git a/drivers/md/persistent-data/dm-space-map-metadata.c b/drivers/md/persistent-data/dm-space-map-metadata.c index 7e44005595c1..20557e2c60c6 100644 --- a/drivers/md/persistent-data/dm-space-map-metadata.c +++ b/drivers/md/persistent-data/dm-space-map-metadata.c | |||
| @@ -775,17 +775,15 @@ int dm_sm_metadata_create(struct dm_space_map *sm, | |||
| 775 | memcpy(&smm->sm, &bootstrap_ops, sizeof(smm->sm)); | 775 | memcpy(&smm->sm, &bootstrap_ops, sizeof(smm->sm)); |
| 776 | 776 | ||
| 777 | r = sm_ll_new_metadata(&smm->ll, tm); | 777 | r = sm_ll_new_metadata(&smm->ll, tm); |
| 778 | if (!r) { | ||
| 779 | if (nr_blocks > DM_SM_METADATA_MAX_BLOCKS) | ||
| 780 | nr_blocks = DM_SM_METADATA_MAX_BLOCKS; | ||
| 781 | r = sm_ll_extend(&smm->ll, nr_blocks); | ||
| 782 | } | ||
| 783 | memcpy(&smm->sm, &ops, sizeof(smm->sm)); | ||
| 778 | if (r) | 784 | if (r) |
| 779 | return r; | 785 | return r; |
| 780 | 786 | ||
| 781 | if (nr_blocks > DM_SM_METADATA_MAX_BLOCKS) | ||
| 782 | nr_blocks = DM_SM_METADATA_MAX_BLOCKS; | ||
| 783 | r = sm_ll_extend(&smm->ll, nr_blocks); | ||
| 784 | if (r) | ||
| 785 | return r; | ||
| 786 | |||
| 787 | memcpy(&smm->sm, &ops, sizeof(smm->sm)); | ||
| 788 | |||
| 789 | /* | 787 | /* |
| 790 | * Now we need to update the newly created data structures with the | 788 | * Now we need to update the newly created data structures with the |
| 791 | * allocated blocks that they were built from. | 789 | * allocated blocks that they were built from. |
diff --git a/include/uapi/linux/dm-log-userspace.h b/include/uapi/linux/dm-log-userspace.h index 0fa0d9ef06a5..05e91e14c501 100644 --- a/include/uapi/linux/dm-log-userspace.h +++ b/include/uapi/linux/dm-log-userspace.h | |||
| @@ -7,6 +7,7 @@ | |||
| 7 | #ifndef __DM_LOG_USERSPACE_H__ | 7 | #ifndef __DM_LOG_USERSPACE_H__ |
| 8 | #define __DM_LOG_USERSPACE_H__ | 8 | #define __DM_LOG_USERSPACE_H__ |
| 9 | 9 | ||
| 10 | #include <linux/types.h> | ||
| 10 | #include <linux/dm-ioctl.h> /* For DM_UUID_LEN */ | 11 | #include <linux/dm-ioctl.h> /* For DM_UUID_LEN */ |
| 11 | 12 | ||
| 12 | /* | 13 | /* |
| @@ -147,12 +148,12 @@ | |||
| 147 | 148 | ||
| 148 | /* | 149 | /* |
| 149 | * DM_ULOG_GET_REGION_SIZE corresponds to (found in dm-dirty-log.h): | 150 | * DM_ULOG_GET_REGION_SIZE corresponds to (found in dm-dirty-log.h): |
| 150 | * uint32_t (*get_region_size)(struct dm_dirty_log *log); | 151 | * __u32 (*get_region_size)(struct dm_dirty_log *log); |
| 151 | * | 152 | * |
| 152 | * Payload-to-userspace: | 153 | * Payload-to-userspace: |
| 153 | * None. | 154 | * None. |
| 154 | * Payload-to-kernel: | 155 | * Payload-to-kernel: |
| 155 | * uint64_t - contains the region size | 156 | * __u64 - contains the region size |
| 156 | * | 157 | * |
| 157 | * The region size is something that was determined at constructor time. | 158 | * The region size is something that was determined at constructor time. |
| 158 | * It is returned in the payload area and 'data_size' is set to | 159 | * It is returned in the payload area and 'data_size' is set to |
| @@ -168,11 +169,11 @@ | |||
| 168 | * int (*is_clean)(struct dm_dirty_log *log, region_t region); | 169 | * int (*is_clean)(struct dm_dirty_log *log, region_t region); |
| 169 | * | 170 | * |
| 170 | * Payload-to-userspace: | 171 | * Payload-to-userspace: |
| 171 | * uint64_t - the region to get clean status on | 172 | * __u64 - the region to get clean status on |
| 172 | * Payload-to-kernel: | 173 | * Payload-to-kernel: |
| 173 | * int64_t - 1 if clean, 0 otherwise | 174 | * __s64 - 1 if clean, 0 otherwise |
| 174 | * | 175 | * |
| 175 | * Payload is sizeof(uint64_t) and contains the region for which the clean | 176 | * Payload is sizeof(__u64) and contains the region for which the clean |
| 176 | * status is being made. | 177 | * status is being made. |
| 177 | * | 178 | * |
| 178 | * When the request has been processed, user-space must return the | 179 | * When the request has been processed, user-space must return the |
| @@ -187,9 +188,9 @@ | |||
| 187 | * int can_block); | 188 | * int can_block); |
| 188 | * | 189 | * |
| 189 | * Payload-to-userspace: | 190 | * Payload-to-userspace: |
| 190 | * uint64_t - the region to get sync status on | 191 | * __u64 - the region to get sync status on |
| 191 | * Payload-to-kernel: | 192 | * Payload-to-kernel: |
| 192 | * int64_t - 1 if in-sync, 0 otherwise | 193 | * __s64 - 1 if in-sync, 0 otherwise |
| 193 | * | 194 | * |
| 194 | * Exactly the same as 'is_clean' above, except this time asking "has the | 195 | * Exactly the same as 'is_clean' above, except this time asking "has the |
| 195 | * region been recovered?" vs. "is the region not being modified?" | 196 | * region been recovered?" vs. "is the region not being modified?" |
| @@ -203,7 +204,7 @@ | |||
| 203 | * Payload-to-userspace: | 204 | * Payload-to-userspace: |
| 204 | * If the 'integrated_flush' directive is present in the constructor | 205 | * If the 'integrated_flush' directive is present in the constructor |
| 205 | * table, the payload is as same as DM_ULOG_MARK_REGION: | 206 | * table, the payload is as same as DM_ULOG_MARK_REGION: |
| 206 | * uint64_t [] - region(s) to mark | 207 | * __u64 [] - region(s) to mark |
| 207 | * else | 208 | * else |
| 208 | * None | 209 | * None |
| 209 | * Payload-to-kernel: | 210 | * Payload-to-kernel: |
| @@ -225,13 +226,13 @@ | |||
| 225 | * void (*mark_region)(struct dm_dirty_log *log, region_t region); | 226 | * void (*mark_region)(struct dm_dirty_log *log, region_t region); |
| 226 | * | 227 | * |
| 227 | * Payload-to-userspace: | 228 | * Payload-to-userspace: |
| 228 | * uint64_t [] - region(s) to mark | 229 | * __u64 [] - region(s) to mark |
| 229 | * Payload-to-kernel: | 230 | * Payload-to-kernel: |
| 230 | * None. | 231 | * None. |
| 231 | * | 232 | * |
| 232 | * Incoming payload contains the one or more regions to mark dirty. | 233 | * Incoming payload contains the one or more regions to mark dirty. |
| 233 | * The number of regions contained in the payload can be determined from | 234 | * The number of regions contained in the payload can be determined from |
| 234 | * 'data_size/sizeof(uint64_t)'. | 235 | * 'data_size/sizeof(__u64)'. |
| 235 | * | 236 | * |
| 236 | * When the request has been processed, user-space must return the | 237 | * When the request has been processed, user-space must return the |
| 237 | * dm_ulog_request to the kernel - setting the 'error' field and clearing | 238 | * dm_ulog_request to the kernel - setting the 'error' field and clearing |
| @@ -244,13 +245,13 @@ | |||
| 244 | * void (*clear_region)(struct dm_dirty_log *log, region_t region); | 245 | * void (*clear_region)(struct dm_dirty_log *log, region_t region); |
| 245 | * | 246 | * |
| 246 | * Payload-to-userspace: | 247 | * Payload-to-userspace: |
| 247 | * uint64_t [] - region(s) to clear | 248 | * __u64 [] - region(s) to clear |
| 248 | * Payload-to-kernel: | 249 | * Payload-to-kernel: |
| 249 | * None. | 250 | * None. |
| 250 | * | 251 | * |
| 251 | * Incoming payload contains the one or more regions to mark clean. | 252 | * Incoming payload contains the one or more regions to mark clean. |
| 252 | * The number of regions contained in the payload can be determined from | 253 | * The number of regions contained in the payload can be determined from |
| 253 | * 'data_size/sizeof(uint64_t)'. | 254 | * 'data_size/sizeof(__u64)'. |
| 254 | * | 255 | * |
| 255 | * When the request has been processed, user-space must return the | 256 | * When the request has been processed, user-space must return the |
| 256 | * dm_ulog_request to the kernel - setting the 'error' field and clearing | 257 | * dm_ulog_request to the kernel - setting the 'error' field and clearing |
| @@ -266,8 +267,8 @@ | |||
| 266 | * None. | 267 | * None. |
| 267 | * Payload-to-kernel: | 268 | * Payload-to-kernel: |
| 268 | * { | 269 | * { |
| 269 | * int64_t i; -- 1 if recovery necessary, 0 otherwise | 270 | * __s64 i; -- 1 if recovery necessary, 0 otherwise |
| 270 | * uint64_t r; -- The region to recover if i=1 | 271 | * __u64 r; -- The region to recover if i=1 |
| 271 | * } | 272 | * } |
| 272 | * 'data_size' should be set appropriately. | 273 | * 'data_size' should be set appropriately. |
| 273 | * | 274 | * |
| @@ -283,8 +284,8 @@ | |||
| 283 | * | 284 | * |
| 284 | * Payload-to-userspace: | 285 | * Payload-to-userspace: |
| 285 | * { | 286 | * { |
| 286 | * uint64_t - region to set sync state on | 287 | * __u64 - region to set sync state on |
| 287 | * int64_t - 0 if not-in-sync, 1 if in-sync | 288 | * __s64 - 0 if not-in-sync, 1 if in-sync |
| 288 | * } | 289 | * } |
| 289 | * Payload-to-kernel: | 290 | * Payload-to-kernel: |
| 290 | * None. | 291 | * None. |
| @@ -302,7 +303,7 @@ | |||
| 302 | * Payload-to-userspace: | 303 | * Payload-to-userspace: |
| 303 | * None. | 304 | * None. |
| 304 | * Payload-to-kernel: | 305 | * Payload-to-kernel: |
| 305 | * uint64_t - the number of in-sync regions | 306 | * __u64 - the number of in-sync regions |
| 306 | * | 307 | * |
| 307 | * No incoming payload. Kernel-bound payload contains the number of | 308 | * No incoming payload. Kernel-bound payload contains the number of |
| 308 | * regions that are in-sync (in a size_t). | 309 | * regions that are in-sync (in a size_t). |
| @@ -350,11 +351,11 @@ | |||
| 350 | * int (*is_remote_recovering)(struct dm_dirty_log *log, region_t region); | 351 | * int (*is_remote_recovering)(struct dm_dirty_log *log, region_t region); |
| 351 | * | 352 | * |
| 352 | * Payload-to-userspace: | 353 | * Payload-to-userspace: |
| 353 | * uint64_t - region to determine recovery status on | 354 | * __u64 - region to determine recovery status on |
| 354 | * Payload-to-kernel: | 355 | * Payload-to-kernel: |
| 355 | * { | 356 | * { |
| 356 | * int64_t is_recovering; -- 0 if no, 1 if yes | 357 | * __s64 is_recovering; -- 0 if no, 1 if yes |
| 357 | * uint64_t in_sync_hint; -- lowest region still needing resync | 358 | * __u64 in_sync_hint; -- lowest region still needing resync |
| 358 | * } | 359 | * } |
| 359 | * | 360 | * |
| 360 | * When the request has been processed, user-space must return the | 361 | * When the request has been processed, user-space must return the |
| @@ -413,16 +414,16 @@ struct dm_ulog_request { | |||
| 413 | * differentiate between logs that are being swapped and have the | 414 | * differentiate between logs that are being swapped and have the |
| 414 | * same 'uuid'. (Think "live" and "inactive" device-mapper tables.) | 415 | * same 'uuid'. (Think "live" and "inactive" device-mapper tables.) |
| 415 | */ | 416 | */ |
| 416 | uint64_t luid; | 417 | __u64 luid; |
| 417 | char uuid[DM_UUID_LEN]; | 418 | char uuid[DM_UUID_LEN]; |
| 418 | char padding[3]; /* Padding because DM_UUID_LEN = 129 */ | 419 | char padding[3]; /* Padding because DM_UUID_LEN = 129 */ |
| 419 | 420 | ||
| 420 | uint32_t version; /* See DM_ULOG_REQUEST_VERSION */ | 421 | __u32 version; /* See DM_ULOG_REQUEST_VERSION */ |
| 421 | int32_t error; /* Used to report back processing errors */ | 422 | __s32 error; /* Used to report back processing errors */ |
| 422 | 423 | ||
| 423 | uint32_t seq; /* Sequence number for request */ | 424 | __u32 seq; /* Sequence number for request */ |
| 424 | uint32_t request_type; /* DM_ULOG_* defined above */ | 425 | __u32 request_type; /* DM_ULOG_* defined above */ |
| 425 | uint32_t data_size; /* How much data (not including this struct) */ | 426 | __u32 data_size; /* How much data (not including this struct) */ |
| 426 | 427 | ||
| 427 | char data[0]; | 428 | char data[0]; |
| 428 | }; | 429 | }; |
