diff options
-rw-r--r-- | drivers/nvdimm/blk.c | 174 | ||||
-rw-r--r-- | drivers/nvdimm/btt.h | 1 | ||||
-rw-r--r-- | drivers/nvdimm/core.c | 3 | ||||
-rw-r--r-- | drivers/nvdimm/namespace_devs.c | 3 | ||||
-rw-r--r-- | drivers/nvdimm/nd.h | 1 |
5 files changed, 159 insertions, 23 deletions
diff --git a/drivers/nvdimm/blk.c b/drivers/nvdimm/blk.c index 9ac0c266c15c..5c44e067652f 100644 --- a/drivers/nvdimm/blk.c +++ b/drivers/nvdimm/blk.c | |||
@@ -27,10 +27,17 @@ struct nd_blk_device { | |||
27 | struct nd_namespace_blk *nsblk; | 27 | struct nd_namespace_blk *nsblk; |
28 | struct nd_blk_region *ndbr; | 28 | struct nd_blk_region *ndbr; |
29 | size_t disk_size; | 29 | size_t disk_size; |
30 | u32 sector_size; | ||
31 | u32 internal_lbasize; | ||
30 | }; | 32 | }; |
31 | 33 | ||
32 | static int nd_blk_major; | 34 | static int nd_blk_major; |
33 | 35 | ||
36 | static u32 nd_blk_meta_size(struct nd_blk_device *blk_dev) | ||
37 | { | ||
38 | return blk_dev->nsblk->lbasize - blk_dev->sector_size; | ||
39 | } | ||
40 | |||
34 | static resource_size_t to_dev_offset(struct nd_namespace_blk *nsblk, | 41 | static resource_size_t to_dev_offset(struct nd_namespace_blk *nsblk, |
35 | resource_size_t ns_offset, unsigned int len) | 42 | resource_size_t ns_offset, unsigned int len) |
36 | { | 43 | { |
@@ -52,41 +59,145 @@ static resource_size_t to_dev_offset(struct nd_namespace_blk *nsblk, | |||
52 | return SIZE_MAX; | 59 | return SIZE_MAX; |
53 | } | 60 | } |
54 | 61 | ||
62 | #ifdef CONFIG_BLK_DEV_INTEGRITY | ||
63 | static int nd_blk_rw_integrity(struct nd_blk_device *blk_dev, | ||
64 | struct bio_integrity_payload *bip, u64 lba, | ||
65 | int rw) | ||
66 | { | ||
67 | unsigned int len = nd_blk_meta_size(blk_dev); | ||
68 | resource_size_t dev_offset, ns_offset; | ||
69 | struct nd_namespace_blk *nsblk; | ||
70 | struct nd_blk_region *ndbr; | ||
71 | int err = 0; | ||
72 | |||
73 | nsblk = blk_dev->nsblk; | ||
74 | ndbr = blk_dev->ndbr; | ||
75 | ns_offset = lba * blk_dev->internal_lbasize + blk_dev->sector_size; | ||
76 | dev_offset = to_dev_offset(nsblk, ns_offset, len); | ||
77 | if (dev_offset == SIZE_MAX) | ||
78 | return -EIO; | ||
79 | |||
80 | while (len) { | ||
81 | unsigned int cur_len; | ||
82 | struct bio_vec bv; | ||
83 | void *iobuf; | ||
84 | |||
85 | bv = bvec_iter_bvec(bip->bip_vec, bip->bip_iter); | ||
86 | /* | ||
87 | * The 'bv' obtained from bvec_iter_bvec has its .bv_len and | ||
88 | * .bv_offset already adjusted for iter->bi_bvec_done, and we | ||
89 | * can use those directly | ||
90 | */ | ||
91 | |||
92 | cur_len = min(len, bv.bv_len); | ||
93 | iobuf = kmap_atomic(bv.bv_page); | ||
94 | err = ndbr->do_io(ndbr, dev_offset, iobuf + bv.bv_offset, | ||
95 | cur_len, rw); | ||
96 | kunmap_atomic(iobuf); | ||
97 | if (err) | ||
98 | return err; | ||
99 | |||
100 | len -= cur_len; | ||
101 | dev_offset += cur_len; | ||
102 | bvec_iter_advance(bip->bip_vec, &bip->bip_iter, cur_len); | ||
103 | } | ||
104 | |||
105 | return err; | ||
106 | } | ||
107 | |||
108 | #else /* CONFIG_BLK_DEV_INTEGRITY */ | ||
109 | static int nd_blk_rw_integrity(struct nd_blk_device *blk_dev, | ||
110 | struct bio_integrity_payload *bip, u64 lba, | ||
111 | int rw) | ||
112 | { | ||
113 | return 0; | ||
114 | } | ||
115 | #endif | ||
116 | |||
117 | static int nd_blk_do_bvec(struct nd_blk_device *blk_dev, | ||
118 | struct bio_integrity_payload *bip, struct page *page, | ||
119 | unsigned int len, unsigned int off, int rw, | ||
120 | sector_t sector) | ||
121 | { | ||
122 | struct nd_blk_region *ndbr = blk_dev->ndbr; | ||
123 | resource_size_t dev_offset, ns_offset; | ||
124 | int err = 0; | ||
125 | void *iobuf; | ||
126 | u64 lba; | ||
127 | |||
128 | while (len) { | ||
129 | unsigned int cur_len; | ||
130 | |||
131 | /* | ||
132 | * If we don't have an integrity payload, we don't have to | ||
133 | * split the bvec into sectors, as this would cause unnecessary | ||
134 | * Block Window setup/move steps. the do_io routine is capable | ||
135 | * of handling len <= PAGE_SIZE. | ||
136 | */ | ||
137 | cur_len = bip ? min(len, blk_dev->sector_size) : len; | ||
138 | |||
139 | lba = div_u64(sector << SECTOR_SHIFT, blk_dev->sector_size); | ||
140 | ns_offset = lba * blk_dev->internal_lbasize; | ||
141 | dev_offset = to_dev_offset(blk_dev->nsblk, ns_offset, cur_len); | ||
142 | if (dev_offset == SIZE_MAX) | ||
143 | return -EIO; | ||
144 | |||
145 | iobuf = kmap_atomic(page); | ||
146 | err = ndbr->do_io(ndbr, dev_offset, iobuf + off, cur_len, rw); | ||
147 | kunmap_atomic(iobuf); | ||
148 | if (err) | ||
149 | return err; | ||
150 | |||
151 | if (bip) { | ||
152 | err = nd_blk_rw_integrity(blk_dev, bip, lba, rw); | ||
153 | if (err) | ||
154 | return err; | ||
155 | } | ||
156 | len -= cur_len; | ||
157 | off += cur_len; | ||
158 | sector += blk_dev->sector_size >> SECTOR_SHIFT; | ||
159 | } | ||
160 | |||
161 | return err; | ||
162 | } | ||
163 | |||
55 | static void nd_blk_make_request(struct request_queue *q, struct bio *bio) | 164 | static void nd_blk_make_request(struct request_queue *q, struct bio *bio) |
56 | { | 165 | { |
57 | struct block_device *bdev = bio->bi_bdev; | 166 | struct block_device *bdev = bio->bi_bdev; |
58 | struct gendisk *disk = bdev->bd_disk; | 167 | struct gendisk *disk = bdev->bd_disk; |
59 | struct nd_namespace_blk *nsblk; | 168 | struct bio_integrity_payload *bip; |
60 | struct nd_blk_device *blk_dev; | 169 | struct nd_blk_device *blk_dev; |
61 | struct nd_blk_region *ndbr; | ||
62 | struct bvec_iter iter; | 170 | struct bvec_iter iter; |
63 | struct bio_vec bvec; | 171 | struct bio_vec bvec; |
64 | int err = 0, rw; | 172 | int err = 0, rw; |
65 | 173 | ||
174 | /* | ||
175 | * bio_integrity_enabled also checks if the bio already has an | ||
176 | * integrity payload attached. If it does, we *don't* do a | ||
177 | * bio_integrity_prep here - the payload has been generated by | ||
178 | * another kernel subsystem, and we just pass it through. | ||
179 | */ | ||
180 | if (bio_integrity_enabled(bio) && bio_integrity_prep(bio)) { | ||
181 | err = -EIO; | ||
182 | goto out; | ||
183 | } | ||
184 | |||
185 | bip = bio_integrity(bio); | ||
66 | blk_dev = disk->private_data; | 186 | blk_dev = disk->private_data; |
67 | nsblk = blk_dev->nsblk; | ||
68 | ndbr = blk_dev->ndbr; | ||
69 | rw = bio_data_dir(bio); | 187 | rw = bio_data_dir(bio); |
70 | bio_for_each_segment(bvec, bio, iter) { | 188 | bio_for_each_segment(bvec, bio, iter) { |
71 | unsigned int len = bvec.bv_len; | 189 | unsigned int len = bvec.bv_len; |
72 | resource_size_t dev_offset; | ||
73 | void *iobuf; | ||
74 | 190 | ||
75 | BUG_ON(len > PAGE_SIZE); | 191 | BUG_ON(len > PAGE_SIZE); |
76 | 192 | err = nd_blk_do_bvec(blk_dev, bip, bvec.bv_page, len, | |
77 | dev_offset = to_dev_offset(nsblk, | 193 | bvec.bv_offset, rw, iter.bi_sector); |
78 | iter.bi_sector << SECTOR_SHIFT, len); | 194 | if (err) { |
79 | if (dev_offset == SIZE_MAX) { | 195 | dev_info(&blk_dev->nsblk->common.dev, |
80 | err = -EIO; | 196 | "io error in %s sector %lld, len %d,\n", |
197 | (rw == READ) ? "READ" : "WRITE", | ||
198 | (unsigned long long) iter.bi_sector, len); | ||
81 | goto out; | 199 | goto out; |
82 | } | 200 | } |
83 | |||
84 | iobuf = kmap_atomic(bvec.bv_page); | ||
85 | err = ndbr->do_io(ndbr, dev_offset, iobuf + bvec.bv_offset, | ||
86 | len, rw); | ||
87 | kunmap_atomic(iobuf); | ||
88 | if (err) | ||
89 | goto out; | ||
90 | } | 201 | } |
91 | 202 | ||
92 | out: | 203 | out: |
@@ -121,8 +232,12 @@ static const struct block_device_operations nd_blk_fops = { | |||
121 | static int nd_blk_attach_disk(struct nd_namespace_common *ndns, | 232 | static int nd_blk_attach_disk(struct nd_namespace_common *ndns, |
122 | struct nd_blk_device *blk_dev) | 233 | struct nd_blk_device *blk_dev) |
123 | { | 234 | { |
124 | struct nd_namespace_blk *nsblk = to_nd_namespace_blk(&ndns->dev); | 235 | resource_size_t available_disk_size; |
125 | struct gendisk *disk; | 236 | struct gendisk *disk; |
237 | u64 internal_nlba; | ||
238 | |||
239 | internal_nlba = div_u64(blk_dev->disk_size, blk_dev->internal_lbasize); | ||
240 | available_disk_size = internal_nlba * blk_dev->sector_size; | ||
126 | 241 | ||
127 | blk_dev->queue = blk_alloc_queue(GFP_KERNEL); | 242 | blk_dev->queue = blk_alloc_queue(GFP_KERNEL); |
128 | if (!blk_dev->queue) | 243 | if (!blk_dev->queue) |
@@ -131,7 +246,7 @@ static int nd_blk_attach_disk(struct nd_namespace_common *ndns, | |||
131 | blk_queue_make_request(blk_dev->queue, nd_blk_make_request); | 246 | blk_queue_make_request(blk_dev->queue, nd_blk_make_request); |
132 | blk_queue_max_hw_sectors(blk_dev->queue, UINT_MAX); | 247 | blk_queue_max_hw_sectors(blk_dev->queue, UINT_MAX); |
133 | blk_queue_bounce_limit(blk_dev->queue, BLK_BOUNCE_ANY); | 248 | blk_queue_bounce_limit(blk_dev->queue, BLK_BOUNCE_ANY); |
134 | blk_queue_logical_block_size(blk_dev->queue, nsblk->lbasize); | 249 | blk_queue_logical_block_size(blk_dev->queue, blk_dev->sector_size); |
135 | queue_flag_set_unlocked(QUEUE_FLAG_NONROT, blk_dev->queue); | 250 | queue_flag_set_unlocked(QUEUE_FLAG_NONROT, blk_dev->queue); |
136 | 251 | ||
137 | disk = blk_dev->disk = alloc_disk(0); | 252 | disk = blk_dev->disk = alloc_disk(0); |
@@ -148,15 +263,28 @@ static int nd_blk_attach_disk(struct nd_namespace_common *ndns, | |||
148 | disk->queue = blk_dev->queue; | 263 | disk->queue = blk_dev->queue; |
149 | disk->flags = GENHD_FL_EXT_DEVT; | 264 | disk->flags = GENHD_FL_EXT_DEVT; |
150 | nvdimm_namespace_disk_name(ndns, disk->disk_name); | 265 | nvdimm_namespace_disk_name(ndns, disk->disk_name); |
151 | set_capacity(disk, blk_dev->disk_size >> SECTOR_SHIFT); | 266 | set_capacity(disk, 0); |
152 | add_disk(disk); | 267 | add_disk(disk); |
153 | 268 | ||
269 | if (nd_blk_meta_size(blk_dev)) { | ||
270 | int rc = nd_integrity_init(disk, nd_blk_meta_size(blk_dev)); | ||
271 | |||
272 | if (rc) { | ||
273 | del_gendisk(disk); | ||
274 | put_disk(disk); | ||
275 | blk_cleanup_queue(blk_dev->queue); | ||
276 | return rc; | ||
277 | } | ||
278 | } | ||
279 | |||
280 | set_capacity(disk, available_disk_size >> SECTOR_SHIFT); | ||
154 | return 0; | 281 | return 0; |
155 | } | 282 | } |
156 | 283 | ||
157 | static int nd_blk_probe(struct device *dev) | 284 | static int nd_blk_probe(struct device *dev) |
158 | { | 285 | { |
159 | struct nd_namespace_common *ndns; | 286 | struct nd_namespace_common *ndns; |
287 | struct nd_namespace_blk *nsblk; | ||
160 | struct nd_blk_device *blk_dev; | 288 | struct nd_blk_device *blk_dev; |
161 | int rc; | 289 | int rc; |
162 | 290 | ||
@@ -168,9 +296,13 @@ static int nd_blk_probe(struct device *dev) | |||
168 | if (!blk_dev) | 296 | if (!blk_dev) |
169 | return -ENOMEM; | 297 | return -ENOMEM; |
170 | 298 | ||
299 | nsblk = to_nd_namespace_blk(&ndns->dev); | ||
171 | blk_dev->disk_size = nvdimm_namespace_capacity(ndns); | 300 | blk_dev->disk_size = nvdimm_namespace_capacity(ndns); |
172 | blk_dev->ndbr = to_nd_blk_region(dev->parent); | 301 | blk_dev->ndbr = to_nd_blk_region(dev->parent); |
173 | blk_dev->nsblk = to_nd_namespace_blk(&ndns->dev); | 302 | blk_dev->nsblk = to_nd_namespace_blk(&ndns->dev); |
303 | blk_dev->internal_lbasize = roundup(nsblk->lbasize, | ||
304 | INT_LBASIZE_ALIGNMENT); | ||
305 | blk_dev->sector_size = ((nsblk->lbasize >= 4096) ? 4096 : 512); | ||
174 | dev_set_drvdata(dev, blk_dev); | 306 | dev_set_drvdata(dev, blk_dev); |
175 | 307 | ||
176 | ndns->rw_bytes = nd_blk_rw_bytes; | 308 | ndns->rw_bytes = nd_blk_rw_bytes; |
diff --git a/drivers/nvdimm/btt.h b/drivers/nvdimm/btt.h index 2caa0ef7e67a..75b0d80a6bd9 100644 --- a/drivers/nvdimm/btt.h +++ b/drivers/nvdimm/btt.h | |||
@@ -31,7 +31,6 @@ | |||
31 | #define ARENA_MAX_SIZE (1ULL << 39) /* 512 GB */ | 31 | #define ARENA_MAX_SIZE (1ULL << 39) /* 512 GB */ |
32 | #define RTT_VALID (1UL << 31) | 32 | #define RTT_VALID (1UL << 31) |
33 | #define RTT_INVALID 0 | 33 | #define RTT_INVALID 0 |
34 | #define INT_LBASIZE_ALIGNMENT 64 | ||
35 | #define BTT_PG_SIZE 4096 | 34 | #define BTT_PG_SIZE 4096 |
36 | #define BTT_DEFAULT_NFREE ND_MAX_LANES | 35 | #define BTT_DEFAULT_NFREE ND_MAX_LANES |
37 | #define LOG_SEQ_INIT 1 | 36 | #define LOG_SEQ_INIT 1 |
diff --git a/drivers/nvdimm/core.c b/drivers/nvdimm/core.c index 1d96b9a6e4cc..4288169432de 100644 --- a/drivers/nvdimm/core.c +++ b/drivers/nvdimm/core.c | |||
@@ -379,6 +379,9 @@ int nd_integrity_init(struct gendisk *disk, unsigned long meta_size) | |||
379 | }; | 379 | }; |
380 | int ret; | 380 | int ret; |
381 | 381 | ||
382 | if (meta_size == 0) | ||
383 | return 0; | ||
384 | |||
382 | ret = blk_integrity_register(disk, &integrity); | 385 | ret = blk_integrity_register(disk, &integrity); |
383 | if (ret) | 386 | if (ret) |
384 | return ret; | 387 | return ret; |
diff --git a/drivers/nvdimm/namespace_devs.c b/drivers/nvdimm/namespace_devs.c index 1ce1e70de44a..27d69bd3b4d6 100644 --- a/drivers/nvdimm/namespace_devs.c +++ b/drivers/nvdimm/namespace_devs.c | |||
@@ -1059,7 +1059,8 @@ static ssize_t resource_show(struct device *dev, | |||
1059 | } | 1059 | } |
1060 | static DEVICE_ATTR_RO(resource); | 1060 | static DEVICE_ATTR_RO(resource); |
1061 | 1061 | ||
1062 | static const unsigned long ns_lbasize_supported[] = { 512, 0 }; | 1062 | static const unsigned long ns_lbasize_supported[] = { 512, 520, 528, |
1063 | 4096, 4104, 4160, 4224, 0 }; | ||
1063 | 1064 | ||
1064 | static ssize_t sector_size_show(struct device *dev, | 1065 | static ssize_t sector_size_show(struct device *dev, |
1065 | struct device_attribute *attr, char *buf) | 1066 | struct device_attribute *attr, char *buf) |
diff --git a/drivers/nvdimm/nd.h b/drivers/nvdimm/nd.h index f4459faa456c..6f916b65b7d6 100644 --- a/drivers/nvdimm/nd.h +++ b/drivers/nvdimm/nd.h | |||
@@ -27,6 +27,7 @@ enum { | |||
27 | */ | 27 | */ |
28 | ND_MAX_LANES = 256, | 28 | ND_MAX_LANES = 256, |
29 | SECTOR_SHIFT = 9, | 29 | SECTOR_SHIFT = 9, |
30 | INT_LBASIZE_ALIGNMENT = 64, | ||
30 | }; | 31 | }; |
31 | 32 | ||
32 | struct nvdimm_drvdata { | 33 | struct nvdimm_drvdata { |