diff options
Diffstat (limited to 'drivers/mmc/card/mmc_test.c')
| -rw-r--r-- | drivers/mmc/card/mmc_test.c | 811 |
1 files changed, 809 insertions, 2 deletions
diff --git a/drivers/mmc/card/mmc_test.c b/drivers/mmc/card/mmc_test.c index 445d7db2277e..5dd8576b5c18 100644 --- a/drivers/mmc/card/mmc_test.c +++ b/drivers/mmc/card/mmc_test.c | |||
| @@ -16,6 +16,7 @@ | |||
| 16 | #include <linux/slab.h> | 16 | #include <linux/slab.h> |
| 17 | 17 | ||
| 18 | #include <linux/scatterlist.h> | 18 | #include <linux/scatterlist.h> |
| 19 | #include <linux/swap.h> /* For nr_free_buffer_pages() */ | ||
| 19 | 20 | ||
| 20 | #define RESULT_OK 0 | 21 | #define RESULT_OK 0 |
| 21 | #define RESULT_FAIL 1 | 22 | #define RESULT_FAIL 1 |
| @@ -25,6 +26,60 @@ | |||
| 25 | #define BUFFER_ORDER 2 | 26 | #define BUFFER_ORDER 2 |
| 26 | #define BUFFER_SIZE (PAGE_SIZE << BUFFER_ORDER) | 27 | #define BUFFER_SIZE (PAGE_SIZE << BUFFER_ORDER) |
| 27 | 28 | ||
| 29 | /* | ||
| 30 | * Limit the test area size to the maximum MMC HC erase group size. Note that | ||
| 31 | * the maximum SD allocation unit size is just 4MiB. | ||
| 32 | */ | ||
| 33 | #define TEST_AREA_MAX_SIZE (128 * 1024 * 1024) | ||
| 34 | |||
| 35 | /** | ||
| 36 | * struct mmc_test_pages - pages allocated by 'alloc_pages()'. | ||
| 37 | * @page: first page in the allocation | ||
| 38 | * @order: order of the number of pages allocated | ||
| 39 | */ | ||
| 40 | struct mmc_test_pages { | ||
| 41 | struct page *page; | ||
| 42 | unsigned int order; | ||
| 43 | }; | ||
| 44 | |||
| 45 | /** | ||
| 46 | * struct mmc_test_mem - allocated memory. | ||
| 47 | * @arr: array of allocations | ||
| 48 | * @cnt: number of allocations | ||
| 49 | */ | ||
| 50 | struct mmc_test_mem { | ||
| 51 | struct mmc_test_pages *arr; | ||
| 52 | unsigned int cnt; | ||
| 53 | }; | ||
| 54 | |||
| 55 | /** | ||
| 56 | * struct mmc_test_area - information for performance tests. | ||
| 57 | * @max_sz: test area size (in bytes) | ||
| 58 | * @dev_addr: address on card at which to do performance tests | ||
| 59 | * @max_segs: maximum segments in scatterlist @sg | ||
| 60 | * @blocks: number of (512 byte) blocks currently mapped by @sg | ||
| 61 | * @sg_len: length of currently mapped scatterlist @sg | ||
| 62 | * @mem: allocated memory | ||
| 63 | * @sg: scatterlist | ||
| 64 | */ | ||
| 65 | struct mmc_test_area { | ||
| 66 | unsigned long max_sz; | ||
| 67 | unsigned int dev_addr; | ||
| 68 | unsigned int max_segs; | ||
| 69 | unsigned int blocks; | ||
| 70 | unsigned int sg_len; | ||
| 71 | struct mmc_test_mem *mem; | ||
| 72 | struct scatterlist *sg; | ||
| 73 | }; | ||
| 74 | |||
| 75 | /** | ||
| 76 | * struct mmc_test_card - test information. | ||
| 77 | * @card: card under test | ||
| 78 | * @scratch: transfer buffer | ||
| 79 | * @buffer: transfer buffer | ||
| 80 | * @highmem: buffer for highmem tests | ||
| 81 | * @area: information for performance tests | ||
| 82 | */ | ||
| 28 | struct mmc_test_card { | 83 | struct mmc_test_card { |
| 29 | struct mmc_card *card; | 84 | struct mmc_card *card; |
| 30 | 85 | ||
| @@ -33,6 +88,7 @@ struct mmc_test_card { | |||
| 33 | #ifdef CONFIG_HIGHMEM | 88 | #ifdef CONFIG_HIGHMEM |
| 34 | struct page *highmem; | 89 | struct page *highmem; |
| 35 | #endif | 90 | #endif |
| 91 | struct mmc_test_area area; | ||
| 36 | }; | 92 | }; |
| 37 | 93 | ||
| 38 | /*******************************************************************/ | 94 | /*******************************************************************/ |
| @@ -97,6 +153,12 @@ static void mmc_test_prepare_mrq(struct mmc_test_card *test, | |||
| 97 | mmc_set_data_timeout(mrq->data, test->card); | 153 | mmc_set_data_timeout(mrq->data, test->card); |
| 98 | } | 154 | } |
| 99 | 155 | ||
| 156 | static int mmc_test_busy(struct mmc_command *cmd) | ||
| 157 | { | ||
| 158 | return !(cmd->resp[0] & R1_READY_FOR_DATA) || | ||
| 159 | (R1_CURRENT_STATE(cmd->resp[0]) == 7); | ||
| 160 | } | ||
| 161 | |||
| 100 | /* | 162 | /* |
| 101 | * Wait for the card to finish the busy state | 163 | * Wait for the card to finish the busy state |
| 102 | */ | 164 | */ |
| @@ -117,13 +179,13 @@ static int mmc_test_wait_busy(struct mmc_test_card *test) | |||
| 117 | if (ret) | 179 | if (ret) |
| 118 | break; | 180 | break; |
| 119 | 181 | ||
| 120 | if (!busy && !(cmd.resp[0] & R1_READY_FOR_DATA)) { | 182 | if (!busy && mmc_test_busy(&cmd)) { |
| 121 | busy = 1; | 183 | busy = 1; |
| 122 | printk(KERN_INFO "%s: Warning: Host did not " | 184 | printk(KERN_INFO "%s: Warning: Host did not " |
| 123 | "wait for busy state to end.\n", | 185 | "wait for busy state to end.\n", |
| 124 | mmc_hostname(test->card->host)); | 186 | mmc_hostname(test->card->host)); |
| 125 | } | 187 | } |
| 126 | } while (!(cmd.resp[0] & R1_READY_FOR_DATA)); | 188 | } while (mmc_test_busy(&cmd)); |
| 127 | 189 | ||
| 128 | return ret; | 190 | return ret; |
| 129 | } | 191 | } |
| @@ -170,6 +232,248 @@ static int mmc_test_buffer_transfer(struct mmc_test_card *test, | |||
| 170 | return 0; | 232 | return 0; |
| 171 | } | 233 | } |
| 172 | 234 | ||
| 235 | static void mmc_test_free_mem(struct mmc_test_mem *mem) | ||
| 236 | { | ||
| 237 | if (!mem) | ||
| 238 | return; | ||
| 239 | while (mem->cnt--) | ||
| 240 | __free_pages(mem->arr[mem->cnt].page, | ||
| 241 | mem->arr[mem->cnt].order); | ||
| 242 | kfree(mem->arr); | ||
| 243 | kfree(mem); | ||
| 244 | } | ||
| 245 | |||
| 246 | /* | ||
| 247 | * Allocate a lot of memory, preferrably max_sz but at least min_sz. In case | ||
| 248 | * there isn't much memory do not exceed 1/16th total lowmem pages. | ||
| 249 | */ | ||
| 250 | static struct mmc_test_mem *mmc_test_alloc_mem(unsigned long min_sz, | ||
| 251 | unsigned long max_sz) | ||
| 252 | { | ||
| 253 | unsigned long max_page_cnt = DIV_ROUND_UP(max_sz, PAGE_SIZE); | ||
| 254 | unsigned long min_page_cnt = DIV_ROUND_UP(min_sz, PAGE_SIZE); | ||
| 255 | unsigned long page_cnt = 0; | ||
| 256 | unsigned long limit = nr_free_buffer_pages() >> 4; | ||
| 257 | struct mmc_test_mem *mem; | ||
| 258 | |||
| 259 | if (max_page_cnt > limit) | ||
| 260 | max_page_cnt = limit; | ||
| 261 | if (max_page_cnt < min_page_cnt) | ||
| 262 | max_page_cnt = min_page_cnt; | ||
| 263 | |||
| 264 | mem = kzalloc(sizeof(struct mmc_test_mem), GFP_KERNEL); | ||
| 265 | if (!mem) | ||
| 266 | return NULL; | ||
| 267 | |||
| 268 | mem->arr = kzalloc(sizeof(struct mmc_test_pages) * max_page_cnt, | ||
| 269 | GFP_KERNEL); | ||
| 270 | if (!mem->arr) | ||
| 271 | goto out_free; | ||
| 272 | |||
| 273 | while (max_page_cnt) { | ||
| 274 | struct page *page; | ||
| 275 | unsigned int order; | ||
| 276 | gfp_t flags = GFP_KERNEL | GFP_DMA | __GFP_NOWARN | | ||
| 277 | __GFP_NORETRY; | ||
| 278 | |||
| 279 | order = get_order(max_page_cnt << PAGE_SHIFT); | ||
| 280 | while (1) { | ||
| 281 | page = alloc_pages(flags, order); | ||
| 282 | if (page || !order) | ||
| 283 | break; | ||
| 284 | order -= 1; | ||
| 285 | } | ||
| 286 | if (!page) { | ||
| 287 | if (page_cnt < min_page_cnt) | ||
| 288 | goto out_free; | ||
| 289 | break; | ||
| 290 | } | ||
| 291 | mem->arr[mem->cnt].page = page; | ||
| 292 | mem->arr[mem->cnt].order = order; | ||
| 293 | mem->cnt += 1; | ||
| 294 | if (max_page_cnt <= (1UL << order)) | ||
| 295 | break; | ||
| 296 | max_page_cnt -= 1UL << order; | ||
| 297 | page_cnt += 1UL << order; | ||
| 298 | } | ||
| 299 | |||
| 300 | return mem; | ||
| 301 | |||
| 302 | out_free: | ||
| 303 | mmc_test_free_mem(mem); | ||
| 304 | return NULL; | ||
| 305 | } | ||
| 306 | |||
| 307 | /* | ||
| 308 | * Map memory into a scatterlist. Optionally allow the same memory to be | ||
| 309 | * mapped more than once. | ||
| 310 | */ | ||
| 311 | static int mmc_test_map_sg(struct mmc_test_mem *mem, unsigned long sz, | ||
| 312 | struct scatterlist *sglist, int repeat, | ||
| 313 | unsigned int max_segs, unsigned int *sg_len) | ||
| 314 | { | ||
| 315 | struct scatterlist *sg = NULL; | ||
| 316 | unsigned int i; | ||
| 317 | |||
| 318 | sg_init_table(sglist, max_segs); | ||
| 319 | |||
| 320 | *sg_len = 0; | ||
| 321 | do { | ||
| 322 | for (i = 0; i < mem->cnt; i++) { | ||
| 323 | unsigned long len = PAGE_SIZE << mem->arr[i].order; | ||
| 324 | |||
| 325 | if (sz < len) | ||
| 326 | len = sz; | ||
| 327 | if (sg) | ||
| 328 | sg = sg_next(sg); | ||
| 329 | else | ||
| 330 | sg = sglist; | ||
| 331 | if (!sg) | ||
| 332 | return -EINVAL; | ||
| 333 | sg_set_page(sg, mem->arr[i].page, len, 0); | ||
| 334 | sz -= len; | ||
| 335 | *sg_len += 1; | ||
| 336 | if (!sz) | ||
| 337 | break; | ||
| 338 | } | ||
| 339 | } while (sz && repeat); | ||
| 340 | |||
| 341 | if (sz) | ||
| 342 | return -EINVAL; | ||
| 343 | |||
| 344 | if (sg) | ||
| 345 | sg_mark_end(sg); | ||
| 346 | |||
| 347 | return 0; | ||
| 348 | } | ||
| 349 | |||
| 350 | /* | ||
| 351 | * Map memory into a scatterlist so that no pages are contiguous. Allow the | ||
| 352 | * same memory to be mapped more than once. | ||
| 353 | */ | ||
| 354 | static int mmc_test_map_sg_max_scatter(struct mmc_test_mem *mem, | ||
| 355 | unsigned long sz, | ||
| 356 | struct scatterlist *sglist, | ||
| 357 | unsigned int max_segs, | ||
| 358 | unsigned int *sg_len) | ||
| 359 | { | ||
| 360 | struct scatterlist *sg = NULL; | ||
| 361 | unsigned int i = mem->cnt, cnt; | ||
| 362 | unsigned long len; | ||
| 363 | void *base, *addr, *last_addr = NULL; | ||
| 364 | |||
| 365 | sg_init_table(sglist, max_segs); | ||
| 366 | |||
| 367 | *sg_len = 0; | ||
| 368 | while (sz && i) { | ||
| 369 | base = page_address(mem->arr[--i].page); | ||
| 370 | cnt = 1 << mem->arr[i].order; | ||
| 371 | while (sz && cnt) { | ||
| 372 | addr = base + PAGE_SIZE * --cnt; | ||
| 373 | if (last_addr && last_addr + PAGE_SIZE == addr) | ||
| 374 | continue; | ||
| 375 | last_addr = addr; | ||
| 376 | len = PAGE_SIZE; | ||
| 377 | if (sz < len) | ||
| 378 | len = sz; | ||
| 379 | if (sg) | ||
| 380 | sg = sg_next(sg); | ||
| 381 | else | ||
| 382 | sg = sglist; | ||
| 383 | if (!sg) | ||
| 384 | return -EINVAL; | ||
| 385 | sg_set_page(sg, virt_to_page(addr), len, 0); | ||
| 386 | sz -= len; | ||
| 387 | *sg_len += 1; | ||
| 388 | } | ||
| 389 | } | ||
| 390 | |||
| 391 | if (sg) | ||
| 392 | sg_mark_end(sg); | ||
| 393 | |||
| 394 | return 0; | ||
| 395 | } | ||
| 396 | |||
| 397 | /* | ||
| 398 | * Calculate transfer rate in bytes per second. | ||
| 399 | */ | ||
| 400 | static unsigned int mmc_test_rate(uint64_t bytes, struct timespec *ts) | ||
| 401 | { | ||
| 402 | uint64_t ns; | ||
| 403 | |||
| 404 | ns = ts->tv_sec; | ||
| 405 | ns *= 1000000000; | ||
| 406 | ns += ts->tv_nsec; | ||
| 407 | |||
| 408 | bytes *= 1000000000; | ||
| 409 | |||
| 410 | while (ns > UINT_MAX) { | ||
| 411 | bytes >>= 1; | ||
| 412 | ns >>= 1; | ||
| 413 | } | ||
| 414 | |||
| 415 | if (!ns) | ||
| 416 | return 0; | ||
| 417 | |||
| 418 | do_div(bytes, (uint32_t)ns); | ||
| 419 | |||
| 420 | return bytes; | ||
| 421 | } | ||
| 422 | |||
| 423 | /* | ||
| 424 | * Print the transfer rate. | ||
| 425 | */ | ||
| 426 | static void mmc_test_print_rate(struct mmc_test_card *test, uint64_t bytes, | ||
| 427 | struct timespec *ts1, struct timespec *ts2) | ||
| 428 | { | ||
| 429 | unsigned int rate, sectors = bytes >> 9; | ||
| 430 | struct timespec ts; | ||
| 431 | |||
| 432 | ts = timespec_sub(*ts2, *ts1); | ||
| 433 | |||
| 434 | rate = mmc_test_rate(bytes, &ts); | ||
| 435 | |||
| 436 | printk(KERN_INFO "%s: Transfer of %u sectors (%u%s KiB) took %lu.%09lu " | ||
| 437 | "seconds (%u kB/s, %u KiB/s)\n", | ||
| 438 | mmc_hostname(test->card->host), sectors, sectors >> 1, | ||
| 439 | (sectors == 1 ? ".5" : ""), (unsigned long)ts.tv_sec, | ||
| 440 | (unsigned long)ts.tv_nsec, rate / 1000, rate / 1024); | ||
| 441 | } | ||
| 442 | |||
| 443 | /* | ||
| 444 | * Print the average transfer rate. | ||
| 445 | */ | ||
| 446 | static void mmc_test_print_avg_rate(struct mmc_test_card *test, uint64_t bytes, | ||
| 447 | unsigned int count, struct timespec *ts1, | ||
| 448 | struct timespec *ts2) | ||
| 449 | { | ||
| 450 | unsigned int rate, sectors = bytes >> 9; | ||
| 451 | uint64_t tot = bytes * count; | ||
| 452 | struct timespec ts; | ||
| 453 | |||
| 454 | ts = timespec_sub(*ts2, *ts1); | ||
| 455 | |||
| 456 | rate = mmc_test_rate(tot, &ts); | ||
| 457 | |||
| 458 | printk(KERN_INFO "%s: Transfer of %u x %u sectors (%u x %u%s KiB) took " | ||
| 459 | "%lu.%09lu seconds (%u kB/s, %u KiB/s)\n", | ||
| 460 | mmc_hostname(test->card->host), count, sectors, count, | ||
| 461 | sectors >> 1, (sectors == 1 ? ".5" : ""), | ||
| 462 | (unsigned long)ts.tv_sec, (unsigned long)ts.tv_nsec, | ||
| 463 | rate / 1000, rate / 1024); | ||
| 464 | } | ||
| 465 | |||
| 466 | /* | ||
| 467 | * Return the card size in sectors. | ||
| 468 | */ | ||
| 469 | static unsigned int mmc_test_capacity(struct mmc_card *card) | ||
| 470 | { | ||
| 471 | if (!mmc_card_sd(card) && mmc_card_blockaddr(card)) | ||
| 472 | return card->ext_csd.sectors; | ||
| 473 | else | ||
| 474 | return card->csd.capacity << (card->csd.read_blkbits - 9); | ||
| 475 | } | ||
| 476 | |||
| 173 | /*******************************************************************/ | 477 | /*******************************************************************/ |
| 174 | /* Test preparation and cleanup */ | 478 | /* Test preparation and cleanup */ |
| 175 | /*******************************************************************/ | 479 | /*******************************************************************/ |
| @@ -893,8 +1197,419 @@ static int mmc_test_multi_read_high(struct mmc_test_card *test) | |||
| 893 | return 0; | 1197 | return 0; |
| 894 | } | 1198 | } |
| 895 | 1199 | ||
| 1200 | #else | ||
| 1201 | |||
| 1202 | static int mmc_test_no_highmem(struct mmc_test_card *test) | ||
| 1203 | { | ||
| 1204 | printk(KERN_INFO "%s: Highmem not configured - test skipped\n", | ||
| 1205 | mmc_hostname(test->card->host)); | ||
| 1206 | return 0; | ||
| 1207 | } | ||
| 1208 | |||
| 896 | #endif /* CONFIG_HIGHMEM */ | 1209 | #endif /* CONFIG_HIGHMEM */ |
| 897 | 1210 | ||
| 1211 | /* | ||
| 1212 | * Map sz bytes so that it can be transferred. | ||
| 1213 | */ | ||
| 1214 | static int mmc_test_area_map(struct mmc_test_card *test, unsigned long sz, | ||
| 1215 | int max_scatter) | ||
| 1216 | { | ||
| 1217 | struct mmc_test_area *t = &test->area; | ||
| 1218 | |||
| 1219 | t->blocks = sz >> 9; | ||
| 1220 | |||
| 1221 | if (max_scatter) { | ||
| 1222 | return mmc_test_map_sg_max_scatter(t->mem, sz, t->sg, | ||
| 1223 | t->max_segs, &t->sg_len); | ||
| 1224 | } else { | ||
| 1225 | return mmc_test_map_sg(t->mem, sz, t->sg, 1, t->max_segs, | ||
| 1226 | &t->sg_len); | ||
| 1227 | } | ||
| 1228 | } | ||
| 1229 | |||
| 1230 | /* | ||
| 1231 | * Transfer bytes mapped by mmc_test_area_map(). | ||
| 1232 | */ | ||
| 1233 | static int mmc_test_area_transfer(struct mmc_test_card *test, | ||
| 1234 | unsigned int dev_addr, int write) | ||
| 1235 | { | ||
| 1236 | struct mmc_test_area *t = &test->area; | ||
| 1237 | |||
| 1238 | return mmc_test_simple_transfer(test, t->sg, t->sg_len, dev_addr, | ||
| 1239 | t->blocks, 512, write); | ||
| 1240 | } | ||
| 1241 | |||
| 1242 | /* | ||
| 1243 | * Map and transfer bytes. | ||
| 1244 | */ | ||
| 1245 | static int mmc_test_area_io(struct mmc_test_card *test, unsigned long sz, | ||
| 1246 | unsigned int dev_addr, int write, int max_scatter, | ||
| 1247 | int timed) | ||
| 1248 | { | ||
| 1249 | struct timespec ts1, ts2; | ||
| 1250 | int ret; | ||
| 1251 | |||
| 1252 | ret = mmc_test_area_map(test, sz, max_scatter); | ||
| 1253 | if (ret) | ||
| 1254 | return ret; | ||
| 1255 | |||
| 1256 | if (timed) | ||
| 1257 | getnstimeofday(&ts1); | ||
| 1258 | |||
| 1259 | ret = mmc_test_area_transfer(test, dev_addr, write); | ||
| 1260 | if (ret) | ||
| 1261 | return ret; | ||
| 1262 | |||
| 1263 | if (timed) | ||
| 1264 | getnstimeofday(&ts2); | ||
| 1265 | |||
| 1266 | if (timed) | ||
| 1267 | mmc_test_print_rate(test, sz, &ts1, &ts2); | ||
| 1268 | |||
| 1269 | return 0; | ||
| 1270 | } | ||
| 1271 | |||
| 1272 | /* | ||
| 1273 | * Write the test area entirely. | ||
| 1274 | */ | ||
| 1275 | static int mmc_test_area_fill(struct mmc_test_card *test) | ||
| 1276 | { | ||
| 1277 | return mmc_test_area_io(test, test->area.max_sz, test->area.dev_addr, | ||
| 1278 | 1, 0, 0); | ||
| 1279 | } | ||
| 1280 | |||
| 1281 | /* | ||
| 1282 | * Erase the test area entirely. | ||
| 1283 | */ | ||
| 1284 | static int mmc_test_area_erase(struct mmc_test_card *test) | ||
| 1285 | { | ||
| 1286 | struct mmc_test_area *t = &test->area; | ||
| 1287 | |||
| 1288 | if (!mmc_can_erase(test->card)) | ||
| 1289 | return 0; | ||
| 1290 | |||
| 1291 | return mmc_erase(test->card, t->dev_addr, test->area.max_sz >> 9, | ||
| 1292 | MMC_ERASE_ARG); | ||
| 1293 | } | ||
| 1294 | |||
| 1295 | /* | ||
| 1296 | * Cleanup struct mmc_test_area. | ||
| 1297 | */ | ||
| 1298 | static int mmc_test_area_cleanup(struct mmc_test_card *test) | ||
| 1299 | { | ||
| 1300 | struct mmc_test_area *t = &test->area; | ||
| 1301 | |||
| 1302 | kfree(t->sg); | ||
| 1303 | mmc_test_free_mem(t->mem); | ||
| 1304 | |||
| 1305 | return 0; | ||
| 1306 | } | ||
| 1307 | |||
| 1308 | /* | ||
| 1309 | * Initialize an area for testing large transfers. The size of the area is the | ||
| 1310 | * preferred erase size which is a good size for optimal transfer speed. Note | ||
| 1311 | * that is typically 4MiB for modern cards. The test area is set to the middle | ||
| 1312 | * of the card because cards may have different charateristics at the front | ||
| 1313 | * (for FAT file system optimization). Optionally, the area is erased (if the | ||
| 1314 | * card supports it) which may improve write performance. Optionally, the area | ||
| 1315 | * is filled with data for subsequent read tests. | ||
| 1316 | */ | ||
| 1317 | static int mmc_test_area_init(struct mmc_test_card *test, int erase, int fill) | ||
| 1318 | { | ||
| 1319 | struct mmc_test_area *t = &test->area; | ||
| 1320 | unsigned long min_sz = 64 * 1024; | ||
| 1321 | int ret; | ||
| 1322 | |||
| 1323 | ret = mmc_test_set_blksize(test, 512); | ||
| 1324 | if (ret) | ||
| 1325 | return ret; | ||
| 1326 | |||
| 1327 | if (test->card->pref_erase > TEST_AREA_MAX_SIZE >> 9) | ||
| 1328 | t->max_sz = TEST_AREA_MAX_SIZE; | ||
| 1329 | else | ||
| 1330 | t->max_sz = (unsigned long)test->card->pref_erase << 9; | ||
| 1331 | /* | ||
| 1332 | * Try to allocate enough memory for the whole area. Less is OK | ||
| 1333 | * because the same memory can be mapped into the scatterlist more than | ||
| 1334 | * once. | ||
| 1335 | */ | ||
| 1336 | t->mem = mmc_test_alloc_mem(min_sz, t->max_sz); | ||
| 1337 | if (!t->mem) | ||
| 1338 | return -ENOMEM; | ||
| 1339 | |||
| 1340 | t->max_segs = DIV_ROUND_UP(t->max_sz, PAGE_SIZE); | ||
| 1341 | t->sg = kmalloc(sizeof(struct scatterlist) * t->max_segs, GFP_KERNEL); | ||
| 1342 | if (!t->sg) { | ||
| 1343 | ret = -ENOMEM; | ||
| 1344 | goto out_free; | ||
| 1345 | } | ||
| 1346 | |||
| 1347 | t->dev_addr = mmc_test_capacity(test->card) / 2; | ||
| 1348 | t->dev_addr -= t->dev_addr % (t->max_sz >> 9); | ||
| 1349 | |||
| 1350 | if (erase) { | ||
| 1351 | ret = mmc_test_area_erase(test); | ||
| 1352 | if (ret) | ||
| 1353 | goto out_free; | ||
| 1354 | } | ||
| 1355 | |||
| 1356 | if (fill) { | ||
| 1357 | ret = mmc_test_area_fill(test); | ||
| 1358 | if (ret) | ||
| 1359 | goto out_free; | ||
| 1360 | } | ||
| 1361 | |||
| 1362 | return 0; | ||
| 1363 | |||
| 1364 | out_free: | ||
| 1365 | mmc_test_area_cleanup(test); | ||
| 1366 | return ret; | ||
| 1367 | } | ||
| 1368 | |||
| 1369 | /* | ||
| 1370 | * Prepare for large transfers. Do not erase the test area. | ||
| 1371 | */ | ||
| 1372 | static int mmc_test_area_prepare(struct mmc_test_card *test) | ||
| 1373 | { | ||
| 1374 | return mmc_test_area_init(test, 0, 0); | ||
| 1375 | } | ||
| 1376 | |||
| 1377 | /* | ||
| 1378 | * Prepare for large transfers. Do erase the test area. | ||
| 1379 | */ | ||
| 1380 | static int mmc_test_area_prepare_erase(struct mmc_test_card *test) | ||
| 1381 | { | ||
| 1382 | return mmc_test_area_init(test, 1, 0); | ||
| 1383 | } | ||
| 1384 | |||
| 1385 | /* | ||
| 1386 | * Prepare for large transfers. Erase and fill the test area. | ||
| 1387 | */ | ||
| 1388 | static int mmc_test_area_prepare_fill(struct mmc_test_card *test) | ||
| 1389 | { | ||
| 1390 | return mmc_test_area_init(test, 1, 1); | ||
| 1391 | } | ||
| 1392 | |||
| 1393 | /* | ||
| 1394 | * Test best-case performance. Best-case performance is expected from | ||
| 1395 | * a single large transfer. | ||
| 1396 | * | ||
| 1397 | * An additional option (max_scatter) allows the measurement of the same | ||
| 1398 | * transfer but with no contiguous pages in the scatter list. This tests | ||
| 1399 | * the efficiency of DMA to handle scattered pages. | ||
| 1400 | */ | ||
| 1401 | static int mmc_test_best_performance(struct mmc_test_card *test, int write, | ||
| 1402 | int max_scatter) | ||
| 1403 | { | ||
| 1404 | return mmc_test_area_io(test, test->area.max_sz, test->area.dev_addr, | ||
| 1405 | write, max_scatter, 1); | ||
| 1406 | } | ||
| 1407 | |||
| 1408 | /* | ||
| 1409 | * Best-case read performance. | ||
| 1410 | */ | ||
| 1411 | static int mmc_test_best_read_performance(struct mmc_test_card *test) | ||
| 1412 | { | ||
| 1413 | return mmc_test_best_performance(test, 0, 0); | ||
| 1414 | } | ||
| 1415 | |||
| 1416 | /* | ||
| 1417 | * Best-case write performance. | ||
| 1418 | */ | ||
| 1419 | static int mmc_test_best_write_performance(struct mmc_test_card *test) | ||
| 1420 | { | ||
| 1421 | return mmc_test_best_performance(test, 1, 0); | ||
| 1422 | } | ||
| 1423 | |||
| 1424 | /* | ||
| 1425 | * Best-case read performance into scattered pages. | ||
| 1426 | */ | ||
| 1427 | static int mmc_test_best_read_perf_max_scatter(struct mmc_test_card *test) | ||
| 1428 | { | ||
| 1429 | return mmc_test_best_performance(test, 0, 1); | ||
| 1430 | } | ||
| 1431 | |||
| 1432 | /* | ||
| 1433 | * Best-case write performance from scattered pages. | ||
| 1434 | */ | ||
| 1435 | static int mmc_test_best_write_perf_max_scatter(struct mmc_test_card *test) | ||
| 1436 | { | ||
| 1437 | return mmc_test_best_performance(test, 1, 1); | ||
| 1438 | } | ||
| 1439 | |||
| 1440 | /* | ||
| 1441 | * Single read performance by transfer size. | ||
| 1442 | */ | ||
| 1443 | static int mmc_test_profile_read_perf(struct mmc_test_card *test) | ||
| 1444 | { | ||
| 1445 | unsigned long sz; | ||
| 1446 | unsigned int dev_addr; | ||
| 1447 | int ret; | ||
| 1448 | |||
| 1449 | for (sz = 512; sz < test->area.max_sz; sz <<= 1) { | ||
| 1450 | dev_addr = test->area.dev_addr + (sz >> 9); | ||
| 1451 | ret = mmc_test_area_io(test, sz, dev_addr, 0, 0, 1); | ||
| 1452 | if (ret) | ||
| 1453 | return ret; | ||
| 1454 | } | ||
| 1455 | dev_addr = test->area.dev_addr; | ||
| 1456 | return mmc_test_area_io(test, sz, dev_addr, 0, 0, 1); | ||
| 1457 | } | ||
| 1458 | |||
| 1459 | /* | ||
| 1460 | * Single write performance by transfer size. | ||
| 1461 | */ | ||
| 1462 | static int mmc_test_profile_write_perf(struct mmc_test_card *test) | ||
| 1463 | { | ||
| 1464 | unsigned long sz; | ||
| 1465 | unsigned int dev_addr; | ||
| 1466 | int ret; | ||
| 1467 | |||
| 1468 | ret = mmc_test_area_erase(test); | ||
| 1469 | if (ret) | ||
| 1470 | return ret; | ||
| 1471 | for (sz = 512; sz < test->area.max_sz; sz <<= 1) { | ||
| 1472 | dev_addr = test->area.dev_addr + (sz >> 9); | ||
| 1473 | ret = mmc_test_area_io(test, sz, dev_addr, 1, 0, 1); | ||
| 1474 | if (ret) | ||
| 1475 | return ret; | ||
| 1476 | } | ||
| 1477 | ret = mmc_test_area_erase(test); | ||
| 1478 | if (ret) | ||
| 1479 | return ret; | ||
| 1480 | dev_addr = test->area.dev_addr; | ||
| 1481 | return mmc_test_area_io(test, sz, dev_addr, 1, 0, 1); | ||
| 1482 | } | ||
| 1483 | |||
| 1484 | /* | ||
| 1485 | * Single trim performance by transfer size. | ||
| 1486 | */ | ||
| 1487 | static int mmc_test_profile_trim_perf(struct mmc_test_card *test) | ||
| 1488 | { | ||
| 1489 | unsigned long sz; | ||
| 1490 | unsigned int dev_addr; | ||
| 1491 | struct timespec ts1, ts2; | ||
| 1492 | int ret; | ||
| 1493 | |||
| 1494 | if (!mmc_can_trim(test->card)) | ||
| 1495 | return RESULT_UNSUP_CARD; | ||
| 1496 | |||
| 1497 | if (!mmc_can_erase(test->card)) | ||
| 1498 | return RESULT_UNSUP_HOST; | ||
| 1499 | |||
| 1500 | for (sz = 512; sz < test->area.max_sz; sz <<= 1) { | ||
| 1501 | dev_addr = test->area.dev_addr + (sz >> 9); | ||
| 1502 | getnstimeofday(&ts1); | ||
| 1503 | ret = mmc_erase(test->card, dev_addr, sz >> 9, MMC_TRIM_ARG); | ||
| 1504 | if (ret) | ||
| 1505 | return ret; | ||
| 1506 | getnstimeofday(&ts2); | ||
| 1507 | mmc_test_print_rate(test, sz, &ts1, &ts2); | ||
| 1508 | } | ||
| 1509 | dev_addr = test->area.dev_addr; | ||
| 1510 | getnstimeofday(&ts1); | ||
| 1511 | ret = mmc_erase(test->card, dev_addr, sz >> 9, MMC_TRIM_ARG); | ||
| 1512 | if (ret) | ||
| 1513 | return ret; | ||
| 1514 | getnstimeofday(&ts2); | ||
| 1515 | mmc_test_print_rate(test, sz, &ts1, &ts2); | ||
| 1516 | return 0; | ||
| 1517 | } | ||
| 1518 | |||
| 1519 | /* | ||
| 1520 | * Consecutive read performance by transfer size. | ||
| 1521 | */ | ||
| 1522 | static int mmc_test_profile_seq_read_perf(struct mmc_test_card *test) | ||
| 1523 | { | ||
| 1524 | unsigned long sz; | ||
| 1525 | unsigned int dev_addr, i, cnt; | ||
| 1526 | struct timespec ts1, ts2; | ||
| 1527 | int ret; | ||
| 1528 | |||
| 1529 | for (sz = 512; sz <= test->area.max_sz; sz <<= 1) { | ||
| 1530 | cnt = test->area.max_sz / sz; | ||
| 1531 | dev_addr = test->area.dev_addr; | ||
| 1532 | getnstimeofday(&ts1); | ||
| 1533 | for (i = 0; i < cnt; i++) { | ||
| 1534 | ret = mmc_test_area_io(test, sz, dev_addr, 0, 0, 0); | ||
| 1535 | if (ret) | ||
| 1536 | return ret; | ||
| 1537 | dev_addr += (sz >> 9); | ||
| 1538 | } | ||
| 1539 | getnstimeofday(&ts2); | ||
| 1540 | mmc_test_print_avg_rate(test, sz, cnt, &ts1, &ts2); | ||
| 1541 | } | ||
| 1542 | return 0; | ||
| 1543 | } | ||
| 1544 | |||
| 1545 | /* | ||
| 1546 | * Consecutive write performance by transfer size. | ||
| 1547 | */ | ||
| 1548 | static int mmc_test_profile_seq_write_perf(struct mmc_test_card *test) | ||
| 1549 | { | ||
| 1550 | unsigned long sz; | ||
| 1551 | unsigned int dev_addr, i, cnt; | ||
| 1552 | struct timespec ts1, ts2; | ||
| 1553 | int ret; | ||
| 1554 | |||
| 1555 | for (sz = 512; sz <= test->area.max_sz; sz <<= 1) { | ||
| 1556 | ret = mmc_test_area_erase(test); | ||
| 1557 | if (ret) | ||
| 1558 | return ret; | ||
| 1559 | cnt = test->area.max_sz / sz; | ||
| 1560 | dev_addr = test->area.dev_addr; | ||
| 1561 | getnstimeofday(&ts1); | ||
| 1562 | for (i = 0; i < cnt; i++) { | ||
| 1563 | ret = mmc_test_area_io(test, sz, dev_addr, 1, 0, 0); | ||
| 1564 | if (ret) | ||
| 1565 | return ret; | ||
| 1566 | dev_addr += (sz >> 9); | ||
| 1567 | } | ||
| 1568 | getnstimeofday(&ts2); | ||
| 1569 | mmc_test_print_avg_rate(test, sz, cnt, &ts1, &ts2); | ||
| 1570 | } | ||
| 1571 | return 0; | ||
| 1572 | } | ||
| 1573 | |||
| 1574 | /* | ||
| 1575 | * Consecutive trim performance by transfer size. | ||
| 1576 | */ | ||
| 1577 | static int mmc_test_profile_seq_trim_perf(struct mmc_test_card *test) | ||
| 1578 | { | ||
| 1579 | unsigned long sz; | ||
| 1580 | unsigned int dev_addr, i, cnt; | ||
| 1581 | struct timespec ts1, ts2; | ||
| 1582 | int ret; | ||
| 1583 | |||
| 1584 | if (!mmc_can_trim(test->card)) | ||
| 1585 | return RESULT_UNSUP_CARD; | ||
| 1586 | |||
| 1587 | if (!mmc_can_erase(test->card)) | ||
| 1588 | return RESULT_UNSUP_HOST; | ||
| 1589 | |||
| 1590 | for (sz = 512; sz <= test->area.max_sz; sz <<= 1) { | ||
| 1591 | ret = mmc_test_area_erase(test); | ||
| 1592 | if (ret) | ||
| 1593 | return ret; | ||
| 1594 | ret = mmc_test_area_fill(test); | ||
| 1595 | if (ret) | ||
| 1596 | return ret; | ||
| 1597 | cnt = test->area.max_sz / sz; | ||
| 1598 | dev_addr = test->area.dev_addr; | ||
| 1599 | getnstimeofday(&ts1); | ||
| 1600 | for (i = 0; i < cnt; i++) { | ||
| 1601 | ret = mmc_erase(test->card, dev_addr, sz >> 9, | ||
| 1602 | MMC_TRIM_ARG); | ||
| 1603 | if (ret) | ||
| 1604 | return ret; | ||
| 1605 | dev_addr += (sz >> 9); | ||
| 1606 | } | ||
| 1607 | getnstimeofday(&ts2); | ||
| 1608 | mmc_test_print_avg_rate(test, sz, cnt, &ts1, &ts2); | ||
| 1609 | } | ||
| 1610 | return 0; | ||
| 1611 | } | ||
| 1612 | |||
| 898 | static const struct mmc_test_case mmc_test_cases[] = { | 1613 | static const struct mmc_test_case mmc_test_cases[] = { |
| 899 | { | 1614 | { |
| 900 | .name = "Basic write (no data verification)", | 1615 | .name = "Basic write (no data verification)", |
| @@ -1040,8 +1755,100 @@ static const struct mmc_test_case mmc_test_cases[] = { | |||
| 1040 | .cleanup = mmc_test_cleanup, | 1755 | .cleanup = mmc_test_cleanup, |
| 1041 | }, | 1756 | }, |
| 1042 | 1757 | ||
| 1758 | #else | ||
| 1759 | |||
| 1760 | { | ||
| 1761 | .name = "Highmem write", | ||
| 1762 | .run = mmc_test_no_highmem, | ||
| 1763 | }, | ||
| 1764 | |||
| 1765 | { | ||
| 1766 | .name = "Highmem read", | ||
| 1767 | .run = mmc_test_no_highmem, | ||
| 1768 | }, | ||
| 1769 | |||
| 1770 | { | ||
| 1771 | .name = "Multi-block highmem write", | ||
| 1772 | .run = mmc_test_no_highmem, | ||
| 1773 | }, | ||
| 1774 | |||
| 1775 | { | ||
| 1776 | .name = "Multi-block highmem read", | ||
| 1777 | .run = mmc_test_no_highmem, | ||
| 1778 | }, | ||
| 1779 | |||
| 1043 | #endif /* CONFIG_HIGHMEM */ | 1780 | #endif /* CONFIG_HIGHMEM */ |
| 1044 | 1781 | ||
| 1782 | { | ||
| 1783 | .name = "Best-case read performance", | ||
| 1784 | .prepare = mmc_test_area_prepare_fill, | ||
| 1785 | .run = mmc_test_best_read_performance, | ||
| 1786 | .cleanup = mmc_test_area_cleanup, | ||
| 1787 | }, | ||
| 1788 | |||
| 1789 | { | ||
| 1790 | .name = "Best-case write performance", | ||
| 1791 | .prepare = mmc_test_area_prepare_erase, | ||
| 1792 | .run = mmc_test_best_write_performance, | ||
| 1793 | .cleanup = mmc_test_area_cleanup, | ||
| 1794 | }, | ||
| 1795 | |||
| 1796 | { | ||
| 1797 | .name = "Best-case read performance into scattered pages", | ||
| 1798 | .prepare = mmc_test_area_prepare_fill, | ||
| 1799 | .run = mmc_test_best_read_perf_max_scatter, | ||
| 1800 | .cleanup = mmc_test_area_cleanup, | ||
| 1801 | }, | ||
| 1802 | |||
| 1803 | { | ||
| 1804 | .name = "Best-case write performance from scattered pages", | ||
| 1805 | .prepare = mmc_test_area_prepare_erase, | ||
| 1806 | .run = mmc_test_best_write_perf_max_scatter, | ||
| 1807 | .cleanup = mmc_test_area_cleanup, | ||
| 1808 | }, | ||
| 1809 | |||
| 1810 | { | ||
| 1811 | .name = "Single read performance by transfer size", | ||
| 1812 | .prepare = mmc_test_area_prepare_fill, | ||
| 1813 | .run = mmc_test_profile_read_perf, | ||
| 1814 | .cleanup = mmc_test_area_cleanup, | ||
| 1815 | }, | ||
| 1816 | |||
| 1817 | { | ||
| 1818 | .name = "Single write performance by transfer size", | ||
| 1819 | .prepare = mmc_test_area_prepare, | ||
| 1820 | .run = mmc_test_profile_write_perf, | ||
| 1821 | .cleanup = mmc_test_area_cleanup, | ||
| 1822 | }, | ||
| 1823 | |||
| 1824 | { | ||
| 1825 | .name = "Single trim performance by transfer size", | ||
| 1826 | .prepare = mmc_test_area_prepare_fill, | ||
| 1827 | .run = mmc_test_profile_trim_perf, | ||
| 1828 | .cleanup = mmc_test_area_cleanup, | ||
| 1829 | }, | ||
| 1830 | |||
| 1831 | { | ||
| 1832 | .name = "Consecutive read performance by transfer size", | ||
| 1833 | .prepare = mmc_test_area_prepare_fill, | ||
| 1834 | .run = mmc_test_profile_seq_read_perf, | ||
| 1835 | .cleanup = mmc_test_area_cleanup, | ||
| 1836 | }, | ||
| 1837 | |||
| 1838 | { | ||
| 1839 | .name = "Consecutive write performance by transfer size", | ||
| 1840 | .prepare = mmc_test_area_prepare, | ||
| 1841 | .run = mmc_test_profile_seq_write_perf, | ||
| 1842 | .cleanup = mmc_test_area_cleanup, | ||
| 1843 | }, | ||
| 1844 | |||
| 1845 | { | ||
| 1846 | .name = "Consecutive trim performance by transfer size", | ||
| 1847 | .prepare = mmc_test_area_prepare, | ||
| 1848 | .run = mmc_test_profile_seq_trim_perf, | ||
| 1849 | .cleanup = mmc_test_area_cleanup, | ||
| 1850 | }, | ||
| 1851 | |||
| 1045 | }; | 1852 | }; |
| 1046 | 1853 | ||
| 1047 | static DEFINE_MUTEX(mmc_test_lock); | 1854 | static DEFINE_MUTEX(mmc_test_lock); |
