aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/mtd/ubi/wl.c
diff options
context:
space:
mode:
authorArtem Bityutskiy <Artem.Bityutskiy@nokia.com>2007-12-16 05:49:01 -0500
committerArtem Bityutskiy <Artem.Bityutskiy@nokia.com>2007-12-26 12:15:14 -0500
commit06b68ba15671f32a3aa3bbddf04b0d2dd7fbf902 (patch)
tree67024857173069023835148077775e8c9e381d8a /drivers/mtd/ubi/wl.c
parent3a8d4642861fb69b62401949e490c0bcb19ceb40 (diff)
UBI: create ubi_wl_entry slab on initialization
Similarly to ltree_entry_slab, it makes more sense to create and destroy ubi_wl_entry slab on module initialization/exit. Signed-off-by: Artem Bityutskiy <Artem.Bityutskiy@nokia.com>
Diffstat (limited to 'drivers/mtd/ubi/wl.c')
-rw-r--r--drivers/mtd/ubi/wl.c58
1 files changed, 14 insertions, 44 deletions
diff --git a/drivers/mtd/ubi/wl.c b/drivers/mtd/ubi/wl.c
index 6330c8cc72b5..a405d40faa23 100644
--- a/drivers/mtd/ubi/wl.c
+++ b/drivers/mtd/ubi/wl.c
@@ -117,21 +117,6 @@
117#define WL_MAX_FAILURES 32 117#define WL_MAX_FAILURES 32
118 118
119/** 119/**
120 * struct ubi_wl_entry - wear-leveling entry.
121 * @rb: link in the corresponding RB-tree
122 * @ec: erase counter
123 * @pnum: physical eraseblock number
124 *
125 * Each physical eraseblock has a corresponding &struct wl_entry object which
126 * may be kept in different RB-trees.
127 */
128struct ubi_wl_entry {
129 struct rb_node rb;
130 int ec;
131 int pnum;
132};
133
134/**
135 * struct ubi_wl_prot_entry - PEB protection entry. 120 * struct ubi_wl_prot_entry - PEB protection entry.
136 * @rb_pnum: link in the @wl->prot.pnum RB-tree 121 * @rb_pnum: link in the @wl->prot.pnum RB-tree
137 * @rb_aec: link in the @wl->prot.aec RB-tree 122 * @rb_aec: link in the @wl->prot.aec RB-tree
@@ -216,9 +201,6 @@ static int paranoid_check_in_wl_tree(struct ubi_wl_entry *e,
216#define paranoid_check_in_wl_tree(e, root) 201#define paranoid_check_in_wl_tree(e, root)
217#endif 202#endif
218 203
219/* Slab cache for wear-leveling entries */
220static struct kmem_cache *wl_entries_slab;
221
222/** 204/**
223 * wl_tree_add - add a wear-leveling entry to a WL RB-tree. 205 * wl_tree_add - add a wear-leveling entry to a WL RB-tree.
224 * @e: the wear-leveling entry to add 206 * @e: the wear-leveling entry to add
@@ -878,14 +860,14 @@ static int wear_leveling_worker(struct ubi_device *ubi, struct ubi_work *wrk,
878 dbg_wl("PEB %d was put meanwhile, erase", e2->pnum); 860 dbg_wl("PEB %d was put meanwhile, erase", e2->pnum);
879 err = schedule_erase(ubi, e2, 0); 861 err = schedule_erase(ubi, e2, 0);
880 if (err) { 862 if (err) {
881 kmem_cache_free(wl_entries_slab, e2); 863 kmem_cache_free(ubi_wl_entry_slab, e2);
882 ubi_ro_mode(ubi); 864 ubi_ro_mode(ubi);
883 } 865 }
884 } 866 }
885 867
886 err = schedule_erase(ubi, e1, 0); 868 err = schedule_erase(ubi, e1, 0);
887 if (err) { 869 if (err) {
888 kmem_cache_free(wl_entries_slab, e1); 870 kmem_cache_free(ubi_wl_entry_slab, e1);
889 ubi_ro_mode(ubi); 871 ubi_ro_mode(ubi);
890 } 872 }
891 873
@@ -920,14 +902,14 @@ error:
920 dbg_wl("PEB %d was put meanwhile, erase", e1->pnum); 902 dbg_wl("PEB %d was put meanwhile, erase", e1->pnum);
921 err = schedule_erase(ubi, e1, 0); 903 err = schedule_erase(ubi, e1, 0);
922 if (err) { 904 if (err) {
923 kmem_cache_free(wl_entries_slab, e1); 905 kmem_cache_free(ubi_wl_entry_slab, e1);
924 ubi_ro_mode(ubi); 906 ubi_ro_mode(ubi);
925 } 907 }
926 } 908 }
927 909
928 err = schedule_erase(ubi, e2, 0); 910 err = schedule_erase(ubi, e2, 0);
929 if (err) { 911 if (err) {
930 kmem_cache_free(wl_entries_slab, e2); 912 kmem_cache_free(ubi_wl_entry_slab, e2);
931 ubi_ro_mode(ubi); 913 ubi_ro_mode(ubi);
932 } 914 }
933 915
@@ -1020,7 +1002,7 @@ static int erase_worker(struct ubi_device *ubi, struct ubi_work *wl_wrk,
1020 if (cancel) { 1002 if (cancel) {
1021 dbg_wl("cancel erasure of PEB %d EC %d", pnum, e->ec); 1003 dbg_wl("cancel erasure of PEB %d EC %d", pnum, e->ec);
1022 kfree(wl_wrk); 1004 kfree(wl_wrk);
1023 kmem_cache_free(wl_entries_slab, e); 1005 kmem_cache_free(ubi_wl_entry_slab, e);
1024 return 0; 1006 return 0;
1025 } 1007 }
1026 1008
@@ -1049,7 +1031,7 @@ static int erase_worker(struct ubi_device *ubi, struct ubi_work *wl_wrk,
1049 1031
1050 ubi_err("failed to erase PEB %d, error %d", pnum, err); 1032 ubi_err("failed to erase PEB %d, error %d", pnum, err);
1051 kfree(wl_wrk); 1033 kfree(wl_wrk);
1052 kmem_cache_free(wl_entries_slab, e); 1034 kmem_cache_free(ubi_wl_entry_slab, e);
1053 1035
1054 if (err == -EINTR || err == -ENOMEM || err == -EAGAIN || 1036 if (err == -EINTR || err == -ENOMEM || err == -EAGAIN ||
1055 err == -EBUSY) { 1037 err == -EBUSY) {
@@ -1294,7 +1276,7 @@ static void tree_destroy(struct rb_root *root)
1294 rb->rb_right = NULL; 1276 rb->rb_right = NULL;
1295 } 1277 }
1296 1278
1297 kmem_cache_free(wl_entries_slab, e); 1279 kmem_cache_free(ubi_wl_entry_slab, e);
1298 } 1280 }
1299 } 1281 }
1300} 1282}
@@ -1407,14 +1389,6 @@ int ubi_wl_init_scan(struct ubi_device *ubi, struct ubi_scan_info *si)
1407 return err; 1389 return err;
1408 } 1390 }
1409 1391
1410 if (ubi_devices_cnt == 0) {
1411 wl_entries_slab = kmem_cache_create("ubi_wl_entry_slab",
1412 sizeof(struct ubi_wl_entry),
1413 0, 0, NULL);
1414 if (!wl_entries_slab)
1415 return -ENOMEM;
1416 }
1417
1418 err = -ENOMEM; 1392 err = -ENOMEM;
1419 ubi->lookuptbl = kzalloc(ubi->peb_count * sizeof(void *), GFP_KERNEL); 1393 ubi->lookuptbl = kzalloc(ubi->peb_count * sizeof(void *), GFP_KERNEL);
1420 if (!ubi->lookuptbl) 1394 if (!ubi->lookuptbl)
@@ -1423,7 +1397,7 @@ int ubi_wl_init_scan(struct ubi_device *ubi, struct ubi_scan_info *si)
1423 list_for_each_entry_safe(seb, tmp, &si->erase, u.list) { 1397 list_for_each_entry_safe(seb, tmp, &si->erase, u.list) {
1424 cond_resched(); 1398 cond_resched();
1425 1399
1426 e = kmem_cache_alloc(wl_entries_slab, GFP_KERNEL); 1400 e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL);
1427 if (!e) 1401 if (!e)
1428 goto out_free; 1402 goto out_free;
1429 1403
@@ -1431,7 +1405,7 @@ int ubi_wl_init_scan(struct ubi_device *ubi, struct ubi_scan_info *si)
1431 e->ec = seb->ec; 1405 e->ec = seb->ec;
1432 ubi->lookuptbl[e->pnum] = e; 1406 ubi->lookuptbl[e->pnum] = e;
1433 if (schedule_erase(ubi, e, 0)) { 1407 if (schedule_erase(ubi, e, 0)) {
1434 kmem_cache_free(wl_entries_slab, e); 1408 kmem_cache_free(ubi_wl_entry_slab, e);
1435 goto out_free; 1409 goto out_free;
1436 } 1410 }
1437 } 1411 }
@@ -1439,7 +1413,7 @@ int ubi_wl_init_scan(struct ubi_device *ubi, struct ubi_scan_info *si)
1439 list_for_each_entry(seb, &si->free, u.list) { 1413 list_for_each_entry(seb, &si->free, u.list) {
1440 cond_resched(); 1414 cond_resched();
1441 1415
1442 e = kmem_cache_alloc(wl_entries_slab, GFP_KERNEL); 1416 e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL);
1443 if (!e) 1417 if (!e)
1444 goto out_free; 1418 goto out_free;
1445 1419
@@ -1453,7 +1427,7 @@ int ubi_wl_init_scan(struct ubi_device *ubi, struct ubi_scan_info *si)
1453 list_for_each_entry(seb, &si->corr, u.list) { 1427 list_for_each_entry(seb, &si->corr, u.list) {
1454 cond_resched(); 1428 cond_resched();
1455 1429
1456 e = kmem_cache_alloc(wl_entries_slab, GFP_KERNEL); 1430 e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL);
1457 if (!e) 1431 if (!e)
1458 goto out_free; 1432 goto out_free;
1459 1433
@@ -1461,7 +1435,7 @@ int ubi_wl_init_scan(struct ubi_device *ubi, struct ubi_scan_info *si)
1461 e->ec = seb->ec; 1435 e->ec = seb->ec;
1462 ubi->lookuptbl[e->pnum] = e; 1436 ubi->lookuptbl[e->pnum] = e;
1463 if (schedule_erase(ubi, e, 0)) { 1437 if (schedule_erase(ubi, e, 0)) {
1464 kmem_cache_free(wl_entries_slab, e); 1438 kmem_cache_free(ubi_wl_entry_slab, e);
1465 goto out_free; 1439 goto out_free;
1466 } 1440 }
1467 } 1441 }
@@ -1470,7 +1444,7 @@ int ubi_wl_init_scan(struct ubi_device *ubi, struct ubi_scan_info *si)
1470 ubi_rb_for_each_entry(rb2, seb, &sv->root, u.rb) { 1444 ubi_rb_for_each_entry(rb2, seb, &sv->root, u.rb) {
1471 cond_resched(); 1445 cond_resched();
1472 1446
1473 e = kmem_cache_alloc(wl_entries_slab, GFP_KERNEL); 1447 e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL);
1474 if (!e) 1448 if (!e)
1475 goto out_free; 1449 goto out_free;
1476 1450
@@ -1510,8 +1484,6 @@ out_free:
1510 tree_destroy(&ubi->free); 1484 tree_destroy(&ubi->free);
1511 tree_destroy(&ubi->scrub); 1485 tree_destroy(&ubi->scrub);
1512 kfree(ubi->lookuptbl); 1486 kfree(ubi->lookuptbl);
1513 if (ubi_devices_cnt == 0)
1514 kmem_cache_destroy(wl_entries_slab);
1515 return err; 1487 return err;
1516} 1488}
1517 1489
@@ -1541,7 +1513,7 @@ static void protection_trees_destroy(struct ubi_device *ubi)
1541 rb->rb_right = NULL; 1513 rb->rb_right = NULL;
1542 } 1514 }
1543 1515
1544 kmem_cache_free(wl_entries_slab, pe->e); 1516 kmem_cache_free(ubi_wl_entry_slab, pe->e);
1545 kfree(pe); 1517 kfree(pe);
1546 } 1518 }
1547 } 1519 }
@@ -1565,8 +1537,6 @@ void ubi_wl_close(struct ubi_device *ubi)
1565 tree_destroy(&ubi->free); 1537 tree_destroy(&ubi->free);
1566 tree_destroy(&ubi->scrub); 1538 tree_destroy(&ubi->scrub);
1567 kfree(ubi->lookuptbl); 1539 kfree(ubi->lookuptbl);
1568 if (ubi_devices_cnt == 1)
1569 kmem_cache_destroy(wl_entries_slab);
1570} 1540}
1571 1541
1572#ifdef CONFIG_MTD_UBI_DEBUG_PARANOID 1542#ifdef CONFIG_MTD_UBI_DEBUG_PARANOID