diff options
-rw-r--r-- | drivers/mtd/ubi/eba.c | 145 | ||||
-rw-r--r-- | drivers/mtd/ubi/ubi.h | 10 | ||||
-rw-r--r-- | drivers/mtd/ubi/wl.c | 219 |
3 files changed, 239 insertions, 135 deletions
diff --git a/drivers/mtd/ubi/eba.c b/drivers/mtd/ubi/eba.c index 84f7dc9fd3ac..c94f475758de 100644 --- a/drivers/mtd/ubi/eba.c +++ b/drivers/mtd/ubi/eba.c | |||
@@ -260,6 +260,44 @@ static int leb_write_lock(struct ubi_device *ubi, int vol_id, int lnum) | |||
260 | } | 260 | } |
261 | 261 | ||
262 | /** | 262 | /** |
263 | * leb_write_lock - lock logical eraseblock for writing. | ||
264 | * @ubi: UBI device description object | ||
265 | * @vol_id: volume ID | ||
266 | * @lnum: logical eraseblock number | ||
267 | * | ||
268 | * This function locks a logical eraseblock for writing if there is no | ||
269 | * contention and does nothing if there is contention. Returns %0 in case of | ||
270 | * success, %1 in case of contention, and and a negative error code in case of | ||
271 | * failure. | ||
272 | */ | ||
273 | static int leb_write_trylock(struct ubi_device *ubi, int vol_id, int lnum) | ||
274 | { | ||
275 | int free; | ||
276 | struct ubi_ltree_entry *le; | ||
277 | |||
278 | le = ltree_add_entry(ubi, vol_id, lnum); | ||
279 | if (IS_ERR(le)) | ||
280 | return PTR_ERR(le); | ||
281 | if (down_write_trylock(&le->mutex)) | ||
282 | return 0; | ||
283 | |||
284 | /* Contention, cancel */ | ||
285 | spin_lock(&ubi->ltree_lock); | ||
286 | le->users -= 1; | ||
287 | ubi_assert(le->users >= 0); | ||
288 | if (le->users == 0) { | ||
289 | rb_erase(&le->rb, &ubi->ltree); | ||
290 | free = 1; | ||
291 | } else | ||
292 | free = 0; | ||
293 | spin_unlock(&ubi->ltree_lock); | ||
294 | if (free) | ||
295 | kmem_cache_free(ubi_ltree_slab, le); | ||
296 | |||
297 | return 1; | ||
298 | } | ||
299 | |||
300 | /** | ||
263 | * leb_write_unlock - unlock logical eraseblock. | 301 | * leb_write_unlock - unlock logical eraseblock. |
264 | * @ubi: UBI device description object | 302 | * @ubi: UBI device description object |
265 | * @vol_id: volume ID | 303 | * @vol_id: volume ID |
@@ -923,14 +961,16 @@ write_error: | |||
923 | * | 961 | * |
924 | * This function copies logical eraseblock from physical eraseblock @from to | 962 | * This function copies logical eraseblock from physical eraseblock @from to |
925 | * physical eraseblock @to. The @vid_hdr buffer may be changed by this | 963 | * physical eraseblock @to. The @vid_hdr buffer may be changed by this |
926 | * function. Returns zero in case of success, %UBI_IO_BITFLIPS if the operation | 964 | * function. Returns: |
927 | * was canceled because bit-flips were detected at the target PEB, and a | 965 | * o %0 in case of success; |
928 | * negative error code in case of failure. | 966 | * o %1 if the operation was canceled and should be tried later (e.g., |
967 | * because a bit-flip was detected at the target PEB); | ||
968 | * o %2 if the volume is being deleted and this LEB should not be moved. | ||
929 | */ | 969 | */ |
930 | int ubi_eba_copy_leb(struct ubi_device *ubi, int from, int to, | 970 | int ubi_eba_copy_leb(struct ubi_device *ubi, int from, int to, |
931 | struct ubi_vid_hdr *vid_hdr) | 971 | struct ubi_vid_hdr *vid_hdr) |
932 | { | 972 | { |
933 | int err, vol_id, lnum, data_size, aldata_size, pnum, idx; | 973 | int err, vol_id, lnum, data_size, aldata_size, idx; |
934 | struct ubi_volume *vol; | 974 | struct ubi_volume *vol; |
935 | uint32_t crc; | 975 | uint32_t crc; |
936 | 976 | ||
@@ -946,57 +986,67 @@ int ubi_eba_copy_leb(struct ubi_device *ubi, int from, int to, | |||
946 | data_size = aldata_size = | 986 | data_size = aldata_size = |
947 | ubi->leb_size - be32_to_cpu(vid_hdr->data_pad); | 987 | ubi->leb_size - be32_to_cpu(vid_hdr->data_pad); |
948 | 988 | ||
949 | /* | ||
950 | * We do not want anybody to write to this logical eraseblock while we | ||
951 | * are moving it, so we lock it. | ||
952 | */ | ||
953 | err = leb_write_lock(ubi, vol_id, lnum); | ||
954 | if (err) | ||
955 | return err; | ||
956 | |||
957 | mutex_lock(&ubi->buf_mutex); | ||
958 | |||
959 | /* | ||
960 | * But the logical eraseblock might have been put by this time. | ||
961 | * Cancel if it is true. | ||
962 | */ | ||
963 | idx = vol_id2idx(ubi, vol_id); | 989 | idx = vol_id2idx(ubi, vol_id); |
964 | 990 | spin_lock(&ubi->volumes_lock); | |
965 | /* | 991 | /* |
966 | * We may race with volume deletion/re-size, so we have to hold | 992 | * Note, we may race with volume deletion, which means that the volume |
967 | * @ubi->volumes_lock. | 993 | * this logical eraseblock belongs to might be being deleted. Since the |
968 | * | 994 | * volume deletion unmaps all the volume's logical eraseblocks, it will |
969 | * Note, it is not a problem if we race with volume deletion or re-size | 995 | * be locked in 'ubi_wl_put_peb()' and wait for the WL worker to finish. |
970 | * here. If the volume is deleted or re-sized while we are moving an | ||
971 | * eraseblock which belongs to this volume, we'll end up with finding | ||
972 | * out that this LEB was unmapped at the end (see WL), and drop this | ||
973 | * PEB. | ||
974 | */ | 996 | */ |
975 | spin_lock(&ubi->volumes_lock); | ||
976 | vol = ubi->volumes[idx]; | 997 | vol = ubi->volumes[idx]; |
977 | if (!vol) { | 998 | if (!vol) { |
978 | dbg_eba("volume %d was removed meanwhile", vol_id); | 999 | /* No need to do further work, cancel */ |
1000 | dbg_eba("volume %d is being removed, cancel", vol_id); | ||
979 | spin_unlock(&ubi->volumes_lock); | 1001 | spin_unlock(&ubi->volumes_lock); |
980 | goto out_unlock; | 1002 | return 2; |
981 | } | 1003 | } |
1004 | spin_unlock(&ubi->volumes_lock); | ||
982 | 1005 | ||
983 | pnum = vol->eba_tbl[lnum]; | 1006 | /* |
984 | if (pnum != from) { | 1007 | * We do not want anybody to write to this logical eraseblock while we |
985 | dbg_eba("LEB %d:%d is no longer mapped to PEB %d, mapped to " | 1008 | * are moving it, so lock it. |
986 | "PEB %d, cancel", vol_id, lnum, from, pnum); | 1009 | * |
987 | spin_unlock(&ubi->volumes_lock); | 1010 | * Note, we are using non-waiting locking here, because we cannot sleep |
988 | goto out_unlock; | 1011 | * on the LEB, since it may cause deadlocks. Indeed, imagine a task is |
1012 | * unmapping the LEB which is mapped to the PEB we are going to move | ||
1013 | * (@from). This task locks the LEB and goes sleep in the | ||
1014 | * 'ubi_wl_put_peb()' function on the @ubi->move_mutex. In turn, we are | ||
1015 | * holding @ubi->move_mutex and go sleep on the LEB lock. So, if the | ||
1016 | * LEB is already locked, we just do not move it and return %1. | ||
1017 | */ | ||
1018 | err = leb_write_trylock(ubi, vol_id, lnum); | ||
1019 | if (err) { | ||
1020 | dbg_eba("contention on LEB %d:%d, cancel", vol_id, lnum); | ||
1021 | return err; | ||
989 | } | 1022 | } |
990 | spin_unlock(&ubi->volumes_lock); | ||
991 | 1023 | ||
992 | /* OK, now the LEB is locked and we can safely start moving it */ | 1024 | /* |
1025 | * The LEB might have been put meanwhile, and the task which put it is | ||
1026 | * probably waiting on @ubi->move_mutex. No need to continue the work, | ||
1027 | * cancel it. | ||
1028 | */ | ||
1029 | if (vol->eba_tbl[lnum] != from) { | ||
1030 | dbg_eba("LEB %d:%d is no longer mapped to PEB %d, mapped to " | ||
1031 | "PEB %d, cancel", vol_id, lnum, from, | ||
1032 | vol->eba_tbl[lnum]); | ||
1033 | err = 1; | ||
1034 | goto out_unlock_leb; | ||
1035 | } | ||
993 | 1036 | ||
1037 | /* | ||
1038 | * OK, now the LEB is locked and we can safely start moving iy. Since | ||
1039 | * this function utilizes thie @ubi->peb1_buf buffer which is shared | ||
1040 | * with some other functions, so lock the buffer by taking the | ||
1041 | * @ubi->buf_mutex. | ||
1042 | */ | ||
1043 | mutex_lock(&ubi->buf_mutex); | ||
994 | dbg_eba("read %d bytes of data", aldata_size); | 1044 | dbg_eba("read %d bytes of data", aldata_size); |
995 | err = ubi_io_read_data(ubi, ubi->peb_buf1, from, 0, aldata_size); | 1045 | err = ubi_io_read_data(ubi, ubi->peb_buf1, from, 0, aldata_size); |
996 | if (err && err != UBI_IO_BITFLIPS) { | 1046 | if (err && err != UBI_IO_BITFLIPS) { |
997 | ubi_warn("error %d while reading data from PEB %d", | 1047 | ubi_warn("error %d while reading data from PEB %d", |
998 | err, from); | 1048 | err, from); |
999 | goto out_unlock; | 1049 | goto out_unlock_buf; |
1000 | } | 1050 | } |
1001 | 1051 | ||
1002 | /* | 1052 | /* |
@@ -1032,7 +1082,7 @@ int ubi_eba_copy_leb(struct ubi_device *ubi, int from, int to, | |||
1032 | 1082 | ||
1033 | err = ubi_io_write_vid_hdr(ubi, to, vid_hdr); | 1083 | err = ubi_io_write_vid_hdr(ubi, to, vid_hdr); |
1034 | if (err) | 1084 | if (err) |
1035 | goto out_unlock; | 1085 | goto out_unlock_buf; |
1036 | 1086 | ||
1037 | cond_resched(); | 1087 | cond_resched(); |
1038 | 1088 | ||
@@ -1041,13 +1091,15 @@ int ubi_eba_copy_leb(struct ubi_device *ubi, int from, int to, | |||
1041 | if (err) { | 1091 | if (err) { |
1042 | if (err != UBI_IO_BITFLIPS) | 1092 | if (err != UBI_IO_BITFLIPS) |
1043 | ubi_warn("cannot read VID header back from PEB %d", to); | 1093 | ubi_warn("cannot read VID header back from PEB %d", to); |
1044 | goto out_unlock; | 1094 | else |
1095 | err = 1; | ||
1096 | goto out_unlock_buf; | ||
1045 | } | 1097 | } |
1046 | 1098 | ||
1047 | if (data_size > 0) { | 1099 | if (data_size > 0) { |
1048 | err = ubi_io_write_data(ubi, ubi->peb_buf1, to, 0, aldata_size); | 1100 | err = ubi_io_write_data(ubi, ubi->peb_buf1, to, 0, aldata_size); |
1049 | if (err) | 1101 | if (err) |
1050 | goto out_unlock; | 1102 | goto out_unlock_buf; |
1051 | 1103 | ||
1052 | cond_resched(); | 1104 | cond_resched(); |
1053 | 1105 | ||
@@ -1061,7 +1113,9 @@ int ubi_eba_copy_leb(struct ubi_device *ubi, int from, int to, | |||
1061 | if (err != UBI_IO_BITFLIPS) | 1113 | if (err != UBI_IO_BITFLIPS) |
1062 | ubi_warn("cannot read data back from PEB %d", | 1114 | ubi_warn("cannot read data back from PEB %d", |
1063 | to); | 1115 | to); |
1064 | goto out_unlock; | 1116 | else |
1117 | err = 1; | ||
1118 | goto out_unlock_buf; | ||
1065 | } | 1119 | } |
1066 | 1120 | ||
1067 | cond_resched(); | 1121 | cond_resched(); |
@@ -1069,15 +1123,16 @@ int ubi_eba_copy_leb(struct ubi_device *ubi, int from, int to, | |||
1069 | if (memcmp(ubi->peb_buf1, ubi->peb_buf2, aldata_size)) { | 1123 | if (memcmp(ubi->peb_buf1, ubi->peb_buf2, aldata_size)) { |
1070 | ubi_warn("read data back from PEB %d - it is different", | 1124 | ubi_warn("read data back from PEB %d - it is different", |
1071 | to); | 1125 | to); |
1072 | goto out_unlock; | 1126 | goto out_unlock_buf; |
1073 | } | 1127 | } |
1074 | } | 1128 | } |
1075 | 1129 | ||
1076 | ubi_assert(vol->eba_tbl[lnum] == from); | 1130 | ubi_assert(vol->eba_tbl[lnum] == from); |
1077 | vol->eba_tbl[lnum] = to; | 1131 | vol->eba_tbl[lnum] = to; |
1078 | 1132 | ||
1079 | out_unlock: | 1133 | out_unlock_buf: |
1080 | mutex_unlock(&ubi->buf_mutex); | 1134 | mutex_unlock(&ubi->buf_mutex); |
1135 | out_unlock_leb: | ||
1081 | leb_write_unlock(ubi, vol_id, lnum); | 1136 | leb_write_unlock(ubi, vol_id, lnum); |
1082 | return err; | 1137 | return err; |
1083 | } | 1138 | } |
diff --git a/drivers/mtd/ubi/ubi.h b/drivers/mtd/ubi/ubi.h index f782d5aa849a..ea9a6990a4dc 100644 --- a/drivers/mtd/ubi/ubi.h +++ b/drivers/mtd/ubi/ubi.h | |||
@@ -275,13 +275,13 @@ struct ubi_wl_entry; | |||
275 | * @wl_lock: protects the @used, @free, @prot, @lookuptbl, @abs_ec, @move_from, | 275 | * @wl_lock: protects the @used, @free, @prot, @lookuptbl, @abs_ec, @move_from, |
276 | * @move_to, @move_to_put @erase_pending, @wl_scheduled, and @works | 276 | * @move_to, @move_to_put @erase_pending, @wl_scheduled, and @works |
277 | * fields | 277 | * fields |
278 | * @move_mutex: serializes eraseblock moves | ||
278 | * @wl_scheduled: non-zero if the wear-leveling was scheduled | 279 | * @wl_scheduled: non-zero if the wear-leveling was scheduled |
279 | * @lookuptbl: a table to quickly find a &struct ubi_wl_entry object for any | 280 | * @lookuptbl: a table to quickly find a &struct ubi_wl_entry object for any |
280 | * physical eraseblock | 281 | * physical eraseblock |
281 | * @abs_ec: absolute erase counter | 282 | * @abs_ec: absolute erase counter |
282 | * @move_from: physical eraseblock from where the data is being moved | 283 | * @move_from: physical eraseblock from where the data is being moved |
283 | * @move_to: physical eraseblock where the data is being moved to | 284 | * @move_to: physical eraseblock where the data is being moved to |
284 | * @move_from_put: if the "from" PEB was put | ||
285 | * @move_to_put: if the "to" PEB was put | 285 | * @move_to_put: if the "to" PEB was put |
286 | * @works: list of pending works | 286 | * @works: list of pending works |
287 | * @works_count: count of pending works | 287 | * @works_count: count of pending works |
@@ -354,12 +354,12 @@ struct ubi_device { | |||
354 | struct rb_root aec; | 354 | struct rb_root aec; |
355 | } prot; | 355 | } prot; |
356 | spinlock_t wl_lock; | 356 | spinlock_t wl_lock; |
357 | struct mutex move_mutex; | ||
357 | int wl_scheduled; | 358 | int wl_scheduled; |
358 | struct ubi_wl_entry **lookuptbl; | 359 | struct ubi_wl_entry **lookuptbl; |
359 | unsigned long long abs_ec; | 360 | unsigned long long abs_ec; |
360 | struct ubi_wl_entry *move_from; | 361 | struct ubi_wl_entry *move_from; |
361 | struct ubi_wl_entry *move_to; | 362 | struct ubi_wl_entry *move_to; |
362 | int move_from_put; | ||
363 | int move_to_put; | 363 | int move_to_put; |
364 | struct list_head works; | 364 | struct list_head works; |
365 | int works_count; | 365 | int works_count; |
@@ -561,8 +561,10 @@ static inline int ubi_io_write_data(struct ubi_device *ubi, const void *buf, | |||
561 | */ | 561 | */ |
562 | static inline void ubi_ro_mode(struct ubi_device *ubi) | 562 | static inline void ubi_ro_mode(struct ubi_device *ubi) |
563 | { | 563 | { |
564 | ubi->ro_mode = 1; | 564 | if (!ubi->ro_mode) { |
565 | ubi_warn("switch to read-only mode"); | 565 | ubi->ro_mode = 1; |
566 | ubi_warn("switch to read-only mode"); | ||
567 | } | ||
566 | } | 568 | } |
567 | 569 | ||
568 | /** | 570 | /** |
diff --git a/drivers/mtd/ubi/wl.c b/drivers/mtd/ubi/wl.c index 36aa097203f9..a60f9425ab13 100644 --- a/drivers/mtd/ubi/wl.c +++ b/drivers/mtd/ubi/wl.c | |||
@@ -249,6 +249,8 @@ static int do_work(struct ubi_device *ubi) | |||
249 | int err; | 249 | int err; |
250 | struct ubi_work *wrk; | 250 | struct ubi_work *wrk; |
251 | 251 | ||
252 | cond_resched(); | ||
253 | |||
252 | spin_lock(&ubi->wl_lock); | 254 | spin_lock(&ubi->wl_lock); |
253 | 255 | ||
254 | if (list_empty(&ubi->works)) { | 256 | if (list_empty(&ubi->works)) { |
@@ -531,8 +533,12 @@ retry: | |||
531 | * prot_tree_del - remove a physical eraseblock from the protection trees | 533 | * prot_tree_del - remove a physical eraseblock from the protection trees |
532 | * @ubi: UBI device description object | 534 | * @ubi: UBI device description object |
533 | * @pnum: the physical eraseblock to remove | 535 | * @pnum: the physical eraseblock to remove |
536 | * | ||
537 | * This function returns PEB @pnum from the protection trees and returns zero | ||
538 | * in case of success and %-ENODEV if the PEB was not found in the protection | ||
539 | * trees. | ||
534 | */ | 540 | */ |
535 | static void prot_tree_del(struct ubi_device *ubi, int pnum) | 541 | static int prot_tree_del(struct ubi_device *ubi, int pnum) |
536 | { | 542 | { |
537 | struct rb_node *p; | 543 | struct rb_node *p; |
538 | struct ubi_wl_prot_entry *pe = NULL; | 544 | struct ubi_wl_prot_entry *pe = NULL; |
@@ -543,7 +549,7 @@ static void prot_tree_del(struct ubi_device *ubi, int pnum) | |||
543 | pe = rb_entry(p, struct ubi_wl_prot_entry, rb_pnum); | 549 | pe = rb_entry(p, struct ubi_wl_prot_entry, rb_pnum); |
544 | 550 | ||
545 | if (pnum == pe->e->pnum) | 551 | if (pnum == pe->e->pnum) |
546 | break; | 552 | goto found; |
547 | 553 | ||
548 | if (pnum < pe->e->pnum) | 554 | if (pnum < pe->e->pnum) |
549 | p = p->rb_left; | 555 | p = p->rb_left; |
@@ -551,10 +557,14 @@ static void prot_tree_del(struct ubi_device *ubi, int pnum) | |||
551 | p = p->rb_right; | 557 | p = p->rb_right; |
552 | } | 558 | } |
553 | 559 | ||
560 | return -ENODEV; | ||
561 | |||
562 | found: | ||
554 | ubi_assert(pe->e->pnum == pnum); | 563 | ubi_assert(pe->e->pnum == pnum); |
555 | rb_erase(&pe->rb_aec, &ubi->prot.aec); | 564 | rb_erase(&pe->rb_aec, &ubi->prot.aec); |
556 | rb_erase(&pe->rb_pnum, &ubi->prot.pnum); | 565 | rb_erase(&pe->rb_pnum, &ubi->prot.pnum); |
557 | kfree(pe); | 566 | kfree(pe); |
567 | return 0; | ||
558 | } | 568 | } |
559 | 569 | ||
560 | /** | 570 | /** |
@@ -726,7 +736,8 @@ static int schedule_erase(struct ubi_device *ubi, struct ubi_wl_entry *e, | |||
726 | static int wear_leveling_worker(struct ubi_device *ubi, struct ubi_work *wrk, | 736 | static int wear_leveling_worker(struct ubi_device *ubi, struct ubi_work *wrk, |
727 | int cancel) | 737 | int cancel) |
728 | { | 738 | { |
729 | int err, put = 0; | 739 | int err, put = 0, scrubbing = 0, protect = 0; |
740 | struct ubi_wl_prot_entry *pe; | ||
730 | struct ubi_wl_entry *e1, *e2; | 741 | struct ubi_wl_entry *e1, *e2; |
731 | struct ubi_vid_hdr *vid_hdr; | 742 | struct ubi_vid_hdr *vid_hdr; |
732 | 743 | ||
@@ -739,21 +750,17 @@ static int wear_leveling_worker(struct ubi_device *ubi, struct ubi_work *wrk, | |||
739 | if (!vid_hdr) | 750 | if (!vid_hdr) |
740 | return -ENOMEM; | 751 | return -ENOMEM; |
741 | 752 | ||
753 | mutex_lock(&ubi->move_mutex); | ||
742 | spin_lock(&ubi->wl_lock); | 754 | spin_lock(&ubi->wl_lock); |
755 | ubi_assert(!ubi->move_from && !ubi->move_to); | ||
756 | ubi_assert(!ubi->move_to_put); | ||
743 | 757 | ||
744 | /* | 758 | if (!ubi->free.rb_node || |
745 | * Only one WL worker at a time is supported at this implementation, so | ||
746 | * make sure a PEB is not being moved already. | ||
747 | */ | ||
748 | if (ubi->move_to || !ubi->free.rb_node || | ||
749 | (!ubi->used.rb_node && !ubi->scrub.rb_node)) { | 759 | (!ubi->used.rb_node && !ubi->scrub.rb_node)) { |
750 | /* | 760 | /* |
751 | * Only one WL worker at a time is supported at this | 761 | * No free physical eraseblocks? Well, they must be waiting in |
752 | * implementation, so if a LEB is already being moved, cancel. | 762 | * the queue to be erased. Cancel movement - it will be |
753 | * | 763 | * triggered again when a free physical eraseblock appears. |
754 | * No free physical eraseblocks? Well, we cancel wear-leveling | ||
755 | * then. It will be triggered again when a free physical | ||
756 | * eraseblock appears. | ||
757 | * | 764 | * |
758 | * No used physical eraseblocks? They must be temporarily | 765 | * No used physical eraseblocks? They must be temporarily |
759 | * protected from being moved. They will be moved to the | 766 | * protected from being moved. They will be moved to the |
@@ -762,10 +769,7 @@ static int wear_leveling_worker(struct ubi_device *ubi, struct ubi_work *wrk, | |||
762 | */ | 769 | */ |
763 | dbg_wl("cancel WL, a list is empty: free %d, used %d", | 770 | dbg_wl("cancel WL, a list is empty: free %d, used %d", |
764 | !ubi->free.rb_node, !ubi->used.rb_node); | 771 | !ubi->free.rb_node, !ubi->used.rb_node); |
765 | ubi->wl_scheduled = 0; | 772 | goto out_cancel; |
766 | spin_unlock(&ubi->wl_lock); | ||
767 | ubi_free_vid_hdr(ubi, vid_hdr); | ||
768 | return 0; | ||
769 | } | 773 | } |
770 | 774 | ||
771 | if (!ubi->scrub.rb_node) { | 775 | if (!ubi->scrub.rb_node) { |
@@ -780,16 +784,15 @@ static int wear_leveling_worker(struct ubi_device *ubi, struct ubi_work *wrk, | |||
780 | if (!(e2->ec - e1->ec >= UBI_WL_THRESHOLD)) { | 784 | if (!(e2->ec - e1->ec >= UBI_WL_THRESHOLD)) { |
781 | dbg_wl("no WL needed: min used EC %d, max free EC %d", | 785 | dbg_wl("no WL needed: min used EC %d, max free EC %d", |
782 | e1->ec, e2->ec); | 786 | e1->ec, e2->ec); |
783 | ubi->wl_scheduled = 0; | 787 | goto out_cancel; |
784 | spin_unlock(&ubi->wl_lock); | ||
785 | ubi_free_vid_hdr(ubi, vid_hdr); | ||
786 | return 0; | ||
787 | } | 788 | } |
788 | paranoid_check_in_wl_tree(e1, &ubi->used); | 789 | paranoid_check_in_wl_tree(e1, &ubi->used); |
789 | rb_erase(&e1->rb, &ubi->used); | 790 | rb_erase(&e1->rb, &ubi->used); |
790 | dbg_wl("move PEB %d EC %d to PEB %d EC %d", | 791 | dbg_wl("move PEB %d EC %d to PEB %d EC %d", |
791 | e1->pnum, e1->ec, e2->pnum, e2->ec); | 792 | e1->pnum, e1->ec, e2->pnum, e2->ec); |
792 | } else { | 793 | } else { |
794 | /* Perform scrubbing */ | ||
795 | scrubbing = 1; | ||
793 | e1 = rb_entry(rb_first(&ubi->scrub), struct ubi_wl_entry, rb); | 796 | e1 = rb_entry(rb_first(&ubi->scrub), struct ubi_wl_entry, rb); |
794 | e2 = find_wl_entry(&ubi->free, WL_FREE_MAX_DIFF); | 797 | e2 = find_wl_entry(&ubi->free, WL_FREE_MAX_DIFF); |
795 | paranoid_check_in_wl_tree(e1, &ubi->scrub); | 798 | paranoid_check_in_wl_tree(e1, &ubi->scrub); |
@@ -799,8 +802,6 @@ static int wear_leveling_worker(struct ubi_device *ubi, struct ubi_work *wrk, | |||
799 | 802 | ||
800 | paranoid_check_in_wl_tree(e2, &ubi->free); | 803 | paranoid_check_in_wl_tree(e2, &ubi->free); |
801 | rb_erase(&e2->rb, &ubi->free); | 804 | rb_erase(&e2->rb, &ubi->free); |
802 | ubi_assert(!ubi->move_from && !ubi->move_to); | ||
803 | ubi_assert(!ubi->move_to_put && !ubi->move_from_put); | ||
804 | ubi->move_from = e1; | 805 | ubi->move_from = e1; |
805 | ubi->move_to = e2; | 806 | ubi->move_to = e2; |
806 | spin_unlock(&ubi->wl_lock); | 807 | spin_unlock(&ubi->wl_lock); |
@@ -810,6 +811,10 @@ static int wear_leveling_worker(struct ubi_device *ubi, struct ubi_work *wrk, | |||
810 | * We so far do not know which logical eraseblock our physical | 811 | * We so far do not know which logical eraseblock our physical |
811 | * eraseblock (@e1) belongs to. We have to read the volume identifier | 812 | * eraseblock (@e1) belongs to. We have to read the volume identifier |
812 | * header first. | 813 | * header first. |
814 | * | ||
815 | * Note, we are protected from this PEB being unmapped and erased. The | ||
816 | * 'ubi_wl_put_peb()' would wait for moving to be finished if the PEB | ||
817 | * which is being moved was unmapped. | ||
813 | */ | 818 | */ |
814 | 819 | ||
815 | err = ubi_io_read_vid_hdr(ubi, e1->pnum, vid_hdr, 0); | 820 | err = ubi_io_read_vid_hdr(ubi, e1->pnum, vid_hdr, 0); |
@@ -824,32 +829,51 @@ static int wear_leveling_worker(struct ubi_device *ubi, struct ubi_work *wrk, | |||
824 | * likely have the VID header in place. | 829 | * likely have the VID header in place. |
825 | */ | 830 | */ |
826 | dbg_wl("PEB %d has no VID header", e1->pnum); | 831 | dbg_wl("PEB %d has no VID header", e1->pnum); |
827 | err = 0; | 832 | goto out_not_moved; |
828 | } else { | ||
829 | ubi_err("error %d while reading VID header from PEB %d", | ||
830 | err, e1->pnum); | ||
831 | if (err > 0) | ||
832 | err = -EIO; | ||
833 | } | 833 | } |
834 | goto error; | 834 | |
835 | ubi_err("error %d while reading VID header from PEB %d", | ||
836 | err, e1->pnum); | ||
837 | if (err > 0) | ||
838 | err = -EIO; | ||
839 | goto out_error; | ||
835 | } | 840 | } |
836 | 841 | ||
837 | err = ubi_eba_copy_leb(ubi, e1->pnum, e2->pnum, vid_hdr); | 842 | err = ubi_eba_copy_leb(ubi, e1->pnum, e2->pnum, vid_hdr); |
838 | if (err) { | 843 | if (err) { |
839 | if (err == UBI_IO_BITFLIPS) | 844 | |
840 | err = 0; | 845 | if (err < 0) |
841 | goto error; | 846 | goto out_error; |
847 | if (err == 1) | ||
848 | goto out_not_moved; | ||
849 | |||
850 | /* | ||
851 | * For some reason the LEB was not moved - it might be because | ||
852 | * the volume is being deleted. We should prevent this PEB from | ||
853 | * being selected for wear-levelling movement for some "time", | ||
854 | * so put it to the protection tree. | ||
855 | */ | ||
856 | |||
857 | dbg_wl("cancelled moving PEB %d", e1->pnum); | ||
858 | pe = kmalloc(sizeof(struct ubi_wl_prot_entry), GFP_NOFS); | ||
859 | if (!pe) { | ||
860 | err = -ENOMEM; | ||
861 | goto out_error; | ||
862 | } | ||
863 | |||
864 | protect = 1; | ||
842 | } | 865 | } |
843 | 866 | ||
844 | ubi_free_vid_hdr(ubi, vid_hdr); | 867 | ubi_free_vid_hdr(ubi, vid_hdr); |
845 | spin_lock(&ubi->wl_lock); | 868 | spin_lock(&ubi->wl_lock); |
869 | if (protect) | ||
870 | prot_tree_add(ubi, e1, pe, protect); | ||
846 | if (!ubi->move_to_put) | 871 | if (!ubi->move_to_put) |
847 | wl_tree_add(e2, &ubi->used); | 872 | wl_tree_add(e2, &ubi->used); |
848 | else | 873 | else |
849 | put = 1; | 874 | put = 1; |
850 | ubi->move_from = ubi->move_to = NULL; | 875 | ubi->move_from = ubi->move_to = NULL; |
851 | ubi->move_from_put = ubi->move_to_put = 0; | 876 | ubi->move_to_put = ubi->wl_scheduled = 0; |
852 | ubi->wl_scheduled = 0; | ||
853 | spin_unlock(&ubi->wl_lock); | 877 | spin_unlock(&ubi->wl_lock); |
854 | 878 | ||
855 | if (put) { | 879 | if (put) { |
@@ -859,62 +883,67 @@ static int wear_leveling_worker(struct ubi_device *ubi, struct ubi_work *wrk, | |||
859 | */ | 883 | */ |
860 | dbg_wl("PEB %d was put meanwhile, erase", e2->pnum); | 884 | dbg_wl("PEB %d was put meanwhile, erase", e2->pnum); |
861 | err = schedule_erase(ubi, e2, 0); | 885 | err = schedule_erase(ubi, e2, 0); |
862 | if (err) { | 886 | if (err) |
863 | kmem_cache_free(ubi_wl_entry_slab, e2); | 887 | goto out_error; |
864 | ubi_ro_mode(ubi); | ||
865 | } | ||
866 | } | 888 | } |
867 | 889 | ||
868 | err = schedule_erase(ubi, e1, 0); | 890 | if (!protect) { |
869 | if (err) { | 891 | err = schedule_erase(ubi, e1, 0); |
870 | kmem_cache_free(ubi_wl_entry_slab, e1); | 892 | if (err) |
871 | ubi_ro_mode(ubi); | 893 | goto out_error; |
872 | } | 894 | } |
873 | 895 | ||
896 | |||
874 | dbg_wl("done"); | 897 | dbg_wl("done"); |
875 | return err; | 898 | mutex_unlock(&ubi->move_mutex); |
899 | return 0; | ||
876 | 900 | ||
877 | /* | 901 | /* |
878 | * Some error occurred. @e1 was not changed, so return it back. @e2 | 902 | * For some reasons the LEB was not moved, might be an error, might be |
879 | * might be changed, schedule it for erasure. | 903 | * something else. @e1 was not changed, so return it back. @e2 might |
904 | * be changed, schedule it for erasure. | ||
880 | */ | 905 | */ |
881 | error: | 906 | out_not_moved: |
882 | if (err) | ||
883 | dbg_wl("error %d occurred, cancel operation", err); | ||
884 | ubi_assert(err <= 0); | ||
885 | |||
886 | ubi_free_vid_hdr(ubi, vid_hdr); | 907 | ubi_free_vid_hdr(ubi, vid_hdr); |
887 | spin_lock(&ubi->wl_lock); | 908 | spin_lock(&ubi->wl_lock); |
888 | ubi->wl_scheduled = 0; | 909 | if (scrubbing) |
889 | if (ubi->move_from_put) | 910 | wl_tree_add(e1, &ubi->scrub); |
890 | put = 1; | ||
891 | else | 911 | else |
892 | wl_tree_add(e1, &ubi->used); | 912 | wl_tree_add(e1, &ubi->used); |
893 | ubi->move_from = ubi->move_to = NULL; | 913 | ubi->move_from = ubi->move_to = NULL; |
894 | ubi->move_from_put = ubi->move_to_put = 0; | 914 | ubi->move_to_put = ubi->wl_scheduled = 0; |
895 | spin_unlock(&ubi->wl_lock); | 915 | spin_unlock(&ubi->wl_lock); |
896 | 916 | ||
897 | if (put) { | ||
898 | /* | ||
899 | * Well, the target PEB was put meanwhile, schedule it for | ||
900 | * erasure. | ||
901 | */ | ||
902 | dbg_wl("PEB %d was put meanwhile, erase", e1->pnum); | ||
903 | err = schedule_erase(ubi, e1, 0); | ||
904 | if (err) { | ||
905 | kmem_cache_free(ubi_wl_entry_slab, e1); | ||
906 | ubi_ro_mode(ubi); | ||
907 | } | ||
908 | } | ||
909 | |||
910 | err = schedule_erase(ubi, e2, 0); | 917 | err = schedule_erase(ubi, e2, 0); |
911 | if (err) { | 918 | if (err) |
912 | kmem_cache_free(ubi_wl_entry_slab, e2); | 919 | goto out_error; |
913 | ubi_ro_mode(ubi); | 920 | |
914 | } | 921 | mutex_unlock(&ubi->move_mutex); |
922 | return 0; | ||
923 | |||
924 | out_error: | ||
925 | ubi_err("error %d while moving PEB %d to PEB %d", | ||
926 | err, e1->pnum, e2->pnum); | ||
915 | 927 | ||
916 | yield(); | 928 | ubi_free_vid_hdr(ubi, vid_hdr); |
929 | spin_lock(&ubi->wl_lock); | ||
930 | ubi->move_from = ubi->move_to = NULL; | ||
931 | ubi->move_to_put = ubi->wl_scheduled = 0; | ||
932 | spin_unlock(&ubi->wl_lock); | ||
933 | |||
934 | kmem_cache_free(ubi_wl_entry_slab, e1); | ||
935 | kmem_cache_free(ubi_wl_entry_slab, e2); | ||
936 | ubi_ro_mode(ubi); | ||
937 | |||
938 | mutex_unlock(&ubi->move_mutex); | ||
917 | return err; | 939 | return err; |
940 | |||
941 | out_cancel: | ||
942 | ubi->wl_scheduled = 0; | ||
943 | spin_unlock(&ubi->wl_lock); | ||
944 | mutex_unlock(&ubi->move_mutex); | ||
945 | ubi_free_vid_hdr(ubi, vid_hdr); | ||
946 | return 0; | ||
918 | } | 947 | } |
919 | 948 | ||
920 | /** | 949 | /** |
@@ -1101,8 +1130,7 @@ out_ro: | |||
1101 | } | 1130 | } |
1102 | 1131 | ||
1103 | /** | 1132 | /** |
1104 | * ubi_wl_put_peb - return a physical eraseblock to the wear-leveling | 1133 | * ubi_wl_put_peb - return a physical eraseblock to the wear-leveling unit. |
1105 | * unit. | ||
1106 | * @ubi: UBI device description object | 1134 | * @ubi: UBI device description object |
1107 | * @pnum: physical eraseblock to return | 1135 | * @pnum: physical eraseblock to return |
1108 | * @torture: if this physical eraseblock has to be tortured | 1136 | * @torture: if this physical eraseblock has to be tortured |
@@ -1110,7 +1138,7 @@ out_ro: | |||
1110 | * This function is called to return physical eraseblock @pnum to the pool of | 1138 | * This function is called to return physical eraseblock @pnum to the pool of |
1111 | * free physical eraseblocks. The @torture flag has to be set if an I/O error | 1139 | * free physical eraseblocks. The @torture flag has to be set if an I/O error |
1112 | * occurred to this @pnum and it has to be tested. This function returns zero | 1140 | * occurred to this @pnum and it has to be tested. This function returns zero |
1113 | * in case of success and a negative error code in case of failure. | 1141 | * in case of success, and a negative error code in case of failure. |
1114 | */ | 1142 | */ |
1115 | int ubi_wl_put_peb(struct ubi_device *ubi, int pnum, int torture) | 1143 | int ubi_wl_put_peb(struct ubi_device *ubi, int pnum, int torture) |
1116 | { | 1144 | { |
@@ -1121,8 +1149,8 @@ int ubi_wl_put_peb(struct ubi_device *ubi, int pnum, int torture) | |||
1121 | ubi_assert(pnum >= 0); | 1149 | ubi_assert(pnum >= 0); |
1122 | ubi_assert(pnum < ubi->peb_count); | 1150 | ubi_assert(pnum < ubi->peb_count); |
1123 | 1151 | ||
1152 | retry: | ||
1124 | spin_lock(&ubi->wl_lock); | 1153 | spin_lock(&ubi->wl_lock); |
1125 | |||
1126 | e = ubi->lookuptbl[pnum]; | 1154 | e = ubi->lookuptbl[pnum]; |
1127 | if (e == ubi->move_from) { | 1155 | if (e == ubi->move_from) { |
1128 | /* | 1156 | /* |
@@ -1130,11 +1158,13 @@ int ubi_wl_put_peb(struct ubi_device *ubi, int pnum, int torture) | |||
1130 | * be moved. It will be scheduled for erasure in the | 1158 | * be moved. It will be scheduled for erasure in the |
1131 | * wear-leveling worker. | 1159 | * wear-leveling worker. |
1132 | */ | 1160 | */ |
1133 | dbg_wl("PEB %d is being moved", pnum); | 1161 | dbg_wl("PEB %d is being moved, wait", pnum); |
1134 | ubi_assert(!ubi->move_from_put); | ||
1135 | ubi->move_from_put = 1; | ||
1136 | spin_unlock(&ubi->wl_lock); | 1162 | spin_unlock(&ubi->wl_lock); |
1137 | return 0; | 1163 | |
1164 | /* Wait for the WL worker by taking the @ubi->move_mutex */ | ||
1165 | mutex_lock(&ubi->move_mutex); | ||
1166 | mutex_unlock(&ubi->move_mutex); | ||
1167 | goto retry; | ||
1138 | } else if (e == ubi->move_to) { | 1168 | } else if (e == ubi->move_to) { |
1139 | /* | 1169 | /* |
1140 | * User is putting the physical eraseblock which was selected | 1170 | * User is putting the physical eraseblock which was selected |
@@ -1157,8 +1187,15 @@ int ubi_wl_put_peb(struct ubi_device *ubi, int pnum, int torture) | |||
1157 | } else if (in_wl_tree(e, &ubi->scrub)) { | 1187 | } else if (in_wl_tree(e, &ubi->scrub)) { |
1158 | paranoid_check_in_wl_tree(e, &ubi->scrub); | 1188 | paranoid_check_in_wl_tree(e, &ubi->scrub); |
1159 | rb_erase(&e->rb, &ubi->scrub); | 1189 | rb_erase(&e->rb, &ubi->scrub); |
1160 | } else | 1190 | } else { |
1161 | prot_tree_del(ubi, e->pnum); | 1191 | err = prot_tree_del(ubi, e->pnum); |
1192 | if (err) { | ||
1193 | ubi_err("PEB %d not found", pnum); | ||
1194 | ubi_ro_mode(ubi); | ||
1195 | spin_unlock(&ubi->wl_lock); | ||
1196 | return err; | ||
1197 | } | ||
1198 | } | ||
1162 | } | 1199 | } |
1163 | spin_unlock(&ubi->wl_lock); | 1200 | spin_unlock(&ubi->wl_lock); |
1164 | 1201 | ||
@@ -1212,8 +1249,17 @@ retry: | |||
1212 | if (in_wl_tree(e, &ubi->used)) { | 1249 | if (in_wl_tree(e, &ubi->used)) { |
1213 | paranoid_check_in_wl_tree(e, &ubi->used); | 1250 | paranoid_check_in_wl_tree(e, &ubi->used); |
1214 | rb_erase(&e->rb, &ubi->used); | 1251 | rb_erase(&e->rb, &ubi->used); |
1215 | } else | 1252 | } else { |
1216 | prot_tree_del(ubi, pnum); | 1253 | int err; |
1254 | |||
1255 | err = prot_tree_del(ubi, e->pnum); | ||
1256 | if (err) { | ||
1257 | ubi_err("PEB %d not found", pnum); | ||
1258 | ubi_ro_mode(ubi); | ||
1259 | spin_unlock(&ubi->wl_lock); | ||
1260 | return err; | ||
1261 | } | ||
1262 | } | ||
1217 | 1263 | ||
1218 | wl_tree_add(e, &ubi->scrub); | 1264 | wl_tree_add(e, &ubi->scrub); |
1219 | spin_unlock(&ubi->wl_lock); | 1265 | spin_unlock(&ubi->wl_lock); |
@@ -1379,6 +1425,7 @@ int ubi_wl_init_scan(struct ubi_device *ubi, struct ubi_scan_info *si) | |||
1379 | ubi->used = ubi->free = ubi->scrub = RB_ROOT; | 1425 | ubi->used = ubi->free = ubi->scrub = RB_ROOT; |
1380 | ubi->prot.pnum = ubi->prot.aec = RB_ROOT; | 1426 | ubi->prot.pnum = ubi->prot.aec = RB_ROOT; |
1381 | spin_lock_init(&ubi->wl_lock); | 1427 | spin_lock_init(&ubi->wl_lock); |
1428 | mutex_init(&ubi->move_mutex); | ||
1382 | ubi->max_ec = si->max_ec; | 1429 | ubi->max_ec = si->max_ec; |
1383 | INIT_LIST_HEAD(&ubi->works); | 1430 | INIT_LIST_HEAD(&ubi->works); |
1384 | 1431 | ||