aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/mtd
diff options
context:
space:
mode:
authorStefani Seibold <stefani@seibold.net>2010-04-18 16:46:44 -0400
committerDavid Woodhouse <David.Woodhouse@intel.com>2010-05-10 09:22:30 -0400
commitc4e773764cead9358fd4b036d1b883fff3968513 (patch)
treeceb30e53d7ba33a071653c2bc05c06293d84575f /drivers/mtd
parent67026418f534045525a7c39f506006cd7fbd197f (diff)
mtd: fix a huge latency problem in the MTD CFI and LPDDR flash drivers.
The use of a memcpy() during a spinlock operation will cause very long thread context switch delays if the flash chip bandwidth is low and the data to be copied large, because a spinlock will disable preemption. For example: A flash with 6,5 MB/s bandwidth will cause under ubifs, which request sometimes 128 KiB (the flash erase size), a preemption delay of 20 milliseconds. High priority threads will not be served during this time, regardless whether this threads access the flash or not. This behavior breaks real time. The patch changes all the use of spin_lock operations for xxxx->mutex into mutex operations, which is exact what the name says and means. I have checked the code of the drivers and there is no use of atomic pathes like interrupt or timers. The mtdoops facility will also not be used by this drivers. So it is dave to replace the spin_lock against mutex. There is no performance regression since the mutex is normally not acquired. Changelog: 06.03.2010 First release 26.03.2010 Fix mutex[1] issue and tested it for compile failure Signed-off-by: Stefani Seibold <stefani@seibold.net> Signed-off-by: Artem Bityutskiy <Artem.Bityutskiy@nokia.com> Signed-off-by: David Woodhouse <David.Woodhouse@intel.com>
Diffstat (limited to 'drivers/mtd')
-rw-r--r--drivers/mtd/chips/cfi_cmdset_0001.c131
-rw-r--r--drivers/mtd/chips/cfi_cmdset_0002.c122
-rw-r--r--drivers/mtd/chips/cfi_cmdset_0020.c136
-rw-r--r--drivers/mtd/chips/fwh_lock.h6
-rw-r--r--drivers/mtd/chips/gen_probe.c3
-rw-r--r--drivers/mtd/lpddr/lpddr_cmds.c79
6 files changed, 237 insertions, 240 deletions
diff --git a/drivers/mtd/chips/cfi_cmdset_0001.c b/drivers/mtd/chips/cfi_cmdset_0001.c
index 92530433c11c..62f3ea9de848 100644
--- a/drivers/mtd/chips/cfi_cmdset_0001.c
+++ b/drivers/mtd/chips/cfi_cmdset_0001.c
@@ -725,8 +725,7 @@ static int cfi_intelext_partition_fixup(struct mtd_info *mtd,
725 /* those should be reset too since 725 /* those should be reset too since
726 they create memory references. */ 726 they create memory references. */
727 init_waitqueue_head(&chip->wq); 727 init_waitqueue_head(&chip->wq);
728 spin_lock_init(&chip->_spinlock); 728 mutex_init(&chip->mutex);
729 chip->mutex = &chip->_spinlock;
730 chip++; 729 chip++;
731 } 730 }
732 } 731 }
@@ -772,9 +771,9 @@ static int chip_ready (struct map_info *map, struct flchip *chip, unsigned long
772 if (chip->priv && map_word_andequal(map, status, status_PWS, status_PWS)) 771 if (chip->priv && map_word_andequal(map, status, status_PWS, status_PWS))
773 break; 772 break;
774 773
775 spin_unlock(chip->mutex); 774 mutex_unlock(&chip->mutex);
776 cfi_udelay(1); 775 cfi_udelay(1);
777 spin_lock(chip->mutex); 776 mutex_lock(&chip->mutex);
778 /* Someone else might have been playing with it. */ 777 /* Someone else might have been playing with it. */
779 return -EAGAIN; 778 return -EAGAIN;
780 } 779 }
@@ -821,9 +820,9 @@ static int chip_ready (struct map_info *map, struct flchip *chip, unsigned long
821 return -EIO; 820 return -EIO;
822 } 821 }
823 822
824 spin_unlock(chip->mutex); 823 mutex_unlock(&chip->mutex);
825 cfi_udelay(1); 824 cfi_udelay(1);
826 spin_lock(chip->mutex); 825 mutex_lock(&chip->mutex);
827 /* Nobody will touch it while it's in state FL_ERASE_SUSPENDING. 826 /* Nobody will touch it while it's in state FL_ERASE_SUSPENDING.
828 So we can just loop here. */ 827 So we can just loop here. */
829 } 828 }
@@ -850,10 +849,10 @@ static int chip_ready (struct map_info *map, struct flchip *chip, unsigned long
850 sleep: 849 sleep:
851 set_current_state(TASK_UNINTERRUPTIBLE); 850 set_current_state(TASK_UNINTERRUPTIBLE);
852 add_wait_queue(&chip->wq, &wait); 851 add_wait_queue(&chip->wq, &wait);
853 spin_unlock(chip->mutex); 852 mutex_unlock(&chip->mutex);
854 schedule(); 853 schedule();
855 remove_wait_queue(&chip->wq, &wait); 854 remove_wait_queue(&chip->wq, &wait);
856 spin_lock(chip->mutex); 855 mutex_lock(&chip->mutex);
857 return -EAGAIN; 856 return -EAGAIN;
858 } 857 }
859} 858}
@@ -899,20 +898,20 @@ static int get_chip(struct map_info *map, struct flchip *chip, unsigned long adr
899 * it'll happily send us to sleep. In any case, when 898 * it'll happily send us to sleep. In any case, when
900 * get_chip returns success we're clear to go ahead. 899 * get_chip returns success we're clear to go ahead.
901 */ 900 */
902 ret = spin_trylock(contender->mutex); 901 ret = mutex_trylock(&contender->mutex);
903 spin_unlock(&shared->lock); 902 spin_unlock(&shared->lock);
904 if (!ret) 903 if (!ret)
905 goto retry; 904 goto retry;
906 spin_unlock(chip->mutex); 905 mutex_unlock(&chip->mutex);
907 ret = chip_ready(map, contender, contender->start, mode); 906 ret = chip_ready(map, contender, contender->start, mode);
908 spin_lock(chip->mutex); 907 mutex_lock(&chip->mutex);
909 908
910 if (ret == -EAGAIN) { 909 if (ret == -EAGAIN) {
911 spin_unlock(contender->mutex); 910 mutex_unlock(&contender->mutex);
912 goto retry; 911 goto retry;
913 } 912 }
914 if (ret) { 913 if (ret) {
915 spin_unlock(contender->mutex); 914 mutex_unlock(&contender->mutex);
916 return ret; 915 return ret;
917 } 916 }
918 spin_lock(&shared->lock); 917 spin_lock(&shared->lock);
@@ -921,10 +920,10 @@ static int get_chip(struct map_info *map, struct flchip *chip, unsigned long adr
921 * in FL_SYNCING state. Put contender and retry. */ 920 * in FL_SYNCING state. Put contender and retry. */
922 if (chip->state == FL_SYNCING) { 921 if (chip->state == FL_SYNCING) {
923 put_chip(map, contender, contender->start); 922 put_chip(map, contender, contender->start);
924 spin_unlock(contender->mutex); 923 mutex_unlock(&contender->mutex);
925 goto retry; 924 goto retry;
926 } 925 }
927 spin_unlock(contender->mutex); 926 mutex_unlock(&contender->mutex);
928 } 927 }
929 928
930 /* Check if we already have suspended erase 929 /* Check if we already have suspended erase
@@ -934,10 +933,10 @@ static int get_chip(struct map_info *map, struct flchip *chip, unsigned long adr
934 spin_unlock(&shared->lock); 933 spin_unlock(&shared->lock);
935 set_current_state(TASK_UNINTERRUPTIBLE); 934 set_current_state(TASK_UNINTERRUPTIBLE);
936 add_wait_queue(&chip->wq, &wait); 935 add_wait_queue(&chip->wq, &wait);
937 spin_unlock(chip->mutex); 936 mutex_unlock(&chip->mutex);
938 schedule(); 937 schedule();
939 remove_wait_queue(&chip->wq, &wait); 938 remove_wait_queue(&chip->wq, &wait);
940 spin_lock(chip->mutex); 939 mutex_lock(&chip->mutex);
941 goto retry; 940 goto retry;
942 } 941 }
943 942
@@ -967,12 +966,12 @@ static void put_chip(struct map_info *map, struct flchip *chip, unsigned long ad
967 if (shared->writing && shared->writing != chip) { 966 if (shared->writing && shared->writing != chip) {
968 /* give back ownership to who we loaned it from */ 967 /* give back ownership to who we loaned it from */
969 struct flchip *loaner = shared->writing; 968 struct flchip *loaner = shared->writing;
970 spin_lock(loaner->mutex); 969 mutex_lock(&loaner->mutex);
971 spin_unlock(&shared->lock); 970 spin_unlock(&shared->lock);
972 spin_unlock(chip->mutex); 971 mutex_unlock(&chip->mutex);
973 put_chip(map, loaner, loaner->start); 972 put_chip(map, loaner, loaner->start);
974 spin_lock(chip->mutex); 973 mutex_lock(&chip->mutex);
975 spin_unlock(loaner->mutex); 974 mutex_unlock(&loaner->mutex);
976 wake_up(&chip->wq); 975 wake_up(&chip->wq);
977 return; 976 return;
978 } 977 }
@@ -1142,7 +1141,7 @@ static int __xipram xip_wait_for_operation(
1142 (void) map_read(map, adr); 1141 (void) map_read(map, adr);
1143 xip_iprefetch(); 1142 xip_iprefetch();
1144 local_irq_enable(); 1143 local_irq_enable();
1145 spin_unlock(chip->mutex); 1144 mutex_unlock(&chip->mutex);
1146 xip_iprefetch(); 1145 xip_iprefetch();
1147 cond_resched(); 1146 cond_resched();
1148 1147
@@ -1152,15 +1151,15 @@ static int __xipram xip_wait_for_operation(
1152 * a suspended erase state. If so let's wait 1151 * a suspended erase state. If so let's wait
1153 * until it's done. 1152 * until it's done.
1154 */ 1153 */
1155 spin_lock(chip->mutex); 1154 mutex_lock(&chip->mutex);
1156 while (chip->state != newstate) { 1155 while (chip->state != newstate) {
1157 DECLARE_WAITQUEUE(wait, current); 1156 DECLARE_WAITQUEUE(wait, current);
1158 set_current_state(TASK_UNINTERRUPTIBLE); 1157 set_current_state(TASK_UNINTERRUPTIBLE);
1159 add_wait_queue(&chip->wq, &wait); 1158 add_wait_queue(&chip->wq, &wait);
1160 spin_unlock(chip->mutex); 1159 mutex_unlock(&chip->mutex);
1161 schedule(); 1160 schedule();
1162 remove_wait_queue(&chip->wq, &wait); 1161 remove_wait_queue(&chip->wq, &wait);
1163 spin_lock(chip->mutex); 1162 mutex_lock(&chip->mutex);
1164 } 1163 }
1165 /* Disallow XIP again */ 1164 /* Disallow XIP again */
1166 local_irq_disable(); 1165 local_irq_disable();
@@ -1216,10 +1215,10 @@ static int inval_cache_and_wait_for_operation(
1216 int chip_state = chip->state; 1215 int chip_state = chip->state;
1217 unsigned int timeo, sleep_time, reset_timeo; 1216 unsigned int timeo, sleep_time, reset_timeo;
1218 1217
1219 spin_unlock(chip->mutex); 1218 mutex_unlock(&chip->mutex);
1220 if (inval_len) 1219 if (inval_len)
1221 INVALIDATE_CACHED_RANGE(map, inval_adr, inval_len); 1220 INVALIDATE_CACHED_RANGE(map, inval_adr, inval_len);
1222 spin_lock(chip->mutex); 1221 mutex_lock(&chip->mutex);
1223 1222
1224 timeo = chip_op_time_max; 1223 timeo = chip_op_time_max;
1225 if (!timeo) 1224 if (!timeo)
@@ -1239,7 +1238,7 @@ static int inval_cache_and_wait_for_operation(
1239 } 1238 }
1240 1239
1241 /* OK Still waiting. Drop the lock, wait a while and retry. */ 1240 /* OK Still waiting. Drop the lock, wait a while and retry. */
1242 spin_unlock(chip->mutex); 1241 mutex_unlock(&chip->mutex);
1243 if (sleep_time >= 1000000/HZ) { 1242 if (sleep_time >= 1000000/HZ) {
1244 /* 1243 /*
1245 * Half of the normal delay still remaining 1244 * Half of the normal delay still remaining
@@ -1254,17 +1253,17 @@ static int inval_cache_and_wait_for_operation(
1254 cond_resched(); 1253 cond_resched();
1255 timeo--; 1254 timeo--;
1256 } 1255 }
1257 spin_lock(chip->mutex); 1256 mutex_lock(&chip->mutex);
1258 1257
1259 while (chip->state != chip_state) { 1258 while (chip->state != chip_state) {
1260 /* Someone's suspended the operation: sleep */ 1259 /* Someone's suspended the operation: sleep */
1261 DECLARE_WAITQUEUE(wait, current); 1260 DECLARE_WAITQUEUE(wait, current);
1262 set_current_state(TASK_UNINTERRUPTIBLE); 1261 set_current_state(TASK_UNINTERRUPTIBLE);
1263 add_wait_queue(&chip->wq, &wait); 1262 add_wait_queue(&chip->wq, &wait);
1264 spin_unlock(chip->mutex); 1263 mutex_unlock(&chip->mutex);
1265 schedule(); 1264 schedule();
1266 remove_wait_queue(&chip->wq, &wait); 1265 remove_wait_queue(&chip->wq, &wait);
1267 spin_lock(chip->mutex); 1266 mutex_lock(&chip->mutex);
1268 } 1267 }
1269 if (chip->erase_suspended && chip_state == FL_ERASING) { 1268 if (chip->erase_suspended && chip_state == FL_ERASING) {
1270 /* Erase suspend occured while sleep: reset timeout */ 1269 /* Erase suspend occured while sleep: reset timeout */
@@ -1300,7 +1299,7 @@ static int do_point_onechip (struct map_info *map, struct flchip *chip, loff_t a
1300 /* Ensure cmd read/writes are aligned. */ 1299 /* Ensure cmd read/writes are aligned. */
1301 cmd_addr = adr & ~(map_bankwidth(map)-1); 1300 cmd_addr = adr & ~(map_bankwidth(map)-1);
1302 1301
1303 spin_lock(chip->mutex); 1302 mutex_lock(&chip->mutex);
1304 1303
1305 ret = get_chip(map, chip, cmd_addr, FL_POINT); 1304 ret = get_chip(map, chip, cmd_addr, FL_POINT);
1306 1305
@@ -1311,7 +1310,7 @@ static int do_point_onechip (struct map_info *map, struct flchip *chip, loff_t a
1311 chip->state = FL_POINT; 1310 chip->state = FL_POINT;
1312 chip->ref_point_counter++; 1311 chip->ref_point_counter++;
1313 } 1312 }
1314 spin_unlock(chip->mutex); 1313 mutex_unlock(&chip->mutex);
1315 1314
1316 return ret; 1315 return ret;
1317} 1316}
@@ -1396,7 +1395,7 @@ static void cfi_intelext_unpoint(struct mtd_info *mtd, loff_t from, size_t len)
1396 else 1395 else
1397 thislen = len; 1396 thislen = len;
1398 1397
1399 spin_lock(chip->mutex); 1398 mutex_lock(&chip->mutex);
1400 if (chip->state == FL_POINT) { 1399 if (chip->state == FL_POINT) {
1401 chip->ref_point_counter--; 1400 chip->ref_point_counter--;
1402 if(chip->ref_point_counter == 0) 1401 if(chip->ref_point_counter == 0)
@@ -1405,7 +1404,7 @@ static void cfi_intelext_unpoint(struct mtd_info *mtd, loff_t from, size_t len)
1405 printk(KERN_ERR "%s: Warning: unpoint called on non pointed region\n", map->name); /* Should this give an error? */ 1404 printk(KERN_ERR "%s: Warning: unpoint called on non pointed region\n", map->name); /* Should this give an error? */
1406 1405
1407 put_chip(map, chip, chip->start); 1406 put_chip(map, chip, chip->start);
1408 spin_unlock(chip->mutex); 1407 mutex_unlock(&chip->mutex);
1409 1408
1410 len -= thislen; 1409 len -= thislen;
1411 ofs = 0; 1410 ofs = 0;
@@ -1424,10 +1423,10 @@ static inline int do_read_onechip(struct map_info *map, struct flchip *chip, lof
1424 /* Ensure cmd read/writes are aligned. */ 1423 /* Ensure cmd read/writes are aligned. */
1425 cmd_addr = adr & ~(map_bankwidth(map)-1); 1424 cmd_addr = adr & ~(map_bankwidth(map)-1);
1426 1425
1427 spin_lock(chip->mutex); 1426 mutex_lock(&chip->mutex);
1428 ret = get_chip(map, chip, cmd_addr, FL_READY); 1427 ret = get_chip(map, chip, cmd_addr, FL_READY);
1429 if (ret) { 1428 if (ret) {
1430 spin_unlock(chip->mutex); 1429 mutex_unlock(&chip->mutex);
1431 return ret; 1430 return ret;
1432 } 1431 }
1433 1432
@@ -1441,7 +1440,7 @@ static inline int do_read_onechip(struct map_info *map, struct flchip *chip, lof
1441 1440
1442 put_chip(map, chip, cmd_addr); 1441 put_chip(map, chip, cmd_addr);
1443 1442
1444 spin_unlock(chip->mutex); 1443 mutex_unlock(&chip->mutex);
1445 return 0; 1444 return 0;
1446} 1445}
1447 1446
@@ -1504,10 +1503,10 @@ static int __xipram do_write_oneword(struct map_info *map, struct flchip *chip,
1504 return -EINVAL; 1503 return -EINVAL;
1505 } 1504 }
1506 1505
1507 spin_lock(chip->mutex); 1506 mutex_lock(&chip->mutex);
1508 ret = get_chip(map, chip, adr, mode); 1507 ret = get_chip(map, chip, adr, mode);
1509 if (ret) { 1508 if (ret) {
1510 spin_unlock(chip->mutex); 1509 mutex_unlock(&chip->mutex);
1511 return ret; 1510 return ret;
1512 } 1511 }
1513 1512
@@ -1553,7 +1552,7 @@ static int __xipram do_write_oneword(struct map_info *map, struct flchip *chip,
1553 1552
1554 xip_enable(map, chip, adr); 1553 xip_enable(map, chip, adr);
1555 out: put_chip(map, chip, adr); 1554 out: put_chip(map, chip, adr);
1556 spin_unlock(chip->mutex); 1555 mutex_unlock(&chip->mutex);
1557 return ret; 1556 return ret;
1558} 1557}
1559 1558
@@ -1662,10 +1661,10 @@ static int __xipram do_write_buffer(struct map_info *map, struct flchip *chip,
1662 /* Let's determine this according to the interleave only once */ 1661 /* Let's determine this according to the interleave only once */
1663 write_cmd = (cfi->cfiq->P_ID != 0x0200) ? CMD(0xe8) : CMD(0xe9); 1662 write_cmd = (cfi->cfiq->P_ID != 0x0200) ? CMD(0xe8) : CMD(0xe9);
1664 1663
1665 spin_lock(chip->mutex); 1664 mutex_lock(&chip->mutex);
1666 ret = get_chip(map, chip, cmd_adr, FL_WRITING); 1665 ret = get_chip(map, chip, cmd_adr, FL_WRITING);
1667 if (ret) { 1666 if (ret) {
1668 spin_unlock(chip->mutex); 1667 mutex_unlock(&chip->mutex);
1669 return ret; 1668 return ret;
1670 } 1669 }
1671 1670
@@ -1796,7 +1795,7 @@ static int __xipram do_write_buffer(struct map_info *map, struct flchip *chip,
1796 1795
1797 xip_enable(map, chip, cmd_adr); 1796 xip_enable(map, chip, cmd_adr);
1798 out: put_chip(map, chip, cmd_adr); 1797 out: put_chip(map, chip, cmd_adr);
1799 spin_unlock(chip->mutex); 1798 mutex_unlock(&chip->mutex);
1800 return ret; 1799 return ret;
1801} 1800}
1802 1801
@@ -1875,10 +1874,10 @@ static int __xipram do_erase_oneblock(struct map_info *map, struct flchip *chip,
1875 adr += chip->start; 1874 adr += chip->start;
1876 1875
1877 retry: 1876 retry:
1878 spin_lock(chip->mutex); 1877 mutex_lock(&chip->mutex);
1879 ret = get_chip(map, chip, adr, FL_ERASING); 1878 ret = get_chip(map, chip, adr, FL_ERASING);
1880 if (ret) { 1879 if (ret) {
1881 spin_unlock(chip->mutex); 1880 mutex_unlock(&chip->mutex);
1882 return ret; 1881 return ret;
1883 } 1882 }
1884 1883
@@ -1934,7 +1933,7 @@ static int __xipram do_erase_oneblock(struct map_info *map, struct flchip *chip,
1934 } else if (chipstatus & 0x20 && retries--) { 1933 } else if (chipstatus & 0x20 && retries--) {
1935 printk(KERN_DEBUG "block erase failed at 0x%08lx: status 0x%lx. Retrying...\n", adr, chipstatus); 1934 printk(KERN_DEBUG "block erase failed at 0x%08lx: status 0x%lx. Retrying...\n", adr, chipstatus);
1936 put_chip(map, chip, adr); 1935 put_chip(map, chip, adr);
1937 spin_unlock(chip->mutex); 1936 mutex_unlock(&chip->mutex);
1938 goto retry; 1937 goto retry;
1939 } else { 1938 } else {
1940 printk(KERN_ERR "%s: block erase failed at 0x%08lx (status 0x%lx)\n", map->name, adr, chipstatus); 1939 printk(KERN_ERR "%s: block erase failed at 0x%08lx (status 0x%lx)\n", map->name, adr, chipstatus);
@@ -1946,7 +1945,7 @@ static int __xipram do_erase_oneblock(struct map_info *map, struct flchip *chip,
1946 1945
1947 xip_enable(map, chip, adr); 1946 xip_enable(map, chip, adr);
1948 out: put_chip(map, chip, adr); 1947 out: put_chip(map, chip, adr);
1949 spin_unlock(chip->mutex); 1948 mutex_unlock(&chip->mutex);
1950 return ret; 1949 return ret;
1951} 1950}
1952 1951
@@ -1979,7 +1978,7 @@ static void cfi_intelext_sync (struct mtd_info *mtd)
1979 for (i=0; !ret && i<cfi->numchips; i++) { 1978 for (i=0; !ret && i<cfi->numchips; i++) {
1980 chip = &cfi->chips[i]; 1979 chip = &cfi->chips[i];
1981 1980
1982 spin_lock(chip->mutex); 1981 mutex_lock(&chip->mutex);
1983 ret = get_chip(map, chip, chip->start, FL_SYNCING); 1982 ret = get_chip(map, chip, chip->start, FL_SYNCING);
1984 1983
1985 if (!ret) { 1984 if (!ret) {
@@ -1990,7 +1989,7 @@ static void cfi_intelext_sync (struct mtd_info *mtd)
1990 * with the chip now anyway. 1989 * with the chip now anyway.
1991 */ 1990 */
1992 } 1991 }
1993 spin_unlock(chip->mutex); 1992 mutex_unlock(&chip->mutex);
1994 } 1993 }
1995 1994
1996 /* Unlock the chips again */ 1995 /* Unlock the chips again */
@@ -1998,14 +1997,14 @@ static void cfi_intelext_sync (struct mtd_info *mtd)
1998 for (i--; i >=0; i--) { 1997 for (i--; i >=0; i--) {
1999 chip = &cfi->chips[i]; 1998 chip = &cfi->chips[i];
2000 1999
2001 spin_lock(chip->mutex); 2000 mutex_lock(&chip->mutex);
2002 2001
2003 if (chip->state == FL_SYNCING) { 2002 if (chip->state == FL_SYNCING) {
2004 chip->state = chip->oldstate; 2003 chip->state = chip->oldstate;
2005 chip->oldstate = FL_READY; 2004 chip->oldstate = FL_READY;
2006 wake_up(&chip->wq); 2005 wake_up(&chip->wq);
2007 } 2006 }
2008 spin_unlock(chip->mutex); 2007 mutex_unlock(&chip->mutex);
2009 } 2008 }
2010} 2009}
2011 2010
@@ -2051,10 +2050,10 @@ static int __xipram do_xxlock_oneblock(struct map_info *map, struct flchip *chip
2051 2050
2052 adr += chip->start; 2051 adr += chip->start;
2053 2052
2054 spin_lock(chip->mutex); 2053 mutex_lock(&chip->mutex);
2055 ret = get_chip(map, chip, adr, FL_LOCKING); 2054 ret = get_chip(map, chip, adr, FL_LOCKING);
2056 if (ret) { 2055 if (ret) {
2057 spin_unlock(chip->mutex); 2056 mutex_unlock(&chip->mutex);
2058 return ret; 2057 return ret;
2059 } 2058 }
2060 2059
@@ -2088,7 +2087,7 @@ static int __xipram do_xxlock_oneblock(struct map_info *map, struct flchip *chip
2088 2087
2089 xip_enable(map, chip, adr); 2088 xip_enable(map, chip, adr);
2090out: put_chip(map, chip, adr); 2089out: put_chip(map, chip, adr);
2091 spin_unlock(chip->mutex); 2090 mutex_unlock(&chip->mutex);
2092 return ret; 2091 return ret;
2093} 2092}
2094 2093
@@ -2153,10 +2152,10 @@ do_otp_read(struct map_info *map, struct flchip *chip, u_long offset,
2153 struct cfi_private *cfi = map->fldrv_priv; 2152 struct cfi_private *cfi = map->fldrv_priv;
2154 int ret; 2153 int ret;
2155 2154
2156 spin_lock(chip->mutex); 2155 mutex_lock(&chip->mutex);
2157 ret = get_chip(map, chip, chip->start, FL_JEDEC_QUERY); 2156 ret = get_chip(map, chip, chip->start, FL_JEDEC_QUERY);
2158 if (ret) { 2157 if (ret) {
2159 spin_unlock(chip->mutex); 2158 mutex_unlock(&chip->mutex);
2160 return ret; 2159 return ret;
2161 } 2160 }
2162 2161
@@ -2175,7 +2174,7 @@ do_otp_read(struct map_info *map, struct flchip *chip, u_long offset,
2175 INVALIDATE_CACHED_RANGE(map, chip->start + offset, size); 2174 INVALIDATE_CACHED_RANGE(map, chip->start + offset, size);
2176 2175
2177 put_chip(map, chip, chip->start); 2176 put_chip(map, chip, chip->start);
2178 spin_unlock(chip->mutex); 2177 mutex_unlock(&chip->mutex);
2179 return 0; 2178 return 0;
2180} 2179}
2181 2180
@@ -2450,7 +2449,7 @@ static int cfi_intelext_suspend(struct mtd_info *mtd)
2450 for (i=0; !ret && i<cfi->numchips; i++) { 2449 for (i=0; !ret && i<cfi->numchips; i++) {
2451 chip = &cfi->chips[i]; 2450 chip = &cfi->chips[i];
2452 2451
2453 spin_lock(chip->mutex); 2452 mutex_lock(&chip->mutex);
2454 2453
2455 switch (chip->state) { 2454 switch (chip->state) {
2456 case FL_READY: 2455 case FL_READY:
@@ -2482,7 +2481,7 @@ static int cfi_intelext_suspend(struct mtd_info *mtd)
2482 case FL_PM_SUSPENDED: 2481 case FL_PM_SUSPENDED:
2483 break; 2482 break;
2484 } 2483 }
2485 spin_unlock(chip->mutex); 2484 mutex_unlock(&chip->mutex);
2486 } 2485 }
2487 2486
2488 /* Unlock the chips again */ 2487 /* Unlock the chips again */
@@ -2491,7 +2490,7 @@ static int cfi_intelext_suspend(struct mtd_info *mtd)
2491 for (i--; i >=0; i--) { 2490 for (i--; i >=0; i--) {
2492 chip = &cfi->chips[i]; 2491 chip = &cfi->chips[i];
2493 2492
2494 spin_lock(chip->mutex); 2493 mutex_lock(&chip->mutex);
2495 2494
2496 if (chip->state == FL_PM_SUSPENDED) { 2495 if (chip->state == FL_PM_SUSPENDED) {
2497 /* No need to force it into a known state here, 2496 /* No need to force it into a known state here,
@@ -2501,7 +2500,7 @@ static int cfi_intelext_suspend(struct mtd_info *mtd)
2501 chip->oldstate = FL_READY; 2500 chip->oldstate = FL_READY;
2502 wake_up(&chip->wq); 2501 wake_up(&chip->wq);
2503 } 2502 }
2504 spin_unlock(chip->mutex); 2503 mutex_unlock(&chip->mutex);
2505 } 2504 }
2506 } 2505 }
2507 2506
@@ -2542,7 +2541,7 @@ static void cfi_intelext_resume(struct mtd_info *mtd)
2542 2541
2543 chip = &cfi->chips[i]; 2542 chip = &cfi->chips[i];
2544 2543
2545 spin_lock(chip->mutex); 2544 mutex_lock(&chip->mutex);
2546 2545
2547 /* Go to known state. Chip may have been power cycled */ 2546 /* Go to known state. Chip may have been power cycled */
2548 if (chip->state == FL_PM_SUSPENDED) { 2547 if (chip->state == FL_PM_SUSPENDED) {
@@ -2551,7 +2550,7 @@ static void cfi_intelext_resume(struct mtd_info *mtd)
2551 wake_up(&chip->wq); 2550 wake_up(&chip->wq);
2552 } 2551 }
2553 2552
2554 spin_unlock(chip->mutex); 2553 mutex_unlock(&chip->mutex);
2555 } 2554 }
2556 2555
2557 if ((mtd->flags & MTD_POWERUP_LOCK) 2556 if ((mtd->flags & MTD_POWERUP_LOCK)
@@ -2571,14 +2570,14 @@ static int cfi_intelext_reset(struct mtd_info *mtd)
2571 /* force the completion of any ongoing operation 2570 /* force the completion of any ongoing operation
2572 and switch to array mode so any bootloader in 2571 and switch to array mode so any bootloader in
2573 flash is accessible for soft reboot. */ 2572 flash is accessible for soft reboot. */
2574 spin_lock(chip->mutex); 2573 mutex_lock(&chip->mutex);
2575 ret = get_chip(map, chip, chip->start, FL_SHUTDOWN); 2574 ret = get_chip(map, chip, chip->start, FL_SHUTDOWN);
2576 if (!ret) { 2575 if (!ret) {
2577 map_write(map, CMD(0xff), chip->start); 2576 map_write(map, CMD(0xff), chip->start);
2578 chip->state = FL_SHUTDOWN; 2577 chip->state = FL_SHUTDOWN;
2579 put_chip(map, chip, chip->start); 2578 put_chip(map, chip, chip->start);
2580 } 2579 }
2581 spin_unlock(chip->mutex); 2580 mutex_unlock(&chip->mutex);
2582 } 2581 }
2583 2582
2584 return 0; 2583 return 0;
diff --git a/drivers/mtd/chips/cfi_cmdset_0002.c b/drivers/mtd/chips/cfi_cmdset_0002.c
index ea2a7f66ddf9..c93e47d21ce0 100644
--- a/drivers/mtd/chips/cfi_cmdset_0002.c
+++ b/drivers/mtd/chips/cfi_cmdset_0002.c
@@ -565,9 +565,9 @@ static int get_chip(struct map_info *map, struct flchip *chip, unsigned long adr
565 printk(KERN_ERR "Waiting for chip to be ready timed out.\n"); 565 printk(KERN_ERR "Waiting for chip to be ready timed out.\n");
566 return -EIO; 566 return -EIO;
567 } 567 }
568 spin_unlock(chip->mutex); 568 mutex_unlock(&chip->mutex);
569 cfi_udelay(1); 569 cfi_udelay(1);
570 spin_lock(chip->mutex); 570 mutex_lock(&chip->mutex);
571 /* Someone else might have been playing with it. */ 571 /* Someone else might have been playing with it. */
572 goto retry; 572 goto retry;
573 } 573 }
@@ -611,9 +611,9 @@ static int get_chip(struct map_info *map, struct flchip *chip, unsigned long adr
611 return -EIO; 611 return -EIO;
612 } 612 }
613 613
614 spin_unlock(chip->mutex); 614 mutex_unlock(&chip->mutex);
615 cfi_udelay(1); 615 cfi_udelay(1);
616 spin_lock(chip->mutex); 616 mutex_lock(&chip->mutex);
617 /* Nobody will touch it while it's in state FL_ERASE_SUSPENDING. 617 /* Nobody will touch it while it's in state FL_ERASE_SUSPENDING.
618 So we can just loop here. */ 618 So we can just loop here. */
619 } 619 }
@@ -637,10 +637,10 @@ static int get_chip(struct map_info *map, struct flchip *chip, unsigned long adr
637 sleep: 637 sleep:
638 set_current_state(TASK_UNINTERRUPTIBLE); 638 set_current_state(TASK_UNINTERRUPTIBLE);
639 add_wait_queue(&chip->wq, &wait); 639 add_wait_queue(&chip->wq, &wait);
640 spin_unlock(chip->mutex); 640 mutex_unlock(&chip->mutex);
641 schedule(); 641 schedule();
642 remove_wait_queue(&chip->wq, &wait); 642 remove_wait_queue(&chip->wq, &wait);
643 spin_lock(chip->mutex); 643 mutex_lock(&chip->mutex);
644 goto resettime; 644 goto resettime;
645 } 645 }
646} 646}
@@ -772,7 +772,7 @@ static void __xipram xip_udelay(struct map_info *map, struct flchip *chip,
772 (void) map_read(map, adr); 772 (void) map_read(map, adr);
773 xip_iprefetch(); 773 xip_iprefetch();
774 local_irq_enable(); 774 local_irq_enable();
775 spin_unlock(chip->mutex); 775 mutex_unlock(&chip->mutex);
776 xip_iprefetch(); 776 xip_iprefetch();
777 cond_resched(); 777 cond_resched();
778 778
@@ -782,15 +782,15 @@ static void __xipram xip_udelay(struct map_info *map, struct flchip *chip,
782 * a suspended erase state. If so let's wait 782 * a suspended erase state. If so let's wait
783 * until it's done. 783 * until it's done.
784 */ 784 */
785 spin_lock(chip->mutex); 785 mutex_lock(&chip->mutex);
786 while (chip->state != FL_XIP_WHILE_ERASING) { 786 while (chip->state != FL_XIP_WHILE_ERASING) {
787 DECLARE_WAITQUEUE(wait, current); 787 DECLARE_WAITQUEUE(wait, current);
788 set_current_state(TASK_UNINTERRUPTIBLE); 788 set_current_state(TASK_UNINTERRUPTIBLE);
789 add_wait_queue(&chip->wq, &wait); 789 add_wait_queue(&chip->wq, &wait);
790 spin_unlock(chip->mutex); 790 mutex_unlock(&chip->mutex);
791 schedule(); 791 schedule();
792 remove_wait_queue(&chip->wq, &wait); 792 remove_wait_queue(&chip->wq, &wait);
793 spin_lock(chip->mutex); 793 mutex_lock(&chip->mutex);
794 } 794 }
795 /* Disallow XIP again */ 795 /* Disallow XIP again */
796 local_irq_disable(); 796 local_irq_disable();
@@ -852,17 +852,17 @@ static void __xipram xip_udelay(struct map_info *map, struct flchip *chip,
852 852
853#define UDELAY(map, chip, adr, usec) \ 853#define UDELAY(map, chip, adr, usec) \
854do { \ 854do { \
855 spin_unlock(chip->mutex); \ 855 mutex_unlock(&chip->mutex); \
856 cfi_udelay(usec); \ 856 cfi_udelay(usec); \
857 spin_lock(chip->mutex); \ 857 mutex_lock(&chip->mutex); \
858} while (0) 858} while (0)
859 859
860#define INVALIDATE_CACHE_UDELAY(map, chip, adr, len, usec) \ 860#define INVALIDATE_CACHE_UDELAY(map, chip, adr, len, usec) \
861do { \ 861do { \
862 spin_unlock(chip->mutex); \ 862 mutex_unlock(&chip->mutex); \
863 INVALIDATE_CACHED_RANGE(map, adr, len); \ 863 INVALIDATE_CACHED_RANGE(map, adr, len); \
864 cfi_udelay(usec); \ 864 cfi_udelay(usec); \
865 spin_lock(chip->mutex); \ 865 mutex_lock(&chip->mutex); \
866} while (0) 866} while (0)
867 867
868#endif 868#endif
@@ -878,10 +878,10 @@ static inline int do_read_onechip(struct map_info *map, struct flchip *chip, lof
878 /* Ensure cmd read/writes are aligned. */ 878 /* Ensure cmd read/writes are aligned. */
879 cmd_addr = adr & ~(map_bankwidth(map)-1); 879 cmd_addr = adr & ~(map_bankwidth(map)-1);
880 880
881 spin_lock(chip->mutex); 881 mutex_lock(&chip->mutex);
882 ret = get_chip(map, chip, cmd_addr, FL_READY); 882 ret = get_chip(map, chip, cmd_addr, FL_READY);
883 if (ret) { 883 if (ret) {
884 spin_unlock(chip->mutex); 884 mutex_unlock(&chip->mutex);
885 return ret; 885 return ret;
886 } 886 }
887 887
@@ -894,7 +894,7 @@ static inline int do_read_onechip(struct map_info *map, struct flchip *chip, lof
894 894
895 put_chip(map, chip, cmd_addr); 895 put_chip(map, chip, cmd_addr);
896 896
897 spin_unlock(chip->mutex); 897 mutex_unlock(&chip->mutex);
898 return 0; 898 return 0;
899} 899}
900 900
@@ -948,7 +948,7 @@ static inline int do_read_secsi_onechip(struct map_info *map, struct flchip *chi
948 struct cfi_private *cfi = map->fldrv_priv; 948 struct cfi_private *cfi = map->fldrv_priv;
949 949
950 retry: 950 retry:
951 spin_lock(chip->mutex); 951 mutex_lock(&chip->mutex);
952 952
953 if (chip->state != FL_READY){ 953 if (chip->state != FL_READY){
954#if 0 954#if 0
@@ -957,7 +957,7 @@ static inline int do_read_secsi_onechip(struct map_info *map, struct flchip *chi
957 set_current_state(TASK_UNINTERRUPTIBLE); 957 set_current_state(TASK_UNINTERRUPTIBLE);
958 add_wait_queue(&chip->wq, &wait); 958 add_wait_queue(&chip->wq, &wait);
959 959
960 spin_unlock(chip->mutex); 960 mutex_unlock(&chip->mutex);
961 961
962 schedule(); 962 schedule();
963 remove_wait_queue(&chip->wq, &wait); 963 remove_wait_queue(&chip->wq, &wait);
@@ -986,7 +986,7 @@ static inline int do_read_secsi_onechip(struct map_info *map, struct flchip *chi
986 cfi_send_gen_cmd(0x00, cfi->addr_unlock1, chip->start, map, cfi, cfi->device_type, NULL); 986 cfi_send_gen_cmd(0x00, cfi->addr_unlock1, chip->start, map, cfi, cfi->device_type, NULL);
987 987
988 wake_up(&chip->wq); 988 wake_up(&chip->wq);
989 spin_unlock(chip->mutex); 989 mutex_unlock(&chip->mutex);
990 990
991 return 0; 991 return 0;
992} 992}
@@ -1055,10 +1055,10 @@ static int __xipram do_write_oneword(struct map_info *map, struct flchip *chip,
1055 1055
1056 adr += chip->start; 1056 adr += chip->start;
1057 1057
1058 spin_lock(chip->mutex); 1058 mutex_lock(&chip->mutex);
1059 ret = get_chip(map, chip, adr, FL_WRITING); 1059 ret = get_chip(map, chip, adr, FL_WRITING);
1060 if (ret) { 1060 if (ret) {
1061 spin_unlock(chip->mutex); 1061 mutex_unlock(&chip->mutex);
1062 return ret; 1062 return ret;
1063 } 1063 }
1064 1064
@@ -1101,11 +1101,11 @@ static int __xipram do_write_oneword(struct map_info *map, struct flchip *chip,
1101 1101
1102 set_current_state(TASK_UNINTERRUPTIBLE); 1102 set_current_state(TASK_UNINTERRUPTIBLE);
1103 add_wait_queue(&chip->wq, &wait); 1103 add_wait_queue(&chip->wq, &wait);
1104 spin_unlock(chip->mutex); 1104 mutex_unlock(&chip->mutex);
1105 schedule(); 1105 schedule();
1106 remove_wait_queue(&chip->wq, &wait); 1106 remove_wait_queue(&chip->wq, &wait);
1107 timeo = jiffies + (HZ / 2); /* FIXME */ 1107 timeo = jiffies + (HZ / 2); /* FIXME */
1108 spin_lock(chip->mutex); 1108 mutex_lock(&chip->mutex);
1109 continue; 1109 continue;
1110 } 1110 }
1111 1111
@@ -1137,7 +1137,7 @@ static int __xipram do_write_oneword(struct map_info *map, struct flchip *chip,
1137 op_done: 1137 op_done:
1138 chip->state = FL_READY; 1138 chip->state = FL_READY;
1139 put_chip(map, chip, adr); 1139 put_chip(map, chip, adr);
1140 spin_unlock(chip->mutex); 1140 mutex_unlock(&chip->mutex);
1141 1141
1142 return ret; 1142 return ret;
1143} 1143}
@@ -1169,7 +1169,7 @@ static int cfi_amdstd_write_words(struct mtd_info *mtd, loff_t to, size_t len,
1169 map_word tmp_buf; 1169 map_word tmp_buf;
1170 1170
1171 retry: 1171 retry:
1172 spin_lock(cfi->chips[chipnum].mutex); 1172 mutex_lock(&cfi->chips[chipnum].mutex);
1173 1173
1174 if (cfi->chips[chipnum].state != FL_READY) { 1174 if (cfi->chips[chipnum].state != FL_READY) {
1175#if 0 1175#if 0
@@ -1178,7 +1178,7 @@ static int cfi_amdstd_write_words(struct mtd_info *mtd, loff_t to, size_t len,
1178 set_current_state(TASK_UNINTERRUPTIBLE); 1178 set_current_state(TASK_UNINTERRUPTIBLE);
1179 add_wait_queue(&cfi->chips[chipnum].wq, &wait); 1179 add_wait_queue(&cfi->chips[chipnum].wq, &wait);
1180 1180
1181 spin_unlock(cfi->chips[chipnum].mutex); 1181 mutex_unlock(&cfi->chips[chipnum].mutex);
1182 1182
1183 schedule(); 1183 schedule();
1184 remove_wait_queue(&cfi->chips[chipnum].wq, &wait); 1184 remove_wait_queue(&cfi->chips[chipnum].wq, &wait);
@@ -1192,7 +1192,7 @@ static int cfi_amdstd_write_words(struct mtd_info *mtd, loff_t to, size_t len,
1192 /* Load 'tmp_buf' with old contents of flash */ 1192 /* Load 'tmp_buf' with old contents of flash */
1193 tmp_buf = map_read(map, bus_ofs+chipstart); 1193 tmp_buf = map_read(map, bus_ofs+chipstart);
1194 1194
1195 spin_unlock(cfi->chips[chipnum].mutex); 1195 mutex_unlock(&cfi->chips[chipnum].mutex);
1196 1196
1197 /* Number of bytes to copy from buffer */ 1197 /* Number of bytes to copy from buffer */
1198 n = min_t(int, len, map_bankwidth(map)-i); 1198 n = min_t(int, len, map_bankwidth(map)-i);
@@ -1247,7 +1247,7 @@ static int cfi_amdstd_write_words(struct mtd_info *mtd, loff_t to, size_t len,
1247 map_word tmp_buf; 1247 map_word tmp_buf;
1248 1248
1249 retry1: 1249 retry1:
1250 spin_lock(cfi->chips[chipnum].mutex); 1250 mutex_lock(&cfi->chips[chipnum].mutex);
1251 1251
1252 if (cfi->chips[chipnum].state != FL_READY) { 1252 if (cfi->chips[chipnum].state != FL_READY) {
1253#if 0 1253#if 0
@@ -1256,7 +1256,7 @@ static int cfi_amdstd_write_words(struct mtd_info *mtd, loff_t to, size_t len,
1256 set_current_state(TASK_UNINTERRUPTIBLE); 1256 set_current_state(TASK_UNINTERRUPTIBLE);
1257 add_wait_queue(&cfi->chips[chipnum].wq, &wait); 1257 add_wait_queue(&cfi->chips[chipnum].wq, &wait);
1258 1258
1259 spin_unlock(cfi->chips[chipnum].mutex); 1259 mutex_unlock(&cfi->chips[chipnum].mutex);
1260 1260
1261 schedule(); 1261 schedule();
1262 remove_wait_queue(&cfi->chips[chipnum].wq, &wait); 1262 remove_wait_queue(&cfi->chips[chipnum].wq, &wait);
@@ -1269,7 +1269,7 @@ static int cfi_amdstd_write_words(struct mtd_info *mtd, loff_t to, size_t len,
1269 1269
1270 tmp_buf = map_read(map, ofs + chipstart); 1270 tmp_buf = map_read(map, ofs + chipstart);
1271 1271
1272 spin_unlock(cfi->chips[chipnum].mutex); 1272 mutex_unlock(&cfi->chips[chipnum].mutex);
1273 1273
1274 tmp_buf = map_word_load_partial(map, tmp_buf, buf, 0, len); 1274 tmp_buf = map_word_load_partial(map, tmp_buf, buf, 0, len);
1275 1275
@@ -1304,10 +1304,10 @@ static int __xipram do_write_buffer(struct map_info *map, struct flchip *chip,
1304 adr += chip->start; 1304 adr += chip->start;
1305 cmd_adr = adr; 1305 cmd_adr = adr;
1306 1306
1307 spin_lock(chip->mutex); 1307 mutex_lock(&chip->mutex);
1308 ret = get_chip(map, chip, adr, FL_WRITING); 1308 ret = get_chip(map, chip, adr, FL_WRITING);
1309 if (ret) { 1309 if (ret) {
1310 spin_unlock(chip->mutex); 1310 mutex_unlock(&chip->mutex);
1311 return ret; 1311 return ret;
1312 } 1312 }
1313 1313
@@ -1362,11 +1362,11 @@ static int __xipram do_write_buffer(struct map_info *map, struct flchip *chip,
1362 1362
1363 set_current_state(TASK_UNINTERRUPTIBLE); 1363 set_current_state(TASK_UNINTERRUPTIBLE);
1364 add_wait_queue(&chip->wq, &wait); 1364 add_wait_queue(&chip->wq, &wait);
1365 spin_unlock(chip->mutex); 1365 mutex_unlock(&chip->mutex);
1366 schedule(); 1366 schedule();
1367 remove_wait_queue(&chip->wq, &wait); 1367 remove_wait_queue(&chip->wq, &wait);
1368 timeo = jiffies + (HZ / 2); /* FIXME */ 1368 timeo = jiffies + (HZ / 2); /* FIXME */
1369 spin_lock(chip->mutex); 1369 mutex_lock(&chip->mutex);
1370 continue; 1370 continue;
1371 } 1371 }
1372 1372
@@ -1394,7 +1394,7 @@ static int __xipram do_write_buffer(struct map_info *map, struct flchip *chip,
1394 op_done: 1394 op_done:
1395 chip->state = FL_READY; 1395 chip->state = FL_READY;
1396 put_chip(map, chip, adr); 1396 put_chip(map, chip, adr);
1397 spin_unlock(chip->mutex); 1397 mutex_unlock(&chip->mutex);
1398 1398
1399 return ret; 1399 return ret;
1400} 1400}
@@ -1494,10 +1494,10 @@ static int __xipram do_erase_chip(struct map_info *map, struct flchip *chip)
1494 1494
1495 adr = cfi->addr_unlock1; 1495 adr = cfi->addr_unlock1;
1496 1496
1497 spin_lock(chip->mutex); 1497 mutex_lock(&chip->mutex);
1498 ret = get_chip(map, chip, adr, FL_WRITING); 1498 ret = get_chip(map, chip, adr, FL_WRITING);
1499 if (ret) { 1499 if (ret) {
1500 spin_unlock(chip->mutex); 1500 mutex_unlock(&chip->mutex);
1501 return ret; 1501 return ret;
1502 } 1502 }
1503 1503
@@ -1530,10 +1530,10 @@ static int __xipram do_erase_chip(struct map_info *map, struct flchip *chip)
1530 /* Someone's suspended the erase. Sleep */ 1530 /* Someone's suspended the erase. Sleep */
1531 set_current_state(TASK_UNINTERRUPTIBLE); 1531 set_current_state(TASK_UNINTERRUPTIBLE);
1532 add_wait_queue(&chip->wq, &wait); 1532 add_wait_queue(&chip->wq, &wait);
1533 spin_unlock(chip->mutex); 1533 mutex_unlock(&chip->mutex);
1534 schedule(); 1534 schedule();
1535 remove_wait_queue(&chip->wq, &wait); 1535 remove_wait_queue(&chip->wq, &wait);
1536 spin_lock(chip->mutex); 1536 mutex_lock(&chip->mutex);
1537 continue; 1537 continue;
1538 } 1538 }
1539 if (chip->erase_suspended) { 1539 if (chip->erase_suspended) {
@@ -1567,7 +1567,7 @@ static int __xipram do_erase_chip(struct map_info *map, struct flchip *chip)
1567 chip->state = FL_READY; 1567 chip->state = FL_READY;
1568 xip_enable(map, chip, adr); 1568 xip_enable(map, chip, adr);
1569 put_chip(map, chip, adr); 1569 put_chip(map, chip, adr);
1570 spin_unlock(chip->mutex); 1570 mutex_unlock(&chip->mutex);
1571 1571
1572 return ret; 1572 return ret;
1573} 1573}
@@ -1582,10 +1582,10 @@ static int __xipram do_erase_oneblock(struct map_info *map, struct flchip *chip,
1582 1582
1583 adr += chip->start; 1583 adr += chip->start;
1584 1584
1585 spin_lock(chip->mutex); 1585 mutex_lock(&chip->mutex);
1586 ret = get_chip(map, chip, adr, FL_ERASING); 1586 ret = get_chip(map, chip, adr, FL_ERASING);
1587 if (ret) { 1587 if (ret) {
1588 spin_unlock(chip->mutex); 1588 mutex_unlock(&chip->mutex);
1589 return ret; 1589 return ret;
1590 } 1590 }
1591 1591
@@ -1618,10 +1618,10 @@ static int __xipram do_erase_oneblock(struct map_info *map, struct flchip *chip,
1618 /* Someone's suspended the erase. Sleep */ 1618 /* Someone's suspended the erase. Sleep */
1619 set_current_state(TASK_UNINTERRUPTIBLE); 1619 set_current_state(TASK_UNINTERRUPTIBLE);
1620 add_wait_queue(&chip->wq, &wait); 1620 add_wait_queue(&chip->wq, &wait);
1621 spin_unlock(chip->mutex); 1621 mutex_unlock(&chip->mutex);
1622 schedule(); 1622 schedule();
1623 remove_wait_queue(&chip->wq, &wait); 1623 remove_wait_queue(&chip->wq, &wait);
1624 spin_lock(chip->mutex); 1624 mutex_lock(&chip->mutex);
1625 continue; 1625 continue;
1626 } 1626 }
1627 if (chip->erase_suspended) { 1627 if (chip->erase_suspended) {
@@ -1657,7 +1657,7 @@ static int __xipram do_erase_oneblock(struct map_info *map, struct flchip *chip,
1657 1657
1658 chip->state = FL_READY; 1658 chip->state = FL_READY;
1659 put_chip(map, chip, adr); 1659 put_chip(map, chip, adr);
1660 spin_unlock(chip->mutex); 1660 mutex_unlock(&chip->mutex);
1661 return ret; 1661 return ret;
1662} 1662}
1663 1663
@@ -1709,7 +1709,7 @@ static int do_atmel_lock(struct map_info *map, struct flchip *chip,
1709 struct cfi_private *cfi = map->fldrv_priv; 1709 struct cfi_private *cfi = map->fldrv_priv;
1710 int ret; 1710 int ret;
1711 1711
1712 spin_lock(chip->mutex); 1712 mutex_lock(&chip->mutex);
1713 ret = get_chip(map, chip, adr + chip->start, FL_LOCKING); 1713 ret = get_chip(map, chip, adr + chip->start, FL_LOCKING);
1714 if (ret) 1714 if (ret)
1715 goto out_unlock; 1715 goto out_unlock;
@@ -1735,7 +1735,7 @@ static int do_atmel_lock(struct map_info *map, struct flchip *chip,
1735 ret = 0; 1735 ret = 0;
1736 1736
1737out_unlock: 1737out_unlock:
1738 spin_unlock(chip->mutex); 1738 mutex_unlock(&chip->mutex);
1739 return ret; 1739 return ret;
1740} 1740}
1741 1741
@@ -1745,7 +1745,7 @@ static int do_atmel_unlock(struct map_info *map, struct flchip *chip,
1745 struct cfi_private *cfi = map->fldrv_priv; 1745 struct cfi_private *cfi = map->fldrv_priv;
1746 int ret; 1746 int ret;
1747 1747
1748 spin_lock(chip->mutex); 1748 mutex_lock(&chip->mutex);
1749 ret = get_chip(map, chip, adr + chip->start, FL_UNLOCKING); 1749 ret = get_chip(map, chip, adr + chip->start, FL_UNLOCKING);
1750 if (ret) 1750 if (ret)
1751 goto out_unlock; 1751 goto out_unlock;
@@ -1763,7 +1763,7 @@ static int do_atmel_unlock(struct map_info *map, struct flchip *chip,
1763 ret = 0; 1763 ret = 0;
1764 1764
1765out_unlock: 1765out_unlock:
1766 spin_unlock(chip->mutex); 1766 mutex_unlock(&chip->mutex);
1767 return ret; 1767 return ret;
1768} 1768}
1769 1769
@@ -1791,7 +1791,7 @@ static void cfi_amdstd_sync (struct mtd_info *mtd)
1791 chip = &cfi->chips[i]; 1791 chip = &cfi->chips[i];
1792 1792
1793 retry: 1793 retry:
1794 spin_lock(chip->mutex); 1794 mutex_lock(&chip->mutex);
1795 1795
1796 switch(chip->state) { 1796 switch(chip->state) {
1797 case FL_READY: 1797 case FL_READY:
@@ -1805,7 +1805,7 @@ static void cfi_amdstd_sync (struct mtd_info *mtd)
1805 * with the chip now anyway. 1805 * with the chip now anyway.
1806 */ 1806 */
1807 case FL_SYNCING: 1807 case FL_SYNCING:
1808 spin_unlock(chip->mutex); 1808 mutex_unlock(&chip->mutex);
1809 break; 1809 break;
1810 1810
1811 default: 1811 default:
@@ -1813,7 +1813,7 @@ static void cfi_amdstd_sync (struct mtd_info *mtd)
1813 set_current_state(TASK_UNINTERRUPTIBLE); 1813 set_current_state(TASK_UNINTERRUPTIBLE);
1814 add_wait_queue(&chip->wq, &wait); 1814 add_wait_queue(&chip->wq, &wait);
1815 1815
1816 spin_unlock(chip->mutex); 1816 mutex_unlock(&chip->mutex);
1817 1817
1818 schedule(); 1818 schedule();
1819 1819
@@ -1828,13 +1828,13 @@ static void cfi_amdstd_sync (struct mtd_info *mtd)
1828 for (i--; i >=0; i--) { 1828 for (i--; i >=0; i--) {
1829 chip = &cfi->chips[i]; 1829 chip = &cfi->chips[i];
1830 1830
1831 spin_lock(chip->mutex); 1831 mutex_lock(&chip->mutex);
1832 1832
1833 if (chip->state == FL_SYNCING) { 1833 if (chip->state == FL_SYNCING) {
1834 chip->state = chip->oldstate; 1834 chip->state = chip->oldstate;
1835 wake_up(&chip->wq); 1835 wake_up(&chip->wq);
1836 } 1836 }
1837 spin_unlock(chip->mutex); 1837 mutex_unlock(&chip->mutex);
1838 } 1838 }
1839} 1839}
1840 1840
@@ -1850,7 +1850,7 @@ static int cfi_amdstd_suspend(struct mtd_info *mtd)
1850 for (i=0; !ret && i<cfi->numchips; i++) { 1850 for (i=0; !ret && i<cfi->numchips; i++) {
1851 chip = &cfi->chips[i]; 1851 chip = &cfi->chips[i];
1852 1852
1853 spin_lock(chip->mutex); 1853 mutex_lock(&chip->mutex);
1854 1854
1855 switch(chip->state) { 1855 switch(chip->state) {
1856 case FL_READY: 1856 case FL_READY:
@@ -1870,7 +1870,7 @@ static int cfi_amdstd_suspend(struct mtd_info *mtd)
1870 ret = -EAGAIN; 1870 ret = -EAGAIN;
1871 break; 1871 break;
1872 } 1872 }
1873 spin_unlock(chip->mutex); 1873 mutex_unlock(&chip->mutex);
1874 } 1874 }
1875 1875
1876 /* Unlock the chips again */ 1876 /* Unlock the chips again */
@@ -1879,13 +1879,13 @@ static int cfi_amdstd_suspend(struct mtd_info *mtd)
1879 for (i--; i >=0; i--) { 1879 for (i--; i >=0; i--) {
1880 chip = &cfi->chips[i]; 1880 chip = &cfi->chips[i];
1881 1881
1882 spin_lock(chip->mutex); 1882 mutex_lock(&chip->mutex);
1883 1883
1884 if (chip->state == FL_PM_SUSPENDED) { 1884 if (chip->state == FL_PM_SUSPENDED) {
1885 chip->state = chip->oldstate; 1885 chip->state = chip->oldstate;
1886 wake_up(&chip->wq); 1886 wake_up(&chip->wq);
1887 } 1887 }
1888 spin_unlock(chip->mutex); 1888 mutex_unlock(&chip->mutex);
1889 } 1889 }
1890 } 1890 }
1891 1891
@@ -1904,7 +1904,7 @@ static void cfi_amdstd_resume(struct mtd_info *mtd)
1904 1904
1905 chip = &cfi->chips[i]; 1905 chip = &cfi->chips[i];
1906 1906
1907 spin_lock(chip->mutex); 1907 mutex_lock(&chip->mutex);
1908 1908
1909 if (chip->state == FL_PM_SUSPENDED) { 1909 if (chip->state == FL_PM_SUSPENDED) {
1910 chip->state = FL_READY; 1910 chip->state = FL_READY;
@@ -1914,7 +1914,7 @@ static void cfi_amdstd_resume(struct mtd_info *mtd)
1914 else 1914 else
1915 printk(KERN_ERR "Argh. Chip not in PM_SUSPENDED state upon resume()\n"); 1915 printk(KERN_ERR "Argh. Chip not in PM_SUSPENDED state upon resume()\n");
1916 1916
1917 spin_unlock(chip->mutex); 1917 mutex_unlock(&chip->mutex);
1918 } 1918 }
1919} 1919}
1920 1920
diff --git a/drivers/mtd/chips/cfi_cmdset_0020.c b/drivers/mtd/chips/cfi_cmdset_0020.c
index 0667a671525d..e54e8c169d76 100644
--- a/drivers/mtd/chips/cfi_cmdset_0020.c
+++ b/drivers/mtd/chips/cfi_cmdset_0020.c
@@ -265,7 +265,7 @@ static inline int do_read_onechip(struct map_info *map, struct flchip *chip, lof
265 265
266 timeo = jiffies + HZ; 266 timeo = jiffies + HZ;
267 retry: 267 retry:
268 spin_lock_bh(chip->mutex); 268 mutex_lock(&chip->mutex);
269 269
270 /* Check that the chip's ready to talk to us. 270 /* Check that the chip's ready to talk to us.
271 * If it's in FL_ERASING state, suspend it and make it talk now. 271 * If it's in FL_ERASING state, suspend it and make it talk now.
@@ -296,15 +296,15 @@ static inline int do_read_onechip(struct map_info *map, struct flchip *chip, lof
296 /* make sure we're in 'read status' mode */ 296 /* make sure we're in 'read status' mode */
297 map_write(map, CMD(0x70), cmd_addr); 297 map_write(map, CMD(0x70), cmd_addr);
298 chip->state = FL_ERASING; 298 chip->state = FL_ERASING;
299 spin_unlock_bh(chip->mutex); 299 mutex_unlock(&chip->mutex);
300 printk(KERN_ERR "Chip not ready after erase " 300 printk(KERN_ERR "Chip not ready after erase "
301 "suspended: status = 0x%lx\n", status.x[0]); 301 "suspended: status = 0x%lx\n", status.x[0]);
302 return -EIO; 302 return -EIO;
303 } 303 }
304 304
305 spin_unlock_bh(chip->mutex); 305 mutex_unlock(&chip->mutex);
306 cfi_udelay(1); 306 cfi_udelay(1);
307 spin_lock_bh(chip->mutex); 307 mutex_lock(&chip->mutex);
308 } 308 }
309 309
310 suspended = 1; 310 suspended = 1;
@@ -335,13 +335,13 @@ static inline int do_read_onechip(struct map_info *map, struct flchip *chip, lof
335 335
336 /* Urgh. Chip not yet ready to talk to us. */ 336 /* Urgh. Chip not yet ready to talk to us. */
337 if (time_after(jiffies, timeo)) { 337 if (time_after(jiffies, timeo)) {
338 spin_unlock_bh(chip->mutex); 338 mutex_unlock(&chip->mutex);
339 printk(KERN_ERR "waiting for chip to be ready timed out in read. WSM status = %lx\n", status.x[0]); 339 printk(KERN_ERR "waiting for chip to be ready timed out in read. WSM status = %lx\n", status.x[0]);
340 return -EIO; 340 return -EIO;
341 } 341 }
342 342
343 /* Latency issues. Drop the lock, wait a while and retry */ 343 /* Latency issues. Drop the lock, wait a while and retry */
344 spin_unlock_bh(chip->mutex); 344 mutex_unlock(&chip->mutex);
345 cfi_udelay(1); 345 cfi_udelay(1);
346 goto retry; 346 goto retry;
347 347
@@ -351,7 +351,7 @@ static inline int do_read_onechip(struct map_info *map, struct flchip *chip, lof
351 someone changes the status */ 351 someone changes the status */
352 set_current_state(TASK_UNINTERRUPTIBLE); 352 set_current_state(TASK_UNINTERRUPTIBLE);
353 add_wait_queue(&chip->wq, &wait); 353 add_wait_queue(&chip->wq, &wait);
354 spin_unlock_bh(chip->mutex); 354 mutex_unlock(&chip->mutex);
355 schedule(); 355 schedule();
356 remove_wait_queue(&chip->wq, &wait); 356 remove_wait_queue(&chip->wq, &wait);
357 timeo = jiffies + HZ; 357 timeo = jiffies + HZ;
@@ -376,7 +376,7 @@ static inline int do_read_onechip(struct map_info *map, struct flchip *chip, lof
376 } 376 }
377 377
378 wake_up(&chip->wq); 378 wake_up(&chip->wq);
379 spin_unlock_bh(chip->mutex); 379 mutex_unlock(&chip->mutex);
380 return 0; 380 return 0;
381} 381}
382 382
@@ -445,7 +445,7 @@ static inline int do_write_buffer(struct map_info *map, struct flchip *chip,
445#ifdef DEBUG_CFI_FEATURES 445#ifdef DEBUG_CFI_FEATURES
446 printk("%s: chip->state[%d]\n", __func__, chip->state); 446 printk("%s: chip->state[%d]\n", __func__, chip->state);
447#endif 447#endif
448 spin_lock_bh(chip->mutex); 448 mutex_lock(&chip->mutex);
449 449
450 /* Check that the chip's ready to talk to us. 450 /* Check that the chip's ready to talk to us.
451 * Later, we can actually think about interrupting it 451 * Later, we can actually think about interrupting it
@@ -470,14 +470,14 @@ static inline int do_write_buffer(struct map_info *map, struct flchip *chip,
470 break; 470 break;
471 /* Urgh. Chip not yet ready to talk to us. */ 471 /* Urgh. Chip not yet ready to talk to us. */
472 if (time_after(jiffies, timeo)) { 472 if (time_after(jiffies, timeo)) {
473 spin_unlock_bh(chip->mutex); 473 mutex_unlock(&chip->mutex);
474 printk(KERN_ERR "waiting for chip to be ready timed out in buffer write Xstatus = %lx, status = %lx\n", 474 printk(KERN_ERR "waiting for chip to be ready timed out in buffer write Xstatus = %lx, status = %lx\n",
475 status.x[0], map_read(map, cmd_adr).x[0]); 475 status.x[0], map_read(map, cmd_adr).x[0]);
476 return -EIO; 476 return -EIO;
477 } 477 }
478 478
479 /* Latency issues. Drop the lock, wait a while and retry */ 479 /* Latency issues. Drop the lock, wait a while and retry */
480 spin_unlock_bh(chip->mutex); 480 mutex_unlock(&chip->mutex);
481 cfi_udelay(1); 481 cfi_udelay(1);
482 goto retry; 482 goto retry;
483 483
@@ -486,7 +486,7 @@ static inline int do_write_buffer(struct map_info *map, struct flchip *chip,
486 someone changes the status */ 486 someone changes the status */
487 set_current_state(TASK_UNINTERRUPTIBLE); 487 set_current_state(TASK_UNINTERRUPTIBLE);
488 add_wait_queue(&chip->wq, &wait); 488 add_wait_queue(&chip->wq, &wait);
489 spin_unlock_bh(chip->mutex); 489 mutex_unlock(&chip->mutex);
490 schedule(); 490 schedule();
491 remove_wait_queue(&chip->wq, &wait); 491 remove_wait_queue(&chip->wq, &wait);
492 timeo = jiffies + HZ; 492 timeo = jiffies + HZ;
@@ -503,16 +503,16 @@ static inline int do_write_buffer(struct map_info *map, struct flchip *chip,
503 if (map_word_andequal(map, status, status_OK, status_OK)) 503 if (map_word_andequal(map, status, status_OK, status_OK))
504 break; 504 break;
505 505
506 spin_unlock_bh(chip->mutex); 506 mutex_unlock(&chip->mutex);
507 cfi_udelay(1); 507 cfi_udelay(1);
508 spin_lock_bh(chip->mutex); 508 mutex_lock(&chip->mutex);
509 509
510 if (++z > 100) { 510 if (++z > 100) {
511 /* Argh. Not ready for write to buffer */ 511 /* Argh. Not ready for write to buffer */
512 DISABLE_VPP(map); 512 DISABLE_VPP(map);
513 map_write(map, CMD(0x70), cmd_adr); 513 map_write(map, CMD(0x70), cmd_adr);
514 chip->state = FL_STATUS; 514 chip->state = FL_STATUS;
515 spin_unlock_bh(chip->mutex); 515 mutex_unlock(&chip->mutex);
516 printk(KERN_ERR "Chip not ready for buffer write. Xstatus = %lx\n", status.x[0]); 516 printk(KERN_ERR "Chip not ready for buffer write. Xstatus = %lx\n", status.x[0]);
517 return -EIO; 517 return -EIO;
518 } 518 }
@@ -532,9 +532,9 @@ static inline int do_write_buffer(struct map_info *map, struct flchip *chip,
532 map_write(map, CMD(0xd0), cmd_adr); 532 map_write(map, CMD(0xd0), cmd_adr);
533 chip->state = FL_WRITING; 533 chip->state = FL_WRITING;
534 534
535 spin_unlock_bh(chip->mutex); 535 mutex_unlock(&chip->mutex);
536 cfi_udelay(chip->buffer_write_time); 536 cfi_udelay(chip->buffer_write_time);
537 spin_lock_bh(chip->mutex); 537 mutex_lock(&chip->mutex);
538 538
539 timeo = jiffies + (HZ/2); 539 timeo = jiffies + (HZ/2);
540 z = 0; 540 z = 0;
@@ -543,11 +543,11 @@ static inline int do_write_buffer(struct map_info *map, struct flchip *chip,
543 /* Someone's suspended the write. Sleep */ 543 /* Someone's suspended the write. Sleep */
544 set_current_state(TASK_UNINTERRUPTIBLE); 544 set_current_state(TASK_UNINTERRUPTIBLE);
545 add_wait_queue(&chip->wq, &wait); 545 add_wait_queue(&chip->wq, &wait);
546 spin_unlock_bh(chip->mutex); 546 mutex_unlock(&chip->mutex);
547 schedule(); 547 schedule();
548 remove_wait_queue(&chip->wq, &wait); 548 remove_wait_queue(&chip->wq, &wait);
549 timeo = jiffies + (HZ / 2); /* FIXME */ 549 timeo = jiffies + (HZ / 2); /* FIXME */
550 spin_lock_bh(chip->mutex); 550 mutex_lock(&chip->mutex);
551 continue; 551 continue;
552 } 552 }
553 553
@@ -563,16 +563,16 @@ static inline int do_write_buffer(struct map_info *map, struct flchip *chip,
563 map_write(map, CMD(0x70), adr); 563 map_write(map, CMD(0x70), adr);
564 chip->state = FL_STATUS; 564 chip->state = FL_STATUS;
565 DISABLE_VPP(map); 565 DISABLE_VPP(map);
566 spin_unlock_bh(chip->mutex); 566 mutex_unlock(&chip->mutex);
567 printk(KERN_ERR "waiting for chip to be ready timed out in bufwrite\n"); 567 printk(KERN_ERR "waiting for chip to be ready timed out in bufwrite\n");
568 return -EIO; 568 return -EIO;
569 } 569 }
570 570
571 /* Latency issues. Drop the lock, wait a while and retry */ 571 /* Latency issues. Drop the lock, wait a while and retry */
572 spin_unlock_bh(chip->mutex); 572 mutex_unlock(&chip->mutex);
573 cfi_udelay(1); 573 cfi_udelay(1);
574 z++; 574 z++;
575 spin_lock_bh(chip->mutex); 575 mutex_lock(&chip->mutex);
576 } 576 }
577 if (!z) { 577 if (!z) {
578 chip->buffer_write_time--; 578 chip->buffer_write_time--;
@@ -596,11 +596,11 @@ static inline int do_write_buffer(struct map_info *map, struct flchip *chip,
596 /* put back into read status register mode */ 596 /* put back into read status register mode */
597 map_write(map, CMD(0x70), adr); 597 map_write(map, CMD(0x70), adr);
598 wake_up(&chip->wq); 598 wake_up(&chip->wq);
599 spin_unlock_bh(chip->mutex); 599 mutex_unlock(&chip->mutex);
600 return map_word_bitsset(map, status, CMD(0x02)) ? -EROFS : -EIO; 600 return map_word_bitsset(map, status, CMD(0x02)) ? -EROFS : -EIO;
601 } 601 }
602 wake_up(&chip->wq); 602 wake_up(&chip->wq);
603 spin_unlock_bh(chip->mutex); 603 mutex_unlock(&chip->mutex);
604 604
605 return 0; 605 return 0;
606} 606}
@@ -749,7 +749,7 @@ static inline int do_erase_oneblock(struct map_info *map, struct flchip *chip, u
749 749
750 timeo = jiffies + HZ; 750 timeo = jiffies + HZ;
751retry: 751retry:
752 spin_lock_bh(chip->mutex); 752 mutex_lock(&chip->mutex);
753 753
754 /* Check that the chip's ready to talk to us. */ 754 /* Check that the chip's ready to talk to us. */
755 switch (chip->state) { 755 switch (chip->state) {
@@ -766,13 +766,13 @@ retry:
766 766
767 /* Urgh. Chip not yet ready to talk to us. */ 767 /* Urgh. Chip not yet ready to talk to us. */
768 if (time_after(jiffies, timeo)) { 768 if (time_after(jiffies, timeo)) {
769 spin_unlock_bh(chip->mutex); 769 mutex_unlock(&chip->mutex);
770 printk(KERN_ERR "waiting for chip to be ready timed out in erase\n"); 770 printk(KERN_ERR "waiting for chip to be ready timed out in erase\n");
771 return -EIO; 771 return -EIO;
772 } 772 }
773 773
774 /* Latency issues. Drop the lock, wait a while and retry */ 774 /* Latency issues. Drop the lock, wait a while and retry */
775 spin_unlock_bh(chip->mutex); 775 mutex_unlock(&chip->mutex);
776 cfi_udelay(1); 776 cfi_udelay(1);
777 goto retry; 777 goto retry;
778 778
@@ -781,7 +781,7 @@ retry:
781 someone changes the status */ 781 someone changes the status */
782 set_current_state(TASK_UNINTERRUPTIBLE); 782 set_current_state(TASK_UNINTERRUPTIBLE);
783 add_wait_queue(&chip->wq, &wait); 783 add_wait_queue(&chip->wq, &wait);
784 spin_unlock_bh(chip->mutex); 784 mutex_unlock(&chip->mutex);
785 schedule(); 785 schedule();
786 remove_wait_queue(&chip->wq, &wait); 786 remove_wait_queue(&chip->wq, &wait);
787 timeo = jiffies + HZ; 787 timeo = jiffies + HZ;
@@ -797,9 +797,9 @@ retry:
797 map_write(map, CMD(0xD0), adr); 797 map_write(map, CMD(0xD0), adr);
798 chip->state = FL_ERASING; 798 chip->state = FL_ERASING;
799 799
800 spin_unlock_bh(chip->mutex); 800 mutex_unlock(&chip->mutex);
801 msleep(1000); 801 msleep(1000);
802 spin_lock_bh(chip->mutex); 802 mutex_lock(&chip->mutex);
803 803
804 /* FIXME. Use a timer to check this, and return immediately. */ 804 /* FIXME. Use a timer to check this, and return immediately. */
805 /* Once the state machine's known to be working I'll do that */ 805 /* Once the state machine's known to be working I'll do that */
@@ -810,11 +810,11 @@ retry:
810 /* Someone's suspended the erase. Sleep */ 810 /* Someone's suspended the erase. Sleep */
811 set_current_state(TASK_UNINTERRUPTIBLE); 811 set_current_state(TASK_UNINTERRUPTIBLE);
812 add_wait_queue(&chip->wq, &wait); 812 add_wait_queue(&chip->wq, &wait);
813 spin_unlock_bh(chip->mutex); 813 mutex_unlock(&chip->mutex);
814 schedule(); 814 schedule();
815 remove_wait_queue(&chip->wq, &wait); 815 remove_wait_queue(&chip->wq, &wait);
816 timeo = jiffies + (HZ*20); /* FIXME */ 816 timeo = jiffies + (HZ*20); /* FIXME */
817 spin_lock_bh(chip->mutex); 817 mutex_lock(&chip->mutex);
818 continue; 818 continue;
819 } 819 }
820 820
@@ -828,14 +828,14 @@ retry:
828 chip->state = FL_STATUS; 828 chip->state = FL_STATUS;
829 printk(KERN_ERR "waiting for erase to complete timed out. Xstatus = %lx, status = %lx.\n", status.x[0], map_read(map, adr).x[0]); 829 printk(KERN_ERR "waiting for erase to complete timed out. Xstatus = %lx, status = %lx.\n", status.x[0], map_read(map, adr).x[0]);
830 DISABLE_VPP(map); 830 DISABLE_VPP(map);
831 spin_unlock_bh(chip->mutex); 831 mutex_unlock(&chip->mutex);
832 return -EIO; 832 return -EIO;
833 } 833 }
834 834
835 /* Latency issues. Drop the lock, wait a while and retry */ 835 /* Latency issues. Drop the lock, wait a while and retry */
836 spin_unlock_bh(chip->mutex); 836 mutex_unlock(&chip->mutex);
837 cfi_udelay(1); 837 cfi_udelay(1);
838 spin_lock_bh(chip->mutex); 838 mutex_lock(&chip->mutex);
839 } 839 }
840 840
841 DISABLE_VPP(map); 841 DISABLE_VPP(map);
@@ -878,7 +878,7 @@ retry:
878 printk(KERN_DEBUG "Chip erase failed at 0x%08lx: status 0x%x. Retrying...\n", adr, chipstatus); 878 printk(KERN_DEBUG "Chip erase failed at 0x%08lx: status 0x%x. Retrying...\n", adr, chipstatus);
879 timeo = jiffies + HZ; 879 timeo = jiffies + HZ;
880 chip->state = FL_STATUS; 880 chip->state = FL_STATUS;
881 spin_unlock_bh(chip->mutex); 881 mutex_unlock(&chip->mutex);
882 goto retry; 882 goto retry;
883 } 883 }
884 printk(KERN_DEBUG "Chip erase failed at 0x%08lx: status 0x%x\n", adr, chipstatus); 884 printk(KERN_DEBUG "Chip erase failed at 0x%08lx: status 0x%x\n", adr, chipstatus);
@@ -887,7 +887,7 @@ retry:
887 } 887 }
888 888
889 wake_up(&chip->wq); 889 wake_up(&chip->wq);
890 spin_unlock_bh(chip->mutex); 890 mutex_unlock(&chip->mutex);
891 return ret; 891 return ret;
892} 892}
893 893
@@ -995,7 +995,7 @@ static void cfi_staa_sync (struct mtd_info *mtd)
995 chip = &cfi->chips[i]; 995 chip = &cfi->chips[i];
996 996
997 retry: 997 retry:
998 spin_lock_bh(chip->mutex); 998 mutex_lock(&chip->mutex);
999 999
1000 switch(chip->state) { 1000 switch(chip->state) {
1001 case FL_READY: 1001 case FL_READY:
@@ -1009,7 +1009,7 @@ static void cfi_staa_sync (struct mtd_info *mtd)
1009 * with the chip now anyway. 1009 * with the chip now anyway.
1010 */ 1010 */
1011 case FL_SYNCING: 1011 case FL_SYNCING:
1012 spin_unlock_bh(chip->mutex); 1012 mutex_unlock(&chip->mutex);
1013 break; 1013 break;
1014 1014
1015 default: 1015 default:
@@ -1017,7 +1017,7 @@ static void cfi_staa_sync (struct mtd_info *mtd)
1017 set_current_state(TASK_UNINTERRUPTIBLE); 1017 set_current_state(TASK_UNINTERRUPTIBLE);
1018 add_wait_queue(&chip->wq, &wait); 1018 add_wait_queue(&chip->wq, &wait);
1019 1019
1020 spin_unlock_bh(chip->mutex); 1020 mutex_unlock(&chip->mutex);
1021 schedule(); 1021 schedule();
1022 remove_wait_queue(&chip->wq, &wait); 1022 remove_wait_queue(&chip->wq, &wait);
1023 1023
@@ -1030,13 +1030,13 @@ static void cfi_staa_sync (struct mtd_info *mtd)
1030 for (i--; i >=0; i--) { 1030 for (i--; i >=0; i--) {
1031 chip = &cfi->chips[i]; 1031 chip = &cfi->chips[i];
1032 1032
1033 spin_lock_bh(chip->mutex); 1033 mutex_lock(&chip->mutex);
1034 1034
1035 if (chip->state == FL_SYNCING) { 1035 if (chip->state == FL_SYNCING) {
1036 chip->state = chip->oldstate; 1036 chip->state = chip->oldstate;
1037 wake_up(&chip->wq); 1037 wake_up(&chip->wq);
1038 } 1038 }
1039 spin_unlock_bh(chip->mutex); 1039 mutex_unlock(&chip->mutex);
1040 } 1040 }
1041} 1041}
1042 1042
@@ -1054,7 +1054,7 @@ static inline int do_lock_oneblock(struct map_info *map, struct flchip *chip, un
1054 1054
1055 timeo = jiffies + HZ; 1055 timeo = jiffies + HZ;
1056retry: 1056retry:
1057 spin_lock_bh(chip->mutex); 1057 mutex_lock(&chip->mutex);
1058 1058
1059 /* Check that the chip's ready to talk to us. */ 1059 /* Check that the chip's ready to talk to us. */
1060 switch (chip->state) { 1060 switch (chip->state) {
@@ -1071,13 +1071,13 @@ retry:
1071 1071
1072 /* Urgh. Chip not yet ready to talk to us. */ 1072 /* Urgh. Chip not yet ready to talk to us. */
1073 if (time_after(jiffies, timeo)) { 1073 if (time_after(jiffies, timeo)) {
1074 spin_unlock_bh(chip->mutex); 1074 mutex_unlock(&chip->mutex);
1075 printk(KERN_ERR "waiting for chip to be ready timed out in lock\n"); 1075 printk(KERN_ERR "waiting for chip to be ready timed out in lock\n");
1076 return -EIO; 1076 return -EIO;
1077 } 1077 }
1078 1078
1079 /* Latency issues. Drop the lock, wait a while and retry */ 1079 /* Latency issues. Drop the lock, wait a while and retry */
1080 spin_unlock_bh(chip->mutex); 1080 mutex_unlock(&chip->mutex);
1081 cfi_udelay(1); 1081 cfi_udelay(1);
1082 goto retry; 1082 goto retry;
1083 1083
@@ -1086,7 +1086,7 @@ retry:
1086 someone changes the status */ 1086 someone changes the status */
1087 set_current_state(TASK_UNINTERRUPTIBLE); 1087 set_current_state(TASK_UNINTERRUPTIBLE);
1088 add_wait_queue(&chip->wq, &wait); 1088 add_wait_queue(&chip->wq, &wait);
1089 spin_unlock_bh(chip->mutex); 1089 mutex_unlock(&chip->mutex);
1090 schedule(); 1090 schedule();
1091 remove_wait_queue(&chip->wq, &wait); 1091 remove_wait_queue(&chip->wq, &wait);
1092 timeo = jiffies + HZ; 1092 timeo = jiffies + HZ;
@@ -1098,9 +1098,9 @@ retry:
1098 map_write(map, CMD(0x01), adr); 1098 map_write(map, CMD(0x01), adr);
1099 chip->state = FL_LOCKING; 1099 chip->state = FL_LOCKING;
1100 1100
1101 spin_unlock_bh(chip->mutex); 1101 mutex_unlock(&chip->mutex);
1102 msleep(1000); 1102 msleep(1000);
1103 spin_lock_bh(chip->mutex); 1103 mutex_lock(&chip->mutex);
1104 1104
1105 /* FIXME. Use a timer to check this, and return immediately. */ 1105 /* FIXME. Use a timer to check this, and return immediately. */
1106 /* Once the state machine's known to be working I'll do that */ 1106 /* Once the state machine's known to be working I'll do that */
@@ -1118,21 +1118,21 @@ retry:
1118 chip->state = FL_STATUS; 1118 chip->state = FL_STATUS;
1119 printk(KERN_ERR "waiting for lock to complete timed out. Xstatus = %lx, status = %lx.\n", status.x[0], map_read(map, adr).x[0]); 1119 printk(KERN_ERR "waiting for lock to complete timed out. Xstatus = %lx, status = %lx.\n", status.x[0], map_read(map, adr).x[0]);
1120 DISABLE_VPP(map); 1120 DISABLE_VPP(map);
1121 spin_unlock_bh(chip->mutex); 1121 mutex_unlock(&chip->mutex);
1122 return -EIO; 1122 return -EIO;
1123 } 1123 }
1124 1124
1125 /* Latency issues. Drop the lock, wait a while and retry */ 1125 /* Latency issues. Drop the lock, wait a while and retry */
1126 spin_unlock_bh(chip->mutex); 1126 mutex_unlock(&chip->mutex);
1127 cfi_udelay(1); 1127 cfi_udelay(1);
1128 spin_lock_bh(chip->mutex); 1128 mutex_lock(&chip->mutex);
1129 } 1129 }
1130 1130
1131 /* Done and happy. */ 1131 /* Done and happy. */
1132 chip->state = FL_STATUS; 1132 chip->state = FL_STATUS;
1133 DISABLE_VPP(map); 1133 DISABLE_VPP(map);
1134 wake_up(&chip->wq); 1134 wake_up(&chip->wq);
1135 spin_unlock_bh(chip->mutex); 1135 mutex_unlock(&chip->mutex);
1136 return 0; 1136 return 0;
1137} 1137}
1138static int cfi_staa_lock(struct mtd_info *mtd, loff_t ofs, uint64_t len) 1138static int cfi_staa_lock(struct mtd_info *mtd, loff_t ofs, uint64_t len)
@@ -1203,7 +1203,7 @@ static inline int do_unlock_oneblock(struct map_info *map, struct flchip *chip,
1203 1203
1204 timeo = jiffies + HZ; 1204 timeo = jiffies + HZ;
1205retry: 1205retry:
1206 spin_lock_bh(chip->mutex); 1206 mutex_lock(&chip->mutex);
1207 1207
1208 /* Check that the chip's ready to talk to us. */ 1208 /* Check that the chip's ready to talk to us. */
1209 switch (chip->state) { 1209 switch (chip->state) {
@@ -1220,13 +1220,13 @@ retry:
1220 1220
1221 /* Urgh. Chip not yet ready to talk to us. */ 1221 /* Urgh. Chip not yet ready to talk to us. */
1222 if (time_after(jiffies, timeo)) { 1222 if (time_after(jiffies, timeo)) {
1223 spin_unlock_bh(chip->mutex); 1223 mutex_unlock(&chip->mutex);
1224 printk(KERN_ERR "waiting for chip to be ready timed out in unlock\n"); 1224 printk(KERN_ERR "waiting for chip to be ready timed out in unlock\n");
1225 return -EIO; 1225 return -EIO;
1226 } 1226 }
1227 1227
1228 /* Latency issues. Drop the lock, wait a while and retry */ 1228 /* Latency issues. Drop the lock, wait a while and retry */
1229 spin_unlock_bh(chip->mutex); 1229 mutex_unlock(&chip->mutex);
1230 cfi_udelay(1); 1230 cfi_udelay(1);
1231 goto retry; 1231 goto retry;
1232 1232
@@ -1235,7 +1235,7 @@ retry:
1235 someone changes the status */ 1235 someone changes the status */
1236 set_current_state(TASK_UNINTERRUPTIBLE); 1236 set_current_state(TASK_UNINTERRUPTIBLE);
1237 add_wait_queue(&chip->wq, &wait); 1237 add_wait_queue(&chip->wq, &wait);
1238 spin_unlock_bh(chip->mutex); 1238 mutex_unlock(&chip->mutex);
1239 schedule(); 1239 schedule();
1240 remove_wait_queue(&chip->wq, &wait); 1240 remove_wait_queue(&chip->wq, &wait);
1241 timeo = jiffies + HZ; 1241 timeo = jiffies + HZ;
@@ -1247,9 +1247,9 @@ retry:
1247 map_write(map, CMD(0xD0), adr); 1247 map_write(map, CMD(0xD0), adr);
1248 chip->state = FL_UNLOCKING; 1248 chip->state = FL_UNLOCKING;
1249 1249
1250 spin_unlock_bh(chip->mutex); 1250 mutex_unlock(&chip->mutex);
1251 msleep(1000); 1251 msleep(1000);
1252 spin_lock_bh(chip->mutex); 1252 mutex_lock(&chip->mutex);
1253 1253
1254 /* FIXME. Use a timer to check this, and return immediately. */ 1254 /* FIXME. Use a timer to check this, and return immediately. */
1255 /* Once the state machine's known to be working I'll do that */ 1255 /* Once the state machine's known to be working I'll do that */
@@ -1267,21 +1267,21 @@ retry:
1267 chip->state = FL_STATUS; 1267 chip->state = FL_STATUS;
1268 printk(KERN_ERR "waiting for unlock to complete timed out. Xstatus = %lx, status = %lx.\n", status.x[0], map_read(map, adr).x[0]); 1268 printk(KERN_ERR "waiting for unlock to complete timed out. Xstatus = %lx, status = %lx.\n", status.x[0], map_read(map, adr).x[0]);
1269 DISABLE_VPP(map); 1269 DISABLE_VPP(map);
1270 spin_unlock_bh(chip->mutex); 1270 mutex_unlock(&chip->mutex);
1271 return -EIO; 1271 return -EIO;
1272 } 1272 }
1273 1273
1274 /* Latency issues. Drop the unlock, wait a while and retry */ 1274 /* Latency issues. Drop the unlock, wait a while and retry */
1275 spin_unlock_bh(chip->mutex); 1275 mutex_unlock(&chip->mutex);
1276 cfi_udelay(1); 1276 cfi_udelay(1);
1277 spin_lock_bh(chip->mutex); 1277 mutex_lock(&chip->mutex);
1278 } 1278 }
1279 1279
1280 /* Done and happy. */ 1280 /* Done and happy. */
1281 chip->state = FL_STATUS; 1281 chip->state = FL_STATUS;
1282 DISABLE_VPP(map); 1282 DISABLE_VPP(map);
1283 wake_up(&chip->wq); 1283 wake_up(&chip->wq);
1284 spin_unlock_bh(chip->mutex); 1284 mutex_unlock(&chip->mutex);
1285 return 0; 1285 return 0;
1286} 1286}
1287static int cfi_staa_unlock(struct mtd_info *mtd, loff_t ofs, uint64_t len) 1287static int cfi_staa_unlock(struct mtd_info *mtd, loff_t ofs, uint64_t len)
@@ -1334,7 +1334,7 @@ static int cfi_staa_suspend(struct mtd_info *mtd)
1334 for (i=0; !ret && i<cfi->numchips; i++) { 1334 for (i=0; !ret && i<cfi->numchips; i++) {
1335 chip = &cfi->chips[i]; 1335 chip = &cfi->chips[i];
1336 1336
1337 spin_lock_bh(chip->mutex); 1337 mutex_lock(&chip->mutex);
1338 1338
1339 switch(chip->state) { 1339 switch(chip->state) {
1340 case FL_READY: 1340 case FL_READY:
@@ -1354,7 +1354,7 @@ static int cfi_staa_suspend(struct mtd_info *mtd)
1354 ret = -EAGAIN; 1354 ret = -EAGAIN;
1355 break; 1355 break;
1356 } 1356 }
1357 spin_unlock_bh(chip->mutex); 1357 mutex_unlock(&chip->mutex);
1358 } 1358 }
1359 1359
1360 /* Unlock the chips again */ 1360 /* Unlock the chips again */
@@ -1363,7 +1363,7 @@ static int cfi_staa_suspend(struct mtd_info *mtd)
1363 for (i--; i >=0; i--) { 1363 for (i--; i >=0; i--) {
1364 chip = &cfi->chips[i]; 1364 chip = &cfi->chips[i];
1365 1365
1366 spin_lock_bh(chip->mutex); 1366 mutex_lock(&chip->mutex);
1367 1367
1368 if (chip->state == FL_PM_SUSPENDED) { 1368 if (chip->state == FL_PM_SUSPENDED) {
1369 /* No need to force it into a known state here, 1369 /* No need to force it into a known state here,
@@ -1372,7 +1372,7 @@ static int cfi_staa_suspend(struct mtd_info *mtd)
1372 chip->state = chip->oldstate; 1372 chip->state = chip->oldstate;
1373 wake_up(&chip->wq); 1373 wake_up(&chip->wq);
1374 } 1374 }
1375 spin_unlock_bh(chip->mutex); 1375 mutex_unlock(&chip->mutex);
1376 } 1376 }
1377 } 1377 }
1378 1378
@@ -1390,7 +1390,7 @@ static void cfi_staa_resume(struct mtd_info *mtd)
1390 1390
1391 chip = &cfi->chips[i]; 1391 chip = &cfi->chips[i];
1392 1392
1393 spin_lock_bh(chip->mutex); 1393 mutex_lock(&chip->mutex);
1394 1394
1395 /* Go to known state. Chip may have been power cycled */ 1395 /* Go to known state. Chip may have been power cycled */
1396 if (chip->state == FL_PM_SUSPENDED) { 1396 if (chip->state == FL_PM_SUSPENDED) {
@@ -1399,7 +1399,7 @@ static void cfi_staa_resume(struct mtd_info *mtd)
1399 wake_up(&chip->wq); 1399 wake_up(&chip->wq);
1400 } 1400 }
1401 1401
1402 spin_unlock_bh(chip->mutex); 1402 mutex_unlock(&chip->mutex);
1403 } 1403 }
1404} 1404}
1405 1405
diff --git a/drivers/mtd/chips/fwh_lock.h b/drivers/mtd/chips/fwh_lock.h
index 57e0e4e921f9..d18064977192 100644
--- a/drivers/mtd/chips/fwh_lock.h
+++ b/drivers/mtd/chips/fwh_lock.h
@@ -58,10 +58,10 @@ static int fwh_xxlock_oneblock(struct map_info *map, struct flchip *chip,
58 * to flash memory - that means that we don't have to check status 58 * to flash memory - that means that we don't have to check status
59 * and timeout. 59 * and timeout.
60 */ 60 */
61 spin_lock(chip->mutex); 61 mutex_lock(&chip->mutex);
62 ret = get_chip(map, chip, adr, FL_LOCKING); 62 ret = get_chip(map, chip, adr, FL_LOCKING);
63 if (ret) { 63 if (ret) {
64 spin_unlock(chip->mutex); 64 mutex_unlock(&chip->mutex);
65 return ret; 65 return ret;
66 } 66 }
67 67
@@ -72,7 +72,7 @@ static int fwh_xxlock_oneblock(struct map_info *map, struct flchip *chip,
72 /* Done and happy. */ 72 /* Done and happy. */
73 chip->state = chip->oldstate; 73 chip->state = chip->oldstate;
74 put_chip(map, chip, adr); 74 put_chip(map, chip, adr);
75 spin_unlock(chip->mutex); 75 mutex_unlock(&chip->mutex);
76 return 0; 76 return 0;
77} 77}
78 78
diff --git a/drivers/mtd/chips/gen_probe.c b/drivers/mtd/chips/gen_probe.c
index e2dc96441e05..fcc1bc02c8a2 100644
--- a/drivers/mtd/chips/gen_probe.c
+++ b/drivers/mtd/chips/gen_probe.c
@@ -155,8 +155,7 @@ static struct cfi_private *genprobe_ident_chips(struct map_info *map, struct chi
155 pchip->start = (i << cfi.chipshift); 155 pchip->start = (i << cfi.chipshift);
156 pchip->state = FL_READY; 156 pchip->state = FL_READY;
157 init_waitqueue_head(&pchip->wq); 157 init_waitqueue_head(&pchip->wq);
158 spin_lock_init(&pchip->_spinlock); 158 mutex_init(&pchip->mutex);
159 pchip->mutex = &pchip->_spinlock;
160 } 159 }
161 } 160 }
162 161
diff --git a/drivers/mtd/lpddr/lpddr_cmds.c b/drivers/mtd/lpddr/lpddr_cmds.c
index e22ca49583e7..eb6f437ca9ec 100644
--- a/drivers/mtd/lpddr/lpddr_cmds.c
+++ b/drivers/mtd/lpddr/lpddr_cmds.c
@@ -106,8 +106,7 @@ struct mtd_info *lpddr_cmdset(struct map_info *map)
106 /* those should be reset too since 106 /* those should be reset too since
107 they create memory references. */ 107 they create memory references. */
108 init_waitqueue_head(&chip->wq); 108 init_waitqueue_head(&chip->wq);
109 spin_lock_init(&chip->_spinlock); 109 mutex_init(&chip->mutex);
110 chip->mutex = &chip->_spinlock;
111 chip++; 110 chip++;
112 } 111 }
113 } 112 }
@@ -143,7 +142,7 @@ static int wait_for_ready(struct map_info *map, struct flchip *chip,
143 } 142 }
144 143
145 /* OK Still waiting. Drop the lock, wait a while and retry. */ 144 /* OK Still waiting. Drop the lock, wait a while and retry. */
146 spin_unlock(chip->mutex); 145 mutex_unlock(&chip->mutex);
147 if (sleep_time >= 1000000/HZ) { 146 if (sleep_time >= 1000000/HZ) {
148 /* 147 /*
149 * Half of the normal delay still remaining 148 * Half of the normal delay still remaining
@@ -158,17 +157,17 @@ static int wait_for_ready(struct map_info *map, struct flchip *chip,
158 cond_resched(); 157 cond_resched();
159 timeo--; 158 timeo--;
160 } 159 }
161 spin_lock(chip->mutex); 160 mutex_lock(&chip->mutex);
162 161
163 while (chip->state != chip_state) { 162 while (chip->state != chip_state) {
164 /* Someone's suspended the operation: sleep */ 163 /* Someone's suspended the operation: sleep */
165 DECLARE_WAITQUEUE(wait, current); 164 DECLARE_WAITQUEUE(wait, current);
166 set_current_state(TASK_UNINTERRUPTIBLE); 165 set_current_state(TASK_UNINTERRUPTIBLE);
167 add_wait_queue(&chip->wq, &wait); 166 add_wait_queue(&chip->wq, &wait);
168 spin_unlock(chip->mutex); 167 mutex_unlock(&chip->mutex);
169 schedule(); 168 schedule();
170 remove_wait_queue(&chip->wq, &wait); 169 remove_wait_queue(&chip->wq, &wait);
171 spin_lock(chip->mutex); 170 mutex_lock(&chip->mutex);
172 } 171 }
173 if (chip->erase_suspended || chip->write_suspended) { 172 if (chip->erase_suspended || chip->write_suspended) {
174 /* Suspend has occured while sleep: reset timeout */ 173 /* Suspend has occured while sleep: reset timeout */
@@ -229,20 +228,20 @@ static int get_chip(struct map_info *map, struct flchip *chip, int mode)
229 * it'll happily send us to sleep. In any case, when 228 * it'll happily send us to sleep. In any case, when
230 * get_chip returns success we're clear to go ahead. 229 * get_chip returns success we're clear to go ahead.
231 */ 230 */
232 ret = spin_trylock(contender->mutex); 231 ret = mutex_trylock(&contender->mutex);
233 spin_unlock(&shared->lock); 232 spin_unlock(&shared->lock);
234 if (!ret) 233 if (!ret)
235 goto retry; 234 goto retry;
236 spin_unlock(chip->mutex); 235 mutex_unlock(&chip->mutex);
237 ret = chip_ready(map, contender, mode); 236 ret = chip_ready(map, contender, mode);
238 spin_lock(chip->mutex); 237 mutex_lock(&chip->mutex);
239 238
240 if (ret == -EAGAIN) { 239 if (ret == -EAGAIN) {
241 spin_unlock(contender->mutex); 240 mutex_unlock(&contender->mutex);
242 goto retry; 241 goto retry;
243 } 242 }
244 if (ret) { 243 if (ret) {
245 spin_unlock(contender->mutex); 244 mutex_unlock(&contender->mutex);
246 return ret; 245 return ret;
247 } 246 }
248 spin_lock(&shared->lock); 247 spin_lock(&shared->lock);
@@ -251,10 +250,10 @@ static int get_chip(struct map_info *map, struct flchip *chip, int mode)
251 * state. Put contender and retry. */ 250 * state. Put contender and retry. */
252 if (chip->state == FL_SYNCING) { 251 if (chip->state == FL_SYNCING) {
253 put_chip(map, contender); 252 put_chip(map, contender);
254 spin_unlock(contender->mutex); 253 mutex_unlock(&contender->mutex);
255 goto retry; 254 goto retry;
256 } 255 }
257 spin_unlock(contender->mutex); 256 mutex_unlock(&contender->mutex);
258 } 257 }
259 258
260 /* Check if we have suspended erase on this chip. 259 /* Check if we have suspended erase on this chip.
@@ -264,10 +263,10 @@ static int get_chip(struct map_info *map, struct flchip *chip, int mode)
264 spin_unlock(&shared->lock); 263 spin_unlock(&shared->lock);
265 set_current_state(TASK_UNINTERRUPTIBLE); 264 set_current_state(TASK_UNINTERRUPTIBLE);
266 add_wait_queue(&chip->wq, &wait); 265 add_wait_queue(&chip->wq, &wait);
267 spin_unlock(chip->mutex); 266 mutex_unlock(&chip->mutex);
268 schedule(); 267 schedule();
269 remove_wait_queue(&chip->wq, &wait); 268 remove_wait_queue(&chip->wq, &wait);
270 spin_lock(chip->mutex); 269 mutex_lock(&chip->mutex);
271 goto retry; 270 goto retry;
272 } 271 }
273 272
@@ -336,10 +335,10 @@ static int chip_ready(struct map_info *map, struct flchip *chip, int mode)
336sleep: 335sleep:
337 set_current_state(TASK_UNINTERRUPTIBLE); 336 set_current_state(TASK_UNINTERRUPTIBLE);
338 add_wait_queue(&chip->wq, &wait); 337 add_wait_queue(&chip->wq, &wait);
339 spin_unlock(chip->mutex); 338 mutex_unlock(&chip->mutex);
340 schedule(); 339 schedule();
341 remove_wait_queue(&chip->wq, &wait); 340 remove_wait_queue(&chip->wq, &wait);
342 spin_lock(chip->mutex); 341 mutex_lock(&chip->mutex);
343 return -EAGAIN; 342 return -EAGAIN;
344 } 343 }
345} 344}
@@ -355,12 +354,12 @@ static void put_chip(struct map_info *map, struct flchip *chip)
355 if (shared->writing && shared->writing != chip) { 354 if (shared->writing && shared->writing != chip) {
356 /* give back the ownership */ 355 /* give back the ownership */
357 struct flchip *loaner = shared->writing; 356 struct flchip *loaner = shared->writing;
358 spin_lock(loaner->mutex); 357 mutex_lock(&loaner->mutex);
359 spin_unlock(&shared->lock); 358 spin_unlock(&shared->lock);
360 spin_unlock(chip->mutex); 359 mutex_unlock(&chip->mutex);
361 put_chip(map, loaner); 360 put_chip(map, loaner);
362 spin_lock(chip->mutex); 361 mutex_lock(&chip->mutex);
363 spin_unlock(loaner->mutex); 362 mutex_unlock(&loaner->mutex);
364 wake_up(&chip->wq); 363 wake_up(&chip->wq);
365 return; 364 return;
366 } 365 }
@@ -413,10 +412,10 @@ int do_write_buffer(struct map_info *map, struct flchip *chip,
413 412
414 wbufsize = 1 << lpddr->qinfo->BufSizeShift; 413 wbufsize = 1 << lpddr->qinfo->BufSizeShift;
415 414
416 spin_lock(chip->mutex); 415 mutex_lock(&chip->mutex);
417 ret = get_chip(map, chip, FL_WRITING); 416 ret = get_chip(map, chip, FL_WRITING);
418 if (ret) { 417 if (ret) {
419 spin_unlock(chip->mutex); 418 mutex_unlock(&chip->mutex);
420 return ret; 419 return ret;
421 } 420 }
422 /* Figure out the number of words to write */ 421 /* Figure out the number of words to write */
@@ -477,7 +476,7 @@ int do_write_buffer(struct map_info *map, struct flchip *chip,
477 } 476 }
478 477
479 out: put_chip(map, chip); 478 out: put_chip(map, chip);
480 spin_unlock(chip->mutex); 479 mutex_unlock(&chip->mutex);
481 return ret; 480 return ret;
482} 481}
483 482
@@ -489,10 +488,10 @@ int do_erase_oneblock(struct mtd_info *mtd, loff_t adr)
489 struct flchip *chip = &lpddr->chips[chipnum]; 488 struct flchip *chip = &lpddr->chips[chipnum];
490 int ret; 489 int ret;
491 490
492 spin_lock(chip->mutex); 491 mutex_lock(&chip->mutex);
493 ret = get_chip(map, chip, FL_ERASING); 492 ret = get_chip(map, chip, FL_ERASING);
494 if (ret) { 493 if (ret) {
495 spin_unlock(chip->mutex); 494 mutex_unlock(&chip->mutex);
496 return ret; 495 return ret;
497 } 496 }
498 send_pfow_command(map, LPDDR_BLOCK_ERASE, adr, 0, NULL); 497 send_pfow_command(map, LPDDR_BLOCK_ERASE, adr, 0, NULL);
@@ -504,7 +503,7 @@ int do_erase_oneblock(struct mtd_info *mtd, loff_t adr)
504 goto out; 503 goto out;
505 } 504 }
506 out: put_chip(map, chip); 505 out: put_chip(map, chip);
507 spin_unlock(chip->mutex); 506 mutex_unlock(&chip->mutex);
508 return ret; 507 return ret;
509} 508}
510 509
@@ -517,10 +516,10 @@ static int lpddr_read(struct mtd_info *mtd, loff_t adr, size_t len,
517 struct flchip *chip = &lpddr->chips[chipnum]; 516 struct flchip *chip = &lpddr->chips[chipnum];
518 int ret = 0; 517 int ret = 0;
519 518
520 spin_lock(chip->mutex); 519 mutex_lock(&chip->mutex);
521 ret = get_chip(map, chip, FL_READY); 520 ret = get_chip(map, chip, FL_READY);
522 if (ret) { 521 if (ret) {
523 spin_unlock(chip->mutex); 522 mutex_unlock(&chip->mutex);
524 return ret; 523 return ret;
525 } 524 }
526 525
@@ -528,7 +527,7 @@ static int lpddr_read(struct mtd_info *mtd, loff_t adr, size_t len,
528 *retlen = len; 527 *retlen = len;
529 528
530 put_chip(map, chip); 529 put_chip(map, chip);
531 spin_unlock(chip->mutex); 530 mutex_unlock(&chip->mutex);
532 return ret; 531 return ret;
533} 532}
534 533
@@ -568,9 +567,9 @@ static int lpddr_point(struct mtd_info *mtd, loff_t adr, size_t len,
568 else 567 else
569 thislen = len; 568 thislen = len;
570 /* get the chip */ 569 /* get the chip */
571 spin_lock(chip->mutex); 570 mutex_lock(&chip->mutex);
572 ret = get_chip(map, chip, FL_POINT); 571 ret = get_chip(map, chip, FL_POINT);
573 spin_unlock(chip->mutex); 572 mutex_unlock(&chip->mutex);
574 if (ret) 573 if (ret)
575 break; 574 break;
576 575
@@ -610,7 +609,7 @@ static void lpddr_unpoint (struct mtd_info *mtd, loff_t adr, size_t len)
610 else 609 else
611 thislen = len; 610 thislen = len;
612 611
613 spin_lock(chip->mutex); 612 mutex_lock(&chip->mutex);
614 if (chip->state == FL_POINT) { 613 if (chip->state == FL_POINT) {
615 chip->ref_point_counter--; 614 chip->ref_point_counter--;
616 if (chip->ref_point_counter == 0) 615 if (chip->ref_point_counter == 0)
@@ -620,7 +619,7 @@ static void lpddr_unpoint (struct mtd_info *mtd, loff_t adr, size_t len)
620 "pointed region\n", map->name); 619 "pointed region\n", map->name);
621 620
622 put_chip(map, chip); 621 put_chip(map, chip);
623 spin_unlock(chip->mutex); 622 mutex_unlock(&chip->mutex);
624 623
625 len -= thislen; 624 len -= thislen;
626 ofs = 0; 625 ofs = 0;
@@ -726,10 +725,10 @@ int do_xxlock(struct mtd_info *mtd, loff_t adr, uint32_t len, int thunk)
726 int chipnum = adr >> lpddr->chipshift; 725 int chipnum = adr >> lpddr->chipshift;
727 struct flchip *chip = &lpddr->chips[chipnum]; 726 struct flchip *chip = &lpddr->chips[chipnum];
728 727
729 spin_lock(chip->mutex); 728 mutex_lock(&chip->mutex);
730 ret = get_chip(map, chip, FL_LOCKING); 729 ret = get_chip(map, chip, FL_LOCKING);
731 if (ret) { 730 if (ret) {
732 spin_unlock(chip->mutex); 731 mutex_unlock(&chip->mutex);
733 return ret; 732 return ret;
734 } 733 }
735 734
@@ -749,7 +748,7 @@ int do_xxlock(struct mtd_info *mtd, loff_t adr, uint32_t len, int thunk)
749 goto out; 748 goto out;
750 } 749 }
751out: put_chip(map, chip); 750out: put_chip(map, chip);
752 spin_unlock(chip->mutex); 751 mutex_unlock(&chip->mutex);
753 return ret; 752 return ret;
754} 753}
755 754
@@ -770,10 +769,10 @@ int word_program(struct map_info *map, loff_t adr, uint32_t curval)
770 int chipnum = adr >> lpddr->chipshift; 769 int chipnum = adr >> lpddr->chipshift;
771 struct flchip *chip = &lpddr->chips[chipnum]; 770 struct flchip *chip = &lpddr->chips[chipnum];
772 771
773 spin_lock(chip->mutex); 772 mutex_lock(&chip->mutex);
774 ret = get_chip(map, chip, FL_WRITING); 773 ret = get_chip(map, chip, FL_WRITING);
775 if (ret) { 774 if (ret) {
776 spin_unlock(chip->mutex); 775 mutex_unlock(&chip->mutex);
777 return ret; 776 return ret;
778 } 777 }
779 778
@@ -787,7 +786,7 @@ int word_program(struct map_info *map, loff_t adr, uint32_t curval)
787 } 786 }
788 787
789out: put_chip(map, chip); 788out: put_chip(map, chip);
790 spin_unlock(chip->mutex); 789 mutex_unlock(&chip->mutex);
791 return ret; 790 return ret;
792} 791}
793 792